Numerical Optimization: Difference between revisions

From David's Wiki
Line 26: Line 26:
At each iteration, you solve a constrained optimization subproblem to find the best step <math>p</math>.<br>
At each iteration, you solve a constrained optimization subproblem to find the best step <math>p</math>.<br>
<math>\min_{p \in \mathbb{R}^n} m_k(p)</math> such that <math>\Vert p \Vert < \Delta_k </math>.
<math>\min_{p \in \mathbb{R}^n} m_k(p)</math> such that <math>\Vert p \Vert < \Delta_k </math>.
===Cauchy Point Algorithms===
The Cauchy point <math>p_k^c = \tau_k p_k^s</math><br>
where <math>p_k^s</math> minimizes the linear model in the trust region<br>
<math> p_k^s = argmin_{p \in \mathbb{R}^n} f_k + g_k^Tp </math> s.t. <math>\Vert p \Vert \leq \Delta_k </math><br>
and <math>\tau_k</math> minimizes our quadratic model along the line <math>p_k^s</math>:<br>
<math>\tau_k = argmin_{\tau \geq 0} m_k(\tau p_k^s)</math> s.t. <math>\Vert \tau p_k^s \leq \Delta_k </math><br>
This can be written explicitly as <math>p_k^s = - \frac{\Delta_k}{\Vert g-K \Vert} g_k</math>


==Resources==
==Resources==
* [https://link.springer.com/book/10.1007%2F978-0-387-40065-5 Numerical Optimization by Nocedal and Wright (2006)]
* [https://link.springer.com/book/10.1007%2F978-0-387-40065-5 Numerical Optimization by Nocedal and Wright (2006)]

Revision as of 21:55, 2 November 2019

Numerical Optimization


Line Search Methods

Basic idea:

  • For each iteration
    • Find a direction \(\displaystyle p\).
    • Then find a step length \(\displaystyle \alpha\) which decreases \(\displaystyle f\).
    • Take a step \(\displaystyle \alpha p\).

Trust Region Methods

Basic idea:

  • For each iteration
    • Assume a quadratic model of your objective function near a point.
    • Find a region where you trust your model accurately represents your objective function.
    • Take a step.


Variables:

  • \(\displaystyle f\) is your objective function.
  • \(\displaystyle m_k\) is your quadratic model at iteration k.
  • \(\displaystyle x_k\) is your point at iteration k.

Your model is \(\displaystyle m_k(p) = f_k + g_k^T p + \frac{1}{2}p^T B_k p\) where \(\displaystyle g_k = \nabla f(x_k)\) and \(\displaystyle B_k\) is a symmetric matrix.
At each iteration, you solve a constrained optimization subproblem to find the best step \(\displaystyle p\).
\(\displaystyle \min_{p \in \mathbb{R}^n} m_k(p)\) such that \(\displaystyle \Vert p \Vert \lt \Delta_k \).

Cauchy Point Algorithms

The Cauchy point \(\displaystyle p_k^c = \tau_k p_k^s\)
where \(\displaystyle p_k^s\) minimizes the linear model in the trust region
\(\displaystyle p_k^s = argmin_{p \in \mathbb{R}^n} f_k + g_k^Tp \) s.t. \(\displaystyle \Vert p \Vert \leq \Delta_k \)
and \(\displaystyle \tau_k\) minimizes our quadratic model along the line \(\displaystyle p_k^s\):
\(\displaystyle \tau_k = argmin_{\tau \geq 0} m_k(\tau p_k^s)\) s.t. \(\displaystyle \Vert \tau p_k^s \leq \Delta_k \)
This can be written explicitly as \(\displaystyle p_k^s = - \frac{\Delta_k}{\Vert g-K \Vert} g_k\)

Resources