Numerical Optimization Chapter 1 and 2
Numerical Optimization¶
Introduction¶
Given the following notation:
- x : a vector of variable/unkowns/paramters
- f : objective function
- $c_i$ constraint functions
We can write an optimization problem in the format
$$\underset{x \in \mathbb{R}^N}{\text{ min } f(x)} \text{ subject to } \matrix{ c_i(x) = 0, i \in \epsilon \\ c_i(x) \geq 0, i \in I }$$where I and $\epsilon$ are equality and inequality constraints
Unconstrained optimization¶
When doing an unconstrained optimization, we define the optimization problem as
$$\underset{x \in \mathbb{R}^N}{\text{ min f(x)}}$$Where N $\geq 1 $ and the objective function (f(x)) is smooth. When solving we define the solutions ($x^*$) as follows:
- $x^*$ is a global minimizer if f($x^*$) $ < $ f(x) for all x $\neq x^*$
- $x^*$ is a weak local minimizer if f($x^*$) $ \leq $ f(x) for some x $\in N$ where N is a neighborhood of values
- $x^*$ is a strict/strong local minimizer if f($x^*$) $ < $ f(x) for some x $\in N$ where N is a neighborhood of values and $x^* \neq x$
How do we define a minimimum value?¶
Necessary conditions for optimizatiy are derived by assuming $x^*$ is a local minimizer and then proving facts about $\nabla f$ and $\nabla^2 f$
First-Order Necessary conditions (Theorm 2.2)¶
If $x^*$ is a local minimizer and f is continuously differentiable in an open neighborhood of $x^*$, then $\nabla f(x^*)$ = 0.
In essence:
- given an objective function at a minimum the derivative is zero
Second-Order Necessary conditions (Theorm 2.3)¶
If $x^*$ is a local minimizer of f and $\nabla^2$ f exists and is continuous in an open neighborhood of $x^*$, then $\nabla f (x^*)$ = 0 and $\nabla^2 f (x^*)$ is positive semidefinite.
In essence:
- given an objective function at a minimum the derivative is zero
- given an objective function at a minimum the second derivative is positive semi definite implies a weak local minimizer
Second-Order Sufficient conditions (Theorm 2.4)¶
Suppose that $\nabla^2 f$ is continuous in an open neighborhood of $x^*$ and that $\nabla f(x^*) = 0$ and $\nabla^2 f(x^*)$ is positive definite. Then $x^*$ is a strict local minimizer of f .
In essence:
- given an objective function at a minimum the derivative is zero
- given an objective function at a minimum the second derivative is positive definite implies a strict local minimizer
Examples of Global, local, strict and weak minimizer:¶
Convex functions only have one minimum (Theorm 2.5)¶
Any local minimizer of a convex function is the global minimizer
Calculating the next step¶
Usually when calculating the next step we write this mathematically as follows: $$x_{k+1} = x_k + \alpha p_k$$
Where k is the current iteration, $p_k$ is the search direction at a given iteration and $\alpha$ is how far we are moving in the direction.