2. Optimality Conditions, Convexity, Newton’s Method for Equations
-
728x90
반응형
Optimality Conditions: Preliminaries
Local minima and maxima have one thing in common
First-order Necessary Condition
From now on f is differentiable and that its first and second derivatives are continuous for every x∈X where X is a domain of f
If x∗ is a local minimum, then
Not a sufficient condition, since it could be a saddle point
Second-order Conditions
If x∗ is a local minimum, then ∇2f(x∗) is positive semi-definite
Proof
But you should note that even though ∇f(x∗)=0 and ∇2f(x∗)≥0, we can’t be sure that x∗ must be a local minimum.
Think about,
That means there is no sufficient and necessary condition in general
Second-order sufficient condition
It is sufficient to guarantee that x∗ is a local minimizer. If
then x∗ is a strict local minimizer of f.
Proof
Since ∇f(x∗)=0,
where ζ∈Np(x∗)
Since ∇2f is continuous, ∃δ such that
Take v=min(1,δ)
Moreover, since ∇2f(x∗) is positive definite,
Therefore,
This means that ∇2f(ζ) is also positive definite. So, x∗ is a local minimum.
Convexity
A set S is convex if any x∈S and y∈S,
A function f is convex on a convex set S if for any x∈S and y∈S,
A minimization problem is convex if f is convex and S is convex
Implication of Convexity
For a convex optimization problem, a local optimum x∗ is also a global optimum
If the feasible region is convex, it implies that all the points along the line segment between the current solution and the new feasible solution are also feasible.
Therefore, any smaller step size will still allow us to move along the same direction from the current solution, while remaining within the feasible region. This is because the convexity property guarantees that the entire line segment between the current solution and the new feasible solution lies within the feasible region.
This property is valuable in optimization algorithms, as it ensures that if we have found a feasible solution, we can always refine it further by taking smaller steps while maintaining feasibility. It provides flexibility in adjusting the step size and allows for a more precise exploration of the feasible region.
How to recognize convexity?
One dimension case : A function f is convex if and only if f′′(x)≥0 for all x∈X where X is a domain of f
Multiple dimension case
f is convex if and only if Hessian ∇2f(x)≥0 for all x∈X where X is a domain of f
If ∇2f(x) is positive definite for all x∈X (i.e. ∇2f(x)>0,∀x∈X), then f is strictly convex
→ Note that it is not true for opposite direction
f is convex if and only if it is above the tangent
Convexity and Optimality Condition
Actually if f is convex, local optimum and global optimum is equivalent. In addition,
Newton’s Method for Nonlinear Equations
A sequence of points as approximate solutions
The sequence converges if it approaches a solution of the equation
Especially for one dimension,
This means that
Therefore,
This is the Newton-Raphson method
A related derivative-free method is the secant method
Convergent or not: depends on the function and the initial point