lecture 4 - techniontosca.cs.technion.ac.il/book/handouts/stanford09_optimization.pdf · numerical...

44
Numerical optimization Lecture 4 © Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009

Upload: others

Post on 25-May-2021

6 views

Category:

Documents


0 download

TRANSCRIPT

Microsoft PowerPoint - Stanford09_optimization1Stanford University, Winter 2009
Longest Shortest
Largest Smallest
Optimization problems
The value is the minimum
where
Local vs. global minimum
Local
minimum
Global
minimum
Find minimum by analyzing the local behavior of the cost function
5Numerical geometry of non-rigid shapes Numerical optimization
Local vs. global in real life
Main summit
8,047 m
False summit
8,030 m
6Numerical geometry of non-rigid shapes Numerical optimization
Convex functions
A function defined on a convex set is called convex if
for any and
One-dimensional optimality conditions
.

Approximate a function around as a parabola using Taylor expansion
guarantees
Gradient
In multidimensional case, linearization of the function according to Taylor
gives a multidimensional analogy of the derivative.
The function , denoted as , is called the gradient of
In one-dimensional case, it reduces to standard definition of derivative
9Numerical geometry of non-rigid shapes Numerical optimization
Gradient
in the following way:
Given (space of real matrices) with standard inner
product
an matrix
Compute the gradient of the function where is
an matrix
Hessian
order derivative.
is called the Hessian of
In the standard basis, Hessian is a symmetric matrix of mixed second-order
derivatives
Point is the local minimizer of a -function if
.
matrix (denoted )
Approximate a function around as a parabola using Taylor expansion
guarantees
Optimization algorithms
Descent direction
Step size
Generic optimization algorithm
Start with some
Determine descent direction
Update iterate
16Numerical geometry of non-rigid shapes Numerical optimization
Stopping criteria
Stop when gradient norm becomes small
Stop when step size becomes small
Stop when relative objective change becomes small
17Numerical geometry of non-rigid shapes Numerical optimization
Line search
Optimal step size can be found by solving a one-dimensional optimization
problem
are generically called exact line search
18Numerical geometry of non-rigid shapes Numerical optimization
Armijo [ar-mi-xo] rule
The function sufficiently decreases if
Armijo rule (Larry Armijo, 1966): start with and decrease it by
multiplying by some until the function sufficiently decreases
19Numerical geometry of non-rigid shapes Numerical optimization
Descent direction
How to descend in the fastest way?
Go in the direction in which the height lines are the densest
20Numerical geometry of non-rigid shapes Numerical optimization
Find a unit-length direction minimizing directional
derivative
the direction (negative for a descent direction)
21Numerical geometry of non-rigid shapes Numerical optimization
Steepest descent
Normalized steepest descent
Steepest descent algorithm
Start with some
Update iterate
Condition number
-0.5
0
0.5
1
-0.5
0
0.5
1
Condition number is the ratio of maximal and minimal eigenvalues of the
Hessian ,
Steepest descent convergence rate is slow for ill-conditioned problems
24Numerical geometry of non-rigid shapes Numerical optimization
Q-norm
Function
Gradient
Preconditioning
Using Q-norm for steepest descent can be regarded as a change of
coordinates, called preconditioning
Preconditioner should be chosen to improve the condition number of
the Hessian in the proximity of the solution
In system of coordinates, the Hessian at the solution is
(a dream)
Newton method as optimal preconditioner
Best theoretically possible preconditioner , giving descent
direction
Newton direction: use Hessian as a preconditioner at each iteration
Problem: the solution is unknown in advance
Ideal condition number
Another derivation of the Newton method
(quadratic function in )
Approximate the function as a quadratic function using second-order Taylor
expansion
Close to solution the function looks like a quadratic function; the Newton
method converges fast
Newton method
Update iterate
Frozen Hessian
Observation: close to the optimum, the Hessian does not change
significantly
Reduce the number of Hessian inversions by keeping the Hessian from
previous iterations and update it once in a few iterations
Such a method is called Newton with frozen Hessian
30Numerical geometry of non-rigid shapes Numerical optimization
Cholesky factorization
Forward substitution
31Numerical geometry of non-rigid shapes Numerical optimization
Truncated Newton
Solve the Newton system approximately
A few iterations of conjugate gradients or other algorithm for the solution
of linear systems can be used
Such a method is called truncated or inexact Newton
32Numerical geometry of non-rigid shapes Numerical optimization
Non-convex optimization
MultiresolutionGood initialization
guarantee global convergence!
33Numerical geometry of non-rigid shapes Numerical optimization
Iterative majorization
.
Iterative majorization
Constrained optimization
Constrained optimization problems
are inequality constraints
are equality constraints
A subset of the search space in which the constraints hold is called
feasible set
A point belonging to the feasible set is called a feasible solution
where
37Numerical geometry of non-rigid shapes Numerical optimization
An example
Inequality constraint
Equality constraint
Feasible set
Inequality constraint is active at point if , inactive otherwise
A point is regular if the gradients of equality constraints and of
active inequality constraints are linearly independent
38Numerical geometry of non-rigid shapes Numerical optimization
Lagrange multipliers
Main idea to solve constrained problems: arrange the objective and
constraints into a single function
is called Lagrangian
and minimize it as an unconstrained problem
39Numerical geometry of non-rigid shapes Numerical optimization
KKT conditions
If is a regular point and a local minimum, there exist Lagrange multipliers
and such that
such that for active constraints and zero for
inactive constraints
KKT conditions
If the objective is convex, the inequality constraints are convex
and the equality constraints are affine, the KKT conditions are
sufficient
Sufficient conditions:
In this case, is the solution of the constrained problem (global constrained
minimizer)
Geometric interpretation
The gradient of objective and constraint must line up at the solution
Equality constraint
Penalty methods
where and are parametric penalty functions
For larger values of the parameter , the penalty on the constraint violation
is stronger
Penalty methods
Penalty methods
Find
problem initialized with