+ All Categories
Home > Documents > Using random models in derivative free optimization

Using random models in derivative free optimization

Date post: 23-Feb-2016
Category:
Upload: diella
View: 68 times
Download: 3 times
Share this document with a friend
Description:
Using random models in derivative free optimization. Katya Scheinberg Lehigh University (mainly based on work with A. Bandeira and L.N. Vicente and also with A.R. Conn, Ph.Toint and C. Cartis ). Derivative free optimization. Unconstrained optimization problem - PowerPoint PPT Presentation
Popular Tags:
78
Using random models in derivative free optimization Katya Scheinberg Lehigh University (mainly based on work with A. Bandeira and L.N. Vicente and also with A.R. Conn, Ph.Toint and C. Cartis ) 08/20/2012 ISMP 2012
Transcript

Derivative free optimization

Using random models in derivative free optimizationKatya ScheinbergLehigh University(mainly based on work with A. Bandeira and L.N. Vicente and also with A.R. Conn, Ph.Toint and C. Cartis )

08/20/2012ISMP 201208/20/2012ISMP 2012Derivative free optimizationUnconstrained optimization problem

Function f is computed by a black box, no derivative information is available.Numerical noise is often present, but we do not account for it in this talk!f 2 C1 or C2 and is deterministic.May be expensive to compute.

Black box function evaluation08/20/2012ISMP 2012 x=(x1,x2,x3,,xn) v=f(x1,,xn)vAll we can do is sample the function values at some sample points08/20/2012ISMP 2012Sampling the black box functionSample pointsHow to choose and to use the sample points and the functions values defines different DFO methods 08/20/2012ISMP 2012OutlineReview with illustrations of existing methods as motivation for using models.Polynomial interpolation models and motivation for models based on random sample sets.Structure recovery using random sample sets and compressed sensing in DFO. Algorithms using random models and conditions on these models.Convergence theory for TR framework based on random models. 08/20/2012ISMP 2012Algorithms08/20/2012ISMP 2012Nelder-Mead method (1965)08/20/2012ISMP 2012Nelder-Mead method (1965)08/20/2012ISMP 2012Nelder-Mead method (1965)08/20/2012ISMP 2012Nelder-Mead method (1965)08/20/2012ISMP 2012Nelder-Mead method (1965)08/20/2012ISMP 2012Nelder-Mead method (1965)The simplex changes shape during the algorithm to adapt to curvature. But the shape can deteriorate and NM gets stuckNelder Mead on Rosenbrock

08/20/2012ISMP 2012Surprisingly good, but essentially a heuristic08/20/2012ISMP 2012Direct Search methods (early 1990s)Torczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodsTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodsTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodTorczon, Dennis, Audet, Vicente, Luizzi, many others08/20/2012ISMP 2012Direct Search methodFixed pattern, never deteriorates: theoretically convergent, but slowCompass Search on Rosenbrock

08/20/2012ISMP 2012Very slow because of badly aligned axis directionsRandom directions on Rosenbrock

08/20/2012ISMP 2012Better progress, but very sensitive to step size choicesPolyak, Yuditski, Nesterov, Lan, Nemirovski, Audet & Dennis, etc08/20/2012ISMP 2012Model based trust region methods Powell, Conn, S. Toint, Vicente, Wild, etc.08/20/2012ISMP 2012Model based trust region methodsPowell, Conn, S. Toint, Vicente, Wild, etc.08/20/2012ISMP 2012Model based trust region methodsPowell, Conn, S. Toint, Vicente, Wild, etc.08/20/2012ISMP 2012Model Based trust region methodsExploits curvature, flexible efficient steps, uses second order models.Second order model based TR method on Rosenbrock08/20/2012ISMP 2012

Moral:Building and using models is a good idea.Randomness may offer speed up.Can we combine randomization and models successfully and what would we gain? 08/20/2012ISMP 201208/20/2012ISMP 2012Polynomial models 08/20/2012ISMP 2012Linear Interpolation

08/20/2012ISMP 2012Good vs. bad linear Interpolation

If is nonsingular then linear model exists for any f(x)Better conditioned M => better models08/20/2012ISMP 2012Examples of sample sets for linear interpolation

Badly poised setFinite difference sample setRandom sample set08/20/2012ISMP 2012Polynomial Interpolation

08/20/2012ISMP 2012Specifically for quadratic interpolation

Interpolation model:

08/20/2012ISMP 2012Sample sets and models for f(x)=cos(x)+sin(y)

08/20/2012ISMP 2012Sample sets and models for f(x)=cos(x)+sin(y)

08/20/2012ISMP 2012Sample sets and models for f(x)=cos(x)+sin(y)

08/20/2012ISMP 2012Example that shows that we need to maintain the quality of the sample set

08/20/2012ISMP 2012

08/20/2012ISMP 2012

Observations:Building and maintaining good models is needed.But it requires computational and implementation effort and many function evaluations.Random sample sets usually produce good models, the only effort required is computing the function values.This can be done in parallel and random sample sets can produce good models with fewer points.08/20/2012ISMP 2012How?sparse black box optimization08/20/2012ISMP 2012 x=(x1,x2,x3,,xn) v=f(xS) S{1..n}v08/20/2012ISMP 2012Sparse linear Interpolation

08/20/2012ISMP 2012Sparse linear Interpolation

We have an (underdetermined) system of linear equations with a sparse solutionCan we find correct sparse using less than n+1 sample points in Y?08/20/2012ISMP 2012Using celebrated compressed sensing results (Candes&Tao, Donoho, etc)

has RIPBy solvingWhenever08/20/2012ISMP 2012Using celebrated compressed sensing results and random matrix theory (Candes&Tao, Donoho, Rauhut, etc)

have RIP?Yes, with high prob., when Y is random and p=O(|S|log n)DoesNote: O(|S|log n)


Recommended