Numerical Methods for PDE-Constrained Optimization

Similar documents
Inexact Newton Methods and Nonlinear Constrained Optimization

PDE-Constrained and Nonsmooth Optimization

An Inexact Newton Method for Optimization

An Inexact Newton Method for Nonconvex Equality Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization

1. Introduction. In this paper we discuss an algorithm for equality constrained optimization problems of the form. f(x) s.t.

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

5 Handling Constraints

A New Penalty-SQP Method

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization

Algorithms for Constrained Optimization

Constrained Nonlinear Optimization Algorithms

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Recent Adaptive Methods for Nonlinear Optimization

Large-Scale Nonlinear Optimization with Inexact Step Computations

A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

minimize x subject to (x 2)(x 4) u,

On the use of piecewise linear models in nonlinear programming

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Hot-Starting NLP Solvers

Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

1 Computing with constraints

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Algorithms for constrained local optimization

Flexible Penalty Functions for Nonlinear Constrained Optimization

A Trust-region-based Sequential Quadratic Programming Algorithm

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

A multigrid method for large scale inverse problems


2.3 Linear Programming

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Newton s Method. Javier Peña Convex Optimization /36-725

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

A Trust-Funnel Algorithm for Nonlinear Programming

On the Implementation of an Interior-Point Algorithm for Nonlinear Optimization with Inexact Step Computations

Sub-Sampled Newton Methods

REPORTS IN INFORMATICS

A GLOBALLY CONVERGENT STABILIZED SQP METHOD

What s New in Active-Set Methods for Nonlinear Optimization?

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

A SUFFICIENTLY EXACT INEXACT NEWTON STEP BASED ON REUSING MATRIX INFORMATION

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Numerical Optimization of Partial Differential Equations

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization /36-725

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Iterative regularization of nonlinear ill-posed problems in Banach space

Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming

A SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints

1. Introduction. We consider nonlinear optimization problems of the form. f(x) ce (x) = 0 c I (x) 0,

Lecture 13: Constrained optimization

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Direct Current Resistivity Inversion using Various Objective Functions

4. DATA ASSIMILATION FUNDAMENTALS

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Primal-Dual Interior-Point Methods

POWER SYSTEMS in general are currently operating

Scientific Computing: An Introductory Survey

Lecture 15: SQP methods for equality constrained optimization

On Lagrange multipliers of trust region subproblems

A Regularized Interior-Point Method for Constrained Nonlinear Least Squares

Arc Search Algorithms

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

Multidisciplinary System Design Optimization (MSDO)

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

Higher-Order Methods

4TE3/6TE3. Algorithms for. Continuous Optimization

5.6 Penalty method and augmented Lagrangian method

Nonlinear Optimization: What s important?

The Squared Slacks Transformation in Nonlinear Programming

Stochastic Optimization Algorithms Beyond SG

c 2007 Society for Industrial and Applied Mathematics

PDE Solvers for Fluid Flow

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

LINEAR AND NONLINEAR PROGRAMMING

A stabilized SQP method: superlinear convergence

RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems

Constrained Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization

A Quasi-Newton Algorithm for Nonconvex, Nonsmooth Optimization with Global Convergence Guarantees

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

Integration of Sequential Quadratic Programming and Domain Decomposition Methods for Nonlinear Optimal Control Problems

Constrained Optimization and Lagrangian Duality

Algorithms for Nonsmooth Optimization

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Optimization Problems in Model Predictive Control

Computational Finance

Constrained optimization: direct methods (cont.)

Transcription:

Numerical Methods for PDE-Constrained Optimization Richard H. Byrd 1 Frank E. Curtis 2 Jorge Nocedal 2 1 University of Colorado at Boulder 2 Northwestern University Courant Institute of Mathematical Sciences, 2007

Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Problem Formulation Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Problem Formulation Data assimilation in weather forecasting Goal: up-to-date global weather forecast for the next 7 to 10 days 1 If an entire initial state of the atmosphere (temperatures, pressures, wind patterns, humidities) were known at a certain point in time, then an accurate forecast could be obtained by integrating atmospheric model equations forward in time Flow described by Navier-Stokes and further sophistications of atmospheric physics and dynamics (none of which will be discussed here) 1 (Fisher, Nocedal, Trémolet, and Wright, 2007)

Problem Formulation In reality: partial information Limited amount of data (satellites, buoys, planes, ground-based sensors) Each observation is subject to error Nonuniformly distributed around the globe (satellite paths, densely-populated areas) (Graphics courtesy Yannick Trémolet)

Problem Formulation In reality: partial information Limited amount of data (satellites, buoys, planes, ground-based sensors) Each observation is subject to error Nonuniformly distributed around the globe (satellite paths, densely-populated areas) (Graphics courtesy Yannick Trémolet)

Problem Formulation Data assimilation: defining the unknowns Currently in operational use at the European Centre for Medium-Range Weather Forecasts (ECMWF) We want values for an initial state, call it x 0 For a given x 0, we could integrate our atmospheric models forward to forecast the state of the atmosphere at N time points x i = M(x i 1 ), i = 1,..., N (x i : state of the atmosphere at time i) Observe the atmosphere at these N time points (y i : observed state at time i) y 1,..., y N Let x b (background state) be values at initial time point obtained from previous forecast carry over old information

Problem Formulation Data assimilation as an optimization problem Define the difference 2 x 0 x b 3 f (x 0 x 1 y 1 ) = 6 7 4. 5 x N y N and choose x 0 as the initial state most likely to have given the observed data 1 min f (x 0 ) 2 x 0 2 R = 1 (x 0 x b ) T (R b ) 1 (x 0 x b P )+ 1 N 1 2 2 i=0 (x i y i ) T (R i ) 1 (x i y i ) 1 2 f (x 0 ) 2 R 1 : distance measure between observed and expected behavior R = (R b, R 1,..., R N ): background and observation error covariance matrices (choice of these values is a separate, but important issue) In current forecasts, x 0 contains approximately 3 10 8 unknowns

Solution Techniques Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Solution Techniques Computational issues Data assimilation optimization problem: 1 min f (x 0 ) 2 x 0 2 R = 1 (x 0 x b ) T (R b ) 1 (x 0 x b P )+ 1 N 1 2 2 i=0 (x i y i ) T (R i ) 1 (x i y i ) Difficulties include: problem is very large ( x 0 3 10 8 ) objective is nonconvex (nonlinear operators M i ) exact derivative information not available solutions needed in real-time

Solution Techniques Current algorithm: nonlinear elimination Given a guess of the initial state x 0 1. Apply operators M i to compute the expected state at the N time points x i = M(x i 1 ), i = 1,..., N and evaluate the objective 1 2 f (x 0 ) 2 R 1 2. Compute a step d toward an improved solution a.... derive sensitivities of x 1,..., x N with respect to x 0 to form J(x 0 ) : Jacobian of f (x 0 ) (Note: can only be done inexactly) b.... solve the quadratic problem min d 1 2 f (x 0 ) + J(x 0 )d 2 R 1

Solution Techniques Illustration of solution process

Solution Techniques Recent advances: multilevel schemes We are interested in solving the subproblem min d q(d) = 1 2 f (x 0 ) + J(x 0 )d 2 R 1 Conjugate Gradients or Lanczos method is applicable, but not at high resolutions Define a restriction operator S such that ˆx i = Sx i, i = 0,..., N Solve min ˆq(ˆd) ˆd where ˆq is a lower-resolution analog of q Use a prolongation operator S + to map the computed step into the higher-resolution space x 0 x 0 + S +ˆd

Solution Techniques Future study: weak constraints Previous formulation assumes that model errors can be neglected However, we can lift this questionable assumption by imposing weak constraints x i M(x i 1 ), i = 1,..., N These approximate equalities can be imposed by creating a penalty term in the objective min 1 2 (x 0 x b ) T (R b ) 1 (x 0 x b ) + 1 2 P N i=0 (x i y i ) T (R i ) 1 (x i y i ) + 1 2 P N i=1 (x i M(x i 1 )) T (Q i ) 1 (x i M(x i 1 )) where Q i are model error covariances Note: state vectors x 1,..., x N become unknowns; problem dimension can increase from 3 10 8 to as large as 7 10 9!

Solution Techniques Future study: weak constraints The weak constraint formulation min 1 2 (x 0 x b ) T (R b ) 1 (x 0 x b ) + 1 2 P N i=0 (x i y i ) T (R i ) 1 (x i y i ) + 1 2 P N i=1 (x i M(x i 1 )) T (Q i ) 1 (x i M(x i 1 )) can be seen as a step toward constrained optimization 1 min (x 0 x b ) T (R b ) 1 (x 0 x b P ) + 1 N x 0,...,x N 2 2 i=0 (x i y i ) T (R i ) 1 (x i y i ) s.t. x i = M(x i 1 ), i = 1,..., N (Note: in solving the constrained formulation, constraints are satisfied only in the limit)

Solution Techniques Challenges in PDE-Constrained Optimization Optimization Method Nonlinear elimination Composite-step methods Full space methods DO versus OD Discretize-Optimize Optimize-Discretize Multilevel Schemes Compute steps at low resolution Optimize at low resolution Inexactness Efficiency Robustness Overview of important optimization questions

Solution Techniques Challenges in PDE-Constrained Optimization Optimization Method Nonlinear elimination Composite-step methods Full space methods DO versus OD Discretize-Optimize Optimize-Discretize Multilevel Schemes Compute steps at low resolution Optimize at low resolution Inexactness Efficiency Robustness Methods arising in the data assimilation problem

Solution Techniques Challenges in PDE-Constrained Optimization Optimization Method Nonlinear elimination Composite-step methods Full space methods DO versus OD Discretize-Optimize Optimize-Discretize Multilevel Schemes Compute steps at low resolution Optimize at low resolution Inexactness Efficiency Robustness Our contribution to the field...

Solution Techniques Computational savings with inexact methods Inexact step computations are... effective for unconstrained optimization 2 min x R n f (x)... effective for systems of nonlinear equations 3 F (x) = 0 or min x R n... necessary for many PDE-constrained problems F (x) 2 (Steihaug, 1983) 3 (Dembo, Eisenstat, Steihaug, 1982)

Background and Basics Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Background and Basics Equality Constrained Optimization Minimize an objective subject to mathematical equalities min f (x) x R n s.t. c(x) = 0 (where the constraints c(x) = 0 contain the PDE)

Background and Basics Characterizing Optimal Solutions Defining and then the derivatives L(x, λ) = f (x) + λ T c(x) g(x) = f (x) and A(x) = [ c i (x)], we have the first order optimality conditions» g(x) + A(x) T λ = 0 c(x)

Background and Basics Method of choice: Sequential Quadratic Programming (SQP) At a given iterate x k, formulate the quadratic subproblem min d R n f k + g T k d + 1 2 d T W k d s.t. c k + A k d = 0 for some symmetric positive definite W k ( xxl 2 k ) to compute a step, or, equivalently, solve the primal-dual system»»» Wk A T k dk gk + A T k λ = k A k 0 Measure progress toward a solution with the merit function δ k c k φ π(x) = f (x) + π c(x), π > 0

Background and Basics SQP: for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k A k 0 Set π k π k 1 Line search: endfor δ k φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) c k (It works!)

Background and Basics SQP: for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k A k 0 Set π k π k 1 Line search: endfor δ k φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) c k

Background and Basics SQP: for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k A k 0 Set π k π k 1 Line search: endfor δ k φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) c k

Background and Basics SQP: for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k A k 0 Set π k π k 1 Line search: endfor δ k φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) c k The step is good; Dφ πk (d k ) < 0 Algorithm is well-defined; α k > 0 π will stabilize reducing φ π drives search toward solution of optimization problem

Algorithm Overview Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Algorithm Overview Inexactness is necessary An SQP algorithm requires the exact solution of the system»»» Wk A T k dk gk + A T k λ = k A k 0 δ k However, for many PDE constrained problems we cannot form or factor this iteration matrix Alternatively, we can consider the application of an iterative solver which, during each inner iteration, yields»»»» Wk A T k dk gk + A T k λ = k ρk + A k 0 with residuals (ρ k, r k ) δ k Iterative methods only require mechanisms for computing products with W k and A k and its transpose, so the algorithm is matrix-free c k c k r k

Algorithm Overview Inexact SQP: Idea: terminate when (ρ k, r k ) is small for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k + A k 0 Set π k π k 1 Line search: endfor δ k c k» ρk φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) r k

Algorithm Overview Inexact SQP: Idea: terminate when (ρ k, r k ) is small for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k + A k 0 δ k c k» ρk r k For any level of inexactness: Set π k π k 1 Line search: φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) endfor

Algorithm Overview Inexact SQP: Idea: terminate when (ρ k, r k ) is small for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k + A k 0 Set π k π k 1 Line search: endfor δ k c k» ρk φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) r k For any level of inexactness: Step may be an ascent direction for φ π Penalty parameter may tend to

Algorithm Overview Inexact SQP: We don t know when to terminate the iterative solver for k = 1, 2,... Compute step:»»» Wk A T k dk gk + A T k λ = k + A k 0 Set π k π k 1 Line search: endfor δ k c k» ρk φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) r k For any level of inexactness: Step may be an ascent direction for φ π Penalty parameter may tend to

Algorithm Overview Main idea: inexactness based on models Modern optimization methods work with models For the merit function φ π(x) = f (x) + π c(x), define a model about x k : m π(d) = f k + gk T d + 1 d T W 2 k d + π c k + A k d, For a given d k, we can estimate the reduction in φ π by evaluating mred π(d k ) = m π(0) m π(d k ) = gk T d k 1 d T 2 k W k d k + π( c k r k ) (recall r k = c k + A k d k )

Algorithm Overview Termination Tests We may terminate the iteration on»»» Wk A T k dk gk + A T k λ = k + A k 0 if the component d k : δ k c k» ρk r k

Algorithm Overview Termination Tests We may terminate the iteration on»»» Wk A T k dk gk + A T k λ = k + A k 0 δ k if the component d k : (1) yields a sufficiently large reduction in the model of the merit function for the most recent penalty parameter π k 1 ; i.e. c k mred πk 1 (d k ) σπ k 1 c k» ρk r k

Algorithm Overview Termination Tests We may terminate the iteration on»»» Wk A T k dk gk + A T k λ = k + A k 0 δ k if the component d k : (2) yields a reduction in the linear model of the constraints; i.e., for π k > π k 1 c k mred πk (d k ) σπ k c k» ρk r k

Algorithm Overview An Inexact SQP Algorithm (Byrd, Curtis, and Nocedal, 2007) Given parameters 0 < ɛ, τ, σ, η < 1 and 0 < β Initialize x 0, λ 0, and π 1 > 0 for k = 0, 1, 2,..., until convergence Set π k = π k 1 Iteratively solve the system» Wk A T»» k dk gk + A = T k λ» k ρk + A k 0 δ k c k r k Terminate the inner iterations when mred πk (d k ) ρ k σπ k c k max{β c k, ɛ g k + A T k λ k } or r k ρ k Perform line search φ πk (x k + α k d k ) φ πk (x k ) + ηα k Dφ πk (d k ) Set (x k+1, λ k+1 ) (x k, λ k ) + α k (d k, δ k ) endfor ɛ c k β c k π k g T k d k + 1 2 dt k W kd k (1 τ)( c k r k )

Algorithm Overview Summary: Our contribution SQP is the algorithm of choice for large-scale constrained optimization Introducing inexactness is a difficult issue and naïve approaches can fail to ensure convergence We present two simple termination tests for the iterative solver, where the level of inexactness is based on models of a merit function The algorithm is globally convergent Numerical experiments are considered next...

Optimization Test Problems: Robustness Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Optimization Test Problems: Robustness Question: Is the algorithm robust? 44 problems from standard optimization test sets (CUTEr and COPS) Algorithm A: inexactness based on»» ρk κ gk + A T k λ k, 0 < κ < 1 r k Algorithm B: inexactness based on model reductions (Note: can have κ 1!) c k Algorithm Alg. A Alg. B κ 2 1 2 5 2 10 % Solved 45% 80% 86% 100%

PDE-Constrained Test Problems: Practicality Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

PDE-Constrained Test Problems: Practicality Question: Is the algorithm practical? 2 model inverse problems in PDE-constrained optimization 4 where 1 min Qz (y,z) R n 2 d 2 + γr(y y ref ) s.t. A(y)z = q objective = data fitting + regularization constraints = PDE Example applications: Elliptic PDE groundwater modeling, DC resistivity Parabolic PDE optical tomography, electromagnetic imaging 4 (Haber and Hanson, 2007)

PDE-Constrained Test Problems: Practicality Question: Is the algorithm practical? First problem (n = 8192, t = 4096) Elliptic PDE (groundwater modeling, DC resistivity)... many details... Solution obtained in 9 iterations Sample output: Note: κ k = (ρ k, r k ) / (g k + A T k λ k, c k ) k KKT Inner Iter. κ k T. Test 3 5.22 10 1 2 9.66 10 1 new π 4 1.11 10 1 10 9.39 10 1 curr. π 5 1.04 10 1 1 9.60 10 1 new π Overall, average κ k 5.40 10 1

PDE-Constrained Test Problems: Practicality Question: Is the algorithm practical? Second problem (n = 69632, t = 65536) Parabolic PDE (optical tomography, electromagnetic imaging)... many details... Solution obtained in 11 iterations Sample output: Note: κ k = (ρ k, r k ) / (g k + A T k λ k, c k ) k KKT Inner Iter. κ k T. Test 4 1.63 10 1 1 9.50 10 1 curr. π 5 1.55 10 1 2 9.55 10 1 curr. π 6 1.48 10 1 3 5.65 10 1 new π Overall, average κ k 4.49 10 1

Inexact SQP for nonconvex optimization Outline Motivating Example: Data Assimilation in Weather Forecasting Problem Formulation Solution Techniques Inexact SQP Methods for Equality Constrained Optimization Background and Basics Algorithm Overview Numerical Results Optimization Test Problems: Robustness PDE-Constrained Test Problems: Practicality Related and Future Work Inexact SQP for nonconvex optimization

Inexact SQP for nonconvex optimization Convexity assumption Step computation procedure min d R n f k + g T k d + 1 2 d T W k d s.t. c k + A k d = 0» Wk A T k A k 0»» dk gk + A T k λ = k δ k c k

Inexact SQP for nonconvex optimization Nonconvex optimization If W k is not convex in the null space of the constraints No guarantee of descent Step may be unbounded Without a factorization of the primal-dual matrix» Wk A T k A k 0 we may not know if the problem is convex or not We could always set W k to a simple positive definite matrix, but then we may be distorting second order information (if available)

Inexact SQP for nonconvex optimization First step: recognizing good steps If the problem is convex, then a solution to»»» Wk A T k dk gk + A T k λ = k A k 0 δ k c k (1) is a good step If the problem is nonconvex, then a solution to (1) may still be a good step Idea: characterize d k based on properties of the decomposition d k = u k + v k (Note: d k 2 = u k 2 + v k 2 ) where A k u k = 0 and v k lies in range(a T k )

Inexact SQP for nonconvex optimization Central claim A step is good if either termination test for the inexact SQP algorithm is satisfied and for some small θ > 0 we have Notice that (a) θ u k v k or (b) d T k W k d k θ u k 2 (a) implies that the step d k is sufficiently parallel to v k (b) implies that the curvature is sufficiently positive along d k

Inexact SQP for nonconvex optimization Estimating properties of the step Observing we find and v k A k v k / A k = A k d k / A k, θ d k A k d k / A k θ u k v k d T k W k d k θ( d k 2 A k d k 2 / A k 2 ) d T k W k d k θ u k 2

Inexact SQP for nonconvex optimization Proposed Algorithm Apply an iterative solver to the primal-dual system Stop if a termination test is satisfied and... step is sufficiently parallel to vk θ d k A k d k / A k... or curvature is sufficiently positive d T k W k d k θ ( d k 2 A k d k 2 / A k 2) If θ d k > A k d k / A k, then set W k W k to satisfy θ d k 2 A k d k 2 / A k 2 dk T W k d k and iterate on perturbed system

Conclusion We have: 1. Described an important and challenging sample problem in PDE-constrained optimization 2. Described a globally convergent algorithm in the framework of a powerful algorithm: sequential quadratic programming 3. The algorithm has shown to be robust and efficient for realistic applications 4. Described some ideas for extending the approach to nonconvex problems