Numerical Optimal Control Part 3: Function space methods
|
|
- Bernice Curtis
- 6 years ago
- Views:
Transcription
1 Numerical Optimal Control Part 3: Function space methods SADCO Summer School and Workshop on Optimal and Model Predictive Control OMPC 2013, Bayreuth Institute of Mathematics and Applied Computing Department of Aerospace Engineering Universita t der Bundeswehr Mu nchen (UniBw M) matthias.gerdts@unibw.de Fotos: nchen Magnus Manske (Panorama), Luidger (Theatinerkirche), Kurmis (Chin. Turm), Arad Mojtahedi (Olympiapark), Max-k (Deutsches Museum), Oliver Raupach (Friedensengel), Andreas Praefcke (Nationaltheater)
2 Schedule and Contents Time Topic 9:00-10:30 Introduction, overview, examples, indirect method 10:30-11:00 Coffee break 11:00-12:30 Discretization techniques, structure exploitation, calculation of gradients, extensions: sensitivity analysis, mixedinteger optimal control 12:30-14:00 Lunch break 14:00-15:30 Function space methods: Gradient and Newton type methods 15:30-16:00 Coffee break 16:00-17:30 Numerical experiments
3 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
4 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
5 Overview on Solution Methods DAE Optimal Control Problem Discretization: Approximation by finite dimensional problem Function space approach Direct Approach: Indirect approach based Indirect approach based Direct Approach: Methods for Methods for discretized on finite dimensional on infinite dimensional infinite dimensional optimization problem optimality system optimality system optimization problem Reduced approach (direct shooting) Full discretization (collocation) Methods for finite dimensional complementarity problems and Semi analytical methods (indirect method, boundary value problems) Reduced approach Full approach SQP methods variational inequalities (semismooth Newton, Josephy Newton, fixed point iteration, projection Methods for infinite dimensional complementarity problems and variational inequalities SQP methods Interior Point Methods methods) (semismooth Interior Point Methods Gradient methods Newton, Josephy Gradient methods Penalty methods Newton, fixed point Penalty methods Multiplier methods iteration, projection Multiplier methods Dynamic programming methods) Dynamic programming
6 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? What are the difficulties?
7 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem What are the difficulties?
8 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible What are the difficulties?
9 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible subtle requirements can get lost in discretizations (e.g. smoothing operator for semismooth methods) What are the difficulties?
10 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible subtle requirements can get lost in discretizations (e.g. smoothing operator for semismooth methods) methods can be very fast What are the difficulties?
11 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible subtle requirements can get lost in discretizations (e.g. smoothing operator for semismooth methods) methods can be very fast What are the difficulties? detailed functional analytic background necessary (cannot be expected in an industrial context)
12 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible subtle requirements can get lost in discretizations (e.g. smoothing operator for semismooth methods) methods can be very fast What are the difficulties? detailed functional analytic background necessary (cannot be expected in an industrial context) discretizations become necessary at lower level anyway; so, why not discretize right away?
13 Function Space Methods Paradigm Analyze and develop methods for some infinite dimensional optimization problem (e.g. an optimal control problem) of type Minimize J(z) subject to G(z) K, H(z) = 0 in the same Banach or Hilbert spaces where z, G and H live. Why is this useful? no immediate approximation error: algorithms work in the same spaces as the problem massive exploitation of structure possible subtle requirements can get lost in discretizations (e.g. smoothing operator for semismooth methods) methods can be very fast What are the difficulties? detailed functional analytic background necessary (cannot be expected in an industrial context) discretizations become necessary at lower level anyway; so, why not discretize right away? theoretical difficulties with, e.g., state constraints (multipliers are measures; how to handle them numerically?)
14 Functional Analysis Banach space: complete normed vector space (X, X )
15 Functional Analysis Banach space: complete normed vector space (X, X ) Hilbert space: complete vector space X with inner product, X X
16 Functional Analysis Banach space: complete normed vector space (X, X ) Hilbert space: complete vector space X with inner product, X X Dual space X of a vector space X is defined by X := {f : X R f linear and continuous}, f X = sup f (x) x =1 (X, X ) is a Banach space, if (X, X ) is a Banach space.
17 Functional Analysis Banach space: complete normed vector space (X, X ) Hilbert space: complete vector space X with inner product, X X Dual space X of a vector space X is defined by X := {f : X R f linear and continuous}, f X = sup f (x) x =1 (X, X ) is a Banach space, if (X, X ) is a Banach space. For Banach spaces X and Y, the Fréchet derivative of f : X Y at ˆx is a linear and continuous operator f (ˆx) : X Y with the property f (ˆx + h) f (ˆx) f (ˆx)h Y = o( h X ). The Fréchet derivative of f at ˆx in direction d is denoted by f (ˆx)d.
18 Functional Analysis Banach space: complete normed vector space (X, X ) Hilbert space: complete vector space X with inner product, X X Dual space X of a vector space X is defined by X := {f : X R f linear and continuous}, f X = sup f (x) x =1 (X, X ) is a Banach space, if (X, X ) is a Banach space. For Banach spaces X and Y, the Fréchet derivative of f : X Y at ˆx is a linear and continuous operator f (ˆx) : X Y with the property f (ˆx + h) f (ˆx) f (ˆx)h Y = o( h X ). The Fréchet derivative of f at ˆx in direction d is denoted by f (ˆx)d. The Fréchet derivative of a function f : X U Y at (ˆx, û) is given by f (ˆx, û)(x, u) = f x (ˆx)x + f u(û)u, where f x and f u are the partial derivatives of f w.r.t. x and u, respectively.
19 Functional Analysis Lebesgue spaces For 1 p the Lebesgue spaces are defined by L p (I, R n ) := {f : I R n f p < } with ( f p := ) 1/p f (t) p I f := ess sup f (t). t I (1 p < ), Properties: L 2 (I, R n ) is a Hilbert space with inner product f, g = I f (t) g(t)dt L (I, R n ) is the dual space of L 1 (I, R n ), but not vice versa L p (I, R n ) and L q (I, R n ) are dual to each other, if 1/p + 1/q = 1
20 Functional Analysis Sobolev spaces For 1 p and q N the Sobolev spaces are defined by W q,p (I, R n ) := {f : I R n f q,p < } with f q,p := 1/p q f (j) p j=0 (1 p ) Properties: W 1,2 ([t 0, t f ], R n ) is a Hilbert space with inner product f, g = f (t 0 ) g(t 0 ) + f (t) g(t) dt t 0 W 1,1 (I, R n ) is the space of absolutely continuous functions, i.e functions satisfy t f (t) = f (t 0 ) + f (τ )dτ in I = [t 0, t f ]. t 0 tf
21 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
22 Infinite Optimization Problem Throughout we restrict the discussion to the following class of optimization problems. Problem (Infinite Optimization Problem (NLP)) Given: Banach spaces (Z, Z ), (Y, Y ) mappings J : Z R, H : Z Y convex set S Z Minimize J(z) subject to z S, H(z) = 0
23 Infinite Optimization Problem Theorem (KKT Conditions) Assumptions: ẑ is a local minimum of NLP J is Fréchet-differentiable at ẑ, H is continuously Fréchet-differentiable at ẑ S is convex with non-empty interior Mangasarian-Fromowitz Constraint Qualification (MFCQ): H (ẑ) is surjective and there exists d int(s {ẑ}) with H (ẑ)(d) = 0 Then there exists a multiplier λ Y such that J (ẑ)(z ẑ) + λ (H (ẑ)(z ẑ)) 0 z S For a proof see [1, Theorems 3.1,4.1]. [1] S. Kurcyusz. On the Existence and Nonexistence of Lagrange Multipliers in Banach Spaces. Journal of Optimization Theory and Applications, 20(1):81 110, 1976.
24 Infinite Optimization Problem Lagrange function: Special cases: L(z, λ ) := J(z) + λ (H(z))
25 Infinite Optimization Problem Lagrange function: Special cases: If S = Z, then or equivalently L(z, λ ) := J(z) + λ (H(z)) J (ẑ)(z) + λ (H (ẑ)(z)) = 0 z Z L z(ẑ, λ ) = 0
26 Infinite Optimization Problem Lagrange function: Special cases: If S = Z, then or equivalently If H is not present, then L(z, λ ) := J(z) + λ (H(z)) J (ẑ)(z) + λ (H (ẑ)(z)) = 0 z Z L z(ẑ, λ ) = 0 J (ẑ)(z ẑ) 0 z S If Z is a Hilbert space, then this condition is equivalent with the nonsmooth equation ẑ = Π S (ẑ αj (ẑ) ) (α > 0, Π S projection onto S)
27 Infinite Optimization Problem Lagrange function: Special cases: If S = Z, then or equivalently If H is not present, then L(z, λ ) := J(z) + λ (H(z)) J (ẑ)(z) + λ (H (ẑ)(z)) = 0 z Z L z(ẑ, λ ) = 0 J (ẑ)(z ẑ) 0 z S If Z is a Hilbert space, then this condition is equivalent with the nonsmooth equation ẑ = Π S (ẑ αj (ẑ) ) (α > 0, Π S projection onto S) If S = Z and H is not present, then J (ẑ)(z ẑ) = 0 z S
28 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
29 Adjoint Formalism Equality constrained NLP (EQ-NLP) Minimize J(x, u) subject to Ax = Bu. A R n n, B R n m, J : R n R m R differentiable
30 Adjoint Formalism Equality constrained NLP (EQ-NLP) Minimize J(x, u) subject to Ax = Bu. A R n n, B R n m, J : R n R m R differentiable Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su S : R m R n, u x = Su is called solution operator.
31 Adjoint Formalism Equality constrained NLP (EQ-NLP) Minimize J(x, u) subject to Ax = Bu. A R n n, B R n m, J : R n R m R differentiable Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su S : R m R n, u x = Su is called solution operator. Reduced objective functional: J(x, u) = J(Su, u) =: j(u), j : R m R.
32 Adjoint Formalism Equality constrained NLP (EQ-NLP) Minimize J(x, u) subject to Ax = Bu. A R n n, B R n m, J : R n R m R differentiable Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su S : R m R n, u x = Su is called solution operator. Reduced objective functional: J(x, u) = J(Su, u) =: j(u), j : R m R. Reduced NLP (R-NLP) Minimize j(u) = J(Su, u) w.r.t. u R m.
33 Adjoint Formalism Task Compute gradient of j at a given û, for instance in a gradient based optimization method or for the evaluation of necessary optimality conditions.
34 Adjoint Formalism Task Compute gradient of j at a given û, for instance in a gradient based optimization method or for the evaluation of necessary optimality conditions. Differentiation yields j (û) = J x (Sû, û)s + J u (Sû, û) = J x (Sû, û)a 1 B + J u (Sû, û)
35 Adjoint Formalism Task Compute gradient of j at a given û, for instance in a gradient based optimization method or for the evaluation of necessary optimality conditions. Differentiation yields j (û) = J x (Sû, û)s + J u (Sû, û) = J x (Sû, û)a 1 B + J u (Sû, û) Computation of A 1 is expensive. Try to avoid it!
36 Adjoint Formalism Task Compute gradient of j at a given û, for instance in a gradient based optimization method or for the evaluation of necessary optimality conditions. Differentiation yields j (û) = J x (Sû, û)s + J u (Sû, û) = J x (Sû, û)a 1 B + J u (Sû, û) Computation of A 1 is expensive. Try to avoid it! To this end define adjoint vector λ by λ := J x (Sû, û)a 1
37 Adjoint Formalism Adjoint equation A λ = x J(Sû, û) The gradient of j at û reads j (û) = λ B + J u (Sû, û)
38 Adjoint Formalism Adjoint equation A λ = x J(Sû, û) The gradient of j at û reads j (û) = λ B + J u (Sû, û) Connection to KKT conditions: Lagrange function of EQ-NLP L(x, u, λ) := J(x, u) + λ (Bu Ax) If λ solves adjoint equation, then L x (ˆx, û, λ) = J x (ˆx, û) λ A = 0
39 Adjoint Formalism Adjoint equation A λ = x J(Sû, û) The gradient of j at û reads j (û) = λ B + J u (Sû, û) Connection to KKT conditions: Lagrange function of EQ-NLP L(x, u, λ) := J(x, u) + λ (Bu Ax) If λ solves adjoint equation, then L x (ˆx, û, λ) = J x (ˆx, û) λ A = 0 This choice of λ automatically satisfies one part of the KKT conditions.
40 Adjoint Formalism Adjoint equation A λ = x J(Sû, û) The gradient of j at û reads j (û) = λ B + J u (Sû, û) Connection to KKT conditions: Lagrange function of EQ-NLP If λ solves adjoint equation, then L(x, u, λ) := J(x, u) + λ (Bu Ax) L x (ˆx, û, λ) = J x (ˆx, û) λ A = 0 This choice of λ automatically satisfies one part of the KKT conditions. The second part 0 = L u(ˆx, û, λ) = J u(ˆx, û) + λ B = j (û) is only satisfied at a stationary point û of the reduced problem R-NLP!
41 Adjoint Formalism Let X and Y be Banach spaces and X and Y their topological dual spaces. Adjoint operator Let A : X Y be a bounded linear operator. The operator A : Y X with the property (A y )(x) = y (Ax) x X, y Y is called adjoint operator (it is a bounded linear operator). Notion: (A y, x) = (y, Ax) dual pairing Equality constrained NLP (EQ-NLP) Minimize J(x, u) subject to Ax = Bu. A : X Y, B : U Y bounded linear operators, J : X U R Fréchet differentiable, X, U, Y Banach spaces
42 Adjoint Formalism Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su with solution operator S : U X, u x = Su.
43 Adjoint Formalism Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su with solution operator S : U X, u x = Su. Reduced objective functional: J(x, u) = J(Su, u) =: j(u), j : U R.
44 Adjoint Formalism Solution operator: Let A be nonsingular. Then: Ax = Bu x = A 1 Bu =: Su with solution operator S : U X, u x = Su. Reduced objective functional: J(x, u) = J(Su, u) =: j(u), j : U R. Reduced NLP (R-NLP) Minimize j(u) = J(Su, u) w.r.t. u U.
45 Adjoint Formalism Differentiation at û in direction u yields j (û)u = J x (Sû, û)su + J u (Sû, û)u = J x (Sû, û)a 1 Bu + J u(sû, û)u
46 Adjoint Formalism Differentiation at û in direction u yields j (û)u = J x (Sû, û)su + J u (Sû, û)u = J x (Sû, û)a 1 Bu + J u(sû, û)u Computation of A 1 is expensive. Try to avoid it!
47 Adjoint Formalism Differentiation at û in direction u yields j (û)u = J x (Sû, û)su + J u (Sû, û)u = J x (Sû, û)a 1 Bu + J u(sû, û)u Computation of A 1 is expensive. Try to avoid it! Define adjoint λ Y by y := J x (Sû, û)a 1 y (Ax) }{{} = J x (Sû, û)x =(A y )x Adjoint equation A y = J x (Sû, û) (operator equation in X )
48 Adjoint Formalism Differentiation at û in direction u yields j (û)u = J x (Sû, û)su + J u (Sû, û)u = J x (Sû, û)a 1 Bu + J u(sû, û)u Computation of A 1 is expensive. Try to avoid it! Define adjoint λ Y by y := J x (Sû, û)a 1 y (Ax) }{{} = J x (Sû, û)x =(A y )x Adjoint equation A y = J x (Sû, û) (operator equation in X ) The gradient of j at û in direction u reads j (û)u = y Bu + J u (Sû, û)u
49 Adjoint Formalism Connection to KKT conditions: Lagrange function of EQ-NLP L(x, u, y ) := J(x, u) + y (Bu Ax)
50 Adjoint Formalism Connection to KKT conditions: Lagrange function of EQ-NLP L(x, u, y ) := J(x, u) + y (Bu Ax) y solves adjoint equation = L x (ˆx, û, y )x = J x (ˆx, û)x y (Ax) = J x (ˆx, û)x (A y )x = 0 This choice of y automatically satisfies one part of the KKT conditions.
51 Adjoint Formalism Connection to KKT conditions: Lagrange function of EQ-NLP L(x, u, y ) := J(x, u) + y (Bu Ax) y solves adjoint equation = L x (ˆx, û, y )x = J x (ˆx, û)x y (Ax) = J x (ˆx, û)x (A y )x = 0 This choice of y automatically satisfies one part of the KKT conditions. The second part 0 = L u(ˆx, û, y ) = J u(ˆx, û) + y B = j (û) is only satisfied at a stationary point û of the reduced problem R-NLP!
52 How to compute the adjoint operator? Theorem (Riesz) Given: Hilbert space X, inner product,, dual space X. For every f X there exists a unique f X such that f X = f X and f (x) = f, x x X. Example (linear ODE) Given: I := [0, T ], linear mapping A : W 1,2 (I, R n ) =: X Y := L 2 (I, R n ) R n, ( ) x ( ) A( )x( ) (Ax)( ) =. x(0) X and Y are Hilbert spaces. According to Riesz theorem for y Y and x X there exist (λ, σ) Y and µ X with y (h, k) = λ(t) h(t) dt + σ k (h, k) Y I x (x) = µ(0) x(0) + µ (t) x (t) dt x X I
53 How to compute the adjoint operator? Example (linear ODE, continued) We intend to identify λ, σ, and µ such that y (Ax) = (A y )(x) holds for all x X and y Y. Note, that A y X. By partial integration we obtain y (Ax) = I λ(t) ( x (t) A(t)x(t) ) dt + σ x(0)
54 How to compute the adjoint operator? Example (linear ODE, continued) We intend to identify λ, σ, and µ such that y (Ax) = (A y )(x) holds for all x X and y Y. Note, that A y X. By partial integration we obtain y (Ax) = I λ(t) ( x (t) A(t)x(t) ) dt + σ x(0) [( T ) ] T = λ(s) A(s)ds x(t) t 0 + I ( λ(t) + σ x(0) T ) λ(s) A(s)ds x (t) dt t
55 How to compute the adjoint operator? Example (linear ODE, continued) We intend to identify λ, σ, and µ such that y (Ax) = (A y )(x) holds for all x X and y Y. Note, that A y X. By partial integration we obtain y (Ax) = I λ(t) ( x (t) A(t)x(t) ) dt + σ x(0) [( T ) ] T = λ(s) A(s)ds x(t) t 0 = + σ x(0) ( T ) + λ(t) λ(s) A(s)ds x (t) dt I t ( T ) σ λ(s) A(s)ds x(0) + 0 I }{{} =µ(0) ( T λ(t) t ) λ(s) A(s)ds } {{ } =µ (t) x (t) dt
56 How to compute the adjoint operator? Example (linear ODE, continued) We intend to identify λ, σ, and µ such that y (Ax) = (A y )(x) holds for all x X and y Y. Note, that A y X. By partial integration we obtain y (Ax) = I λ(t) ( x (t) A(t)x(t) ) dt + σ x(0) [( T ) ] T = λ(s) A(s)ds x(t) t 0 = + σ x(0) ( T ) + λ(t) λ(s) A(s)ds x (t) dt I t ( T ) σ λ(s) A(s)ds x(0) + 0 I }{{} =µ(0) = (A y )(x). ( T λ(t) t ) λ(s) A(s)ds } {{ } =µ (t) x (t) dt
57 How to compute the adjoint operator? Example (linear ODE, continued) We intend to identify λ, σ, and µ such that y (Ax) = (A y )(x) holds for all x X and y Y. Note, that A y X. By partial integration we obtain y (Ax) = I λ(t) ( x (t) A(t)x(t) ) dt + σ x(0) [( T ) ] T = λ(s) A(s)ds x(t) t 0 = + σ x(0) ( T ) + λ(t) λ(s) A(s)ds x (t) dt I t ( T ) σ λ(s) A(s)ds x(0) + 0 I }{{} =µ(0) = (A y )(x). ( T λ(t) t ) λ(s) A(s)ds } {{ } =µ (t) The latter equation defines the adjoint operator since y (Ax) = (A y )(x) holds for all x X, y Y. x (t) dt
58 How to compute the adjoint equation? Example (OCP and adjoint equation) Optimal control problem: (assumption: ϕ Fréchet differentiable) ( x (t) A(t)x(t) ) ( B(t)u(t) ) Minimize J(x, u) := ϕ(x(t )) s.t. x(0) = 0 } {{ } } {{ } =A(x)(t) =B(u)(t) Adjoint equation: For every x W 1,2 (I, R n ) we have 0 = (A y )(x) J (ˆx, û)(x)
59 How to compute the adjoint equation? Example (OCP and adjoint equation) Optimal control problem: (assumption: ϕ Fréchet differentiable) ( x (t) A(t)x(t) ) ( B(t)u(t) ) Minimize J(x, u) := ϕ(x(t )) s.t. x(0) = 0 } {{ } } {{ } Adjoint equation: For every x W 1,2 (I, R n ) we have =A(x)(t) =B(u)(t) 0 = (A y )(x) J (ˆx, û)(x) ( T ) ( T ) = σ λ(s) A(s)ds x(0) + λ(t) λ(s) A(s)ds x (t) dt 0 I t ϕ (ˆx(T ))x(t )
60 How to compute the adjoint equation? Example (OCP and adjoint equation) Optimal control problem: (assumption: ϕ Fréchet differentiable) ( x (t) A(t)x(t) ) ( B(t)u(t) ) Minimize J(x, u) := ϕ(x(t )) s.t. x(0) = 0 } {{ } } {{ } Adjoint equation: For every x W 1,2 (I, R n ) we have =A(x)(t) =B(u)(t) 0 = (A y )(x) J (ˆx, û)(x) ( T ) ( T ) = σ λ(s) A(s)ds x(0) + λ(t) λ(s) A(s)ds x (t) dt 0 I t ϕ (ˆx(T ))x(t ) Application of variation lemma (DuBois-Reymond) yields and T λ(t) = λ(t ) + λ(s) A(s)ds t λ(0) = σ, λ(t ) = ϕ (ˆx(T )).
61 How to compute the adjoint equation? Example (OCP and adjoint equation, continued) Adjoint equation: λ (t) = A(t) λ(t), λ(t ) = ϕ(ˆx(t )). Gradient of reduced objective functional j(u) = J(x(u), u): j (û)(u) = y (Bu) + J u (ˆx, û)u = λ(t) B(t)u(t) dt + σ 0 I = λ(t) B(t)u(t) dt (u L 2 (I, R m )) I Stationary point: 0 = j (û)(u) = λ(t) B(t)u(t) dt u L 2 (I, R m ) I implies 0 = B(t) λ(t) a.e. in I
62 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
63 Gradient Method in Finite Dimensions Unconstrained minimization problem J : R n R continuously differentiable Minimize J(u) w.r.t. u R n
64 Gradient Method in Finite Dimensions Unconstrained minimization problem Minimize J(u) w.r.t. u R n J : R n R continuously differentiable Gradient Method for Finite Dimensional Problems (0) Let u (0) R n, β (0, 1), σ (0, 1), and k := 0.
65 Gradient Method in Finite Dimensions Unconstrained minimization problem Minimize J(u) w.r.t. u R n J : R n R continuously differentiable Gradient Method for Finite Dimensional Problems (0) Let u (0) R n, β (0, 1), σ (0, 1), and k := 0. (1) Compute d (k) := J(u (k) ).
66 Gradient Method in Finite Dimensions Unconstrained minimization problem Minimize J(u) w.r.t. u R n J : R n R continuously differentiable Gradient Method for Finite Dimensional Problems (0) Let u (0) R n, β (0, 1), σ (0, 1), and k := 0. (1) Compute d (k) := J(u (k) ). (2) If d (k) 0, STOP.
67 Gradient Method in Finite Dimensions Unconstrained minimization problem Minimize J(u) w.r.t. u R n J : R n R continuously differentiable Gradient Method for Finite Dimensional Problems (0) Let u (0) R n, β (0, 1), σ (0, 1), and k := 0. (1) Compute d (k) := J(u (k) ). (2) If d (k) 0, STOP. (3) Perform line-search: Find smallest j {0, 1, 2,...} with and set α k := β j. J(u (k) + β j d (k) ) J(u (k) ) σβ j J(u (k) ) 2 2
68 Gradient Method in Finite Dimensions Unconstrained minimization problem Minimize J(u) w.r.t. u R n J : R n R continuously differentiable Gradient Method for Finite Dimensional Problems (0) Let u (0) R n, β (0, 1), σ (0, 1), and k := 0. (1) Compute d (k) := J(u (k) ). (2) If d (k) 0, STOP. (3) Perform line-search: Find smallest j {0, 1, 2,...} with and set α k := β j. J(u (k) + β j d (k) ) J(u (k) ) σβ j J(u (k) ) 2 2 (4) Set u (k+1) := u (k) + α k d (k), k := k + 1, and go to (1).
69 Gradient Method Pro s: Con s: requires only first derivatives easy to implement global convergence achieved by Armijo linesearch extension: projected gradient method for simple constraints only linear convergence rate only unconstrained minimization (except simple bounds)
70 Gradient Method for Optimal Control Problems Numerical Optimal Control Part 3: Function space methods Optimal control problem (OCP) Given: I := [t 0, t f ], x R nx, continuously differentiable functions ϕ : R nx R, f 0 : R nx R nu R, f : R nx R nu R nx. Minimize Γ(x, u) := ϕ(x(t f )) + f 0 (x(t), u(t))dt I w.r.t. x W 1, (I, R nx ), u L (I, R nu ) subject to the constraints x (t) = f (x(t), u(t)) a.e. in I, x(t 0 ) = x. Define X := W 1, (I, R nx ) and U := L (I, R nu ).
71 Gradient Method for Optimal Control Problems Numerical Optimal Control Part 3: Function space methods Assumptions: (existence of solution operator) For every u U the initial value problem x (t) = f (x(t), u(t)) a.e. in I, x(t 0 ) = x has a unique solution x = S(u) X. The solution mapping S : U X is continuously Fréchet-differentiable. Reduced optimal control problem (R-OCP) Minimize J(u) := Γ(S(u), u) subject to u U. Gradient method requires gradient of reduced objective function J at some û U.
72 Gradient Method for Optimal Control Problems Numerical Optimal Control Part 3: Function space methods Assumptions: (existence of solution operator) For every u U the initial value problem x (t) = f (x(t), u(t)) a.e. in I, x(t 0 ) = x has a unique solution x = S(u) X. The solution mapping S : U X is continuously Fréchet-differentiable. Reduced optimal control problem (R-OCP) Minimize J(u) := Γ(S(u), u) subject to u U. Gradient method requires gradient of reduced objective function J at some û U. How does the gradient look like?
73 Computation of Gradient In R n : J(û) := J (u) = J u 1 (û). J u n (û), J (û)(u) = J(û) u = J(û), u R n R n,
74 Computation of Gradient In R n : J(û) := J (u) = J u 1 (û). J u n (û), J (û)(u) = J(û) u = J(û), u R n R n, In a Hilbert space setting, i.e J : U R, U Hilbert space: J (û) U Riesz = η(û) U : J (û)(u) = η(û), u U U Hence, J(û) := η(û) U is the gradient of J at û.
75 Computation of Gradient In R n : J(û) := J (u) = J u 1 (û). J u n (û), J (û)(u) = J(û) u = J(û), u R n R n, In a Hilbert space setting, i.e J : U R, U Hilbert space: J (û) U Riesz = η(û) U : J (û)(u) = η(û), u U U Hence, J(û) := η(û) U is the gradient of J at û. But, in our case U = L (I, R nu ) is not a Hilbert space and hence the functional J (û) U does not a priori have such a nice representation as in the above cases.
76 Computation of Gradient In R n : J(û) := J (u) = J u 1 (û). J u n (û), J (û)(u) = J(û) u = J(û), u R n R n, In a Hilbert space setting, i.e J : U R, U Hilbert space: J (û) U Riesz = η(û) U : J (û)(u) = η(û), u U U Hence, J(û) := η(û) U is the gradient of J at û. But, in our case U = L (I, R nu ) is not a Hilbert space and hence the functional J (û) U does not a priori have such a nice representation as in the above cases. How to define the gradient in this case? use formal Lagrange technique (see also metric gradient, see [1]) [1] M. Golomb and R. A. Tapia. The metric gradient in normed linear spaces. Numerische Mathematik, 20: , 1972.
77 Computation of Gradient by Formal Lagrange Technique Hamilton function: H(x, u, λ) := f 0 (x, u) + λ f (x, u) Auxiliary functional: (û, ˆx = S(û) given; λ X to be specified later) J(û) := J(û) + λ, f (ˆx, û) ˆx L 2 (I,R nx ) L2 (I,R nx )
78 Computation of Gradient by Formal Lagrange Technique Hamilton function: H(x, u, λ) := f 0 (x, u) + λ f (x, u) Auxiliary functional: (û, ˆx = S(û) given; λ X to be specified later) J(û) := J(û) + λ, f (ˆx, û) ˆx L 2 (I,R nx ) L2 (I,R nx ) = ϕ(ˆx(t f )) + H(ˆx(t), û(t), λ(t)) λ(t) ˆx (t) dt I
79 Computation of Gradient by Formal Lagrange Technique Hamilton function: H(x, u, λ) := f 0 (x, u) + λ f (x, u) Auxiliary functional: (û, ˆx = S(û) given; λ X to be specified later) J(û) := J(û) + λ, f (ˆx, û) ˆx L 2 (I,R nx ) L2 (I,R nx ) = ϕ(ˆx(t f )) + H(ˆx(t), û(t), λ(t)) λ(t) ˆx (t) dt I [ p.i. = ϕ(ˆx(t f )) λ(t) ˆx(t) ] tf + H(ˆx(t), û(t), λ(t)) + λ (t) ˆx(t) dt. t 0 I
80 Computation of Gradient by Formal Lagrange Technique Hamilton function: H(x, u, λ) := f 0 (x, u) + λ f (x, u) Auxiliary functional: (û, ˆx = S(û) given; λ X to be specified later) J(û) := J(û) + λ, f (ˆx, û) ˆx L 2 (I,R nx ) L2 (I,R nx ) = ϕ(ˆx(t f )) + H(ˆx(t), û(t), λ(t)) λ(t) ˆx (t) dt I [ p.i. = ϕ(ˆx(t f )) λ(t) ˆx(t) ] tf + H(ˆx(t), û(t), λ(t)) + λ (t) ˆx(t) dt. t 0 I Fréchet derivative: (exploit S (û)(t 0 ) = 0) J (û)(u) = ( ϕ (ˆx(t f )) λ(t f ) ) S (û)(t f ) ( + H x [t] + λ (t) ) S (û)(t) + H u [t]u(t) dt. I Computation of S (û) is expensive! Eliminate red terms:
81 Computation of Gradient by Formal Lagrange Technique Hamilton function: H(x, u, λ) := f 0 (x, u) + λ f (x, u) Auxiliary functional: (û, ˆx = S(û) given; λ X to be specified later) J(û) := J(û) + λ, f (ˆx, û) ˆx L 2 (I,R nx ) L2 (I,R nx ) = ϕ(ˆx(t f )) + H(ˆx(t), û(t), λ(t)) λ(t) ˆx (t) dt I [ p.i. = ϕ(ˆx(t f )) λ(t) ˆx(t) ] tf + H(ˆx(t), û(t), λ(t)) + λ (t) ˆx(t) dt. t 0 I Fréchet derivative: (exploit S (û)(t 0 ) = 0) J (û)(u) = ( ϕ (ˆx(t f )) λ(t f ) ) S (û)(t f ) ( + H x [t] + λ (t) ) S (û)(t) + H u [t]u(t) dt. I Computation of S (û) is expensive! Eliminate red terms: } λ (t) = H x (ˆx(t), û(t), λ(t)) λ(t f ) = ϕ (ˆx(t f )) adjoint ODE
82 Computation of Gradient Gradient form : J (û)(u) = I H u [t]u(t)dt = uh, u L 2 (I,R nu ) L2 (I,Rnu ). Theorem (steepest descent) The direction solves d(t) := 1 H u 2 H u [t] Minimize J (û)(u) w.r.t. u subject to u 2 = 1. Proof: For every u with u 2 = 1 we have by the Schwarz inequality, J (û)(u) H u 2 u 2 = H u 2 and for d we have J (û)(d) = H u 2.
83 Computation of Gradient Relation between J (û) and J (û), for a proof see [5, Chapter 8]: Theorem Let û U and ˆx = S(û) be given. Let λ satisfy the adjoint ODE λ = H x (ˆx, û, λ), λ(t f ) = ϕ (ˆx(t f )). Then: J (û)(u) = J (û)(u) u U. Owing to the theorem we may define the gradient of J at û as follows. Definition (Gradient of reduced objective functional) Let û U and ˆx = S(û) be given. Let λ satisfy the adjoint ODE. The gradient J(û) U of J at û is defined by J(û)( ) := H u (ˆx( ), û( ), λ( )).
84 Gradient Method Gradient method for R-OCP (0) Choose u (0) U, β (0, 1), σ (0, 1), and set k := 0. (1) Solve the ODE x = f (x, u (k) ), x(t 0 ) = x, and the adjoint ODE λ = H x (x, u (k), λ), λ(t f ) = ϕ (x(t f )). Denote the solution by x (k) = S(u (k) ) and λ (k). (2) If H u 2 0, STOP. (3) Set d (k) (t) := H u(x (k) (t), u (k) (t), λ (k) (t)). (4) Perform an Armijo line-search: Find smallest j {0, 1, 2,...} with and set α k := β j. J(u (k) + β j d (k) ) J(u (k) ) σβ j H u (x (k), u (k), λ (k) ) 2 2 (5) Set u (k+1) := u (k) + α k d (k), k := k + 1, and go to (1).
85 Gradient Method Theorem (Convergence) Suppose that the gradient method does not terminate. Let u be an accumulation point of the sequence {u (k) } k N generated by the gradient method and x := S(u ). Then: J(u ) 2 = 0.
86 Gradient Method Examples Example Minimize x 2 (1) subject to the constraints x 1 (t) = x 1(t) + 3u(t), x 1 (0) = 2, x 2 (t) = 1 ( x 1 (t) 2 + u(t) 2), 2 x 2 (0) = 0. Output of gradient method: (u (0) 0, β = 0.9, σ = 0.1, symplectic Euler, N = 100) k α k J(u (k) ) H u H u E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E-12..
87 Gradient Method Examples Example (continued) Some iterates (red) and converged solution (blue): x1(t) state x time t [s] u(t) control u time t [s] lambda1(t) adjoint lambda time t [s]
88 Gradient Method Examples Example Minimize x 2 (1) + 2.5(x 1 (1) 1) 2 subject to the constraints x 1 (t) = u(t) 15 exp( 2t), x 1(0) = 4, x 2 (t) = 1 ( u(t) 2 + x 1 (t) 3), 2 x 2 (0) = 0. Output of gradient method: (u (0) 0, β = 0.9, σ = 0.1, symplectic Euler, N = 100) k α k J(u (k) ) H u H u E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E-12..
89 Gradient Method Examples Example (continued) Some iterates (red) and converged solution (blue): 4 state x1 6 control u x1(t) 1 0 u(t) time t [s] time t [s] 15 adjoint lambda lambda1(t) time t [s]
90 Projected Gradient Method We add simple constraints u U with a convex set U U to the reduced problem. Reduced optimal control problem (R-OCP) Minimize J(u) := Γ(S(u), u) subject to u U U. Assumptions: the projection Π U : U U is easy to compute
91 Projected Gradient Method We add simple constraints u U with a convex set U U to the reduced problem. Reduced optimal control problem (R-OCP) Minimize J(u) := Γ(S(u), u) subject to u U U. Assumptions: the projection Π U : U U is easy to compute Example: For box constraints U = {u R a u b} the projection computes to a, if u < a, Π U (u) = max{a, min{b, u}} = u, if a u b, b, if u > b
92 Projected Gradient Method We add simple constraints u U with a convex set U U to the reduced problem. Reduced optimal control problem (R-OCP) Minimize J(u) := Γ(S(u), u) subject to u U U. Assumptions: the projection Π U : U U is easy to compute Example: For box constraints U = {u R a u b} the projection computes to a, if u < a, Π U (u) = max{a, min{b, u}} = u, if a u b, b, if u > b Optimality: J (û)(u û) 0 u U û = Π U (û αj (û) ) (α > 0)
93 Projected Gradient Method The projected gradient method requires a feasible initial u (0) and differs from the gradient method in one of the following two components: Version 1: In iteration k compute ũ (k) := Π U ( u (k) + d (k)) and use the direction d (k) := ũ (k) u (k) instead of d (k) in steps (4) and (5), i.e. ( J u (k) + β j (k) d ) J(u (k) ) + σβ j J (u (k) ) d (k), u (k+1) = u (k) (k) + α k d
94 Projected Gradient Method The projected gradient method requires a feasible initial u (0) and differs from the gradient method in one of the following two components: Version 1: In iteration k compute ũ (k) := Π U ( u (k) + d (k)) and use the direction d (k) := ũ (k) u (k) instead of d (k) in steps (4) and (5), i.e. ( J u (k) + β j (k) d ) J(u (k) ) + σβ j J (u (k) ) d (k), u (k+1) = u (k) (k) + α k d Version 2: In iteration k use the projection within the Armijo linesearch in step (4), i.e. ( J (Π U u (k) + β j d (k))) J(u (k) ) + σβ j J (u (k) )d (k) and set u (k+1) := Π U ( u (k) + α k d (k)) in step (5)
95 Projected Gradient Method Examples Example Minimize x 2 (1) + 2.5(x 1 (1) 1) 2 subject to the constraints x 1 (t) = u(t) 15 exp( 2t), x 1(0) = 4, x 2 (t) = 1 ( u(t) 2 + x 1 (t) 3), 2 x 2 (0) = 0, u(t) U := {u R 1 u 3}. Output of projected gradient method (version 1): (u (0) 1, β = 0.9, σ = 0.1, symplectic Euler, N = 100) k α k J(u (k) ) u (k) Π U (u (k) J (u (k) )) J (u (k) )d (k) E E E E E E E E E E E E E E E E E E E E E E E E E E E E-13..
96 Projected Gradient Method Examples Example (continued) Some iterates (red) and converged solution (blue): 4 state x1 3 control u x1(t) 1 u(t) time t [s] time t [s] lambda1(t) adjoint lambda time t [s]
97 Gradient Method for the Discretized Problem Numerical Optimal Control Part 3: Function space methods Instead of applying the function space gradient method to OCP, we could first discretize OCP and apply the standard gradient method to the discretized problem. Problem (Discretized optimal control problem (D-OCP)) Given: x R nx, grid G N, and continuously differentiable functions defined by Minimize ϕ : R nx R, f 0 : R nx R nu R, f : R nx R nu R nx, G N := {t i t i = t 0 + ih, i = 0, 1,..., N}, h = (t f t 0 )/N, N N. N 1 Γ N (x, y, u) := ϕ(x N ) + h f 0 (x i, u i ) w.r.t. x = (x 0, x 1,..., x N ) R (N+1)nx, u = (u 0,..., u N 1 ) R Nnu subject to the constraints i=0 x i+1 x i h = f (x i, u i ) i = 0,..., N 1, x 0 = x.
98 Gradient Method for the Discretized Problem Numerical Optimal Control Part 3: Function space methods Denote by S : R Nnu R (N+1)nx, u x = S(u), the solution operator, that maps the control input u to the solution x of the discrete dynamics. Problem (Reduced Discretized Problem (RD-OCP)) Minimize J N (u) := Γ N (S(u), u) w.r.t. u R Nnu.
99 Gradient Method for the Discretized Problem Numerical Optimal Control Part 3: Function space methods Auxiliary functional: (û, ˆx = S(u) given; λ R (N+1)nx to be defined later) J N (û) := ϕ(ˆx N ) + h N 1 i=0 ( H(ˆx i, û i, λ i+1 ) λ i+1 ( )) ˆxi+1 ˆx i h Derivative: (S i (û) denotes the sensitivity matrix ˆx i u ; exploit S 0 (û) = 0) J N (û) = ϕ (ˆx N )S N (û) ( ( N 1 +h H x [t i ]S i (û) + H u [t i ] û i S û i+1 (û) S λ i (û) )) i+1 h i=0 ( ) ( ) = ϕ (ˆx N ) λ N S N (û) + hh x [t 0 ] + λ 1 S 0 (û) +h N 1 i=1 ( H x [t i ] + 1 ( λ i+1 h λ i ) ) S i (û) + h N 1 i=0 H u [t i ] û i û Avoid calculation of sensitivities S i! Choose λ such that red terms vanish.
100 Gradient Method for the Discretized Problem Numerical Optimal Control Part 3: Function space methods Discrete adjoint ODE: λ i+1 λ i h = H x (ˆx i, û i, λ i+1 ), i = 0,..., N 1, λ N = ϕ (ˆx N ) Gradient of auxiliary functional: N 1 J N (û) = h i=0 H u[t i ] û i û = h ( H u [t 0] H u [t 1] H u [t N 1] ).
101 Gradient Method for the Discretized Problem Numerical Optimal Control Part 3: Function space methods Link to gradient of reduced objective functional J N : Theorem Let û R Nnu be given and let λ R (N+1)nx satisfy the discrete adjoint ODE. Then, J N (û) = J N (û). Consequence: The gradient method for RD-OCP uses in iteration k the search direction H u[t 0 ] d (k) = J N (u (k) ) = h.. H u [t N 1] This is the same search direction as in the function space gradient method, except that it is scaled by h. Slower convergence expected!
102 Gradient Method for the Discretized Problem Gradient method for RD-OCP (0) Choose u (0) R Nnu, β (0, 1), σ (0, 1), and set k := 0. (1) Solve (2) Set x i+1 x i h λ i+1 λ i h (3) If d (k) 2 0, STOP. Numerical Optimal Control Part 3: Function space methods = f (x i, u (k) i ) (i = 0,..., N 1), x 0 = x, = H x (ˆx i, û (k) i, λ i+1 ) (i = 0,..., N 1), λ N = ϕ (ˆx N ) H u [t 0] d (k) := J N (u (k) ) = h. H u [t N 1] (4) Perform an Armijo line-search: Find smallest j {0, 1, 2,...} with and set α k := β j. J N (u (k) + β j d (k) ) J N (u (k) ) σβ j d (k) 2 2 (5) Set u (k+1) := u (k) + α k d (k), k := k + 1, and go to (1).
103 Gradient Method for the Discretized Problem Example Example (compare function space equivalent example) Minimize x 2 (1) subject to the constraints Numerical Optimal Control Part 3: Function space methods x 1 (t) = x 1(t) + 3u(t), x 1 (0) = 2, x 2 (t) = 1 ( x 1 (t) 2 + u(t) 2), 2 x 2 (0) = 0. Output of gradient method for RD-OCP: (u (0) = 0, N = 100, β = 0.9, σ = 0.1) k α k J(u (k) ) H u H u E E E E E E E E E E E E E E E E E E E E-11 Observation: Since the search direction in RD-OCP is scaled by h, we need much more iterations compared to the function space gradient method, which required only 26 iterations at the same accuracy...
104 Contents Introduction Necessary Conditions Adjoint Formalism Gradient Method Gradient Method in Finite Dimensions Gradient Method for Optimal Control Problems Extensions Gradient Method for Discrete Problem Examples Lagrange-Newton Method Lagrange-Newton Method in Finite Dimensions Lagrange-Newton Method in Infinite Dimensions Application to Optimal Control Search Direction Examples Extensions
105 Lagrange-Newton Method Pro s: Con s: locally quadratic convergence rate (fast convergence) can handle nonlinear equality constraints global convergence achieved by Armijo linesearch requires second derivatives higher implementation effort compared to gradient method
106 Lagrange-Newton Method in Finite Dimensions Numerical Optimal Control Part 3: Function space methods Consider the equality constrained nonlinear optimization problem: Equality constrained optimization problem (E-NLP) Minimize J(x, u) subject to H(x, u) = 0 J : R nx R nu R, H : R nx R nu R n H twice continuously differentiable Lagrange function; L(x, u, λ) := J(x, u) + λ H(x, u) Theorem (KKT conditions) Let (ˆx, û) be a local minimum of E-NLP and let H (ˆx, û) have full rank. Then there exists a multiplier λ R n H such that (x,u) L(ˆx, û, λ) = 0.
107 Lagrange-Newton Method in Finite Dimensions Numerical Optimal Control Part 3: Function space methods Idea of Lagrange-Newton method: Apply Newton s method to the optimality system T (ˆx, û, λ) = 0 with T (x, u, λ) := ( (x,u) L(x, u, λ) H(x, u) ) = x L(x, u, λ) ul(x, u, λ) H(x, u) Newton system: T (x (k), u (k), λ (k) )d (k) = T (x (k), u (k), λ (k) ) with T (x, u, λ) = L xx (x, u, λ) L xu(x, u, λ) L ux (x, u, λ) L uu (x, u, λ) H H x (x, u) u (x, u) H x (x, u) H u (x, u) 0
108 Lagrange-Newton Method in Finite Dimensions Numerical Optimal Control Part 3: Function space methods Theorem (Nonsingularity) T (x, u, λ) is nonsingular, if the following conditions hold: L (x,u),(x,u) (x, u, λ) is positive definite on the nullspace of H (x, u), that is v L (x,u),(x,u) (x, u, λ)v > 0 v : H (x, u)v = 0. H (x, u) has full rank Note: These conditions are actually sufficient, if (x, u, λ) is a stationary point of L.
109 Lagrange-Newton Method in Infinite Dimensions Consider the equality constrained nonlinear optimization problem: Equality constrained optimization problem (E-NLP) Numerical Optimal Control Part 3: Function space methods Minimize J(x, u) subject to H(x, u) = 0 J : X U R, H : X U Λ twice continuously Fréchet differentiable, X, U, Λ Banach spaces Lagrange function; L : X U Λ R with L(x, u, λ ) := J(x, u) + λ (H(x, u)) Theorem (KKT conditions) Let (ˆx, û) be a local minimum of E-NLP and let H (ˆx, û) be surjective. Then there exists a multiplier λ Λ such that L (x,u) (ˆx, û, λ ) = 0.
110 Lagrange-Newton Method in Infinite Dimensions Numerical Optimal Control Part 3: Function space methods Idea of Lagrange-Newton method: Use Newton s method to find a zero ẑ = (ˆx, û, ˆλ ) Z := X U Λ of the operator T : Z Y defined by T (x, u, λ ) := L x (x, u, λ ) L u(x, u, λ ) H(x, u) Observe: L x (x, u, λ )x = J x (x, u)x + λ (H x (x, u)x) = J x (x, u)x + (H x (x, u) λ )x L u (x, u, λ )u = J u (x, u)u + λ (H u (x, u)u) = J u (x, u)u + (H u (x, u) λ )u where H x (x, u) : Y X and H u (x, u) : Y U denote the respective adjoint operators.
111 Local Lagrange-Newton Method Local Lagrange-Newton Method (0) Choose z (0) Z and set k := 0. (1) If T (z (k) ) Y 0, STOP. (2) Compute the search direction d (k) from T (z (k) )(d (k) ) = T (z (k) ). (3) Set z (k+1) := z (k) + d (k), k := k + 1, and go to (1).
112 Lagrange-Newton Method in Infinite Dimensions Numerical Optimal Control Part 3: Function space methods Theorem (Nonsingularity) T (x, u, λ ) is nonsingular, if the following conditions hold: L (x,u),(x,u) (x, u, λ ) is uniformly positive definite on the nullspace of H (x, u), i.e. there exists C > 0 such that v L (x,u),(x,u) (x, u, λ )v C v 2 X U v X U : H (x, u)v = 0. H (x, u) is surjective Note: These conditions are actually sufficient, if (x, u, λ) is a stationary point of L, see [1, Theorem 5.6], [2, Theorem 2.3]. [1] H. Maurer and J. Zowe. First and Second-Order Necessary and Sufficient Optimality Conditions for Infinite-Dimensional Programming Problems. Mathematical Programming, 16:98 110, [2] H. Maurer. First and Second Order Sufficient Optimality Conditions in Mathematical Programming and Optimal Control. Mathematical Programming Study, 14: , 1981.
113 Local Lagrange-Newton Method Theorem (local convergence) Let z be a zero of T. Suppose there exist constants > 0 and C > 0 such that for every z B (z ) the derivative T (z) is non-singular and T (z) 1 L(Y,Z ) C. (a) If ϕ, f 0, f, ψ are twice continuously differentiable, then there exists δ > 0 such that the local Lagrange-Newton method is well-defined for every z (0) B δ (z ) and the sequence {z (k) } k N converges superlinearly to z for every z (0) B δ (z ). (b) If the second derivatives of ϕ, f 0, f, ψ are locally Lipschitz continuous, then the convergence in (a) is quadratic. (c) If in addition to the assumption in (a), T (z (k) ) 0 for all k, then the residual values converge superlinearly: T (z (k+1) ) Y lim k T (z (k) = 0. ) Y
114 Global Lagrange-Newton Method Merit function for globalization: γ(z) := 1 2 T (z) 2 2 Globalized Lagrange-Newton Method (0) Choose z (0) Z, β (0, 1), σ (0, 1/2). (1) If γ(z (k) ) 0, STOP. (2) Compute the search direction d (k) from T (z (k) )(d (k) ) = T (z (k) ). (3) Find smallest j {0, 1, 2,...} with γ(z (k) + β j d (k) ) γ(z (k) ) + σβ j γ (z (k) )(d (k) ) and set α k := β j. (4) Set z (k+1) := z (k) + α k d (k), k := k + 1, and go to (1). Note: γ : Z R is Fréchet-differentiable with γ (z (k) )(d (k) ) = 2γ(z (k) ) = T (z (k) ) 2 2
115 Lagrange-Newton Method Application to Optimal Control Problem (Optimal control problem (OCP)) Given: I := [t 0, t f ], twice continuously differentiable functions ϕ : R nx R nx R, f 0 : R nx R nu R, f : R nx R nu R nx, ψ : R nx R nx R n ψ. Minimize J(x, u) := ϕ(x(t 0 ), x(t f )) + f 0 (x(t), u(t))dt I with respect to x X := W 1, (I, R nx ) and u U := L (I, R nu ) subject to the constraints x (t) = f (x(t), u(t)) a.e. in I, 0 = ψ(x(t 0 ), x(t f )). Remark: A partially reduced approach is possible, where x = x( ; u, x 0 ) is expressed as a function of initial value x 0 and u. The constraint ψ(x 0, x(t f ; u, x 0 )) = 0 remains. This is the function space equivalent of the direct shooting method.
116 Lagrange-Newton Method Application to Optimal Control Hamilton function: Define H(x, u, λ) := f 0 (x, u) + λ f (x, u) κ := ϕ + σ ψ Theorem (Minimum principle, KKT conditions) Let (x, u ) be a local minimizer of OCP. Let H (x, u ) be surjective. Then there exist multipliers λ W 1, (I, R nx ) and σ R n ψ with x (t) f (x (t), u (t)) = 0 λ (t) + H x (x (t), u (t), λ (t)) = 0 ψ(x (t 0 ), x (t f )) = 0 λ (t 0 ) + κ x 0 (x (t 0 ), x (t f ), σ ) = 0 λ (t f ) κ x f (x (t 0 ), x (t f ), σ ) = 0 H u(x (t), u (t), λ (t)) = 0 Root finding problem: T (z ) = 0, T : Z Y
117 Lagrange-Newton Method Application to Optimal Control Definiton of operator T : T (z)( ) := with x ( ) f (x( ), u( )) λ ( ) + H x (x( ), u( ), λ( )) ψ(x(t 0 ), x(t f )) λ(t 0 ) + κ x 0 (x(t 0 ), x(t f ), σ) λ(t f ) κ x f (x(t 0 ), x(t f ), σ) H u (x( ), u( ), λ( )) z := (x, u, λ, σ) Z := X U X R n ψ Y := L (I, R nx ) L (I, R nx ) R n ψ R nx R nx L (I, R nu )
118 Lagrange-Newton Method Computation of Search Direction Newton direction: T (z (k) )d = T (z (k) ), d = (x, u, λ, σ) Z Fréchet derivative: (evaluated at z (k) ) T (z (k) )(z) = x f x x f uu λ + H xx x + H xu u + H xλ λ ψ x 0 x(t 0 ) + ψ x f x(t f ) λ(t 0 ) + κ x 0 x 0 x(t 0 ) + κ x 0 x f x(t f ) + κ x 0 σσ λ(t f ) κ x f x 0 x(t 0 ) κ x f x f x(t f ) κ x f σσ H ux x + H uu u + H uλ λ.
119 Lagrange-Newton Method Computation of Search Direction Newton direction is equivalent to the linear DAE boundary value problem x f x 0 f u x λ H xx H xλ H xu λ = 0 H ux H uλ H uu u ( x (k) ) f ( λ (k) ) + (H x ) (H u ) with boundary conditions ψ x κ x 0 x 0 Id κ x 0 σ κ x f x 0 0 κ x f σ ψ = λ (k) (t 0 ) + (κ x 0 ) λ (k) (t f ) (κ x f ) x(0) λ(t 0 ) σ. + ψ x f 0 0 κ x 0 x f 0 0 κ x f x f Id 0 x(1) λ(t f ) σ Hence: In each Newton iteration, we need to solve the above linear BVP.
120 Lagrange-Newton Method Computation of Search Direction Theorem The differential-algebraic equation (DAE) in BVP has index one (i.e. the last equality can be solved for u), if the matrix function M(t) := H uu [t] is non-singular for almost every t [0, 1] and M(t) 1 C for some constant C and almost every t [0, 1]. If M( ) is singular: BVP contains a differential-algebraic equation of higher index, which is numerically unstable. boundary conditions may become infeasible
121 Lagrange-Newton Method Examples Example (Trolley) (x 1, z) u x 2 l z m 2 g x
122 Lagrange-Newton Method Examples Example (Trolley, continued) Dynamics: x 1 = x 3 x 2 = x 4 x 3 = x 4 = m 2 2 l3 sin(x 2 )x 2 4 m 2l 2 u + m 2 I y lx 2 4 sin(x 2) I y u m 1 m 2 l 2 m 1 I y m 2 2 l2 m 2 I y + m 2 2 l2 cos(x 2 ) 2 m l2 g cos(x 2 ) sin(x 2 ) m 1 m 2 l 2 m 1 I y m2 2l2 m 2 I y + m2 2l2 cos(x 2 ) 2 ( ) m 2 l m 2 l cos(x 2 )x 2 4 sin(x 2) cos(x 2 )u + g sin(x 2 )(m 1 + m 2 ) m 1 m 2 l 2 m 1 I y m 2 2 l2 m 2 I y + m 2 2 l2 cos(x 2 ) 2 Parameters: g = 9.81, m 1 = 0.3, m 2 = 0.5, l = 0.75, r = 0.1, I y =
123 Lagrange-Newton Method Examples Example (Trolley, continued) Minimize subject to the ODE, the initial conditions and the terminal conditions within the fixed time t f = tf u(t) 2 + 5x 4 (t) 2 dt 2 0 x 1 (0) = x 2 (0) = x 3 (0) = x 4 (0) = 0, x 1 (t f ) = 1, x 2 (t f ) = x 3 (t f ) = x 4 (t f ) = 0
124 Lagrange-Newton Method Examples Example (Trolley, continued) Output of Lagrange-Newton method: (N = 1000, Euler) k α k T (z (k) ) 2 2 d (k) E E E E E E E E E E E E-11 Iterations for different stepsizes N: mesh independence; linear CPU N CPU time [s] Iterations
125 Lagrange-Newton Method Examples Example (Trolley, continued) x4(t) x1(t) state x time t [s] state x time t [s] lambda1(t) x2(t) state x time t [s] adjoint lambda time t [s] lambda2(t) x3(t) state x time t [s] adjoint lambda time t [s] lambda3(t) adjoint lambda3 lambda4(t) adjoint lambda4 u(t) control u time t [s] time t [s] time t [s]
126 Lagrange-Newton Method Example from Chemical Engineering DAE index-1 optimal control problem: (chemical reaction of substances A, B, C, and D) Minimize tf M C (t f ) F B (t) 2 + Q(t) 2 dt 0 w.r.t. controls F B (feed rate of substance B) and cooling power Q subject to the index-1 DAE where M A = V A 1 e E 1 /T R C A C B ) M B = F B V (A 1 e E 1 /T R C A C B + A 2 e E 2 /T R C B C C ) M C = V (A 1 e E 1 /T R C A C B A 2 e E 2 /T R C B C C M D = V A 2 e E 2 /T R C B C C ) H = 20F B Q V ( A 1 e E 1 /T R C A C B 75A 2 e E 2 /T R C B C C 0 = H ( M i α i (T R T ref ) + β ( ) ) i T 2 R 2 T 2 ref i=a,b,c,d V = i=a,b,c,d M i ρ i, C i = M i /V, i = A, B, C, D. Source: V. C. Vassiliades, R. W. H. Sargent, and C. C. Pantelides. Solution of a class of multistage dynamic optimization problems. 2. Problems with path constraints. Industrial & Engineering Chemistry Research, 33: , 1994.
127 Lagrange-Newton Method Example from Chemical Engineering Lagrange-Newton method: (N = intervals) control F_B k α k T (z (k) ) 2 2 d (k) E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E-05 F_B(t) Q(t) time t [s] control Q time t [s]
128 Numerical Optimal Control Part 3: Function space methods Lagrange-Newton Method Example from Chemical Engineering state M_A state M_C state M_B M_C(t) M_B(t) 8800 M_A(t) time t [s] time t [s] state M_D state H time t [s] algebraic variable T_R T_R(t) H(t) M_D(t) time t [s] time t [s] time t [s]
129 Lagrange-Newton Method Example from Chemical Engineering adjoint lambda1 adjoint lambda2 adjoint lambda lambda1(t) lambda2(t) lambda3(t) time t [s] time t [s] time t [s] adjoint lambda4 adjoint lambda5 adjoint lambda_g e-05 5e-05 lambda4(t) lambda5(t) 0-5e lambda_g(t) 0-5e time t [s] time t [s] time t [s]
130 Lagrange-Newton Method Navier-Stokes Example Optimal control problem: Minimize Numerical Optimal Control Part 3: Function space methods 1 y(t, x) y d (t, x) 2 dx dt + δ u(t, x) 2 dx dt 2 Q 2 Q w.r.t. velocity y, pressure p and control u subject to the 2D Navier-Stokes equations y t = 1 Re y (y )y p + u in Q := (0, t f ) Ω, 0 = div(y) in Q, 0 = y(0, x) for x Ω := (0, 1) (0, 1), 0 = y(t, x) for (t, x) (0, t f ) Ω. Given: desired velocity field y d (t, x 1, x 2 ) = ( q(t, x 1 )q x 2 (t, x 2 ), q(t, x 2 )q x 1 (t, x 1 )), q(t, z) = (1 z) 2 (1 cos(2πzt)) M. Gerdts and M. Kunkel. A globally convergent semi-smooth Newton method for control-state constrained DAE optimal control problems. Computational Optimization and Applications, 48(3): , 2011.
131 Lagrange-Newton Method Navier-Stokes Example Numerical Optimal Control Part 3: Function space methods Discretization by method of lines: (details omitted) Minimize subject to the index-2 DAE 1 tf y h (t) y d,h (t) 2 dt + δ tf u h (t) 2 dt ẏ h (t) = 1 Re A hy h (t) = B h y h(t), y h (0) = 0. y h (t) H h,1 y h (t). B hp h (t) + u h (t), y h (t) H h,2(n 1) 2y h (t)
132 Numerical Optimal Control Part 3: Function space methods Lagrange-Newton Method Navier-Stokes Example Pressure p at t = 0.6, t = 1.0, t = 1.4 and t = 1.967: (Parameters: tf = 2, δ = 10 5, Re = 1, N = 31, Nt = 60, nx = 2(N 1)2 = 1800, ny = (N 1)2 = 900, nu = 1800 controls)
133 Numerical Optimal Control Part 3: Function space methods Lagrange-Newton Method Navier-Stokes Example Desired flow (left), controlled flow (middle), control (right) at t = 0.6, t = 1.0, t = 1.4 and t = 1.967:
134 Lagrange-Newton Method Navier-Stokes Example Numerical Optimal Control Part 3: Function space methods Output of Lagrange-Newton method: Solve Stokes problem as initial guess: k 1 0 f 0[t]dt α k 1 T (z (k) ) e e e e-10 Solve Navier-Stokes problem: k 1 0 f 0[t]dt α k 1 T (z (k) ) e e e e e e e e-10
135 Lagrange-Newton Method Extensions Treatment of inequality constraints: sequential-quadratic programming (SQP) Idea: solve linear-quadratic optimization problems to obtain a search direction interior-point methods (IP) Idea: solve a sequence of barrier problems (equivalent: perturbation of complementarity conditions in KKT conditions) semismooth Newton methods Idea: transform complementarity conditions into equivalent (nonsmooth!) equation
136 References. [1] W. Alt. The Lagrange-Newton method for infinite dimensional optimization problems. Numerical Functional Analysis and Optimization, 11: , [2] W. Alt. Sequential Quadratic Programming in Banach Spaces. In W. Oettli and D. Pallaschke, editors, Advances in Optimization, pages , Berlin, Springer. [3] W. Alt and K. Malanowski. The Lagrange-Newton method for nonlinear optimal control problems. Computational Optimization and Applications, 2:77 100, [4] W. Alt and K. Malanowski. The Lagrange-Newton method for state constrained optimal control problems. Computational Optimization and Applications, 4: , [5] M. Gerdts. Optimal Control of ODEs and DAEs. Walter de Gruyter, Berlin/Boston, 2012, [6] M. Hinze, R. Pinnau, M. Ulbrich, and S. Ulbrich. Optimization with PDE constraints. Mathematical Modelling: Theory and Applications 23. Dordrecht: Springer. xi, 270 p., [7] K. Ito and K. Kunisch. Lagrange multiplier approach to variational problems and applications. Advances in Design and Control 15. Philadelphia, PA: Society for Industrial and Applied Mathematics (SIAM), [8] K. C. P. Machielsen. Numerical Solution of Optimal Control Problems with State Constraints by Sequential Quadratic Programming in Function Space. Volume 53 of CWI Tract, Centrum voor Wiskunde en Informatica, Amsterdam, [9] F. Tröltzsch. Optimale Steuerung partieller Differentialgleichungen. Vieweg, Wiesbaden, 2005.
137 Resources Software: (available for registered users; free for academic use) OCPID-DAE1 (optimal control and parameter identification with differential-algebraic equations of index 1): sqpfiltertoolbox: SQP method for dense NLPs WORHP (Büskens/Gerdts): SQP method for sparse large-scale NLPs QPSOL: interior-point and nonsmooth Newton methods for sparse large-scale convex quadratic programs available upon request Robotics lab at UniBw M: research stays with use of lab equipment upon request KUKA youbot robot (2 arm robot on a platform); 3 scale rc cars; LEGO Mindstorms robots; quarter car testbench; quadcopter
138 More Resources Further optimal control software: CasADI, ACADO: M. Diehl et al.; NUDOCCCS: C. Büskens, University of Bremen SOCS: J. Betts, The Boeing Company, Seattle; DIRCOL: O. von Stryk, TU Darmstadt; MUSCOD-II: H.G. Bock et al., IWR Heidelberg; agbock/research/muscod.php MISER: K.L. Teo et al., Curtin University, Perth; les/miser/ PSOPT: Further optimization software: NPSOL (dense problems), SNOPT (sparse large-scale problems): Stanford Business Software; KNITRO (sparse large-scale problems): Ziena Optimization; IPOPT (sparse large-scale problems): A. Wächter: filtersqp: R. Fletcher, S. Leyffer; leyffer/solvers.html ooqp: M. Gertz, S. Wright; swright/ooqp/ qpoases: H.J. Ferreau, A. Potschka, C. Kirches; optec/software/qpoases/... Software for boundary value problems: BOUNDSCO: H. J. Oberle, University of Hamburg; COLDAE: U. Ascher; ascher/coldae.f... Links: Decision Tree for Optimization Software; CUTEr: large collection of optimization test problems; COPS: large-scale optimization test problems; more/cops/ MINTOC: testcases for mixed-integer optimal control;
139 Numerical Optimal Control Part 3: Function space methods Announcement: youbot Robotics Hackathon Description: I I I student programming contest addresses students and Phd s who like to realize projects with the KUKA youbot robot 12 particpants from 5 Universities (UniBw M, Bayreuth, TUM, TU Berlin, Maastricht)
140 Numerical Optimal Control Part 3: Function space methods Thanks for your Attention! Questions? Further information: Fotos: nchen Magnus Manske (Panorama), Luidger (Theatinerkirche), Kurmis (Chin. Turm), Arad Mojtahedi (Olympiapark), Max-k (Deutsches Museum), Oliver Raupach (Friedensengel), Andreas Praefcke (Nationaltheater)
Numerical Optimal Control Part 2: Discretization techniques, structure exploitation, calculation of gradients
Numerical Optimal Control Part 2: Discretization techniques, structure exploitation, calculation of gradients SADCO Summer School and Workshop on Optimal and Model Predictive Control OMPC 2013, Bayreuth
More informationTrajectory Planning and Collision Detection for Robotics Applications
Trajectory Planning and Collision Detection for Robotics Applications joint work with Rene Henrion, Dietmar Homberg, Chantal Landry (WIAS) Institute of Mathematics and Applied Computing Department of Aerospace
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization
E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained
More informationDirect and indirect methods for optimal control problems and applications in engineering
Direct and indirect methods for optimal control problems and applications in engineering Matthias Gerdts Computational Optimisation Group School of Mathematics The University of Birmingham gerdtsm@maths.bham.ac.uk
More information4TE3/6TE3. Algorithms for. Continuous Optimization
4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationAffine covariant Semi-smooth Newton in function space
Affine covariant Semi-smooth Newton in function space Anton Schiela March 14, 2018 These are lecture notes of my talks given for the Winter School Modern Methods in Nonsmooth Optimization that was held
More informationPart 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)
Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization Nick Gould (RAL) x IR n f(x) subject to c(x) = Part C course on continuoue optimization CONSTRAINED MINIMIZATION x
More informationNumerical Optimization
Constrained Optimization Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on Constrained Optimization Constrained Optimization Problem: min h j (x) 0,
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationTutorial on Control and State Constrained Optimal Control Problems
Tutorial on Control and State Constrained Optimal Control Problems To cite this version:. blems. SADCO Summer School 211 - Optimal Control, Sep 211, London, United Kingdom. HAL Id: inria-629518
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationDetermination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study
International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:
More informationOptimality Conditions for Constrained Optimization
72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationRegularity and approximations of generalized equations; applications in optimal control
SWM ORCOS Operations Research and Control Systems Regularity and approximations of generalized equations; applications in optimal control Vladimir M. Veliov (Based on joint works with A. Dontchev, M. Krastanov,
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationLecture 18: Optimization Programming
Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming
More informationDirect Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.
Direct Methods Moritz Diehl Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Leuven Belgium Overview Direct Single Shooting Direct Collocation Direct Multiple
More informationLecture 15 Newton Method and Self-Concordance. October 23, 2008
Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications
More informationSolution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark
Solution Methods Richard Lusby Department of Management Engineering Technical University of Denmark Lecture Overview (jg Unconstrained Several Variables Quadratic Programming Separable Programming SUMT
More informationAn Inexact Newton Method for Nonlinear Constrained Optimization
An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results
More informationInfeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization
Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationConvex Optimization. Newton s method. ENSAE: Optimisation 1/44
Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)
More informationOptimization Problems with Constraints - introduction to theory, numerical Methods and applications
Optimization Problems with Constraints - introduction to theory, numerical Methods and applications Dr. Abebe Geletu Ilmenau University of Technology Department of Simulation and Optimal Processes (SOP)
More informationSF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren
SF2822 Applied Nonlinear Optimization Lecture 9: Sequential quadratic programming Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH / 24 Lecture 9, 207/208 Preparatory question. Try to solve theory
More informationSolving generalized semi-infinite programs by reduction to simpler problems.
Solving generalized semi-infinite programs by reduction to simpler problems. G. Still, University of Twente January 20, 2004 Abstract. The paper intends to give a unifying treatment of different approaches
More informationWeak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations
Weak Convergence of Numerical Methods for Dynamical Systems and, and a relation with Large Deviations for Stochastic Equations Mattias Sandberg KTH CSC 2010-10-21 Outline The error representation for weak
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationTheory and Applications of Constrained Optimal Control Proble
Theory and Applications of Constrained Optimal Control Problems with Delays PART 1 : Mixed Control State Constraints Helmut Maurer 1, Laurenz Göllmann 2 1 Institut für Numerische und Angewandte Mathematik,
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationPDE Constrained Optimization selected Proofs
PDE Constrained Optimization selected Proofs Jeff Snider jeff@snider.com George Mason University Department of Mathematical Sciences April, 214 Outline 1 Prelim 2 Thms 3.9 3.11 3 Thm 3.12 4 Thm 3.13 5
More informationDichotomy, the Closed Range Theorem and Optimal Control
Dichotomy, the Closed Range Theorem and Optimal Control Pavel Brunovský (joint work with Mária Holecyová) Comenius University Bratislava, Slovakia Praha 13. 5. 2016 Brunovsky Praha 13. 5. 2016 Closed Range
More informationConstrained Optimization and Lagrangian Duality
CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationIntroduction to Optimization Techniques. Nonlinear Optimization in Function Spaces
Introduction to Optimization Techniques Nonlinear Optimization in Function Spaces X : T : Gateaux and Fréchet Differentials Gateaux and Fréchet Differentials a vector space, Y : a normed space transformation
More informationLecture 3. Optimization Problems and Iterative Algorithms
Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex
More informationOn the Optimization of Riemann-Stieltjes-Control-Systems with Application in Vehicle Dynamics
On the Optimization of Riemann-Stieltjes-Control-Systems with Application in Vehicle Dynamics Johannes Michael To cite this version: Johannes Michael. On the Optimization of Riemann-Stieltjes-Control-Systems
More informationLecture 15: SQP methods for equality constrained optimization
Lecture 15: SQP methods for equality constrained optimization Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 15: SQP methods for equality constrained
More informationLectures 9 and 10: Constrained optimization problems and their optimality conditions
Lectures 9 and 10: Constrained optimization problems and their optimality conditions Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lectures 9 and 10: Constrained
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods
More informationLecture 13: Constrained optimization
2010-12-03 Basic ideas A nonlinearly constrained problem must somehow be converted relaxed into a problem which we can solve (a linear/quadratic or unconstrained problem) We solve a sequence of such problems
More informationConvex Optimization M2
Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationDouble Smoothing technique for Convex Optimization Problems with Linear Constraints
1 Double Smoothing technique for Convex Optimization Problems with Linear Constraints O. Devolder (F.R.S.-FNRS Research Fellow), F. Glineur and Y. Nesterov Center for Operations Research and Econometrics
More informationConstrained optimization: direct methods (cont.)
Constrained optimization: direct methods (cont.) Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Direct methods Also known as methods of feasible directions Idea in a point x h, generate a
More informationOptimal control problems with PDE constraints
Optimal control problems with PDE constraints Maya Neytcheva CIM, October 2017 General framework Unconstrained optimization problems min f (q) q x R n (real vector) and f : R n R is a smooth function.
More informationJoint work with Nguyen Hoang (Univ. Concepción, Chile) Padova, Italy, May 2018
EXTENDED EULER-LAGRANGE AND HAMILTONIAN CONDITIONS IN OPTIMAL CONTROL OF SWEEPING PROCESSES WITH CONTROLLED MOVING SETS BORIS MORDUKHOVICH Wayne State University Talk given at the conference Optimization,
More informationPart 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)
Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective
More information10 Numerical methods for constrained problems
10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside
More information12. Interior-point methods
12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity
More informationSuboptimal Open-loop Control Using POD. Stefan Volkwein
Institute for Mathematics and Scientific Computing University of Graz, Austria PhD program in Mathematics for Technology Catania, May 22, 2007 Motivation Optimal control of evolution problems: min J(y,
More informationMultidisciplinary System Design Optimization (MSDO)
Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential
More informationISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints
ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints Instructor: Prof. Kevin Ross Scribe: Nitish John October 18, 2011 1 The Basic Goal The main idea is to transform a given constrained
More informationMathematical Economics. Lecture Notes (in extracts)
Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter
More informationNecessary optimality conditions for optimal control problems with nonsmooth mixed state and control constraints
Necessary optimality conditions for optimal control problems with nonsmooth mixed state and control constraints An Li and Jane J. Ye Abstract. In this paper we study an optimal control problem with nonsmooth
More informationTopic # Feedback Control Systems
Topic #17 16.31 Feedback Control Systems Deterministic LQR Optimal control and the Riccati equation Weight Selection Fall 2007 16.31 17 1 Linear Quadratic Regulator (LQR) Have seen the solutions to the
More informationMarch 8, 2010 MATH 408 FINAL EXAM SAMPLE
March 8, 200 MATH 408 FINAL EXAM SAMPLE EXAM OUTLINE The final exam for this course takes place in the regular course classroom (MEB 238) on Monday, March 2, 8:30-0:20 am. You may bring two-sided 8 page
More informationApplications of Linear Programming
Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal
More informationFact Sheet Functional Analysis
Fact Sheet Functional Analysis Literature: Hackbusch, W.: Theorie und Numerik elliptischer Differentialgleichungen. Teubner, 986. Knabner, P., Angermann, L.: Numerik partieller Differentialgleichungen.
More informationSupport Vector Machines
Support Vector Machines Support vector machines (SVMs) are one of the central concepts in all of machine learning. They are simply a combination of two ideas: linear classification via maximum (or optimal
More informationOn the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method
Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical
More informationConvex Optimization Boyd & Vandenberghe. 5. Duality
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More information5. Duality. Lagrangian
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationCONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING
CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global and local convergence results
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationHot-Starting NLP Solvers
Hot-Starting NLP Solvers Andreas Wächter Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu 204 Mixed Integer Programming Workshop Ohio
More information8 Numerical methods for unconstrained problems
8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields
More informationYou should be able to...
Lecture Outline Gradient Projection Algorithm Constant Step Length, Varying Step Length, Diminishing Step Length Complexity Issues Gradient Projection With Exploration Projection Solving QPs: active set
More informationSF2822 Applied nonlinear optimization, final exam Saturday December
SF2822 Applied nonlinear optimization, final exam Saturday December 5 27 8. 3. Examiner: Anders Forsgren, tel. 79 7 27. Allowed tools: Pen/pencil, ruler and rubber; plus a calculator provided by the department.
More informationSequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO
Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems Guidance Professor Masao FUKUSHIMA Hirokazu KATO 2004 Graduate Course in Department of Applied Mathematics and
More informationAn Inexact Newton Method for Optimization
New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)
More informationGeneralization to inequality constrained problem. Maximize
Lecture 11. 26 September 2006 Review of Lecture #10: Second order optimality conditions necessary condition, sufficient condition. If the necessary condition is violated the point cannot be a local minimum
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationA Concise Course on Stochastic Partial Differential Equations
A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original
More informationQuiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006
Quiz Discussion IE417: Nonlinear Programming: Lecture 12 Jeff Linderoth Department of Industrial and Systems Engineering Lehigh University 16th March 2006 Motivation Why do we care? We are interested in
More informationConstrained Optimization
1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange
More informationSolving MPECs Implicit Programming and NLP Methods
Solving MPECs Implicit Programming and NLP Methods Michal Kočvara Academy of Sciences of the Czech Republic September 2005 1 Mathematical Programs with Equilibrium Constraints Mechanical motivation Mechanical
More informationExact Augmented Lagrangian Functions for Nonlinear Semidefinite Programming
Exact Augmented Lagrangian Functions for Nonlinear Semidefinite Programming Ellen H. Fukuda Bruno F. Lourenço June 0, 018 Abstract In this paper, we study augmented Lagrangian functions for nonlinear semidefinite
More informationConstrained Nonlinear Optimization Algorithms
Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu Institute for Mathematics and its Applications University of Minnesota August 4, 2016
More informationA GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE
A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30,
More informationAn Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization
An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns
More informationComputational Optimization. Augmented Lagrangian NW 17.3
Computational Optimization Augmented Lagrangian NW 17.3 Upcoming Schedule No class April 18 Friday, April 25, in class presentations. Projects due unless you present April 25 (free extension until Monday
More informationLecture 13 Newton-type Methods A Newton Method for VIs. October 20, 2008
Lecture 13 Newton-type Methods A Newton Method for VIs October 20, 2008 Outline Quick recap of Newton methods for composite functions Josephy-Newton methods for VIs A special case: mixed complementarity
More informationICS-E4030 Kernel Methods in Machine Learning
ICS-E4030 Kernel Methods in Machine Learning Lecture 3: Convex optimization and duality Juho Rousu 28. September, 2016 Juho Rousu 28. September, 2016 1 / 38 Convex optimization Convex optimisation This
More informationLecture 4 - The Gradient Method Objective: find an optimal solution of the problem
Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...
More informationGENERALIZED second-order cone complementarity
Stochastic Generalized Complementarity Problems in Second-Order Cone: Box-Constrained Minimization Reformulation and Solving Methods Mei-Ju Luo and Yan Zhang Abstract In this paper, we reformulate the
More informationFirst-order optimality conditions for mathematical programs with second-order cone complementarity constraints
First-order optimality conditions for mathematical programs with second-order cone complementarity constraints Jane J. Ye Jinchuan Zhou Abstract In this paper we consider a mathematical program with second-order
More informationComputational Finance
Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples
More informationINTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE
INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationOptimal Control of Partial Differential Equations I+II
About the lecture: Optimal Control of Partial Differential Equations I+II Prof. Dr. H. J. Pesch Winter semester 2011/12, summer semester 2012 (seminar), winter semester 2012/13 University of Bayreuth Preliminary
More informationPenalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.
AMSC 607 / CMSC 878o Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 3: Penalty and Barrier Methods Dianne P. O Leary c 2008 Reference: N&S Chapter 16 Penalty and Barrier
More informationAn optimal control problem for a parabolic PDE with control constraints
An optimal control problem for a parabolic PDE with control constraints PhD Summer School on Reduced Basis Methods, Ulm Martin Gubisch University of Konstanz October 7 Martin Gubisch (University of Konstanz)
More informationA STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE
A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30, 2014 Abstract Regularized
More informationLecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima
B9824 Foundations of Optimization Lecture 1: Introduction Fall 2009 Copyright 2009 Ciamac Moallemi Outline 1. Administrative matters 2. Introduction 3. Existence of optima 4. Local theory of unconstrained
More information