MATH 4211/6211 Optimization Constrained Optimization

Similar documents
MATH 4211/6211 Optimization Basics of Optimization Problems

Math 273a: Optimization Basic concepts

Mechanical Systems II. Method of Lagrange Multipliers

Lecture 3: Basics of set-constrained and unconstrained optimization

Paul Schrimpf. October 17, UBC Economics 526. Constrained optimization. Paul Schrimpf. First order conditions. Second order conditions

ECE580 Solution to Problem Set 6

Math 164-1: Optimization Instructor: Alpár R. Mészáros

Lecture 18: Optimization Programming

March 5, 2012 MATH 408 FINAL EXAM SAMPLE

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

8. Constrained Optimization

MATH 4211/6211 Optimization Linear Programming

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

May 9, 2014 MATH 408 MIDTERM EXAM OUTLINE. Sample Questions

MATH2070 Optimisation

Written Examination

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Lagrange Multipliers

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Constrained Optimization

Recall : Eigenvalues and Eigenvectors

Optimality Conditions for Constrained Optimization

Constrained optimization

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

Symmetric Matrices and Eigendecomposition

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

ECE580 Solution to Problem Set 3: Applications of the FONC, SONC, and SOSC

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Math 5311 Constrained Optimization Notes

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016

Constrained Optimization

Second Order Optimality Conditions for Constrained Nonlinear Programming

Linear and non-linear programming

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Mathematical Economics. Lecture Notes (in extracts)

4TE3/6TE3. Algorithms for. Continuous Optimization

0.1 Tangent Spaces and Lagrange Multipliers

Assignment 1: From the Definition of Convexity to Helley Theorem

Math 164 (Lec 1): Optimization Instructor: Alpár R. Mészáros

Examination paper for TMA4180 Optimization I

Math Camp Notes: Everything Else

1 Computing with constraints

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

Nonlinear Optimization: What s important?

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

MATH 4211/6211 Optimization Quasi-Newton Method

EE364a Review Session 5

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Week 4: Calculus and Optimization (Jehle and Reny, Chapter A2)

Math 489AB Exercises for Chapter 1 Fall Section 1.0

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Lecture 4 Eigenvalue problems

Lagrange duality. The Lagrangian. We consider an optimization program of the form

Algorithms for constrained local optimization

COMP 558 lecture 18 Nov. 15, 2010

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

CHAPTER 2: QUADRATIC PROGRAMMING

5. Duality. Lagrangian

CS 246 Review of Linear Algebra 01/17/19

Computational Optimization. Constrained Optimization Part 2

Review of Some Concepts from Linear Algebra: Part 2

Review of Linear Algebra

DUALITY, OPTIMALITY CONDITIONS AND PERTURBATION ANALYSIS

Numerical Optimization

DEPARTMENT OF MATHEMATICS AND STATISTICS UNIVERSITY OF MASSACHUSETTS. MATH 233 SOME SOLUTIONS TO EXAM 2 Fall 2018

Basic Math for

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Algorithms for nonlinear programming problems II

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

Introduction and Math Preliminaries

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Additional Exercises for Introduction to Nonlinear Optimization Amir Beck March 16, 2017

MAT Linear Algebra Collection of sample exams

Math 24 Spring 2012 Sample Homework Solutions Week 8

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Theorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0).

OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Test #2 Math 2250 Summer 2003

Math 308 Practice Final Exam Page and vector y =

Lecture 2: Convex Sets and Functions

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

Introduction to Matrix Algebra

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

Transpose & Dot Product

MAC College Algebra

Generalization to inequality constrained problem. Maximize

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Math 164-1: Optimization Instructor: Alpár R. Mészáros

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Transcription:

MATH 4211/6211 Optimization Constrained Optimization Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0

Constrained optimization problems are formulated as minimize f(x) subject to g j (x) 0, j = 1,..., p, h i (x) = 0, i = 1,..., m, where g j, h i : R n R are inequality and equality constraint functions, respectively. We can summarize them into vector-valued functions g and h: g(x) = g 1 (x). and h(x) = g p (x) h 1 (x). h m (x) so the constraints can be written as g(x) 0 and h(x) = 0, respectively. Note that g : R n R p and h : R n R m. Xiaojing Ye, Math & Stat, Georgia State University 1

We can write the constrained optimization concisely as minimize f(x) subject to g(x) 0 h(x) = 0 The feasible set is Ω := {x R n : g(x) 0, h(x) = 0}. Exmaple. LP (standard form) is a constrained optimization with f(x) = c T x, g(x) = x and h(x) = Ax b. Xiaojing Ye, Math & Stat, Georgia State University 2

We now focus on constrained optimization problems with equality constraints only, i.e., minimize f(x) subject to h(x) = 0 and the feasible set is Ω = {x R n : h(x) = 0}. Xiaojing Ye, Math & Stat, Georgia State University 3

Some equality-constrained optimization problem can be converted into unconstrained ones. Example. Consider the constrained optimization problem minimize x 2 1 + 2x 1x 2 + 3x 2 2 + 4x 1 + 5x 2 + 6x 3 subject to x 1 + 2x 2 = 3 4x 1 + 5x 3 = 6 The constraints imply that x 2 = 1 2 (3 x 1) and x 3 = 1 5 (6 x 1). Substitute x 2 and x 3 in the objective function to get an unconstrained minimization of x 1 only. Xiaojing Ye, Math & Stat, Georgia State University 4

Example. Consider the constrained optimization problem maximize x 1 x 2 subject to x 2 1 + 4x2 2 = 1 It is equivalent to maximizing x 2 1 x2 2 then substitute x2 1 by 1 4x2 2 an unconstrained problem of x 2. to get Another way to solving this is using 1 = x 2 1 + (2x 2) 2 4x 1 x 2 where the equality holds when x 1 = 2x 2.So x 1 = 2/2 and x 2 = 2/4. However, not all equality-constrained problems can be easily converted into unconstrained ones. Xiaojing Ye, Math & Stat, Georgia State University 5

We need general theory to solve constrained optimization problems with equality constraints: minimize f(x) subject to h(x) = 0 and the feasible set is Ω = {x R n : h(x) = 0}. Recall that h : R n R m (m n) has Jacobian matrix Dh(x) = h 1 (x) T. h m (x) T Rm n Definition. We say a point x Ω is a regular point if rank(dh(x)) = m, i.e., the Jacobian matrix has full row rank. Xiaojing Ye, Math & Stat, Georgia State University 6

Example. Let n = 3 and m = 1. Define h 1 (x) = x 2 x 2 3 constraint. Then the Jacobian matrix is be the only Dh(x) = [ h 1 (x) T ] = [0, 1, 2x 3 ] Note that Dh(x) 0 and hence rank(dh(x)) = 1 everywhere. The feasible set Ω is a surface in R 3 with dimension n m = 3 1 = 2. Xiaojing Ye, Math & Stat, Georgia State University 7

Example. Let n = 3 and m = 2. Define h 1 (x) = x 1 and h 2 (x) = x 2 x 2 3. The Jacobian is [ ] 1 0 0 Dh(x) = 0 1 2x 3 with rank(dh(x)) = 2 everywhere, and the feasible set Ω is a line in R 3 with dimension n m = 3 2 = 1. Xiaojing Ye, Math & Stat, Georgia State University 8

Tangent space and normal space Defintion. We say x : (a, b) R n, a curve in R n, is differentiable if x i (t) exists for all t (a, b). The derivative is defined by x (t) = We say x is twice differentiable if x i x (t) = x 1 (t). x n(t) (t) exists for all t (a, b), and x 1 (t). x n(t) Xiaojing Ye, Math & Stat, Georgia State University 9

Defintion. The tangent space of Ω = {x R n : h(x) = 0} at x is the set T (x ) = {y R n : Dh(x )y = 0}. In other words, T (x ) = Null(Dh(x )). Remark. If x is regular, then rank(dh(x )) = m, and hence dim(t (x )) = dim(null(dh(x ))) = n m. Remark. We sometimes draw the tangent space as a plane tangent to Ω at x, that tangent plane is T P (x ) := x + T (x ) = {x + y : y T (x )}. Xiaojing Ye, Math & Stat, Georgia State University 10

Theorem. Suppose x is regular. Then y T (x ) iff there exists curve x : (a, b) Ω such that x(t) = x and x (t) = y for some t (a, b). Xiaojing Ye, Math & Stat, Georgia State University 11

Defintion. The normal space of Ω at x is defined by N(x ) = {Dh(x ) T z : z R m } In other word, N(x ) = R(Dh(x )), the row space of Dh(x ). Note that dim(n(x )) = dim(r(dh(x ))) = m. Remark. The tangent space T (x ) and the normal space N(x ) form an orthogonal decomposition of R n : where T (x ) N(x ). R n = T (x ) N(x ) Hence, for any x R n, there exist unique y T (x ) and v N(x ), such that x = y + v. Xiaojing Ye, Math & Stat, Georgia State University 12

Now let us see the first-order necessary conditions (FONC) for equality-constrained minimization. Suppose x is a local minimizer of f(x) over Ω = {x : h(x) = 0}, where f, h C 1. Then for any y T (x ), there exists curve x : (a, b) Ω such that x(t) = x and x (t) = y for some t (a, b). Define φ(s) = f(x(s)) (note that φ : (a, b) R), then φ (s) = f(x(s)) T x (s) for all s (a, b). In particular, due to the standard FONC, we have φ (t) = f(x(t)) T x (t) = f(x ) T y = 0. Since y T (x ) is arbitrary, we know f(x ) T (x ), i.e., f(x ) N(x ). This means that λ R m, such that f(x ) + Dh(x ) T λ = 0. Xiaojing Ye, Math & Stat, Georgia State University 13

Now we have showed that, for x to be a local minimizer of f over Ω = {x : h(x) = 0}, it must satisfy f(x ) + Dh(x ) T λ = 0 h(x ) = 0 There are called the first-order necessary conditions (FNOC), or the Lagrange condition, of the equality-constrained minimization problem. λ is called the Lagrange multiplier. Remark. The conditions above are necessary but not sufficient to determine x to be a local minimizer a point satisfying these conditions could be a local maximizer or neither. Xiaojing Ye, Math & Stat, Georgia State University 14

We introduce the Lagrange function L(x, λ) = f(x) + h(x) T λ Then the Lagrange condition becomes x L(x, λ ) = f(x ) + Dh(x ) T λ = 0 λ L(x, λ ) = h(x ) = 0 for a local minimizer x. Xiaojing Ye, Math & Stat, Georgia State University 15

Example. Consider an equality-constrained optimization problem minimize x 2 1 + x2 2 subject to x 2 1 + 2x2 2 = 1 Solution. Here f(x) = x 2 1 + x2 2 and h(x) = x2 1 + 2x2 2 1. The Lagrange function is L(x, λ) = (x 2 1 + x2 2 ) + λ(x2 1 + 2x2 2 1) Then we obtain x1 L(x, λ) = 2x 1 + 2λx 1 = 0 x2 L(x, λ) = 2x 2 + 4λx 2 = 0 λ L(x, λ) = x 2 1 + 2x2 2 1 = 0 Xiaojing Ye, Math & Stat, Georgia State University 16

Solution (cont). Solving this system yields x 1 x 2 = λ 0 1/ 2 1/2, 0 1/ 2 1/2, 1 0 1, and 1 0 1 It is easy to check that x = [0; ±1/ 2] are local minimizers, and x = [0; ±1] are local maximizers. Xiaojing Ye, Math & Stat, Georgia State University 17

Now we consider second-order conditions. We assume f, h C 2. Following the same steps as in FONC, suppose x is a local minimizer, then for any y T (x ), there exists a curve x : (a, b) Ω such that x(t) = x and x (t) = y for some t (a, b). Again define φ(s) = f(x(s)), and hence φ (s) = f(x(s)) T x (s). Then the standard second-order necessary condition (SONC) implies that φ (t) = f(x(t)) T x (t) = f(x ) T y = 0 and φ (t) = y T 2 f(x )y + f(x ) T x (t) 0 Xiaojing Ye, Math & Stat, Georgia State University 18

In addition, since ψ i (s) := h i (x(s)) = 0 for all s (a, b), we have ψ i (t) = 0 which yields for all i = 1,..., m. y T 2 h i (x )y + h i (x ) T x (t) = 0 Together with the Lagrange condition, we know λ R m such that f(x )+ Dh(x ) T λ = 0. Summing the results above, the second-order necessary condition (SONC) for the equality-constrained minimization becomes for all y T (x ). y T [ 2 f(x ) + m i=1 ] λ i 2 h i (x ) y 0 Xiaojing Ye, Math & Stat, Georgia State University 19

We summarize the second-order necessary condition (SONC): Theorem (SONC). Let x be a local minimizer of f : R n R over Ω = {x : h(x) = 0 R m } with m n, where f, h C 2. Suppose x is regular, then λ = [λ 1 ;... ; λ m] R m such that 1. f(x ) + Dh(x ) T λ = 0; 2. for every y T (x ), there is y T [ 2 f(x ) + m i=1 ] λ i 2 h i (x ) y 0 Xiaojing Ye, Math & Stat, Georgia State University 20

We also have the following second-order sufficient condition (SOSC): Theorem (SOSC). Suppose x Ω = {x : h(x) = 0} is regular. If λ = [λ 1 ;... ; λ m] R m such that 1. f(x ) + Dh(x ) T λ = 0; 2. for every nonzero y T (x ), there is y T [ 2 f(x ) + m i=1 ] λ i 2 h i (x ) y > 0, then x is a strict local minimizer of f over Ω. Xiaojing Ye, Math & Stat, Georgia State University 21

Example. Solve the following problem maximize xt Qx x T P x where Q = diag([4, 1]) and P = diag([2, 1]). Solution. Note the objective function is scale-invariant (replacing x by tx for any t 0 yields the same value). This can be converted into the constrained minimization problem minimize x T Qx subject to x T P x 1 = 0 and h(x) = x T P x 1 R is the constraint. Note that Dh(x) = 2P x = [4x 1 ; 2x 2 ]. Xiaojing Ye, Math & Stat, Georgia State University 22

Solution (cont). We first write the Lagrange function Then the Lagrange condition becomes L(x, λ) = x T Qx + λ(x T P x 1) x L(x, λ ) = 2(Q λp )x = 0 λ L(x, λ ) = x T P x 1 = 0 The first equation implies P 1 Qx = λ x, and hence λ is an eigenvalue of P 1 Q = diag([2, 1]). Hence λ = 2 or λ = 1. Xiaojing Ye, Math & Stat, Georgia State University 23

For λ = 2, we know x is the corresponding eigenvector of P 1 Q and satisfies x T P x T = 1. Hence x = [±1/ 2; 0]. The Tangent space is T (x ) = Null(Dh(x )) = Null([± 2; 0]) = {[0; a] : a R}. We also have 2 f(x ) + λ 2 h(x ) = 2Q + 2λ P = [ ] 0 0 0 2 Therefore y T [ 2 f(x ) + λ 2 h(x )]y = 2a 2 > 0 for all y = [0; a] T (x ) with a 0. Therefore x = [±1/ 2; 0] are both strict local minimizers of the constrained optimization problem. Going back to the original problem, any x = [t; 0] with t 0 is local maximizer of xt Qx x T P x. Xiaojing Ye, Math & Stat, Georgia State University 24

For λ = 1, we know x is the corresponding eigenvector of P 1 Q and satisfies x T P x T = 1. Hence x = [0; ±1]. The Tangent space is T (x ) = Null(Dh(x )) = Null([0; ±1]) = {[a; 0] : a R}. We also have 2 f(x ) + λ 2 h(x ) = 2Q + 2λ P = [ ] 4 0 0 0 Therefore y T [ 2 f(x ) + λ 2 h(x )]y = 4a 2 < 0 for all y = [a; 0] T (x ) with a 0. Therefore x = [0; ±1] are both strict local maximizers of the constrained optimization problem. Going back to the original problem, any x = [0; t] with t 0 is local minimizer of xt Qx x T P x. Xiaojing Ye, Math & Stat, Georgia State University 25

Now we consider a special type of constrained minimization problem with linear equality constraints: minimize subject to 1 2 xt Qx Ax = b We have f(x) = 1 2 xt Qx and h(x) = b Ax. The Lagrange function is L(x, λ) = 1 2 xt Qx + λ T (b Ax). Hence the Lagrange condition is x L(x, λ) = Qx A T λ = 0 λ L(x, λ) = b Ax = 0 Xiaojing Ye, Math & Stat, Georgia State University 26

Now we solve the following system for [x ; λ ]: x L(x, λ ) = Qx A T λ = 0 λ L(x, λ ) = b Ax = 0 The first equation implies x = Q 1 A T λ. Plugging this into the second equation and solve for λ to get λ = (AQ 1 A T ) 1 b Hence the solution is x = Q 1 A T λ = Q 1 A T (AQ 1 A T ) 1 b Xiaojing Ye, Math & Stat, Georgia State University 27

Example. Consider the problem of finding the solution of minimal norm to the linear system Ax = b. That is minimize subject to x Ax = b Solution. The problem is equivalent to minimize subject to 1 2 x 2 = 1 2 xt x Ax = b which is the problem above with Q = I. Hence the solution is x = A T (AA T ) 1 b Xiaojing Ye, Math & Stat, Georgia State University 28