Mathematical Economics. Lecture Notes (in extracts)

Similar documents
EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

EC487 Advanced Microeconomics, Part I: Lecture 2

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

EC400 Math for Microeconomics Syllabus The course is based on 6 sixty minutes lectures and on 6 ninety minutes classes.

Mathematical Economics: Lecture 16

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

minimize x subject to (x 2)(x 4) u,

Math 273a: Optimization Basic concepts

September Math Course: First Order Derivative

Constrained Optimization

Numerical Optimization

Summary Notes on Maximization

Session: 09 Aug 2016 (Tue), 10:00am 1:00pm; 10 Aug 2016 (Wed), 10:00am 1:00pm &3:00pm 5:00pm (Revision)

Generalization to inequality constrained problem. Maximize

Optimality Conditions for Constrained Optimization

Lecture Notes for Chapter 12

Lecture 4: Optimization. Maximizing a function of a single variable

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Optimization Theory. Lectures 4-6

Monotone Function. Function f is called monotonically increasing, if. x 1 x 2 f (x 1 ) f (x 2 ) x 1 < x 2 f (x 1 ) < f (x 2 ) x 1 x 2

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

ECON 5111 Mathematical Economics

Optimization. A first course on mathematics for economists

FIRST-ORDER SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS III: Autonomous Planar Systems David Levermore Department of Mathematics University of Maryland

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

Calculus C (ordinary differential equations)

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4.

Review of Optimization Methods

Constrained Optimization and Lagrangian Duality

Convex Functions and Optimization

Mathematical Economics (ECON 471) Lecture 3 Calculus of Several Variables & Implicit Functions

Nonlinear Programming and the Kuhn-Tucker Conditions

ARE211, Fall 2005 CONTENTS. 5. Characteristics of Functions Surjective, Injective and Bijective functions. 5.2.

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

Microeconomics I. September, c Leopold Sögner

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

Math Ordinary Differential Equations

3 Applications of partial differentiation

OR MSc Maths Revision Course

Nonlinear Optimization: What s important?

Week 4: Calculus and Optimization (Jehle and Reny, Chapter A2)

A Brief Review on Convex Optimization

Lecture 18: Optimization Programming

Mathematical Foundations -1- Convexity and quasi-convexity. Convex set Convex function Concave function Quasi-concave function Supporting hyperplane

MATH 215/255 Solutions to Additional Practice Problems April dy dt

5. Duality. Lagrangian

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Mathematical Foundations -1- Constrained Optimization. Constrained Optimization. An intuitive approach 2. First Order Conditions (FOC) 7

1. First and Second Order Conditions for A Local Min or Max.

II. An Application of Derivatives: Optimization

Lagrange Multipliers

Chapter 13. Convex and Concave. Josef Leydold Mathematical Methods WS 2018/19 13 Convex and Concave 1 / 44

Nonlinear Programming (NLP)

Dynamical Systems. August 13, 2013

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

MATH2070 Optimisation

Final Exam Advanced Mathematics for Economics and Finance

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

z = f (x; y) f (x ; y ) f (x; y) f (x; y )

Math Camp Notes: Everything Else

Constrained maxima and Lagrangean saddlepoints

5 Handling Constraints

Chapter 3: Constrained Extrema

REVIEW OF DIFFERENTIAL CALCULUS

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control

CSCI : Optimization and Control of Networks. Review on Convex Optimization

g(x,y) = c. For instance (see Figure 1 on the right), consider the optimization problem maximize subject to

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimality Conditions

CHAPTER 4: HIGHER ORDER DERIVATIVES. Likewise, we may define the higher order derivatives. f(x, y, z) = xy 2 + e zx. y = 2xy.

MA102: Multivariable Calculus

Calculus for the Life Sciences II Assignment 6 solutions. f(x, y) = 3π 3 cos 2x + 2 sin 3y

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Econ Slides from Lecture 14

Math 118, Fall 2014 Final Exam

154 Chapter 9 Hints, Answers, and Solutions The particular trajectories are highlighted in the phase portraits below.

Lecture: Duality.

Problem List MATH 5173 Spring, 2014

Optimality, Duality, Complementarity for Constrained Optimization

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

Fundamentals of Unconstrained Optimization

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112

HW3 - Due 02/06. Each answer must be mathematically justified. Don t forget your name. 1 2, A = 2 2

Gradient Descent. Dr. Xiaowei Huang

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Convex Optimization & Lagrange Duality

Convex Optimization M2

Constrained optimization

Final Exam - Math Camp August 27, 2014

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

BEEM103 UNIVERSITY OF EXETER. BUSINESS School. January 2009 Mock Exam, Part A. OPTIMIZATION TECHNIQUES FOR ECONOMISTS solutions

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

Transcription:

Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter Term 2018/19 Annotation: 1. These lecture notes do not replace your attendance of the lecture. Numerical examples are only presented during the lecture. 2. The symbol points to additional, detailed remarks given in the lecture. 3. I am grateful to Julia Lange for her contribution in editing the lecture notes.

Contents 1 Basic mathematical concepts 1 1.1 Preliminaries...................................... 1 1.2 Convex sets....................................... 6 1.3 Convex and concave functions............................. 7 1.4 Quasi-convex and quasi-concave functions...................... 8 2 Unconstrained and constrained optimization 11 2.1 Extreme points..................................... 11 2.1.1 Global extreme points............................. 11 2.1.2 Local extreme points.............................. 12 2.2 Equality constraints.................................. 14 2.3 Inequality constraints................................. 16 2.4 Non-negativity constraints............................... 18 3 Sensitivity analysis 20 3.1 Preliminaries...................................... 20 3.2 Value functions and envelope results......................... 20 3.2.1 Equality constraints.............................. 20 3.2.2 Properties of the value function for inequality constraints......... 21 3.2.3 Mixed constraints................................ 22 3.3 Some further microeconomic applications...................... 23 3.3.1 Cost minimization problem.......................... 23 3.3.2 Profit maximization problem of a competitive firm............. 24 4 Differential equations 25 4.1 Preliminaries...................................... 25 4.2 Differential equations of the first order........................ 26 4.2.1 Separable equations.............................. 26 4.2.2 First-order linear differential equations.................... 26 4.3 Second-order linear differential equations and systems in the plane........ 28 5 Optimal control theory 35 5.1 Calculus of variations.................................. 35 i

CONTENTS ii 5.2 Control theory..................................... 37 5.2.1 Basic problem.................................. 37 5.2.2 Standard problem............................... 38 5.2.3 Current value formulations.......................... 40

Chapter 1 Basic mathematical concepts 1.1 Preliminaries Quadratic forms and their sign Definition 1: If A = (a ij ) is a matrix of order n n and x T = (x 1, x 2,..., x n ) R n, then the term Q(x) = x T A x is called a quadratic form. Thus: Q(x) = Q(x 1, x 2,..., x n ) = n n a ij x i x j i=1 j=1 Example 1 Definition 2: A matrix A of order n n and its associated quadratic form Q(x) are said to be 1. positive definite, if Q(x) = x T A x > 0 for all x T = (x 1, x 2,..., x n ) (0, 0,..., 0); 2. positive semi-definite, if Q(x) = x T A x 0 for all x R n ; 3. negative definite, if Q(x) = x T A x < 0 for all x T = (x 1, x 2,..., x n ) (0, 0,..., 0); 4. negative semi-definite, if Q(x) = x T A x 0 for all x R n ; 5. indefinite, if it is neither positive semi-definite nor negative semi-definite. Remark: In case 5., there exist vectors x and y such that Q(x ) > 0 and Q(y ) < 0. 1

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 2 Definition 3: The leading principle minors of a matrix A = (a ij ) of order n n are the determinants a 11 a 12 a 1k a D k = 21 a 22 a 2k....., k = 1, 2,..., n. a k1 a k2 a kk (i.e., D k is obtained from A by crossing out the last n k columns and rows). Theorem 1 Let A be a symmetric matrix of order n n. Then: 1. A positive definite D k > 0 for k = 1, 2,..., n. 2. A negative definite ( 1) k D k > 0 for k = 1, 2,..., n. 3. A positive semi-definite = D k 0 for k = 1, 2,..., n. 4. A negative semi-definite = ( 1) k D k 0 for k = 1, 2,..., n. now: necessary and sufficient criterion for positive (negative) semi-definiteness Definition 4: An (arbitrary) principle minor k of order k (1 k n) is the determinant of a submatrix of A obtained by deleting all but k rows and columns in A with the same numbers. Theorem 2 Let A be a symmetric matrix of order n n. Then: 1. A positive semi-definite k 0 for all principle minors of order k = 1, 2,..., n. 2. A negative semi-definite ( 1) k k 0 for all principle minors of order k = 1, 2,..., n. Example 2 alternative criterion for checking the sign of A:

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 3 Theorem 3 Let A be a symmetric matrix of order n n and λ 1, λ 2,..., λ n be the real eigenvalues of A. Then: 1. A positive definite λ 1 > 0, λ 2 > 0,..., λ n > 0. 2. A positive semi-definite λ 1 0, λ 2 0,..., λ n 0. 3. A negative definite λ 1 < 0, λ 2 < 0,..., λ n < 0. 4. A negative semi-definite λ 1 0, λ 2 0,..., λ n 0. 5. A indefinite A has eigenvalues with opposite signs. Example 3 Level curve and tangent line consider: level curve: z = F (x, y) F (x, y) = C with C R = slope of the level curve F (x, y) = C at the point (x, y): y = F x(x, y) F y (x, y) (See Werner/Sotskov(2006): Mathematics of Economics and Business, Theorem 11.6, implicit-function theorem.) equation of the tangent line T : y y 0 = y (x x 0 ) y y 0 = F x(x 0, y 0 ) F y (x 0, y 0 ) (x x 0) = F x (x 0, y 0 ) (x x 0 ) + F y (x 0, y 0 ) (y y 0 ) = 0 Illustration: equation of the tangent line T Remark: The gradient F (x 0, y 0 ) is orthogonal to the tangent line T at (x 0, y 0 ). Example 4

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 4 generalization to R n : let x 0 = (x 0 1, x0 2,..., x0 n) gradient of F at x 0 : F x1 (x 0 ) F (x 0 F ) = x2 (x 0 ). F xn (x 0 ) = equation of the tangent hyperplane T at x 0 : F x1 (x 0 ) (x 1 x 0 1) + F x2 (x 0 ) (x 2 x 0 2) + + F xn (x 0 ) (x n x 0 n) = 0 or, equivalently: [ F (x 0 )] T (x x 0 ) = 0 Directional derivative measures the rate of change of function f in an arbitrary direction r Definition 5: Let function f : D f R, D f R n, be continuously partially differentiable and r = (r 1, r 2,..., r n ) T R n with r = 1. The term [ f(x 0 ) ] T r = fx1 (x 0 ) r 1 + f x2 (x 0 ) r 2 + + f xn (x 0 ) r n is called the directional derivative of function f at the point x 0 = (x 0 1, x0 2,..., x0 n) D f. Example 5 Homogeneous functions and Euler s theorem Definition 6 A function f : D f R, D f R n, is said to be homogeneous of degree k on D f, if t > 0 and (x 1, x 2,..., x n ) D f imply (t x 1, t x 2,..., t x n ) D f and f(t x 1, t x 2,..., t x n ) = t k f(x 1, x 2,..., x n ) for all t > 0, where k can be positive, zero or negative.

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 5 Theorem 4 (Euler s theorem) Let the function f : D f R, D f R n, be continuously partially differentiable, where t > 0 and (x 1, x 2,..., x n ) D f imply (t x 1, t x 2,..., t x n ) D f. Then: f is homogeneous of degree k on D f x 1 f x1 (x) + x 2 f x2 (x) + + x n f xn (x) = k f(x) holds for all (x 1, x 2,..., x n ) D f. Example 6 Linear and quadratic approximations of functions in R 2 known: Taylor s formula for functions of one variable (See Werner/Sotskov (2006), Theorem 4.20.) f(x) = f(x 0 ) + f (x 0 ) 1! R n (x) - remainder (x x 0 ) + f (x 0 ) 2! (x x 0 ) 2 + + f (n) (x 0 ) n! (x x 0 ) n + R n (x) now: n = 2 z = f(x, y) defined around (x 0, y 0 ) D f let: x = x 0 + h, y = y 0 + k Linear approximation of f: f(x 0 + h, y 0 + k) = f(x 0, y 0 ) + f x (x 0, y 0 ) h + f y (x 0, y 0 ) k + R 1 (x, y) Quadratic approximation of f: f(x 0 + h, y 0 + k) = f(x 0, y 0 ) + f x (x 0, y 0 ) h + f y (x 0, y 0 ) k + 2[ 1 fxx (x 0, y 0 ) h 2 + 2f xy (x 0, y 0 ) h k + f yy (x 0, y 0 ) k 2] +R 2 (x, y) often: (x 0, y 0 ) = (0, 0) Example 7 Implicitly defined functions exogenous variables: x 1, x 2,..., x n endogenous variables: y 1, y 2,..., y m

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 6 F 1 (x 1, x 2,..., x n ; y 1, y 2,..., y m ) = 0 F 2 (x 1, x 2,..., x n ; y 1, y 2,..., y m ) = 0. (1) F m (x 1, x 2,..., x n ; y 1, y 2,..., y m ) = 0 (m < n) Is it possible to put this system into its reduced form: y 1 = f 1 (x 1, x 2,..., x n ) y 2 = f 2 (x 1, x 2,..., x n ). (2) y m = f m (x 1, x 2,..., x n ) Theorem 5 Assume that: F 1, F 2,..., F m are continuously partially differentiable; (x 0, y 0 ) = (x 0 1, x0 2,..., x0 n; y 0 1, y0 2,..., y0 m) satisfies (1); ( J(x 0, y 0 ) = det F j (x 0,y 0 ) y k ) 0 (i.e., the Jacobian determinant is regular). Then the system (1) can be put into its reduced form (2). Example 8 1.2 Convex sets Definition 7 A set M is called convex, if for any two points (vectors) x 1, x 2 M, any convex combination λx 1 + (1 λ)x 2 with 0 λ 1 also belongs to M. Illustration: Convex set Remark: The intersection of convex sets is always a convex set, while the union of convex sets is not necessarily a convex set.

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 7 Illustration: Union and intersection of convex sets 1.3 Convex and concave functions Definition 8 Let M R n be a convex set. A function f : M R is called convex on M, if f(λx 1 + (1 λ)x 2 ) λf(x 1 ) + (1 λ)f(x 2 ) for all x 1, x 2 M and all λ [0, 1]. f is called concave, if f(λx 1 + (1 λ)x 2 ) λf(x 1 ) + (1 λ)f(x 2 ) for all x 1, x 2 M and all λ [0, 1]. Illustration: Convex and concave functions Definition 9 The matrix f x1 x 1 (x 0 ) f x1 x 2 (x 0 ) f x1 x n (x 0 ) H f (x 0 ) = (f xi x j (x 0 f )) = x2 x 1 (x 0 ) f x2 x 2 (x 0 ) f x2 x n (x 0 )...... f xnx1 (x 0 ) f xnx2 (x 0 ) f xnxn (x 0 ) is called the Hessian matrix of function f at the point x 0 = (x 0 1, x0 2,..., x0 n) D f R n. Remark: If f has continuous second-order partial derivatives, the Hessian matrix is symmetric.

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 8 Theorem 6 Let f : D f R, D f R n, be twice continuously differentiable and M D f be convex. Then: 1. f is convex on M the Hessian matrix H f (x) is positive semi-definite for all x M; 2. f is concave on M the Hessian matrix H f (x) is negative semi-definite for all x M; 3. the Hessian matrix H f (x) is positive definite for all x M = f is strictly convex on M; 4. the Hessian matrix H f (x) is negative definite for all x M = f is strictly concave on M. Example 9 Theorem 7 Let f : M R, g : M R and M R n be a convex set. Then: 1. f, g are convex on M and a 0, b 0 = a f + b g is convex on M; 2. f, g are concave on M and a 0, b 0 = a f + b g is concave on M. Theorem 8 Let f : M R with M R n being convex and let F : D F R with R f D F. Then: 1. f is convex and F is convex and increasing = (F f)(x) = F (f(x)) is convex; 2. f is convex and F is concave and decreasing = (F f)(x) = F (f(x)) is concave; 3. f is concave and F is concave and increasing = (F f)(x) = F (f(x)) is concave; 4. f is concave and F is convex and decreasing = (F f)(x) = F (f(x)) is convex. Example 10 1.4 Quasi-convex and quasi-concave functions Definition 10 Let M R n be a convex set and f : M R. For any a R, the set P a = {x M f(x) a} is called an upper level set for f. Illustration: Upper level set

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 9 Theorem 9 Let M R n be a convex set and f : M R. Then: 1. If f is concave, then P a = {x M f(x) a} is a convex set for any a R; 2. If f is convex, then the lower level set P a = {x M f(x) a} is a convex set for any a R. Definition 11 Let M R n be a convex set and f : M R. Function f is called quasi-concave, if the upper level set P a = {x M f(x) a} is convex for any number a R. Function f is called quasi-convex, if f is quasi-concave. Remark: f quasi-convex the lower level set P a = {x M f(x) a} is convex for any a R Example 11 Remarks: 1. f convex = f quasi-convex f concave = f quasi-concave 2. The sum of quasi-convex (quasi-concave) functions is not necessarily quasi-convex (quasiconcave). Definition 12 Let M R n be a convex set and f : M R. Function f is called strictly quasi-concave, if f(λx 1 + (1 λ)x 2 ) > min{f(x 1 ), f(x 2 )} for all x 1, x 2 M with x 1 x 2 and λ (0, 1). Function f is strictly quasi-convex, if f is strictly quasi-concave.

CHAPTER 1. BASIC MATHEMATICAL CONCEPTS 10 Remarks: 1. f strictly quasi-concave = f quasi-concave 2. f : D f R, D f R, strictly increasing (decreasing) = f strictly quasi-concave 3. A strictly quasi-concave function cannot have more than one global maximum point. Theorem 10 Let f : D f R, D f R n, be twice continuously differentiable on a convex set M R n and 0 f x1 (x) f xr (x) f B r = x1 (x) f x1 x 1 (x) f x1 x r (x), r = 1, 2,..., n... f xr (x) f xrx1 (x) f xrxr (x) Then: 1. A necessary condition for f to be quasi-concave is that ( 1) r B r (x) 0 for all x M and all r = 1, 2,..., n; 2. A sufficient condition for f to be strictly quasi-concave is that ( 1) r B r (x) > 0 for all x M and all r = 1, 2,..., n. Example 12

Chapter 2 Unconstrained and constrained optimization 2.1 Extreme points Consider: f(x) min! (or max!) s.t. x M, where f : R n R, M R n M - set of feasible solutions x M - feasible solution f - objective function x i, i = 1, 2,..., n - decision variables (choice variables) often: M = {x R n g i (x) 0, i = 1, 2,..., m} where g i : R n R, i = 1, 2,..., m 2.1.1 Global extreme points Definition 1 A point x M is called a global minimum point for f in M if f(x ) f(x) for all x M. The number f := min{f(x) x M} is called the global minimum. 11

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 12 similarly: global maximum point global maximum (global) extreme point: (global) minimum or maximum point Theorem 1 (necessary first-order condition) Let f : M R be differentiable and x = (x 1, x 2,..., x n) be an interior point of M. A necessary condition for x to be an extreme point is f(x ) = 0, i.e., f x1 (x ) = f x2 (x ) = = f xn (x ) = 0. Remark: x is a stationary point for f Theorem 2 (sufficient condition) Let f : M R with M R n being a convex set. Then: 1. If f is convex on M, then: x is a (global) minimum point for f in M x is a stationary point for f; 2. If f is concave on M, then: x is a (global) maximum point for f in M x is a stationary point for f. Example 1 2.1.2 Local extreme points Definition 2 The set U ɛ (x ) := {x R n x x < ɛ} is called an (open) ɛ-neighborhood U ɛ (x ) with ɛ > 0.

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 13 Definition 3 A point x M is called a local minimum point for function f in M if there exists an ɛ > 0 such that f(x ) f(x) for all x M U ɛ (x ). The number f(x ) is called a local minimum. similarly: local maximum point local maximum (local) extreme point: (local) minimum or maximum point Illustration: Global and local minimum points Theorem 3 (necessary optimality condition) Let f be continuously differentiable and x be an interior point of M being a local minimum or maximum point. Then f(x ) = 0. Theorem 4 (sufficient optimality condition) Let f be twice continuously differentiable and x be an interior point of M. Then: 1. If f(x ) = 0 and H(x ) is positive definite, then x is a local minimum point. 2. If f(x ) = 0 and H(x ) is negative definite, then x is a local maximum point. Remarks: 1. If H(x ) is only positive (negative) semi-definite and f(x ) = 0, then the above condition is only necessary. 2. If x is a stationary point and H f (x ) 0 and neither of the conditions in (1) and (2) of Theorem 4 are satisfied, then x is a saddle point. The case H f (x ) = 0 requires further examination. Example 2

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 14 2.2 Equality constraints Consider: z = f(x 1, x 2,..., x n ) min! (or max!) s.t. g 1 (x 1, x 2,..., x n ) = 0 g 2 (x 1, x 2,..., x n ) = 0. g m (x 1, x 2,..., x n ) = 0 (m < n) apply Lagrange multiplier method: L(x; λ) = L(x 1, x 2,..., x n ; λ 1, λ 2,..., λ m ) = f(x 1, x 2,..., x n ) + L - Lagrangian function λ i - Lagrangian multiplier m λ i g i (x 1, x 2,..., x n ) i=1 Theorem 5 (necessary optimality condition, Lagrange s theorem) Let f and g i, i = 1, 2,..., m, be continuously differentiable, x 0 = (x 0 1, x0 2,..., x0 n) be a local extreme point subject to the given constraints and let J(x 0 1, x0 2,..., x0 n) 0. Then there exists a λ 0 = (λ 0 1, λ0 2,..., λ0 m) such that L(x 0 ; λ 0 ) = 0. The condition of Theorem 5 corresponds to L xj (x 0 ; λ 0 ) = 0, j = 1, 2,..., n; L λi (x 0 ; λ 0 ) = g i (x 1, x 2,..., x n ) = 0, i = 1, 2,..., m.

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 15 Theorem 6 (sufficient optimality condition) Let f and g i, i = 1, 2,..., m, be twice continuously differentiable and let (x 0 ; λ 0 ) with x 0 D f be a solution of the system L(x; λ) = 0. Moreover, let 0 0 L λ1 x 1 (x; λ) L λ1 x n (x; λ).... H L (x; λ) = 0 0 L λmx1 (x; λ) L λmxn (x; λ) L x1 λ 1 (x; λ) L x1 λ m (x; λ) L x1 x 1 (x; λ) L x1 x n (x; λ).... L xnλ1 (x; λ) L xnλm (x; λ) L xnx1 (x; λ) L xnxn (x; λ) be the bordered Hessian matrix and consider its leading principle minors D j (x 0 ; λ 0 ) of the order j = 2m + 1, 2m + 2,..., n + m at point (x 0 ; λ 0 ). Then: 1. If all D j (x 0 ; λ 0 ), 2m+1 j n+m, have the sign ( 1) m, then x 0 = (x 0 1, x0 2,..., x0 n) is a local minimum point of function f subject to the given constraints. 2. If all D j (x 0 ; λ 0 ), 2m + 1 j n + m, alternate in sign, the sign of D n+m (x 0 ; λ 0 ) being that of ( 1) n, then x 0 = (x 0 1, x0 2,..., x0 n) is a local maximum point of function f subject to the given constraints. 3. If neither the condition 1. nor those of 2. are satisfied, then x 0 is not a local extreme point of function f subject to the constraints. Here the case when one or several principle minors have value zero is not considered as a violation of condition 1. or 2. special case: n = 2, m = 1 = 2m + 1 = n + m = 3 = consider only D 3 (x 0 ; λ 0 ) D 3 (x 0 ; λ 0 ) < 0 = sign is ( 1) m = ( 1) 1 = 1 = x 0 is a local minimum point according to 1. D 3 (x 0 ; λ 0 ) > 0 = sign is ( 1) n = ( 1) 2 = 1 = x 0 is a local maximum point according to 2. Example 3 Theorem 7 (sufficient condition for global optimality) If there exist numbers (λ 0 1, λ0 2,..., λ0 m) = λ 0 and an x 0 D f such that L(x 0, λ 0 ) = 0, then: 1. If L(x) = f(x) + m i=1 λ 0 i g i(x) is concave in x, then x 0 is a maximum point. 2. If L(x) = f(x) + m λ 0 i g i(x) is convex in x, then x 0 is a minimum point. i=1

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 16 Example 4 2.3 Inequality constraints Consider: f(x 1, x 2,..., x n ) min! s.t. g 1 (x 1, x 2,..., x n ) 0 g 2 (x 1, x 2,..., x n ) 0 (3) g m (x 1, x 2,..., x n ) 0. where m = L(x; λ) = f(x 1, x 2,..., x n ) + λ i g i (x 1, x 2,..., x n ) = f(x) + λ T g(x), λ 1 λ m λ λ = 2. i=1 g 1 (x) g and g(x) = 2 (x). g m (x) Definition 4 A point (x ; λ ) is called a saddle point of the Lagrangian function L, if L(x ; λ) L(x ; λ ) L(x; λ ) (2.1) for all x R n, λ R m +. Theorem 8 If (x ; λ ) with λ 0 is a saddle point of L, then x is an optimal solution of problem (3). Question: Does any optimal solution correspond to a saddle point? additional assumptions required Slater condition (S): There exists a z R n such that for all nonlinear constraints g i inequality g i (z) < 0 is satisfied.

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 17 Remarks: 1. If all constraints g 1,..., g m are nonlinear, the Slater condition implies that the set M of feasible solutions contains interior points. 2. Condition (S) is one of the constraint qualifications. Theorem 9 (Theorem by Kuhn and Tucker) If condition (S) is satisfied, then x f(x) min! is an optimal solution of the convex problem s.t. g i (x) 0, i = 1, 2,..., m (4) f, g 1, g 2,..., g m convex functions if and only if L has a saddle point (x ; λ ) with λ 0. Remark: Condition (2.1) is often difficult to check. It is a global condition on the Lagrangian function. If all functions f, g 1,..., g m are continuously differentiable and convex, then the saddle point condition of Theorem 9 can be replaced by the following equivalent local conditions. Theorem 10 If condition (S) is satisfied and functions f, g 1,..., g m are continuously differentiable and convex, then x is an optimal solution of problem (4) if and only if the following Karush-Kuhn-Tucker (KKT)-conditions are satisfied. m f(x ) + λ i g i (x ) = 0 (2.2) i=1 λ i g i (x ) = 0 (2.3) g i (x ) 0 (2.4) λ i 0 (2.5) i = 1, 2,..., m Remark: Without convexity of the functions f, g 1,..., g m the KKT-conditions are only a necessary optimality condition, i.e.: If x is a local minimum point, condition (S) is satisfied and functions f, g 1,..., g m are continuously differentiable, then the KKT-conditions (2.2)-(2.5) are satisfied.

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 18 Summary: (x ; λ ) satisfies the KKT-conditions, problem is convex = x is a global minimum point x local minimum point, condition (S) is satisfied = KKT-conditions are satisfied Example 5 2.4 Non-negativity constraints Consider a problem with additional non-negativity constraints: f(x) min! s.t. g i (x) 0, i = 1, 2,..., m (5) x 0 Example 6 To find KKT-conditions for problem (5) introduce a Lagrangian multiplier µ j for any nonnegativity constraint x j 0 which corresponds to x j 0. KKT-conditions: m f(x ) + λ i g i (x ) µ = 0 (2.6) i=1 λ i g i (x ) = 0, i = 1, 2,..., m (2.7) µ j x j = 0, j = 1, 2,..., n (2.8) g i (x ) 0 (2.9) x 0, λ 0, µ 0 (2.10)

CHAPTER 2. UNCONSTRAINED AND CONSTRAINED OPTIMIZATION 19 Using (2.6) to (2.10), we can rewrite the KKT-conditions as follows: ( f x j (x ) + x j f(x ) + m λ i g i (x ) 0 i=1 λ i g i (x ) = 0, m i=1 λ i g i x j (x ) g i (x ) 0 x 0, λ 0 i.e., the new Lagrangian multipliers µ j have been eliminated. i = 1, 2,..., m ) = 0, j = 1, 2,..., n Example 7 Some comments on quasi-convex programming Theorem 11 Consider a problem (5), where function f is continuously differentiable and quasi-convex. Assume that there exist numbers λ 1, λ 2,..., λ m and a vector x such that 1. the KKT-conditions are satisfied; 2. f(x ) 0; 3. λ i g i(x) is quasi-convex for i = 1, 2,..., m. Then x is optimal for problem (5). Remark: Theorem 11 holds analogously for problem (3).

Chapter 3 Sensitivity analysis 3.1 Preliminaries Question: How does a change in the parameters affect the solution of an optimization problem? sensitivity analysis (in optimization) comparative statics (or dynamics) (in economics) Example 1 3.2 Value functions and envelope results 3.2.1 Equality constraints Consider: f(x; r) min! s.t. g i (x; r) = 0, i = 1, 2,..., m (6) where r = (r 1, r 2,..., r k ) T - vector of parameters Remark: In (6), we optimize w.r.t. x with r held constant. Notations: x 1 (r), x 2 (r),..., x n (r) - optimal solution in dependence on r f (r) = f(x 1 (r), x 2 (r),..., x n (r)) - (minimum) value function 20

CHAPTER 3. SENSITIVITY ANALYSIS 21 λ i (r) (i = 1, 2,..., m) - Lagrangian multipliers in the necessary optimality condition Lagrangian function: m L(x; λ; r) = f(x; r) + λ i g i (x; r) i=1 = f(x(r); r) + m λ i (r) g i (x(r); r) = L (r) i=1 Theorem 1 (Envelope Theorem for equality constraints) For j = 1, 2,..., k, we have: f (r) r j = ( ) L(x; λ; r) r j ( x(r) λ(r)) = L (r) r j Remark: Notice that L r j measures the total effect of a change in r j on the Lagrangian function, while L r j measures the partial effect of a change in r j on the Lagrangian function with x and λ being held constant. Example 2 3.2.2 Properties of the value function for inequality constraints Consider: f(x, r) min! s.t. g i (x, r) 0, i = 1, 2,..., m minimum value function: b f (b) x(b) - optimal solution f (b) = min{f(x) g i (x) b i 0, i = 1, 2,..., m} λ i (b) - corresponding Lagrangian multipliers Remark: = f (b) b i = λ i (b), i = 1, 2,..., m Function f is not necessarily continuously differentiable.

CHAPTER 3. SENSITIVITY ANALYSIS 22 Theorem 2 If function f(x) is concave and functions g 1 (x), g 2 (x),..., g m (x) are convex, then function f (b) is concave. Example 3: A firm has L units of labour available and produces 3 goods whose values per unit of output are a, b and c, respectively. Producing x, y and z units of the goods requires αx 2, βy 2 and γz 2 units of labour, respectively. We maximize the value of output and determine the value function. 3.2.3 Mixed constraints Consider: f(x, r) min! s.t. x M(r) = {x R n g i (x, r) 0, i = 1, 2,..., m ; g i (x, r) = 0, i = m + 1, m + 2,..., m} (minimum) value function: f (r) = min {f(x, r) = f(x 1 (r), x 2 (r),..., x n (r)) x M(r)} Lagrangian function: m L(x; λ; r) = f(x; r) + λ i g i (x; r) i=1 = f(x(r); r) + m λ i (r) g i (x(r); r) = L (r) i=1 Theorem 3 (Envelope Theorem for mixed constraints) For j = 1, 2,..., k, we have: f (r) r j = ( ) L(x; λ; r) r j ( x(r) λ(r)) = L (r) r j Example 4

CHAPTER 3. SENSITIVITY ANALYSIS 23 3.3 Some further microeconomic applications 3.3.1 Cost minimization problem Consider: C(w, x) = w T x(w, y) min! s.t. y f(x) 0 x 0, y 0 Assume that w > 0 and that the partial derivatives of C are > 0. Let x(w, y) be the optimal input vector and λ(w, y) be the corresponding Lagrangian multiplier. L(x; λ; w, y) = w T x + λ (y f(x)) i.e., λ signifies marginal costs = C y = L = λ = λ(w, y) (3.1) y Shepard(-McKenzie) Lemma: Remark: C w i = x i = x i (w, y), i = 1, 2,..., n (3.2) Assume that C is twice continuously differentiable. Then the Hessian H C is symmetric. Differentiating (3.1) w.r.t. w i and (3.2) w.r.t. y, we obtain Samuelson s reciprocity relation: = x j w i = x i w j and x i y = λ, w i for all i and j Interpretation of the first result: A change in the j-th factor input w.r.t. a change in the i-th factor price (output being constant) must be equal to the change in the i-th factor input w.r.t. a change in the j-th factor price.

CHAPTER 3. SENSITIVITY ANALYSIS 24 3.3.2 Profit maximization problem of a competitive firm Consider: π(x, y) = p T y w T x max! ( π min!) s.t. g(x, y) = y f(x) 0 x 0, y 0, where: p > 0 - output price vector w > 0 - input price vector y R m + - produced vector of output x R n + - used input vector f(x) - production function Let: x(p, w), y(p, w) be the optimal solutions of the problem and π(p, w) = p T y(p, w) w T x(p, w) be the (maximum) profit function. L(x, y; λ; p, w) = p T y + w T x + λ (y f(x)) The Envelope theorem implies Hotelling s lemma: 1. ( π) p i = L p i = y i i.e.: π p i = y i > 0, i = 1, 2,..., m (3.3) 2. ( π) w i = L w i = x i i.e.: π w i = x i < 0, i = 1, 2,..., m (3.4) Interpretation: 1. An increase in the price of any output increases the maximum profit. 2. An increase in the price of any input lowers the maximum profit. Remark: Let π(p, w) be twice continuously differentiable. Using (3.3) and (3.4), we obtain Hotelling s symmetry relation: y j p i = y i p j, x j w i = x i w j, x j p i = y i w j, for all i and j.

Chapter 4 Differential equations 4.1 Preliminaries Definition 1 A relationship F (x, y, y, y,..., y (n) ) = 0 between the independent variable x, a function y(x) and its derivatives is called an ordinary differential equation. The order of the differential equation is determined by the highest order of the derivatives appearing in the differential equation. Explicit representation: y (n) = f(x, y, y, y,..., y (n 1) ) Example 1 Definition 2 A function y(x) for which the relationship F (x, y, y, y,..., y (n) ) = 0 holds for all x D y is called a solution of the differential equation. The set S = {y(x) F (x, y, y, y,..., y (n) ) = 0 for all x D y } is called the set of solutions or the general solution of the differential equation. in economics often: time t is the independent variable, solution x(t) with ẋ = dx dt, ẍ = d2 x dt 2, etc. 25

CHAPTER 4. DIFFERENTIAL EQUATIONS 26 4.2 Differential equations of the first order implicit form: explicit form: F (t, x, ẋ) = 0 ẋ = f(t, x) Graphical solution: given: ẋ = f(t, x) At any point (t 0, x 0 ) the value ẋ = f(t 0, x 0 ) is given, which corresponds to the slope of the tangent at point (t 0, x 0 ). graph the direction field (or slope field) Example 2 4.2.1 Separable equations solve for x (if possible) x(t 0 ) = x 0 given: C is assigned a particular value = x p - particular solution ẋ = f(t, x) = g(t) h(x) dx = h(x) = g(t) dt = H(x) = G(t) + C Example 3 Example 4 4.2.2 First-order linear differential equations ẋ + a(t) x = q(t) q(t) - forcing term

CHAPTER 4. DIFFERENTIAL EQUATIONS 27 (a) a(t) = a and q(t) = q multiply both sides by the integrating factor e at > 0 = ẋe at + axe at = qe at d = dt (x eat ) = qe at = x e at = qe at dt = q a eat + C i.e. ẋ + ax = q x = Ce at + q a (C R) (4.1) C = 0 = x(t) = q a = constant x = q a - equilibrium or stationary state Remark: The equilibrium state can be obtained by letting ẋ = 0 and solving the remaining equation for x. If a > 0, then x = Ce at + q a converges to q a as t, and the equation is said to be stable (every solution converges to an equilibrium as t ). Example 5 (b) a(t) = a and q(t) multiply both sides by the integrating factor e at > 0 = ẋe at + axe at = q(t) e at = d dt (x eat ) = q(t) e at = x e at = q(t) e at dt + C i.e. ẋ + ax = q(t) x = Ce at + e at e at q(t)dt (4.2) (c) General case multiply both sides by e A(t) = ẋe A(t) + a(t)xe A(t) = q(t) e A(t)

CHAPTER 4. DIFFERENTIAL EQUATIONS 28 choose A(t) such that A(t) = a(t)dt because d dt (x ea(t) ) = ẋ e A(t) + x A(t) }{{} a(t) e A(t) = x e A(t) = = x = Ce A(t) + e A(t) q(t) e A(t) dt + C e A(t) q(t) e A(t) dt, where A(t) = a(t)dt Example 6 (d) Stability and phase diagrams Consider an autonomous (i.e. time-independent) equation and a phase diagram: ẋ = F (x) (4.3) Illustration: Phase diagram Definition 3 A point a represents an equilibrium or stationary state for equation (4.3) if F (a) = 0. = x(t) = a is a solution if x(t 0 ) = x 0. = x(t) converges to x = a for any starting point (t 0, x 0 ). Illustration: Stability 4.3 Second-order linear differential equations and systems in the plane Homogeneous differential equation: ẍ + a(t)ẋ + b(t)x q(t) (4.4) q(t) 0 = ẍ + a(t)ẋ + b(t)x = 0 (4.5)

CHAPTER 4. DIFFERENTIAL EQUATIONS 29 Theorem 1 The homogeneous differential equation (4.5) has the general solution x H (t) = C 1 x 1 (t) + C 2 x 2 (t), C 1, C 2 R where x 1 (t), x 2 (t) are two solutions that are not proportional (i.e., linearly independent). The non-homogeneous equation (4.4) has the general solution x(t) = x H (t) + x N (t) = C 1 x 1 (t) + C 2 x 2 (t) + x N (t), where x N (t) is any particular solution of the non-homogeneous equation. (a) Constant coefficients a(t) = a and b(t) = b ẍ + aẋ + bx = q(t) Homogeneous equation: ẍ + aẋ + bx = 0 use the setting x(t) = e λt (λ R) = ẋ(t) = λe λt, ẍ(t) = λ 2 e λt = Characteristic equation: λ 2 + aλ + b = 0 (4.6) 3 cases: 1. (4.6) has two distinct real roots λ 1, λ 2 = x H (t) = C 1 e λ 1t + C 2 e λ 2t 2. (4.6) has a real double root λ 1 = λ 2 = x H (t) = C 1 e λ 1t + C 2 te λ 1t 3. (4.6) has two complex roots λ 1 = α + β i and λ 2 = α β i x H (t) = e αt (C 1 cos βt + C 2 sin βt) Non-homogeneous equation: ẍ + aẋ + bx = q(t)

CHAPTER 4. DIFFERENTIAL EQUATIONS 30 Discussion of special forcing terms: Forcing term q(t) Setting x N (t) 1. q(t) = p e st (a) x N (t) = A e st - if s is not a root of the characteristic equation (b) x N (t) = A t k e st - if s is a root of multiplicity k (k 2) of the characteristic equation 2. q(t) = p n t n +p n 1 t n 1 + +p 1 t+p 0 (a) x N (t) = A n t n + A n 1 t n 1 + + A 1 t + A 0 - if b 0 in the homogeneous equation (b) x N (t) = t k (A n t n + A n 1 t n 1 + + A 1 t + A 0 ) - with k = 1 if a 0, b = 0 and k = 2 if a = b = 0 3. q(t) = p cos st + r sin st (a) x N (t) = A cos st + B sin st - if si is not a root of the characteristic equation (b) x N (t) = t k (A cos st + B sin st) - if si is a root of multiplicity k of the characteristic equation Use the above setting and insert it and the derivatives into the non-homogeneous equation. Determine the coefficients A, B and A i, respectively. Example 7 (b) Stability Consider equation (4.4) Definition 4 Equation (4.4) is called globally asymptotically stable if every solution x H (t) = C 1 x 1 (t) + C 2 x 2 (t) of the associated homogeneous equation tends to 0 as t for all values of C 1 and C 2. Remark: x H (t) 0 as t x 1 (t) 0 and x 2 (t) 0 as t Example 8

CHAPTER 4. DIFFERENTIAL EQUATIONS 31 Theorem 2 Equation ẍ+aẋ+bx = q(t) is globally asymptotically stable if and only if a > 0 and b > 0. (c) Systems of equations in the plane Consider: ẋ = f(t, x, y) ẏ = g(t, x, y) (7) Solution: pair (x(t), y(t)) satisfying (7) Initial value problem: The initial conditions x(t 0 ) = x 0 and y(t 0 ) = y 0 are given. A solution method: Reduce the given system (7) to a second-order differential equation in only one unknown. 1. Use the first equation in (7) to express y as a function of t, x, ẋ. y = h(t, x, ẋ) 2. Differentiate y w.r.t. t and substitute the terms for y and ẏ into the second equation in (7). 3. Solve the resulting second-order differential equation to determine x(t). 4. Determine y(t) = h(t, x(t), ẋ(t)) Example 9 (d) Systems with constant coefficients Consider: ẋ = a 11 x + a 12 y + q 1 (t) Solution of the homogeneous system: ( ) ẋ ẏ = a 21 x + a 22 y + q 2 (t) ẏ ( ) ( ) a 11 a 12 x = a 21 a 22 y

CHAPTER 4. DIFFERENTIAL EQUATIONS 32 we set ( ) ( ) x z 1 = y z 2 ( ) ( ẋ = = λ ẏ = we obtain the eigenvalue problem: ( ) ( a 11 a 12 a 21 a 22 or equivalently ( ) ( a 11 λ a 12 a 21 a 22 λ z 1 z 2 ) e λt z 1 z 2 = λ z 1 z 2 ) ( e λt z 1 z 2 ) ) ( ) 0 = 0 Determine the eigenvalues λ 1, λ 2 and the corresponding eigenvectors ( ) ( ) z 1 = z 1 1 z 1 2 and z 2 = z 2 1 z 2 2. Consider now the cases in a similar way as for a second-order differential equation, e.g. λ 1 R, λ 2 R and λ 1 λ 2. = General solution: ( ) ( x H (t) = C 1 y H (t) ) e λ 1t + C 2 ( z1 1 z1 2 z2 1 z2 2 ) e λ 2t Solution of the non-homogeneous system: A particular solution of the non-homogeneous system can be determined in a similar way as for a second-order differential equation. Note that all occurring specific functions q 1 (t) and q 2 (t) have to be considered in each function x N (t) and y N (t). Example 10

CHAPTER 4. DIFFERENTIAL EQUATIONS 33 (e) Equilibrium points for linear systems with constant coefficients and forcing term Consider: ẋ = a 11 x + a 12 y + q 1 ẏ = a 21 x + a 22 y + q 2 For finding an equilibrium point (state), we set ẋ = ẏ = 0 and obtain a 11 x + a 12 y = q 1 a 21 x + a 22 y = q 2 Cramer s rule = equilibrium point: q 1 a 12 x q 2 a 22 = a 11 a 12 a 21 a 22 a 11 q 1 y a 21 q 2 = a 11 a 12 a 21 a 22 = a 12q 2 a 22 q 1 A = a 21q 1 a 11 q 2 A Example 11 Theorem 3 Suppose that A 0. Then the equilibrium point (x, y ) for the linear system ẋ = a 11 x + a 12 y + q 1 ẏ = a 21 x + a 22 y + q 2 is globally asymptotically stable if and only if a 11 a 12 tr(a) = a 11 + a 22 < 0 and A = a 21 a 22 > 0, where tr(a) is the trace of A (or equivalently, if and only if both eigenvalues of A have negative real parts). Example 12

CHAPTER 4. DIFFERENTIAL EQUATIONS 34 (f) Phase plane analysis Consider an autonomous system: ẋ = f(x, y) ẏ = g(x, y) Rates of change of x(t) and y(t) are given by f(x(t), y(t)) and g(x(t), y(t)), e.g. if f(x(t), y(t)) > 0 and g(x(t), y(t)) < 0 at a point P = (x(t), y(t)), then (as t increases) the system will move from point P down and to the right. = (ẋ(t), ẏ(t)) gives direction of motion, length of (ẋ(t), ẏ(t)) gives speed of motion Illustration: Motion of a system Graph a sample of these vectors. = phase diagram Equilibrium point: point (a, b) with f(a, b) = g(a, b) = 0 equilibrium points are the points of the intersection of the nullclines f(x, y) = 0 and g(x, y) = 0 Graph the nullclines: At point P with f(x, y) = 0, ẋ = 0 and the velocity vector is vertical, it points up if ẏ > 0 and down if ẏ < 0. At point Q with g(x, y) = 0, ẏ = 0 and the velocity vector is horizontal, it points to the right if ẋ > 0 and to the left if ẋ < 0. Continue and graph further arrows. Example 13

Chapter 5 Optimal control theory 5.1 Calculus of variations Consider: s.t. t 1 t 0 F (t, x, ẋ)dt max! x(t 0 ) = x 0, x(t 1 ) = x 1 (8) Illustration necessary optimality condition: Function x(t) can only solve problem (8) if x(t) satisfies the following differential equation. Euler equation: F x d ( ) F = 0 (5.1) dt ẋ we have d dt ( ) F (t, x, ẋ) ẋ = 2 F t ẋ + 2 F x ẋ ẋ + 2 F ẋ ẋ ẍ = (5.1) can be rewritten as 2 F ẋ ẋ ẍ + 2 F x ẋ ẋ + 2 F t ẋ F x = 0 35

CHAPTER 5. OPTIMAL CONTROL THEORY 36 Theorem 1 If F (t, x, ẋ) is concave in (x, ẋ), a feasible x (t) that satisfies the Euler equation solves the maximization problem (8). Example 1 More general terminal conditions Consider: s.t. t 1 F (t, x, ẋ)dt max! t 0 x(t 0 ) = x 0 (a) x(t 1 ) free or (b) x(t 1 ) x 1 (9) Illustration = transversality condition needed to determine the second constant Theorem 2 (Transversality conditions) If x (t) solves problem (9) with either (a) or (b) as the terminal condition, then x (t) must satisfy the Euler equation. With the terminal condition (a), the transversality condition is ( ) F = 0. (5.2) ẋ t=t 1 With the terminal condition (b), the transversality condition is ( ) [ F ( F ) ] 0 = 0, if x (t 1 ) > x 1 ẋ t=t 1 ẋ t=t 1 (5.3) If F (t, x, ẋ) is concave in (x, ẋ), then a feasible x (t) that satisfies both the Euler equation and the appropriate transversality condition will solve problem (9). Example 2

CHAPTER 5. OPTIMAL CONTROL THEORY 37 5.2 Control theory 5.2.1 Basic problem Let: x(t) - characterization of the state of a system u(t) - control function; t t 0 t 1 J = f(t, x(t), u(t))dt - objective function t 0 Given: ẋ(t) = g(t, x(t), u(t)), x(t 0 ) = x 0 (10) Problem: Among all pairs (x(t), u(t)) that obey (10) find one such that t 1 J = f(t, x(t), u(t))dt max! t 0 Example 3 Optimality conditions: Consider: s.t. t 1 J = f(t, x(t), u(t))dt max! (5.4) t 0 ẋ(t) = g(t, x(t), u(t)), x(t 0 ) = x 0, x(t 1 ) free (5.5) Introduce the Hamiltonian function p = p(t) - costate variable (adjoint function) H(t, x, u, p) = f(t, x, u) + p g(t, x, u)

CHAPTER 5. OPTIMAL CONTROL THEORY 38 Theorem 3 (Maximum principle) Suppose that (x (t), u (t)) is an optimal pair for problem (5.4) - (5.5). Then there exists a continuous function p(t) such that 1. u = u (t) maximizes H(t, x (t), u, p(t)) for u (, ) (5.6) 2. ṗ(t) = H x (t, x (t), u (t), p(t)), p(t 1 ) = 0 }{{} transversality condition (5.7) Theorem 4 If the condition H(t, x, u, p(t)) is concave in (x, u) for each t [t 0, t 1 ] (5.8) is added to the conditions in Theorem 3, we obtain a sufficient optimality condition, i.e., if we find a triple (x (t), u (t), p (t)) that satisfies (5.5), (5.6), (5.7) and (5.8), then (x (t), u (t)) is optimal. Example 4 5.2.2 Standard problem Consider the standard end constrained problem : t 1 t 0 f(t, x, u)dt max!, u U R (5.9) s.t. ẋ(t) = g(t, x(t), u(t)), x(t 0 ) = x 0 (5.10) with one of the following terminal conditions (a) x(t 1 ) = x 1, (b) x(t 1 ) x 1 or (c) x(t 1 ) free. (5.11) Define now the Hamiltonian function as follows: H(t, x, u, p) = p 0 f(t, x, u) + p g(t, x, u)

CHAPTER 5. OPTIMAL CONTROL THEORY 39 Theorem 5 (Maximum principle for standard end constraints) Suppose that (x (t), u (t)) is an optimal pair for problem (5.9) - (5.11). Then there exist a continuous function p(t) and a number p 0 {0, 1} such that for all t [t 0, t 1 ] we have (p 0, p(t)) (0, 0) and, moreover: 1. u = u (t) maximizes the Hamiltonian H(t, x (t), u, p(t)) w.r.t. u U, i.e., H(t, x (t), u, p(t)) H(t, x (t), u (t), p(t)) for all u U 2. ṗ(t) = H x (t, x (t), u (t), p(t)) (5.12) 3. Corresponding to each of the terminal conditions (a), (b) and (c) in (5.11), there is a transversality condition on p(t 1 ): (a ) no condition on p(t 1 ) (b ) p(t 1 ) 0 (with p(t 1 ) = 0 if x (t 1 ) > x 1 ) (c ) p(t 1 ) = 0 Theorem 6 (Mangasarian) Suppose that (x (t), u (t)) is a feasible pair with the corresponding costate variable p(t) such that conditions 1. - 3. in Theorem 5 are satisfied with p 0 = 1. Suppose further that the control region U is convex and that H(t, x, u, p(t)) is concave in (x, u) for every t [t 0, t 1 ]. Then (x (t), u (t)) is an optimal pair. General approach: 1. For each triple (t, x, p) maximize H(t, x, u, p) w.r.t. u U (often there exists a unique maximization point u = û(t, x, p)). 2. Insert this function into the differential equations (5.10) and (5.12) to obtain ẋ(t) = g(t, x, û(t, x(t), p(t))) and ṗ(t) = H x (t, x(t), û(t, x(t), p(t))) (i.e., a system of two first-order differential equations) to determine x(t) and p(t). 3. Determine the constants in the general solution (x(t), p(t)) by combining the initial condition x(t 0 ) = x 0 with the terminal conditions and transversality conditions. = state variable x (t), corresponding control variable u (t) = û(t, x (t), p(t))

CHAPTER 5. OPTIMAL CONTROL THEORY 40 Remarks: 1. If the Hamiltonian is not concave, there exists a weaker sufficient condition due to Arrow: If the maximized Hamiltonian Ĥ(t, x, p) = max u H(t, x, u, p) is concave in x for every t [t 0, t 1 ] and conditions 1. - 3. of Theorem 5 are satisfied with p 0 = 1, then (x (t), u (t)) solves problem (5.9) - (5.11). (Arrow s sufficient condition) 2. If the resulting differential equations are non-linear, one may linearize these functions about the equilibrium state, i.e., one can expand the functions into Taylor polynomials with n = 1 (see linear approximation in Section 1.1). Example 5 5.2.3 Current value formulations Consider: max u U R t 1 t 0 f(t, x, u)e rt dt, ẋ = g(t, x, u) x(t 0 ) = x 0 (a) x(t 1 ) = x 1 (b) x(t 1 ) x 1 or (c) x(t 1 ) free (11) e rt - discount factor = Hamiltonian H = p 0 f(t, x, u)e rt + p g(t, x, u) = Current value Hamiltonian (multiply H by e rt ) H c = He rt = p 0 f(t, x, u) + e rt p g(t, x, u) λ = e rt p - current value shadow price, λ 0 = p 0 = H c (t, x, u, λ) = λ 0 f(t, x, u) + λ g(t, x, u)

CHAPTER 5. OPTIMAL CONTROL THEORY 41 Theorem 7 (Maximum principle, current value formulation) Suppose that (x (t), u (t)) is an optimal pair for problem (11) and let H c be the current value Hamiltonian. Then there exist a continuous function λ(t) and a number λ 0 {0, 1} such that for all t [t 0, t 1 ] we have (λ 0, λ(t)) (0, 0) and, moreover: 1. u = u (t) maximizes H c (t, x (t), u, λ(t)) for u U 2. λ(t) rλ(t) = Hc (t, x (t), u (t), λ(t)) x 3. The transversality conditions are: (a ) no condition on λ(t 1 ) (b ) λ(t 1 ) 0 (with λ(t 1 ) = 0 if x (t 1 ) > x 1 ) (c ) λ(t 1 ) = 0 Remark: The conditions in Theorem 7 are sufficient for optimality if λ 0 = 1 and H c (t, x, u, λ(t)) is concave in (x, u) (Mangasarian) or more generally Ĥ c (t, x, λ(t)) = max u U Hc (t, x, u, λ(t)) is concave in x (Arrow). Example 6 Remark: If explicit solutions for the system of differential equations are not obtainable, a phase diagram may be helpful. Illustration: Phase diagram for example 6