ODE Final exam - Solutions

Similar documents
1 Lyapunov theory of stability

Final exam (practice) UCLA: Math 31B, Spring 2017

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Math 273 (51) - Final

Taylor and Maclaurin Series

Principle of Mathematical Induction

Taylor Series. Math114. March 1, Department of Mathematics, University of Kentucky. Math114 Lecture 18 1/ 13

Old Math 330 Exams. David M. McClendon. Department of Mathematics Ferris State University

ENGI 9420 Lecture Notes 4 - Stability Analysis Page Stability Analysis for Non-linear Ordinary Differential Equations

MATH 1A, Complete Lecture Notes. Fedor Duzhin

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

California State University Northridge MATH 280: Applied Differential Equations Midterm Exam 1

Half of Final Exam Name: Practice Problems October 28, 2014

Econ Lecture 14. Outline

Name (print): Question 4. exercise 1.24 (compute the union, then the intersection of two sets)

Math 113 (Calculus 2) Exam 4

Mathematics 104 Fall Term 2006 Solutions to Final Exam. sin(ln t) dt = e x sin(x) dx.

Main topics and some repetition exercises for the course MMG511/MVE161 ODE and mathematical modeling in year Main topics in the course:

INTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES

Asymptotic Stability by Linearization

Short Solutions to Review Material for Test #2 MATH 3200

n=0 ( 1)n /(n + 1) converges, but not

Math 308 Final Exam Practice Problems

Part 3.3 Differentiation Taylor Polynomials

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson

Econ 204 Differential Equations. 1 Existence and Uniqueness of Solutions

Solution of Linear State-space Systems

Solutions to Math 53 Math 53 Practice Final

Chapter 8: Taylor s theorem and L Hospital s rule

Practice Final Exam Solutions for Calculus II, Math 1502, December 5, 2013

Class Meeting # 2: The Diffusion (aka Heat) Equation

8.7 MacLaurin Polynomials

1.4 The Jacobian of a map

Math Ordinary Differential Equations

Elementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4.

3. On the grid below, sketch and label graphs of the following functions: y = sin x, y = cos x, and y = sin(x π/2). π/2 π 3π/2 2π 5π/2

Linear ODE s with periodic coefficients

Richard F. Bass Krzysztof Burdzy University of Washington

Iowa State University. Instructor: Alex Roitershtein Summer Homework #5. Solutions

Course 216: Ordinary Differential Equations

Infinite series, improper integrals, and Taylor series

DIFFERENTIAL EQUATIONS REVIEW. Here are notes to special make-up discussion 35 on November 21, in case you couldn t make it.

Section 8.7. Taylor and MacLaurin Series. (1) Definitions, (2) Common Maclaurin Series, (3) Taylor Polynomials, (4) Applications.

8 Periodic Linear Di erential Equations - Floquet Theory

Chapter 3a Topics in differentiation. Problems in differentiation. Problems in differentiation. LC Abueg: mathematical economics

Infinite series, improper integrals, and Taylor series

First order differential equations

Stability of Stochastic Differential Equations

3. Identify and find the general solution of each of the following first order differential equations.

Math 106: Review for Exam II - SOLUTIONS

1+t 2 (l) y = 2xy 3 (m) x = 2tx + 1 (n) x = 2tx + t (o) y = 1 + y (p) y = ty (q) y =

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

Polytechnic Institute of NYU MA 2132 Final Practice Answers Fall 2012

Math 312 Lecture Notes Linear Two-dimensional Systems of Differential Equations

Math 310 Introduction to Ordinary Differential Equations Final Examination August 9, Instructor: John Stockie

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

2nd-Order Linear Equations

2.3. VECTOR SPACES 25

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

Calculus of Variations. Final Examination

6 Linear Equation. 6.1 Equation with constant coefficients

ODEs Cathal Ormond 1

Lecture 4: Numerical solution of ordinary differential equations

Preliminary Examination, Numerical Analysis, August 2016

LMI Methods in Optimal and Robust Control

C-N M151 Lecture Notes (part 1) Based on Stewart s Calculus (2013) B. A. Starnes

THE NATIONAL UNIVERSITY OF IRELAND, CORK COLÁISTE NA hollscoile, CORCAIGH UNIVERSITY COLLEGE, CORK. Summer Examination 2009.

Math 164 (Lec 1): Optimization Instructor: Alpár R. Mészáros

Calculus C (ordinary differential equations)

The Steepest Descent Algorithm for Unconstrained Optimization

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Chapter #4 EEE8086-EEE8115. Robust and Adaptive Control Systems

MATH 205C: STATIONARY PHASE LEMMA

Final exam (practice) UCLA: Math 31B, Spring 2017

Math 164-1: Optimization Instructor: Alpár R. Mészáros

SECTION A. f(x) = ln(x). Sketch the graph of y = f(x), indicating the coordinates of any points where the graph crosses the axes.

ẋ = f(x, y), ẏ = g(x, y), (x, y) D, can only have periodic solutions if (f,g) changes sign in D or if (f,g)=0in D.

Question: Total. Points:

MATH 23 Exam 2 Review Solutions

Solution: In standard form (i.e. y + P (t)y = Q(t)) we have y t y = cos(t)

Mathematic 108, Fall 2015: Solutions to assignment #7

5.9 Representations of Functions as a Power Series

Nonlinear Control. Nonlinear Control Lecture # 2 Stability of Equilibrium Points

Math Exam 2, October 14, 2008

In these chapter 2A notes write vectors in boldface to reduce the ambiguity of the notation.

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

MB4018 Differential equations

Linear ODEs. Existence of solutions to linear IVPs. Resolvent matrix. Autonomous linear systems

Notes on uniform convergence

MAS113 CALCULUS II SPRING 2008, QUIZ 5 SOLUTIONS. x 2 dx = 3y + y 3 = x 3 + c. It can be easily verified that the differential equation is exact, as

MATH 307 Introduction to Differential Equations Autumn 2017 Midterm Exam Monday November

INFINITE SEQUENCES AND SERIES

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim

ANALYSIS QUALIFYING EXAM FALL 2017: SOLUTIONS. 1 cos(nx) lim. n 2 x 2. g n (x) = 1 cos(nx) n 2 x 2. x 2.

ON SOME CONSTANTS FOR OSCILLATION AND STABILITY OF DELAY EQUATIONS

SOLUTION GUIDE TO MATH GRE FORM GR9367

Assignment # 8, Math 370, Fall 2018 SOLUTIONS:

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote

Math 113 Winter 2005 Key

Transcription:

ODE Final exam - Solutions May 3, 018 1 Computational questions (30 For all the following ODE s with given initial condition, find the expression of the solution as a function of the time variable t You do not have to justify existence, uniqueness, or to worry about the time of existence of the solutions, but you need to explain your computations 1 3 4 5 6 x = t, x(0 = 1 3 + x x x + x = 0, x(0 = 1, x (0 = 0 X = ( 0 1 1 0 ( 1 X, X(0 = 0 x = t + tx, x(0 = 0 x x = e t, x(0 = 1 tx = x + te x/t, x(1 = 1 1 As long as x(s 3 we may write x (3 + x(s = s Integrating between s = 0 and s = t and using the initial condition x(0 = 1, we obtain 3x(t + 1 x (t (3 + 1 = 1 t Hence we get so we must have x (t + 6x(t (7 + t = 0, x(t = 6 ± 36 + 4(7 + t 1

Since x(0 = 1 and x cannot cross 3, the correct solution is x(t = 6 + 36 + 4(7 + t Let us consider the associated characteristic polynomial x x + 1 = 0 its roots are λ ± = 1±i 3 Thus we know that the general solution takes the form x(t = Ae tλ + + Be tλ, where A, B are coefficients to be determined We have x(0 = 1 = A + B, x (0 = 0 = λ + A + λ B We obtain thus A + B = 1, A B = i 3, A = 1 + i 3, B = 1 i 3 3 The eigenvalues of the matrix are easily seen to be ±1, with eigenvectors ( 1 ±1 We do the standard thing: changing basis, computing the exponential of a diagonal matrix, etc 4 We write and integrate to get x 1 + x = t, tan 1 (x(t tan 1 (x(0 = t, thus (since x(0 = 0 we have ( t x(t = tan 5 The associated homogeneous equation has general solution t Ae t for A arbitrary Since the right-hand side is also of this form, we look for a particular solution as t ate t We find a = 1 Hence the general solution of the ODE is (t + Ce t with C arbitrary The initial condition implies that C = 1, thus we obtain x(t = (t + 1e t

6 This is an homogeneous equation (with the other meaning of homogeneous For t 0 we write it as x x t ex/t = 0 Introducing z = x t we have x = z + tz and thus z + tz z e z = 0, so z e z = 1 t, which gives, after integrating between 1 and t e z(t + e z(1 = ln(t We have x(1 = 1 hence z(1 = 1 and we get e z(t = e ln(t, thus z(t = ln(e ln(t and finally x(t = t ln(e ln(t 3

Time of existence (0 We consider the ODE x = t x + (1 + cos (tx We denote by γ the maximal solution of this ODE with initial condition γ(0 = 1, defined on some interval (α, β 1 Show that γ(t is always positive for t in (α, β Show that γ is increasing on (α, β 3 Justify that α = 4 Justify that β is finite You may use the ODE x = x for comparison 1 We observe that x(t 0 is solution to the ODE By the uniqueness part of Cauchy- Lipschitz theorem, any solution that vanishes for a certain time must thus be the zero solution In other words, any non-zero solution has a constant sign Since γ(0 = 1, we deduce that γ stays positive for all times Since γ(t is always positive, the right-hand side of the ODE is positive and thus γ (t is always positive, hence γ is increasing 3 The function t γ(t is continuous, increasing, and bounded below by 0 Thus for any t (α, 0 we have 0 γ(t 1 and this negates the blow up in finite time criterion, thus α = 4 Since γ is positive, we have, for t (α, β γ (t = t γ(t + (1 + cos (tγ (t γ (t In particular, integrating between 0 and t for t (α, β we obtain so γ(t 1 1 t 1 1 γ(t t,, which proves that γ blows-up in finite time 4

3 Qualitative study (30 We consider the ODE x sin(x = 0 (1 1 Re-write (1 as a first-order ODE with unknown function X = (x, x Find a (non-trivial conserved quantity 3 Sketch the allure of the orbits in R with the system of coordinates (x, x (a Near the point (π/, 0 (b Near the point (0, (Briefly justify your drawing 4 Explain why we can find a change of variables that would transform the two sketches drawn in the previous question onto one another CANCELLED Explain why, near the point (0, 0, the flow of this ODE looks like ( e ta 0 1, A = 1 0 5 Sketch the allure of the orbits near (0, 0 1 X = (x, x = (x, sin(x We apply the usual trick, multiplying by x yields x x sin(xx = 0, and thus Q(x, x := (x + cos(x is conserved 3 The orbits are contained in the level sets of Q We can use the fact that to sketch these level sets cos(x (x π/ near π/, cos(x 1 x / near 0 4 We are considering two points that are not stationary points By the straightening of vector fields theorem, the flow near both points can be mapped onto the flow of a constant vector field (ie an ODE with constant speed By transitivity, the two flows can be mapped onto each other 5 That would have been an application of Grobman-Hartman 6 We use again the conserved quantity and the approximation cos(x 1 x / near 0 5

4 Around the gradient descent Let d 1 be the dimension In this problem, E is a function from R d to R of class C 1 such that: E is Lipschitz, with a Lipschitz constant denoted by L By definition, it means x, y R d, E(x E(y L x y You may use the following consequence: x R d, E(x L The gradient E is Lipschitz, with a Lipschitz constant denoted by M By definition, it means x, y R d, E(x E(y M x y E is α-convex for some α product of two vectors By definition, it means (we denote by a, b the scalar Preliminary question x, y R d, E(x E(y, x y α x y 1 Show that, because E is α-convex, then E has at most one critical point 1 By contradiction, if there were two critical points x and y, we would have E(x = E(y = 0 but α-convexity would give us thus x = y α x y 0, In the following, we will denote by X min the unique critical point, we assume that it exists and is the unique global minimizer of E 6

41 Gradient descent in continuous time (40 In this section, we fix X 0 in R d and we study the ODE X (t = E(X(t, X(0 = X 0 ( where X is an unknown function with values in R d This is known as a gradient descent 411 Convergence to the minimizer 1 Explain why the maximal solution to the ODE ( exists, is unique, and is defined for all times t in (, + Are there constant solutions to (? If yes, how many? 3 Show that either the solution is constant, or E(X(t is (strictly decreasing in t 4 Is the equilibrium solution X(t X min stable? 5 Prove that lim t+ X(t = X min (for an initial condition close enough to X min, or, more difficult, for any choice of initial condition 1 The function X E(X is Lipschitz, by assumption The Cauchy-Lipschitz theorem thus ensures existence and uniqueness of maximal solutions Moreover over, it is globally Lipschitz, still by assumption, thus the maximal solutions are global (defined for all times - this is a result from class Constant solutions correspond to zeros of E, thus to critical points, and we know that there is only one such point So there is exactly one constant solution, equal to X min 3 If the solution is not equal to X min, we have d dt E(X(t = E(X(t, X (t = E(X(t, and the right-hand side is always negative, so the quantity E(X(t is decreasing 4 The quantity E(X(t is a Liapounov function for the equilibrium (as shown in the previous question, thus we know the equilibrium is asymptotically stable and, in particular, it is stable 5 Asymptotic stability implies that lim t+ X(t = X min for an initial condition close enough to X min In fact this is true for all initial conditions - we could prove it here by elementary methods, but the following questions will also provide a proof 41 Speed of convergence In this paragraph, we want to quantify the speed at which X(t tends to X min We suppose that the initial condition (at time 0 X 0 is not equal to X min For t 0, we introduce the quantity D(t := X(t X min 7

1 Compute D (t You may use one of the auxiliary results Show that for t 0 we have D (t αd(t 3 Prove that D(t X 0 X min e αt and conclude about the speed of convergence of X(t to X min 1 We have, using the formula given to us and the fact that X satisfies the ODE D (t = X (t, X(t X min = E(X(t, X(t X min We may also write, since E(X min = 0, E(X(t, X(t X min = E(X(t E(X min, X(t X min, and by the α-convexity assumption we get D (t α X(t X min = αd(t (the question is correct buot sharp, by a factor 3 We may write D (t D(t α Integrating this inequality between 0 and t, we obtain D(t D(0e αt = X(0 X min e αt (again, the question is correct buot sharp, by a factor We thus have X(t X min e αt X(0 X min, thus X(t converges to X min as t +, exponentially fast 8

4 Numerical study (40 In this paragraph, we are interested in a numerical approach to gradient descent It can be described as a sequence {X n } n 0 defined as follows: We start at some point X 0 At each step n 0, we chose a step-size s n 0 and we compute X n+1 in terms of X n by X n+1 := X n s n E(X n (3 41 A model case In this paragraph only, we take E(X = X For the two following choices of step-sizes, show that the numerical scheme defined above does not converge to the minimizer of E (here X min = 0 of course, unless we start at this point 1 For a constant step-size s n = 1 For a step-size s n = n 100 It is thus important to chose the step-size carefully 1 In this case, we have X n+1 = X n X n = X n, and thus we have X n = ( 1 n X 0 for all n This does not converge to 0, unless X 0 is 0 In this case, we have We deduce that X n = X 0 X n+1 = X n 1 n 100 X n = X n (1 n 100 n k=1 ( n (1 k 100 = X 0 exp ln(1 k 100 The series + k=1 ln(1 k 100 converges, thus X n converges to CX 0 for some constant C 0, and in particular X n does not converge to X 0 In applied maths classes, the usual heuristics for step-sizes is to chose s n such that { n s n diverges We will try to justify this heuristics n s n converges k=1 9

4 Convergence to the minimizer We let {X n } n be the sequence of points defined as above We let t X(t be the solution to the gradient descent ODE ( with initial condition X(0 = X 0 We let t 0 = 0 and we let X 0 := X(t 0 = X(0 = X 0 For any n 0, we define +1 = + s n, Xn+1 = X(+1 In other words: is the time after n steps, Xn is the value of the real solution at time while X n is the value of the numerical solution after n steps 1 Explain why, if we assume that the series n s n diverges, then X n tends to the minimizer X min as n + In the following, we will always assume that n s n diverges Show that X n satisfies X n+1 = X tn+1 n E(X(sds The next questions are devoted to the analysis of this numerical scheme, and are thus of real analysis spirit 3 Show that we have X n+1 = X n s n E( X n + ε n, with an error term ε n bounded by ε n MLs n, where L, M are the Lipschitz constants defined in the introduction 4 Using α-convexity, show that X n s n E(X n X n + s n E( X n X n X n ( 1 αs n + s nm, where α, M are the constants defined in the introduction 5 For any n 0, we let V n be the difference V n := X n X n Prove that V n+1 V n 1 αs n + s nm + MLs n 6 Using the discrete version of Grönwall s lemma recalled in the Auxiliary results section, show that if n s n converges, then X n tends to X min as n + 10

1 Since = n k=0 s k, the divergence of the series is equivalent to the fact that lim n = +, but we know from previous questions that so, since X n = X(, we have lim t X(t = X min, lim n X n = lim n X( = X min We use the fundamental theorem of calculus and the fact that X is a solution to the ODE: tx n+1 = X(+1 = X( + 3 We need to show that tn+1 tn+1 X (sds = X n tn+1 E(X(sds s n E(X( MLs n E(X(sds We may write tn+1 tn+1 E(X(sds s n E(X( = E(X(s E(X( ds Since E is Lipschitz, we have The mean value theorem gives E(X(s E(X( M X(s X( X(s X( (s sup, X (t and, since E is assumed to be Lipschitz, we have So we obtain tn+1 sup t 4 Expand the square as X (t = sup t E(X(sds s n E(X( E(X(t sup E(x L x tn+s n ML(s ds = MLs n X n s n E(X n X n + s n E( X n = X n X n + s n E( X n E(X n The α-convexity assumption implies that and the Lipschitz-ness of E gives and thus we get + s n X n X n, E( X n E(X n X n X n, E( X n E(X n α X n X n, E( X n E(X n M X n X n X n s n E(X n X n + s n E( X n X n X n (1 αs n + M s n 11

5 We simply combine the results of question 3 and question 4 6 Applying the lemma, we obtain V n exp ( n 1 k=0 ( ( n 1 ln 1 αs k + s k M V 0 + k=0 MLs k We assumed that k s k is finite (and in particular the step-sizes tend to 0 We may thus write ( n 1 MLs k V 0 + C, k=0 for some constant C, and on the other hand a first-order expansion of ln(1 x yields ln ( 1 αs k + s k M = 1 ( αsk + O(s k Since k s k diverges and k s k converges a well-known theorem about comparison of series with positive terms implies lim n n k=0 ln ( 1 αs k + s k M = So since lim x exp(x = 0, we obtain lim V n = 0 n 1

43 A noisy version (0 In this section, we fix the dimension d = 1 and we consider the ODE x ε(t = E (x ε (t + εa(t, x ε (0 = x 0 (4 where t x ε (t is an unknown function with real values, E satisfies the same assumptions as before (but in addition, we assume E to be of class C, ε is some fixed real parameter and A is a continuous function such that T T A(tdt is bounded Let x be the solution to (4 when ε = 0 We look for an expression of x ε as 0 x ε = x + ε x + O(ε 1 Write down (without rigorous justification the ODE satisfied by x Write down an expression for x 3 Show that x(t is bounded as t + 1 We have so x (t + ε x (t + O(ε = E ( x + ε x + O(ε + εa(t, x (t + ε x (t + O(ε = E ( x(t εe ( x(t + εa(t + O(ε Formally, x must satisfy x (t = E ( x(t + A(t The α-convexity assumption translates into the fact that E is positive, and bounded below by α We may thus consider the map x 1 E (x, and we denote by G an antiderivative of this map Since G has a sign, G is a one-to-one continuous map and we denote by G 1 its inverse bijection (defined on the codomain We have G ( x(t x (t = A(t, and thus, integrating between 0 and t, we obtain so we obtain the formal expression G( x(t G( x(0 = x(t = G 1 ( t 0 t 0 A(sds, A(sds + G( x(0 3 Since t 0 A(sds stays bounded as t varies (by assumption, and since G 1 is continuous, we see that x(t stays bounded as t varies 13