SZG Macro 2011 Lecture 3: Dynamic Programming. SZG macro 2011 lecture 3 1

Similar documents
BU Macro BU Macro Fall 2008, Lecture 4

Essential Microeconomics : OPTIMAL CONTROL 1. Consider the following class of optimization problems

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

This document was generated at 1:04 PM, 09/10/13 Copyright 2013 Richard T. Woodward. 4. End points and transversality conditions AGEC

This document was generated at 7:34 PM, 07/27/09 Copyright 2009 Richard T. Woodward

Macroeconomic Theory Ph.D. Qualifying Examination Fall 2005 ANSWER EACH PART IN A SEPARATE BLUE BOOK. PART ONE: ANSWER IN BOOK 1 WEIGHT 1/3

Online Appendix to Solution Methods for Models with Rare Disasters

E β t log (C t ) + M t M t 1. = Y t + B t 1 P t. B t 0 (3) v t = P tc t M t Question 1. Find the FOC s for an optimum in the agent s problem.

1. An introduction to dynamic optimization -- Optimal Control and Dynamic Programming AGEC

THE BELLMAN PRINCIPLE OF OPTIMALITY

ANSWERS TO EVEN NUMBERED EXERCISES IN CHAPTER 6 SECTION 6.1: LIFE CYCLE CONSUMPTION AND WEALTH T 1. . Let ct. ) is a strictly concave function of c

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

RANDOM LAGRANGE MULTIPLIERS AND TRANSVERSALITY

The Brock-Mirman Stochastic Growth Model

10. State Space Methods

Appendix 14.1 The optimal control problem and its solution using

Problem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims

Midterm Exam. Macroeconomic Theory (ECON 8105) Larry Jones. Fall September 27th, Question 1: (55 points)

Linear Dynamic Models

= ( ) ) or a system of differential equations with continuous parametrization (T = R

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

Reserves measures have an economic component eg. what could be extracted at current prices?

1. An introduction to dynamic optimization -- Optimal Control and Dynamic Programming AGEC

Economics 8105 Macroeconomic Theory Recitation 6

( ) a system of differential equations with continuous parametrization ( T = R + These look like, respectively:

Two Coupled Oscillators / Normal Modes

FINM 6900 Finance Theory

Section 7.4 Modeling Changing Amplitude and Midline

Chapter 2. First Order Scalar Equations

O Q L N. Discrete-Time Stochastic Dynamic Programming. I. Notation and basic assumptions. ε t : a px1 random vector of disturbances at time t.

Introduction D P. r = constant discount rate, g = Gordon Model (1962): constant dividend growth rate.

Some Basic Information about M-S-D Systems

Section 3.5 Nonhomogeneous Equations; Method of Undetermined Coefficients

Seminar 4: Hotelling 2

Lecture 20: Riccati Equations and Least Squares Feedback Control

IMPLICIT AND INVERSE FUNCTION THEOREMS PAUL SCHRIMPF 1 OCTOBER 25, 2013

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Solutions Problem Set 3 Macro II (14.452)

Graduate Macro Theory II: Notes on Neoclassical Growth Model

Final Spring 2007

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

dt = C exp (3 ln t 4 ). t 4 W = C exp ( ln(4 t) 3) = C(4 t) 3.

1 Answers to Final Exam, ECN 200E, Spring

The expectation value of the field operator.

Vehicle Arrival Models : Headway

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions

Lecture Notes 3: Quantitative Analysis in DSGE Models: New Keynesian Model

Chapter #1 EEE8013 EEE3001. Linear Controller Design and State Space Analysis

Notes on Kalman Filtering

A Dynamic Model of Economic Fluctuations

Differential Equations

Let us start with a two dimensional case. We consider a vector ( x,

Chapter 6. Systems of First Order Linear Differential Equations

Chapter 4 Introduction to Dynamic Programming

Policy regimes Theory

System of Linear Differential Equations

Introduction to choice over time

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

ODEs II, Lecture 1: Homogeneous Linear Systems - I. Mike Raugh 1. March 8, 2004

Problem Set on Differential Equations

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes

1 Consumption and Risky Assets

Linear Response Theory: The connection between QFT and experiments

23.5. Half-Range Series. Introduction. Prerequisites. Learning Outcomes

Chapter 15: Phenomena. Chapter 15 Chemical Kinetics. Reaction Rates. Reaction Rates R P. Reaction Rates. Rate Laws

8. Basic RL and RC Circuits

Chapter 3 Boundary Value Problem

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

Lecture Notes 2. The Hilbert Space Approach to Time Series

Cooperative Ph.D. Program in School of Economic Sciences and Finance QUALIFYING EXAMINATION IN MACROECONOMICS. August 8, :45 a.m. to 1:00 p.m.

EXERCISES FOR SECTION 1.5

Problem 1 / 25 Problem 2 / 20 Problem 3 / 10 Problem 4 / 15 Problem 5 / 30 TOTAL / 100

Math 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:

Suggested Solutions to Assignment 4 (REQUIRED) Submisson Deadline and Location: March 27 in Class

LAPLACE TRANSFORM AND TRANSFER FUNCTION

Solutions from Chapter 9.1 and 9.2

CHAPTER 12 DIRECT CURRENT CIRCUITS

In this chapter the model of free motion under gravity is extended to objects projected at an angle. When you have completed it, you should

Announcements: Warm-up Exercise:

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

EE 315 Notes. Gürdal Arslan CLASS 1. (Sections ) What is a signal?

This document was generated at 4:25 PM, 03/29/18 Copyright 2018 Richard T. Woodward

Chapter #1 EEE8013 EEE3001. Linear Controller Design and State Space Analysis

ADVANCED MATHEMATICS FOR ECONOMICS /2013 Sheet 3: Di erential equations

Hamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t

EE363 homework 1 solutions

Reading from Young & Freedman: For this topic, read sections 25.4 & 25.5, the introduction to chapter 26 and sections 26.1 to 26.2 & 26.4.

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

Y 0.4Y 0.45Y Y to a proper ARMA specification.

Math Week 14 April 16-20: sections first order systems of linear differential equations; 7.4 mass-spring systems.

1. Consider a pure-exchange economy with stochastic endowments. The state of the economy

The Brock-Mirman Stochastic Growth Model

Math 334 Fall 2011 Homework 11 Solutions

) were both constant and we brought them from under the integral.

d 1 = c 1 b 2 - b 1 c 2 d 2 = c 1 b 3 - b 1 c 3

Echocardiography Project and Finite Fourier Series

Seminar 5 Sustainability

Physics 127b: Statistical Mechanics. Fokker-Planck Equation. Time Evolution

Transcription:

SZG Macro 2011 Lecure 3: Dynamic Programming SZG macro 2011 lecure 3 1

Background Our previous discussion of opimal consumpion over ime and of opimal capial accumulaion sugges sudying he general decision problems on he nex pages, where c = conrol (consumpion) k = sae x = exogenous variable In each case, we migh solve he dynamic opimizaion problem using a discree ime opimal conrol approach as in prior lecures. SZG macro 2011 lecure 3 2

Finie horizon problem T å { c},{ k} b uk x c = 0 max (,, ) u: momenary objecive s.. k - k = g( k, x, c ) g : sae law of moion + 1 k k 0 ³ b( x ) T+ 1 T+ 1 SZG macro 2011 lecure 3 3

Infinie horizon problem å { c},{ k} b uk x c = 0 max (,, ) s.. k + - k = g( k, x, c ) = 0,1,... 1 k 0 SZG macro 2011 lecure 3 4

We migh aack he finie horizon problem by forming L = T å = 0 T å = 0 b u( k, x, c ) + bl[ k + g( k, x, c )-k ] + 1 +Q [ k - b( x )] T+ 1 + 1 + 1 SZG macro 2011 lecure 3 5

We migh aack he infinie horizon problem by forming L = å = 0 å = 0 b u( k, x, c ) + bl[ k + g( k, x, c )-k ] + 1 SZG macro 2011 lecure 3 6

Eiher case Oucome is sequence of opimal conrol {c }, opimal sae {k }, and opimal shadow prices { } ha saisfy FOCs and TC give he pah {x }. Desirable because we have boh real oucomes and a means of deriving supporing prices. SZG macro 2011 lecure 3 7

Why? We will sudy an alernaive, dynamic programming, nex Alernaive ool for oolki DP is beer for cerain problems wih uncerainy DP logic applied in oher, equilibrium conexs LS: he imperialism of recursive mehods SZG macro 2011 lecure 3 8

Ouline 1. Cerainy opimizaion problem used o illusrae: a. Resricions on exogenous variables {x } b. Value funcion c. Policy funcion d. The Bellman equaion and an associaed Lagrangian e. The envelope heorem f. The Euler equaion SZG macro 2011 lecure 3 9

Ouline Con d 2. Adding uncerainy 3. Applicaions opimal consumpion over ime opimal consumpion under uncerainy. SZG macro 2011 lecure 3 10

1. A cerainy dynamic problem and he DP approach Maximize å = 0 b uk (, x, c) Subjec o k+ 1 - k = g( k, x, c) conrolled sae and x = x( V ) exogenous variable V = m( V - 1) exogenous sae SZG macro 2011 lecure 3 11

Wha s differen from background seup? Immediae jump o infinie horizon problem, no essenial bu maches presenaion in LS chaper (noe differences in noaion, hough). The exogenous (x) variable(s) are now funcions of a vecor of exogenous sae variables, which evolve according o a difference equaion (perhaps nonlinear, perhaps in a vecor). The laer is a key par of he vision of Richard Bellman, he invenor of DP: his experience in oher areas (such as difference equaions) led him o hink in erms of describing dynamics in erms of sae variables. SZG macro 2011 lecure 3 12

Recursive policies Suppose conrols are funcions of saes, c ( k, ) policy funcion k 1 k g( k, x, c) k g( k, x( ), ( k, )) Then, he sae vecor evolves according o a recursion k 1 k g( k, x( ), ( k, )) s 1 M( s) 1 m( ) ha can be used o generae fuure saes from given iniial condiions SZG macro 2011 lecure 3 13

Evaluaing he objecive Under any recursive policy, we can see ha all of he erms which ener in he objecive are a funcion of he iniial sae (s 0 ) so ha he objecive is also a funcion of he iniial sae å b = 0 = 0 uk (, x, c) = b uk (, x( V ), p( k, x( V ))) = Wk (, V ) = Ws ( ) å 0 0 0 SZG macro 2011 lecure 3 14

Recursion for objecive under arbirary recursive policy Ws ( ; p) = ws ( ; p) + bws ( ; p) 0 0 1 ws ( ; p) = ukx (, ( V), p( kx, ( V))) s + 1 = M( s ) SZG macro 2011 lecure 3 15

Noice he swich Given ha here is a policy funcion ( ), he objecive is now a funcion of he sae vecor. We have made he change we are now hinking in erms of funcions raher han sequences. Bu we haven opimized ye! We could be calculaing he objecive wih a very bad policy. SZG macro 2011 lecure 3 16

Bellman s core idea Subdivide complicaed ineremporal problems ino many wo period problems, in which he rade-off is beween he presen now and laer. Specifically, he idea was o find he opimal conrol and sae now, aking as given ha laer behavior would iself be opimal. SZG macro 2011 lecure 3 17

The Principle of Opimaliy An opimal policy has he propery ha, whaever he sae and opimal firs decision may be, he remaining decisions consiue an opimal policy wih respec o he sae originaing from he firs decisions Bellman (1957, pg. 83) SZG macro 2011 lecure 3 18

Following he principle, The naural maximizaion problem is max{ uc (, k, x( V )) + bv( k, V )} c, k + 1 + 1 + 1 s.. k = k + g( k, x, c ) V + 1 + 1 = m( V ) Where he righ hand side is he curren momenary objecive (u) plus he consequences (V) for he discouned objecive of behaving opimally in he fuure. SZG macro 2011 lecure 3 19

Noing ha ime does no ener in an essenial way We someimes wrie his as (wih meaning nex period) max{ uckx (,, ( V)) + bv( k', V')} ck, ' s.. k' = k + g( k, x( V), c) V' = m( V) So hen he Bellman equaion is wrien as ck, ' Vk (, V) = max{ uckx (,, ( V)) + bvk ( ', V')} s.. k' = k + g( k, x( V), c) V' = m( V) SZG macro 2011 lecure 3 20

Afer he maximizaion We know he opimal policy (which we will call as above, bu wih he proviso ha i is opimal) and can calculae he associaed value, so ha here is now a Bellman equaion of he form Vk (, V) = {((, up kv), kx, ()) V + bvk ( + gkx (, ( V), p( k, V)), V')} A funcional equaion is defined, colloquially, as an equaion whose unknowns are funcions. In our conex, he unknowns are he policy and value funcions. SZG macro 2011 lecure 3 21

How o do he opimizaion? You are free o choose, depending on he applicaion Someimes we ake he Euler roue, subsiuing in he consrain and maximizing direcly over k Oher imes we wan o use a Lagrange approach, puing a muliplier on he consrain governing k SZG macro 2011 lecure 3 22

The associaed Lagrangian Takes he form L { u( c, k, x( )) V( k', ')} [ k g( k, x( ), c) k'] The opimal policy, sae evoluion and relaed muliplier are obained by maximizing wih respec o c,k and minimizing wih respec o. Hence hese are all funcions of he sae variables. SZG macro 2011 lecure 3 23

For an opimum (off corners) We mus have L u(, c k, x()) g(, k x(),) c 0 c c c L V( k', ') 0 k' k' L [ k g( k, x( ), c) k'] 0 And, a he values which solve hese equaions, V=L SZG macro 2011 lecure 3 24

The envelope heorem (Benvenise-Scheinkman) Quesion: wha is he effec of an infiniessimal change in k on V? Answer: I is given by V u(, c k, x()) g(, k x(),) c [ 1] k k k when we evaluae a he opimal policy and he associaed muliplier. As in LS, his may also be wrien a form which does no involve V u( c, k, x( )) V( k', ') g( k, x( ), c) he muliplier, [ 1] k k k' k SZG macro 2011 lecure 3 25

Ouline of proof Nonrivial o show differeniabiliy of V Bu if we have his (as we will frequenly assume) hen V L u( c, k, x( )) c u( c, k, x( )) { } k k c k k V( k', ') k' k' k [ k g( k, x( ), c) k'] k g( kx, ( ), c) g( kx, ( ), c) c k' [1 ] [ ] k k k k While his looks ugly, all erms involving behavior are muliplied by coefficiens ha are se o zero by he FOCs. SZG macro 2011 lecure 3 26

Deails on ET V u( c, k, x( )) g( k, x( ), c) [1 ] k k k uckx (,, ( )) gckx (,, ( )) c { } c c k V( k', ') k' { } k' k { [ k g( k, x( ), c) k']} k SZG macro 2011 lecure 3 27

Ieraing on he Bellman Equaion Under specific condiions on he funcions u and g, he Bellman equaion has a unique, sricly concave (in k) soluion. Under hese condiions, i can be calculaed by considering he limi Vj 1 ( k, ) max c, k' { u( k, x( ), c) Vj( k', ')} s.. k' k g( k, x( ), c) These ieraions are inerpreable as calculaing he value funcions for a class of finie horizon problems, wih successively longer horizons. SZG macro 2011 lecure 3 28

3. A Sochasic dynamic problem and he DP approach Maximize å = 0 E{ b u( k, x, c )} ( k, V ) 0 0 Subjec o k+ 1 - k = g( k, x, c) and Markovian exogenous sae variables x = x( V ) (, V B) = prob( V Î B V = V) + 1 SZG macro 2011 lecure 3 29

Markov examples Markov chains (LS, Chaper 1) Linear sae space sysems Nonlinear difference equaions wih iid shocks, m(, e ) 1 1 We won be more explici unil necessary. Key poin: saes are enough o compue expecaions. SZG macro 2011 lecure 3 30

Bellman Equaion Uncerainy case is minor modificaion of cerainy case Vk (, V) = max{ uckx (,, ( V)) + bevk ( ', V') ( k, V)} ck, ' s.. k' = k + g( k, x( V), c) SZG macro 2011 lecure 3 31

Proceeding as above Lagrangian L {(, u c k, x()) EV( k', ') (, k )} [ k g( k, x( ), c) k'] FOCs ET is unchanged L u(, c k, x()) g(, k x(),) c c c c L EV( k', ') 0 k' k' L [ k g( k, x( ), c) k'] 0 0 SZG macro 2011 lecure 3 32

Implicaions for opimal policies Funcions of saes and sae evoluion c ( k, ) k k g( k, x( ), ( k, )) 1 ( k, ) Sae evoluion is now a larger Markov process. For example, s k k g( k, x( ), ( k, )) M( s, e ) 1 1 1 1 m(, e 1) SZG macro 2011 lecure 3 33

Value Funcion Since c,k,x depend on saes, he value funcion also is V(s). I is he maximized RHS of he Bellman equaion. SZG macro 2011 lecure 3 34

Wha we ve covered in his lecure Inroducion o DP under cerainy Bellman Equaion Associaed Lagrangian FOCs and he ET DP wih exogenous variables ha are funcions of a Markov process (exogenous sae vecor) Wha follows: Opimal consumpion over ime via dynamic programming: calculaion of policy and value funcions in a simple case Seing up opimal consumpion problem wih uncerain income SZG macro 2011 lecure 3 35

3A. Opimal consumpion over ime Simple case (no k,x in u) å = 0 uc ( ) Accumulaion of asses b a = Ra [ + y -c ] + 1 y = y + r( y -y) + 1 And R=1 (level consumpion) SZG macro 2011 lecure 3 36

Bellman Equaion V( a, y) max { u( c) V( a', y')} ca, ' s.. a' R[ a y c] y' y ( y y) SZG macro 2011 lecure 3 37

Taking an Euler Roue 1 V( a, y) max ca, '{ u( a y a') V( a', y')} R s.. y' y ( y y) 1 1 V( a', y') EE :0 uc( a y a ') R R a' V( a, y) 1 ET : uc ( a y a ') a R SZG macro 2011 lecure 3 38

Learning abou consumpion Updae ET and inser in EE o ge 1 1 uc( a y a') uc( a' y' a'') c c' R R Suppose here is a linear policy funcion c ( y y) a y a c' ( y' y) a' y ( y y) R[ a y c] y a a ( y y) R[ a y ( y y) a] y a y a SZG macro 2011 lecure 3 39

Requiring c=c, we have equaions ha resric undeermined coefficiens ( y y) a y a ( y y) R[ a ( y y) y ( y y) a] y a y a Ry ( ) a R(1 ) R/[1 R] y y a y y a a R 1 a ar[1 a] a ( ) R y y a 1 (1 ) R SZG macro 2011 lecure 3 40

Economic Rules Consume he normal level of income (y) Consume he ineres from asse sock, leaving he asse sock unchanged period o period (excep as noed nex) Consume based on he presen value of deviaions from normal income, reaing his as if i were anoher source of wealh; allow variaions in asse posiion on his basis. SZG macro 2011 lecure 3 41

Could have goen hese rules more direcly 1 j 1 j j ( ) c a ( ) [ y ( y y)] R R j 0 j 0 1 1 1 c a y ( y y ) 1 1 1 1 1 R R R R 1 1 c y [ a ( y y)] R 1 R SZG macro 2011 lecure 3 42

Quesions & Answers If we could have goen hem more easily, hen why do we need DP? Because here are many problems ha we canno solve so easily and DP is a procedure for solving hem. Wha is he value funcion? V 1 1 1 ( a, y ) ( ( )) 1 u a 1 y y y 1 1 R Easy o deermine in his case because c is consan over ime; V inheris properies of u Check: ake his v, inser in Bellman equaion as v, show opimal form c has specified form, show v has his form. R SZG macro 2011 lecure 3 43

3B. Opimal consumpion wih flucuaing income: seing up a DP Simple case (no k,x in u) å = 0 E{ b u( c )} s Accumulaion of asses (don necessarily resric R) a+ 1 = Ra [ + y -c] 0 Income process y( V ) V : Markov SZG macro 2011 lecure 3 44

One version of he Bellman equaion V( a, ) max {( u( c) EV( a', ')} ca, ' 1 s..[ a y( ) c a'] 0 R SZG macro 2011 lecure 3 45

FOCs and ET Make sure you can work hese ou following he recipe above, c: u ( c) 0 c 1 EV ( a ', ') a': E{ } 0 R a' 1 : [ a y( ) c a'] 0 R ET : EV ( a, ) a SZG macro 2011 lecure 3 46

Implicaions for policies Opimal consumpion depends on (a) wealh; and (b) he variables ha are useful for forecasing fuure income. ca (, ) Bu solving for his funcion is no longer easy. Raionalizes SL s discussion of numerical mehods, a opic ha we will consider furher laer. SZG macro 2011 lecure 3 47

Implicaion for value funcion Value funcion is objecive evaluaed a opimal consumpion policy, which is a funcion of a Markov process, so ha 0 V0 = å b p V 0 V0 = 0 Va (, ) E{ u( ( a, ))} ( a, ) Value funcion saisfies he Bellman funcional equaion. Va (, V) = max {( uc ( ) + EVa ( ', V')} ca, ' 1 s..[ a + y( V) -c- a' = 0] R = ( u( p( a, V)) + EV( Ra [ + y( V) -p( a, V)], V') ( a, V) SZG macro 2011 lecure 3 48