Numerical Methods in Economics

Size: px
Start display at page:

Download "Numerical Methods in Economics"

Transcription

1 Numerical Methods in Economics MIT Press, 1998 Chapter 12 Notes Numerical Dynamic Programming Kenneth L. Judd Hoover Institution November 15,

2 Discrete-Time Dynamic Programming Objective: X: set of states E { T t=1 D: thesetofcontrols π(x t,u t,t)+w(x T +1 ) }, (12.1.1) π(x, u, t) payoffs in period t, forx X at the beginning of period t, and control u Dis applied in period t. D(x, t) D: controls which are feasible in state x at time t. F (A; x, u, t) :probability that x t+1 A X conditional on time t control and state Value function { T V (x, t) sup E U(x,t) s=t π(x s,u s,s)+w(x T +1 ) x t = x }. (12.1.2) Bellman equation V (x, t) = sup u D(x,t) π(x, u, t)+e {V (x t+1,t+1) x t = x, u t = u} (12.1.3) Existence: boundedness of π is sufficient 2

3 Autonomous, Infinite-Horizon Problem: Objective: X: set of states D: thesetofcontrols max u t { } E β t π(x t,u t ) t=1 D(x) D: controls which are feasible in state x. (12.1.1) π(x, u) payoff in period t if x X at the beginning of period t, and control u Dis applied in period t. F (A; x, u) :probability that x + A X conditional on current control u and current state x. Value function definition: if U(x) is set of all feasible strategies starting at x. { } V (x) sup E β t π(x t,u t ) U(x) x 0 = x, (12.1.8) t=0 3

4 Bellman equation for V (x) V (x) = sup u D(x) π(x, u)+βe { V (x + ) x, u } (TV)(x), (12.1.9) Optimal policy function, U(x), if it exists, is defined by U(x) arg max π(x, u)+βe{ V (x + ) x, u } u D(x) Standard existence theorem: Theorem 1 If X is compact, β<1, andπ is bounded above and below, then the map TV = sup u D(x) π(x, u)+βe { V (x + ) x, u } ( ) is monotone in V, is a contraction mapping with modulus β in the space of bounded functions, and has a unique fixed point. 4

5 Deterministic Growth Example Problem: Euler equation: V (k 0 )=max ct t=0 βt u(c t ), k t+1 = F (k t ) c t k 0 given u (c t )=βu (c t+1 )F (k t+1 ) ( ) Bellman equation V (k) =max c u(c) +βv (F(k) c). ( ) Solution to ( ) is a policy function C(k) and a value function V (k) satisfying 0=u (C(k))F (k) V (k) ( ) V (k)=u(c(k)) + βv (F(k) C(k)) ( ) ( ) defines the value of an arbitrary policy function C(k), not just for the optimal C(k). The pair ( ) and ( ) expresses the value function given a policy, and a first-order condition for optimality. 5

6 Stochastic Growth Accumulation Problem: { } V (k, θ) =max c t,l t E t=0 β t u(c t ) k t+1 = F(k t,θ t ) c t θ t+1 = g(θ t,ε t ) ε t : i.i.d. random variable k 0 = k, θ 0 = θ. State variables: k: productive capital stock, endogenous θ: productivity state, exogenous The dynamic programming formulation is V (k, θ) =max c u(c)+βe{v (F (k, θ) c, θ + ) θ} ( ) θ + = g(θ, ε) The control law c = C(k, θ) satisfies the first-order conditions 0=u c (C(k, θ)) βe{u c (C(k +,θ + ))F k (k +,θ + ) θ}, ( ) where k + F (k, L(k, θ),θ) C(k, θ), 6

7 General Stochastic Accumulation Problem: { } V (k, θ) =max c t,l t E t=0 β t u(c t,l t ) k t+1 = F (k t,l t,θ t ) c t θ t+1 = g(θ t,ε t ) k 0 = k, θ 0 = θ. State variables: k: productive capital stock, endogenous θ: productivity state, exogenous The dynamic programming formulation is V (k, θ) =max c,l u(c, l)+βe{v (F(k, l, θ) c, θ + ) θ}, ( ) where θ + is next period s θ realization. Control laws c = C(k, θ) and l = L(k, θ) satisfy foc s 0= u c (C(k, θ),l(k, θ))f k (k, L(k, θ),θ) V k (k, θ), 0= u l (C(k, θ),l(k, θ)) + F l (k, θ)u c (C(k, θ),l(k, θ)). Euler equation implies 0=u c (C(k, θ),l(k, θ)) βe{u c (C(k +,θ + ),l + )F k (k +,l +,θ + ) θ}, ( ) where next period s capital stock and labor supply are k + F (k, L(k, θ),θ) C(k, θ), l + L(k +,θ + ), 7

8 Discrete State Space Problems State space X = {x i,i=1,,n} Controls D = {u i i =1,..., m} qij t (u) =Pr(x t+1 = x j x t = x i,u t = u) Q t (u) = ( qij t (u)) : Markov transition matrix at t if u i,j t = u. 8

9 Value Function iteration Terminal value: V T +1 i = W(x i ),i=1,,n. Bellman equation: time t value function is n Vi t =max[π(x i,u,t)+β q t u ij(u) Vj t+1 ], i=1,,n j=1 Bellman equation can be directly implemented. Called value function iteration It is only choice for finite-horizon problems because each period has a different value function. Infinite-horizon problems Bellman equation is now a simultaneous set of equations for V i values: n V i =max π(x i,u)+β q ij (u) V j,i=1,,n u j=1 Value function iteration is now V k+1 i =max u π(x i,u)+β n j=1 q ij (u) V k j,i=1,,n Can use value function iteration with arbitrary Vi 0. Error is given by contraction mapping property: V k V 1 V k+1 V k 1 β and iterate k 9

10 Algorithm 12.1: Value Function Iteration Algorithm Objective: Solve the Bellman equation, (12.3.4). Step 0: Make initial guess V 0 ; choose stopping criterion ɛ>0. Step 1: For i =1,..., n, compute Vi l+1 =max u D π(x i,u)+β n j=1 q ij(u)vj l. Step 2: If V l+1 V l <ɛ, then go to step 3; else go to step 1. Step 3: Compute the final solution, setting U = UV l+1, Pi = π(x i,ui ), i =1,,n, V =(I βq U ) 1 P, and STOP. Output: 10

11 Policy Iteration (a.k.a. Howard improvement) Value function iteration is a slow process Linear convergence at rate β Convergence is particularly slow if β is close to 1. Policy iteration is faster Current guess: V k i,i=1,,n. Iteration: compute optimal policy today if V k is value tomorrow: n Ui k+1 =argmax π(x i,u)+β q ij (u) Vj k,i=1,,n, u Compute the value function if the policy U k+1 is used forever, which is solution to the linear system Vi k+1 = π ( ) n x i,ui k+1 + β q ij (Ui k+1 ) Vj k+1,i=1,,n, j=1 Comments: j=1 Policy iteration depends on only monotonicity Policy iteration is faster than value function iteration If initial guess is above or below solution then policy iteration is between truth and value function iterate Works well even for β close to 1. 11

12 Algorithm 12.2: Policy Function Algorithm Objective: Solve the Bellman equation, (12.3.4). Step 0: Choose stopping criterion ɛ>0. EITHER make initial guess, V 0,forthe value function and go to step 1, OR make initial guess, U 1,forthe policy function and go to step 2. Step 1: U l+1 = UV l Step 2: Pi l+1 = π ( ) x i,ui l+1, i =1,,n ( ) 1 Step 3: V l+1 = I βq U l+1 P l+1 Step 4: If V l+1 V l <ɛ, STOP; else go to step 1. 12

13 Modified policy iteration If n is large, difficult to solve policy iteration step Alternative approximation: Assume policy U l+1 is used for k periods: V l+1 = k ( β t Q U l+1) t ( P l+1 + β k+1 Q U l+1) k+1 V l. (12.4.1) t=0 Theorem 4.1 points out that as the policy function gets close to U, the linear rate of convergence approaches β k+1. Hence convergence accelerates as the iterates converge. Theorem 2 (Putterman and Shin) The successive iterates of modified policy iteration with k steps, (12.4.1), satisfy the error bound V V l+1 [ ] V V l min β(1 β k ) β, U l U +β k+1 (12.4.3) 1 β 13

14 Gaussian acceleration methods for infinite-horizon models Key observation: Bellman equation is a simultaneous set of equations n V i =max π(x i,u)+β q ij (u) V j,i=1,,n u Idea: Treat problem as a large system of nonlinear equations Value function iteration is the pre-gauss-jacobi iteration n Vi k+1 =max π(x i,u)+β q ij (u) Vj k,i=1,,n u True Gauss-Jacobi is V k+1 i =max u pre-gauss-seidel iteration j=1 j=1 [ π(xi,u)+β j i q ij(u) V k j 1 βq ii (u) Value function iteration is a pre-gauss-jacobi scheme. ],i=1,,n Gauss-Seidel alternatives use new information immediately Suppose we have Vi l At each x i,givenvj l+1 for j<i,computevi l+1 Seidel fashion in a pre-gauss- V l+1 i =max u π(x i,u)+β j<i q ij (u)vj l+1 +β j i q ij (u)v l j (12.4.7) Iterate (12.4.7) for i =1,.., n 14

15 Gauss-Seidel iteration Suppose we have V l i If optimal control at state i is u, then Gauss-Seidel iterate would be V l+1 i = π(x i,u)+β j<i q ij(u)v l+1 j 1 βq ii (u) + j>i q ij(u)v l j Gauss-Seidel: At each x i,givenvj l+1 for j<i,computev i l+1 V l+1 i =max u Iterate this for i =1,.., n Gauss-Seidel iteration: better notation π(x i,u)+β j<i q ij(u)vj l+1 + β j>i q ij(u)vj l 1 βq ii (u) No reason to keep track of l, number of iterations At each x i, V i max u π(x i,u)+β j<i q ij(u)v j + β j>i q ij(u)v j 1 βq ij (u) Iterate this for i =1,.., n, 1,..., etc. 15

16 Upwind Gauss-Seidel Gauss-Seidel methods in (12.4.7) and (12.4.8) Sensitive to ordering of the states. Need to find good ordering schemes to enhance convergence. Example: Two states, x 1 and x 2,andtwocontrols,u 1 and u 2 u i causesstatetomovetox i, i =1, 2 Payoffs: π(x 1,u 1 )= 1, π(x 1,u 2 )=0, π(x 2,u 1 )= 0,π(x 2,u 2 )=1. β =0.9. (12.4.9) Solution: Optimal policy: always choose u 2,movingtox 2 Value function: V (x 1 )=9,V(x 2 )=10. x 2 is the unique steady state, and is stable Value iteration with V 0 (x 1 )=V 0 (x 2 )=0converges linearly: V 1 (x 1 )=0,V 1 (x 2 )=1,U 1 (x 1 )=2,U 1 (x 2 )=2, V 2 (x 1 )=0.9, V 2 (x 2 )=1.9, U 2 (x 1 )=2,U 2 (x 2 )=2, V 3 (x 1 )=1.71, V 3 (x 2 )=2.71, U 3 (x 1 )=2,U 3 (x 2 )=2, Policy iteration converges after two iterations V 1 (x 1 )=0,V 1 (x 2 )=1,U 1 (x 1 )=2,U 1 (x 2 )=2, V 2 (x 1 )=9,V 2 (x 2 )=10,U 2 (x 1 )=2,U 2 (x 2 )=2, 16

17 Upwind Gauss-Seidel Value function at absorbing states is trivial to compute Suppose s is absorbing state with control u V (s) =π(s, u)/(1 β). With absorbing state V (s) we compute V (s ) of any s that sends system to s. V (s )=π (s,u)+βv (s) With V (s ), we can compute values of states s that send system to s ;etc. 17

18 Alternating Sweep It may be difficult to find proper order. Idea: alternate between two approaches with different directions. W = V k, W i =max u π(x i,u)+β n j=1 q ij(u)w j,i=1, 2, 3,..., n W i =max u π(x i,u)+β n j=1 q ij(u)w j,i= n, n 1,..., 1 V k+1 = W Will always work well in one-dimensional problems since state moves either right or left, and alternating sweep will exploit this half of the time. In two dimensions, there may still be a natural ordering to be exploited. Simulated Upwind Gauss-Seidel It may be difficult to find proper order in higher dimensions Idea: simulate using latest policy function to find downwind direction Simulate to get an example path, x 1,x 2,x 3,x 4,..., x m Execute Gauss-Seidel with states x m,x m 1,x m 2,..., x 1 18

19 Linear Programming Approach If D is finite, we can reformulate dynamic programming as a linear programming problem. (12.3.4) is equivalent to the linear program min n Vi i=1 V i s.t. V i π(x i,u)+β n j=1 q ij(u)v j, i, u D, ( ) Computational considerations ( ) may be a large problem OR literature does not favor this approach Trick and Zin (1997) pursued an acceleration approach with success. 19

20 Continuous states: discretization Method: Replace continuous X with a finite X = {x i,i=1,,n} X Proceed with a finite-state method. Problems: Sometimes need to alter space of controls to assure landing on an x in X. A fine discretization often necessary to get accurate approximations 20

21 Continuous States: Linear-Quadratic Dynamic Programming Problem: max u t T ( 1 β t 2 x t Q t x t + u t R t x t + 1 ) 2 u t S t u t x T +1W T +1 x T +1 t=0 Bellman equation: x t+1 = A t x t + B t u t, (12.6.1) 1 V (x, t) =max u t 2 x Q t x + u t R t x u t S t u t + βv (A t x + B t u t,t+1). (12.6.2) Finite horizon Key fact: We know solution is quadratic, solve for the unknown coefficients The guess V (x, t) = 1 2 x W t+1 x implies f.o.c. 0=S t u t + R t x + βb t W t+1 (A t x + B t u t ), F.o.c. implies the time t control law u t = (S t + βbt W t+1 B t ) 1 (R t + βbt W t+1 A t )x (12.6.3) U t x. Substitution into Bellman implies Riccati equation for W t : W t = Q t + βa t W t+1 A t +(βb t W t+1 A t + R t )U t. (12.6.4) Value function method iterates (12.6.4) beginning with known W T +1 matrix of coefficients. 21

22 Autonomous, Infinite-horizon case. Assume R t = R, Q t = Q, S t = S, A t = A, andb t = B The guess V (x) 1 2 x Wx implies the algebraic Riccati equation Two convergent procedures: W =Q + βa WA (βb WA+ R ) (12.6.5) (S + βb WB) 1 (βb WB + R ). Value function iteration: W 0 : a negative definite initial guess W k+1 =Q + βa W k A (βb W k A + R ) (S + βb W k B) 1 (βb W k B + R ). (12.6.6) Policy function iteration: Lessons W 0 : initial guess U i+1 = (S + βb W i B) 1 (R + βb W i A):optimal policy for W i 1 2 W i+1 = Q U i+1su i+1 + Ui+1R : value of U i 1 β We used a functional form to solve the dynamic programming problem We solve for unknown coefficients We did not restrict either the state or control set Canwedothisingeneral? 22

23 Continuous Methods for Continuous-State Problems Basic Bellman equation: V (x) = max u D(x) π(u, x)+βe{v (x+ ) x, u)} (TV)(x). (12.7.1) Discretization essentially approximates V with a step function Approximation theory provides better methods to approximate continuous functions. General Task Find good approximation for V Identify parameters 23

24 Continuous States: Parametric Approx. and Simulation General Idea: parameterize critical functions and find parameter values that generates a good approximation. Direct approach: parameterize the control law, Û(x; a), and use simulation to find a that produces highest value. Example: Consider stochastic growth problem: V (k) =max c u(c)+βe{v (k c + θf(k c)) k, c}, (12.8.1) Parameterize savings function, S(k) k C(k). Consider linear rules: Ŝ(k) =a + bk Use simulation to approximate value of a savings rule. Simulate θ t,t=1,,t sequence of productivity shocks. For given k 0, θ t,andŝ(k), compute paths for c t and k t : c t =k t Ŝ(k t ) k t+1 =Ŝ(k t )+θ t f(ŝ(k t )) Compute realized discounted utility is T W(θ; Ŝ) = β t u(c t ). (12.8.2) Repeat for several θ t sequences. Value Ŝ(k 0 ) is V (k 0 ; Ŝ) =E{W(θ; Ŝ)}, approximated by average t=0 1 N N W(θ j ; Ŝ) = 1 N j=1 N j=1 T β t u(c j t). (12.8.3) t=0 Iterate over various a and b to find optimal rule 24

25 General Parametric Approach: Approximating V (x) Choose a finite-dimensional parameterization and a finite number of states V (x). = ˆV (x; a), a R m (12.7.2) X = {x 1,x 2,,x n }, (12.7.3) polynomials with coefficients a and collocation points X splines with coefficients a with uniform nodes X rational function with parameters a and nodes X neural network specially designed functional form Objective: find coefficients a R m such that ˆV (x; a) approximately satisfies the Bellman equation. 25

26 General Parametric Approach: Approximating T For each x j, (TV)(x j ) is defined by v j =(TV)(x j )= max u D(x j ) π(u, x j)+β ˆV (x + ; a)df (x + x j,u) (12.7.5) In practice, we compute the approximation ˆT v j =(ˆTV)(x j ). =(TV)(x j ) Integration step: for ω j and x j for some numerical quadrature formula E{V (x + ; a) x j,u)}= ˆV (x + ; a)df (x + x j,u) = ˆV (g(x j,u,ε); a)df (ε). = ω l ˆV (g(x j,u,ε l ); a) l Maximization step: for x i X, evaluate v i =(T ˆV )(x i ) Hot starts Concave stopping rules Fitting step: Data: (v i,x i ),i=1,,n Objective: find an a R m such that ˆV (x; a) best fits the data Methods: determined by ˆV (x; a) 26

27 Approximating T with Hermite Data Conventional methods just generate data on V (x j ): v j = max π(u, x j)+β ˆV (x + ; a)df (x + x j,u) (12.7.5) u D(x j ) Envelope theorem: If solution u is interior, v j = π x (u, x j )+β ˆV (x + ; a)df x (x + x j,u) If solution u is on boundary v j = µ + π x (u, x j )+β ˆV (x + ; a)df x (x + x j,u) where µ is a Kuhn-Tucker multiplier Since computing v j is cheap, we should include it in data: Data: (v i,v i,x i), i=1,,n Objective: find an a R m such that ˆV (x; a) best fits Hermite data Methods: determined by ˆV (x; a) 27

28 General Parametric Approach: Value Function Iteration Comparison with discretization guess a ˆV (x; a) (v i,x i ),i=1,,n new a This procedure examines only a finite number of points, but does not assume that future points lie in same finite set. Our choices for the x i are guided by systematic numerical considerations. Synergies Smooth interpolation schemes allow us to use Newton s method in the maximization step. They also make it easier to evaluate the integral in (12.7.5). 28

29 Algorithm 12.5: Parametric Dynamic Programming with Value Function Iteration Objective: Solve the Bellman equation, (12.7.1). Step 0: Choose functional form for ˆV (x; a), andchoose the approximation grid, X = {x 1,..., x n }. Make initial guess ˆV (x; a 0 ), and choose stopping criterion ɛ>0. Step 1: Maximization step: Compute v j =(T ˆV ( ; a i ))(x j ) for all x j X. Step 2: Fitting step: Using the appropriate approximation method, compute the a i+1 R m such that ˆV (x; a i+1 ) approximates the (v i,x i ) data. Step 3: If ˆV (x; a i ) ˆV (x; a i+1 ) < ɛ, STOP; else go to step 1. 29

30 Convergence T is a contraction mapping ˆT may be neither monotonic nor a contraction Shape problems An Instructive Example Figure 1: Shape problems may become worse with value function iteration Shape-preserving approximation implies monotonicity 30

31 Comparisons We apply various methods to the deterministic growth model Relative L2 Errors over [0.7,1.3] N (β,γ) : (.95,-10.) (.95,-2.) (.95,-.5) (.99,-10.) (.99,-2.) (.99,-.5) Discrete model e e e e e e e e e e e e-04 Linear Interpolation 4 7.9e e e e e e e e e e e e e e e e e e-05 Cubic Spline 4 6.6e e e e e e e e e e e e e e e e e e e e e e e e-09 Polynomial (without slopes) 4 DNC 5.4e e e e e e e e e e e-09 Shape Preserving Quadratic Hermite Interpolation 4 4.7e e e e e e e e e e e e e e e e e e-08 Shape Preserving Quadratic Interpolation (ignoring slopes) 4 1.1e e e e e e e e e e e e e e e e e e-07 31

32 General Parametric Approach: Policy Iteration Basic Bellman equation: Policy iteration: V (x) = max u D(x) π(u, x)+βe{v (x+ ) x, u)} (TV)(x). Current guess: a finite-dimensional linear parameterization V (x). = ˆV (x; a), a R m Iteration: compute optimal policy today if ˆV (x; a) is value tomorrow U (x) =π u (x i,u(x),t)+β d ( { E ( ˆV x + ; a ) }) x, U (x)) du using some approximation scheme Û(x; b) Compute the value function if the policy Û(x; b) is used forever, which is solution to the linear integral equation ˆV (x; a )=π(û(x; b),x)+βe{ ˆV (x + ; a ) x, Û(x; b))} that can be solved by a projection method 32

33 Summary: Discretization methods Easy to implement Numerically stable Amenable to many accelerations Poor approximation to continuous problems Continuous approximation methods Can exploit smoothness in problems Possible numerical instabilities Acceleration is less possible 33

Numerical Dynamic Programming

Numerical Dynamic Programming Numerical Dynamic Programming Kenneth L. Judd Hoover Institution Prepared for ICE05 July 20, 2005 - lectures 1 Discrete-Time Dynamic Programming Objective: X: set of states E ( TX D: the set of controls

More information

NUMERICAL DYNAMIC PROGRAMMING

NUMERICAL DYNAMIC PROGRAMMING NUMERICAL DYNAMIC PROGRAMMING Kenneth L. Judd Hoover Institution and NBER July 29, 2008 1 Dynamic Programming Foundation of dynamic economic modelling Individual decisionmaking Social planners problems,

More information

Dynamic Programming with Hermite Interpolation

Dynamic Programming with Hermite Interpolation Dynamic Programming with Hermite Interpolation Yongyang Cai Hoover Institution, 424 Galvez Mall, Stanford University, Stanford, CA, 94305 Kenneth L. Judd Hoover Institution, 424 Galvez Mall, Stanford University,

More information

Lecture 1: Dynamic Programming

Lecture 1: Dynamic Programming Lecture 1: Dynamic Programming Fatih Guvenen November 2, 2016 Fatih Guvenen Lecture 1: Dynamic Programming November 2, 2016 1 / 32 Goal Solve V (k, z) =max c,k 0 u(c)+ E(V (k 0, z 0 ) z) c + k 0 =(1 +

More information

Lecture 4: Dynamic Programming

Lecture 4: Dynamic Programming Lecture 4: Dynamic Programming Fatih Guvenen January 10, 2016 Fatih Guvenen Lecture 4: Dynamic Programming January 10, 2016 1 / 30 Goal Solve V (k, z) =max c,k 0 u(c)+ E(V (k 0, z 0 ) z) c + k 0 =(1 +

More information

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION UNIVERSITY OF MARYLAND: ECON 600. Alternative Methods of Discrete Time Intertemporal Optimization We will start by solving a discrete time intertemporal

More information

PROJECTION METHODS FOR DYNAMIC MODELS

PROJECTION METHODS FOR DYNAMIC MODELS PROJECTION METHODS FOR DYNAMIC MODELS Kenneth L. Judd Hoover Institution and NBER June 28, 2006 Functional Problems Many problems involve solving for some unknown function Dynamic programming Consumption

More information

A Quick Introduction to Numerical Methods

A Quick Introduction to Numerical Methods Chapter 5 A Quick Introduction to Numerical Methods One of the main advantages of the recursive approach is that we can use the computer to solve numerically interesting models. There is a wide variety

More information

Projection Methods. Felix Kubler 1. October 10, DBF, University of Zurich and Swiss Finance Institute

Projection Methods. Felix Kubler 1. October 10, DBF, University of Zurich and Swiss Finance Institute Projection Methods Felix Kubler 1 1 DBF, University of Zurich and Swiss Finance Institute October 10, 2017 Felix Kubler Comp.Econ. Gerzensee, Ch5 October 10, 2017 1 / 55 Motivation In many dynamic economic

More information

Chapter 3. Dynamic Programming

Chapter 3. Dynamic Programming Chapter 3. Dynamic Programming This chapter introduces basic ideas and methods of dynamic programming. 1 It sets out the basic elements of a recursive optimization problem, describes the functional equation

More information

ADVANCED MACROECONOMIC TECHNIQUES NOTE 3a

ADVANCED MACROECONOMIC TECHNIQUES NOTE 3a 316-406 ADVANCED MACROECONOMIC TECHNIQUES NOTE 3a Chris Edmond hcpedmond@unimelb.edu.aui Dynamic programming and the growth model Dynamic programming and closely related recursive methods provide an important

More information

A comparison of numerical methods for the. Solution of continuous-time DSGE models. Juan Carlos Parra Alvarez

A comparison of numerical methods for the. Solution of continuous-time DSGE models. Juan Carlos Parra Alvarez A comparison of numerical methods for the solution of continuous-time DSGE models Juan Carlos Parra Alvarez Department of Economics and Business, and CREATES Aarhus University, Denmark November 14, 2012

More information

Contents. An example 5. Mathematical Preliminaries 13. Dynamic programming under certainty 29. Numerical methods 41. Some applications 57

Contents. An example 5. Mathematical Preliminaries 13. Dynamic programming under certainty 29. Numerical methods 41. Some applications 57 T H O M A S D E M U Y N C K DY N A M I C O P T I M I Z AT I O N Contents An example 5 Mathematical Preliminaries 13 Dynamic programming under certainty 29 Numerical methods 41 Some applications 57 Stochastic

More information

ADVANCED MACRO TECHNIQUES Midterm Solutions

ADVANCED MACRO TECHNIQUES Midterm Solutions 36-406 ADVANCED MACRO TECHNIQUES Midterm Solutions Chris Edmond hcpedmond@unimelb.edu.aui This exam lasts 90 minutes and has three questions, each of equal marks. Within each question there are a number

More information

University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming

University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming University of Warwick, EC9A0 Maths for Economists 1 of 63 University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming Peter J. Hammond Autumn 2013, revised 2014 University of

More information

Stochastic Shortest Path Problems

Stochastic Shortest Path Problems Chapter 8 Stochastic Shortest Path Problems 1 In this chapter, we study a stochastic version of the shortest path problem of chapter 2, where only probabilities of transitions along different arcs can

More information

Slides II - Dynamic Programming

Slides II - Dynamic Programming Slides II - Dynamic Programming Julio Garín University of Georgia Macroeconomic Theory II (Ph.D.) Spring 2017 Macroeconomic Theory II Slides II - Dynamic Programming Spring 2017 1 / 32 Outline 1. Lagrangian

More information

Basic Deterministic Dynamic Programming

Basic Deterministic Dynamic Programming Basic Deterministic Dynamic Programming Timothy Kam School of Economics & CAMA Australian National University ECON8022, This version March 17, 2008 Motivation What do we do? Outline Deterministic IHDP

More information

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko

ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko Indirect Utility Recall: static consumer theory; J goods, p j is the price of good j (j = 1; : : : ; J), c j is consumption

More information

Dynamic Programming Theorems

Dynamic Programming Theorems Dynamic Programming Theorems Prof. Lutz Hendricks Econ720 September 11, 2017 1 / 39 Dynamic Programming Theorems Useful theorems to characterize the solution to a DP problem. There is no reason to remember

More information

Projection Methods. Michal Kejak CERGE CERGE-EI ( ) 1 / 29

Projection Methods. Michal Kejak CERGE CERGE-EI ( ) 1 / 29 Projection Methods Michal Kejak CERGE CERGE-EI ( ) 1 / 29 Introduction numerical methods for dynamic economies nite-di erence methods initial value problems (Euler method) two-point boundary value problems

More information

Dynamic Programming. Macro 3: Lecture 2. Mark Huggett 2. 2 Georgetown. September, 2016

Dynamic Programming. Macro 3: Lecture 2. Mark Huggett 2. 2 Georgetown. September, 2016 Macro 3: Lecture 2 Mark Huggett 2 2 Georgetown September, 2016 Three Maps: Review of Lecture 1 X = R 1 + and X grid = {x 1,..., x n } X where x i+1 > x i 1. T (v)(x) = max u(f (x) y) + βv(y) s.t. y Γ 1

More information

HOMEWORK #1 This homework assignment is due at 5PM on Friday, November 3 in Marnix Amand s mailbox.

HOMEWORK #1 This homework assignment is due at 5PM on Friday, November 3 in Marnix Amand s mailbox. Econ 50a (second half) Yale University Fall 2006 Prof. Tony Smith HOMEWORK # This homework assignment is due at 5PM on Friday, November 3 in Marnix Amand s mailbox.. Consider a growth model with capital

More information

Problem Set #4 Answer Key

Problem Set #4 Answer Key Problem Set #4 Answer Key Economics 808: Macroeconomic Theory Fall 2004 The cake-eating problem a) Bellman s equation is: b) If this policy is followed: c) If this policy is followed: V (k) = max {log

More information

Introduction to Recursive Methods

Introduction to Recursive Methods Chapter 1 Introduction to Recursive Methods These notes are targeted to advanced Master and Ph.D. students in economics. They can be of some use to researchers in macroeconomic theory. The material contained

More information

Interpolation. 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter

Interpolation. 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter Key References: Interpolation 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter 6. 2. Press, W. et. al. Numerical Recipes in C, Cambridge: Cambridge University Press. Chapter 3

More information

Projection. Wouter J. Den Haan London School of Economics. c by Wouter J. Den Haan

Projection. Wouter J. Den Haan London School of Economics. c by Wouter J. Den Haan Projection Wouter J. Den Haan London School of Economics c by Wouter J. Den Haan Model [ ] ct ν = E t βct+1 ν αz t+1kt+1 α 1 c t + k t+1 = z t k α t ln(z t+1 ) = ρ ln (z t ) + ε t+1 ε t+1 N(0, σ 2 ) k

More information

Value Function Iteration as a Solution Method for the Ramsey Model

Value Function Iteration as a Solution Method for the Ramsey Model Value Function Iteration as a Solution Method for the Ramsey Model BURKHARD HEER ALFRED MAUßNER CESIFO WORKING PAPER NO. 2278 CATEGORY 10: EMPIRICAL AND THEORETICAL METHODS APRIL 2008 An electronic version

More information

Lecture 4: The Bellman Operator Dynamic Programming

Lecture 4: The Bellman Operator Dynamic Programming Lecture 4: The Bellman Operator Dynamic Programming Jeppe Druedahl Department of Economics 15th of February 2016 Slide 1/19 Infinite horizon, t We know V 0 (M t ) = whatever { } V 1 (M t ) = max u(m t,

More information

A simple macro dynamic model with endogenous saving rate: the representative agent model

A simple macro dynamic model with endogenous saving rate: the representative agent model A simple macro dynamic model with endogenous saving rate: the representative agent model Virginia Sánchez-Marcos Macroeconomics, MIE-UNICAN Macroeconomics (MIE-UNICAN) A simple macro dynamic model with

More information

UNIVERSITY OF VIENNA

UNIVERSITY OF VIENNA WORKING PAPERS Cycles and chaos in the one-sector growth model with elastic labor supply Gerhard Sorger May 2015 Working Paper No: 1505 DEPARTMENT OF ECONOMICS UNIVERSITY OF VIENNA All our working papers

More information

Markov Decision Processes Infinite Horizon Problems

Markov Decision Processes Infinite Horizon Problems Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld 1 What is a solution to an MDP? MDP Planning Problem: Input: an MDP (S,A,R,T)

More information

Lecture Notes 10: Dynamic Programming

Lecture Notes 10: Dynamic Programming University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 81 Lecture Notes 10: Dynamic Programming Peter J. Hammond 2018 September 28th University of Warwick, EC9A0 Maths for Economists Peter

More information

Hoover Institution, Stanford University. Department of Economics, Stanford University. Department of Economics, Santa Clara University

Hoover Institution, Stanford University. Department of Economics, Stanford University. Department of Economics, Santa Clara University Supplementary Material Supplement to How to solve dynamic stochastic models computing expectations just once : Appendices (Quantitative Economics, Vol. 8, No. 3, November 2017, 851 893) Kenneth L. Judd

More information

Lecture notes: Rust (1987) Economics : M. Shum 1

Lecture notes: Rust (1987) Economics : M. Shum 1 Economics 180.672: M. Shum 1 Estimate the parameters of a dynamic optimization problem: when to replace engine of a bus? This is a another practical example of an optimal stopping problem, which is easily

More information

Macro 1: Dynamic Programming 1

Macro 1: Dynamic Programming 1 Macro 1: Dynamic Programming 1 Mark Huggett 2 2 Georgetown September, 2016 DP Warm up: Cake eating problem ( ) max f 1 (y 1 ) + f 2 (y 2 ) s.t. y 1 + y 2 100, y 1 0, y 2 0 1. v 1 (x) max f 1(y 1 ) + f

More information

Lecture Notes On Solution Methods for Representative Agent Dynamic Stochastic Optimization Problems

Lecture Notes On Solution Methods for Representative Agent Dynamic Stochastic Optimization Problems Comments Welcome Lecture Notes On Solution Methods for Representative Agent Dynamic Stochastic Optimization Problems Christopher D. Carroll ccarroll@jhu.edu May 4, 2011 Abstract These lecture notes sketch

More information

Dynamic Optimization Using Lagrange Multipliers

Dynamic Optimization Using Lagrange Multipliers Dynamic Optimization Using Lagrange Multipliers Barbara Annicchiarico barbara.annicchiarico@uniroma2.it Università degli Studi di Roma "Tor Vergata" Presentation #2 Deterministic Infinite-Horizon Ramsey

More information

COURSE Iterative methods for solving linear systems

COURSE Iterative methods for solving linear systems COURSE 0 4.3. Iterative methods for solving linear systems Because of round-off errors, direct methods become less efficient than iterative methods for large systems (>00 000 variables). An iterative scheme

More information

Lecture 7: Linear-Quadratic Dynamic Programming Real Business Cycle Models

Lecture 7: Linear-Quadratic Dynamic Programming Real Business Cycle Models Lecture 7: Linear-Quadratic Dynamic Programming Real Business Cycle Models Shinichi Nishiyama Graduate School of Economics Kyoto University January 10, 2019 Abstract In this lecture, we solve and simulate

More information

Dynamic Programming: Numerical Methods

Dynamic Programming: Numerical Methods Chapter 4 Dynamic Programming: Numerical Methods Many approaches to solving a Bellman equation have their roots in the simple idea of value function iteration and the guess and verify methods. These are

More information

Optimal Control. McGill COMP 765 Oct 3 rd, 2017

Optimal Control. McGill COMP 765 Oct 3 rd, 2017 Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps

More information

Example I: Capital Accumulation

Example I: Capital Accumulation 1 Example I: Capital Accumulation Time t = 0, 1,..., T < Output y, initial output y 0 Fraction of output invested a, capital k = ay Transition (production function) y = g(k) = g(ay) Reward (utility of

More information

1 THE GAME. Two players, i=1, 2 U i : concave, strictly increasing f: concave, continuous, f(0) 0 β (0, 1): discount factor, common

1 THE GAME. Two players, i=1, 2 U i : concave, strictly increasing f: concave, continuous, f(0) 0 β (0, 1): discount factor, common 1 THE GAME Two players, i=1, 2 U i : concave, strictly increasing f: concave, continuous, f(0) 0 β (0, 1): discount factor, common With Law of motion of the state: Payoff: Histories: Strategies: k t+1

More information

Practical Dynamic Programming: An Introduction. Associated programs dpexample.m: deterministic dpexample2.m: stochastic

Practical Dynamic Programming: An Introduction. Associated programs dpexample.m: deterministic dpexample2.m: stochastic Practical Dynamic Programming: An Introduction Associated programs dpexample.m: deterministic dpexample2.m: stochastic Outline 1. Specific problem: stochastic model of accumulation from a DP perspective

More information

Lecture 7: Stochastic Dynamic Programing and Markov Processes

Lecture 7: Stochastic Dynamic Programing and Markov Processes Lecture 7: Stochastic Dynamic Programing and Markov Processes Florian Scheuer References: SLP chapters 9, 10, 11; LS chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology

More information

Topic 2. Consumption/Saving and Productivity shocks

Topic 2. Consumption/Saving and Productivity shocks 14.452. Topic 2. Consumption/Saving and Productivity shocks Olivier Blanchard April 2006 Nr. 1 1. What starting point? Want to start with a model with at least two ingredients: Shocks, so uncertainty.

More information

High-dimensional Problems in Finance and Economics. Thomas M. Mertens

High-dimensional Problems in Finance and Economics. Thomas M. Mertens High-dimensional Problems in Finance and Economics Thomas M. Mertens NYU Stern Risk Economics Lab April 17, 2012 1 / 78 Motivation Many problems in finance and economics are high dimensional. Dynamic Optimization:

More information

Dynamic Optimization Problem. April 2, Graduate School of Economics, University of Tokyo. Math Camp Day 4. Daiki Kishishita.

Dynamic Optimization Problem. April 2, Graduate School of Economics, University of Tokyo. Math Camp Day 4. Daiki Kishishita. Discrete Math Camp Optimization Problem Graduate School of Economics, University of Tokyo April 2, 2016 Goal of day 4 Discrete We discuss methods both in discrete and continuous : Discrete : condition

More information

Suggested Solutions to Homework #3 Econ 511b (Part I), Spring 2004

Suggested Solutions to Homework #3 Econ 511b (Part I), Spring 2004 Suggested Solutions to Homework #3 Econ 5b (Part I), Spring 2004. Consider an exchange economy with two (types of) consumers. Type-A consumers comprise fraction λ of the economy s population and type-b

More information

Some AI Planning Problems

Some AI Planning Problems Course Logistics CS533: Intelligent Agents and Decision Making M, W, F: 1:00 1:50 Instructor: Alan Fern (KEC2071) Office hours: by appointment (see me after class or send email) Emailing me: include CS533

More information

An approximate consumption function

An approximate consumption function An approximate consumption function Mario Padula Very Preliminary and Very Incomplete 8 December 2005 Abstract This notes proposes an approximation to the consumption function in the buffer-stock model.

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

An Application to Growth Theory

An Application to Growth Theory An Application to Growth Theory First let s review the concepts of solution function and value function for a maximization problem. Suppose we have the problem max F (x, α) subject to G(x, β) 0, (P) x

More information

A Computational Method for Multidimensional Continuous-choice. Dynamic Problems

A Computational Method for Multidimensional Continuous-choice. Dynamic Problems A Computational Method for Multidimensional Continuous-choice Dynamic Problems (Preliminary) Xiaolu Zhou School of Economics & Wangyannan Institution for Studies in Economics Xiamen University April 9,

More information

Value Function Iteration

Value Function Iteration Value Function Iteration (Lectures on Solution Methods for Economists II) Jesús Fernández-Villaverde 1 and Pablo Guerrón 2 February 26, 2018 1 University of Pennsylvania 2 Boston College Theoretical Background

More information

Outline Today s Lecture

Outline Today s Lecture Outline Today s Lecture finish Euler Equations and Transversality Condition Principle of Optimality: Bellman s Equation Study of Bellman equation with bounded F contraction mapping and theorem of the maximum

More information

1. Using the model and notations covered in class, the expected returns are:

1. Using the model and notations covered in class, the expected returns are: Econ 510a second half Yale University Fall 2006 Prof. Tony Smith HOMEWORK #5 This homework assignment is due at 5PM on Friday, December 8 in Marnix Amand s mailbox. Solution 1. a In the Mehra-Prescott

More information

Discrete State Space Methods for Dynamic Economies

Discrete State Space Methods for Dynamic Economies Discrete State Space Methods for Dynamic Economies A Brief Introduction Craig Burnside Duke University September 2006 Craig Burnside (Duke University) Discrete State Space Methods September 2006 1 / 42

More information

Neoclassical Growth Model: I

Neoclassical Growth Model: I Neoclassical Growth Model: I Mark Huggett 2 2 Georgetown October, 2017 Growth Model: Introduction Neoclassical Growth Model is the workhorse model in macroeconomics. It comes in two main varieties: infinitely-lived

More information

Lecture 4 The Centralized Economy: Extensions

Lecture 4 The Centralized Economy: Extensions Lecture 4 The Centralized Economy: Extensions Leopold von Thadden University of Mainz and ECB (on leave) Advanced Macroeconomics, Winter Term 2013 1 / 36 I Motivation This Lecture considers some applications

More information

Linear and Loglinear Approximations (Started: July 7, 2005; Revised: February 6, 2009)

Linear and Loglinear Approximations (Started: July 7, 2005; Revised: February 6, 2009) Dave Backus / NYU Linear and Loglinear Approximations (Started: July 7, 2005; Revised: February 6, 2009) Campbell s Inspecting the mechanism (JME, 1994) added some useful intuition to the stochastic growth

More information

Lecture 3: Hamilton-Jacobi-Bellman Equations. Distributional Macroeconomics. Benjamin Moll. Part II of ECON Harvard University, Spring

Lecture 3: Hamilton-Jacobi-Bellman Equations. Distributional Macroeconomics. Benjamin Moll. Part II of ECON Harvard University, Spring Lecture 3: Hamilton-Jacobi-Bellman Equations Distributional Macroeconomics Part II of ECON 2149 Benjamin Moll Harvard University, Spring 2018 1 Outline 1. Hamilton-Jacobi-Bellman equations in deterministic

More information

Advanced Economic Growth: Lecture 21: Stochastic Dynamic Programming and Applications

Advanced Economic Growth: Lecture 21: Stochastic Dynamic Programming and Applications Advanced Economic Growth: Lecture 21: Stochastic Dynamic Programming and Applications Daron Acemoglu MIT November 19, 2007 Daron Acemoglu (MIT) Advanced Growth Lecture 21 November 19, 2007 1 / 79 Stochastic

More information

Economics 210B Due: September 16, Problem Set 10. s.t. k t+1 = R(k t c t ) for all t 0, and k 0 given, lim. and

Economics 210B Due: September 16, Problem Set 10. s.t. k t+1 = R(k t c t ) for all t 0, and k 0 given, lim. and Economics 210B Due: September 16, 2010 Problem 1: Constant returns to saving Consider the following problem. c0,k1,c1,k2,... β t Problem Set 10 1 α c1 α t s.t. k t+1 = R(k t c t ) for all t 0, and k 0

More information

Recall that in finite dynamic optimization we are trying to optimize problems of the form. T β t u(c t ) max. c t. t=0 T. c t = W max. subject to.

Recall that in finite dynamic optimization we are trying to optimize problems of the form. T β t u(c t ) max. c t. t=0 T. c t = W max. subject to. 19 Policy Function Iteration Lab Objective: Learn how iterative methods can be used to solve dynamic optimization problems. Implement value iteration and policy iteration in a pseudo-infinite setting where

More information

problem. max Both k (0) and h (0) are given at time 0. (a) Write down the Hamilton-Jacobi-Bellman (HJB) Equation in the dynamic programming

problem. max Both k (0) and h (0) are given at time 0. (a) Write down the Hamilton-Jacobi-Bellman (HJB) Equation in the dynamic programming 1. Endogenous Growth with Human Capital Consider the following endogenous growth model with both physical capital (k (t)) and human capital (h (t)) in continuous time. The representative household solves

More information

Elements of Reinforcement Learning

Elements of Reinforcement Learning Elements of Reinforcement Learning Policy: way learning algorithm behaves (mapping from state to action) Reward function: Mapping of state action pair to reward or cost Value function: long term reward,

More information

Fixed Points and Contractive Transformations. Ron Goldman Department of Computer Science Rice University

Fixed Points and Contractive Transformations. Ron Goldman Department of Computer Science Rice University Fixed Points and Contractive Transformations Ron Goldman Department of Computer Science Rice University Applications Computer Graphics Fractals Bezier and B-Spline Curves and Surfaces Root Finding Newton

More information

Value and Policy Iteration

Value and Policy Iteration Chapter 7 Value and Policy Iteration 1 For infinite horizon problems, we need to replace our basic computational tool, the DP algorithm, which we used to compute the optimal cost and policy for finite

More information

Macroeconomic Theory II Homework 2 - Solution

Macroeconomic Theory II Homework 2 - Solution Macroeconomic Theory II Homework 2 - Solution Professor Gianluca Violante, TA: Diego Daruich New York University Spring 204 Problem The household has preferences over the stochastic processes of a single

More information

5. Solving the Bellman Equation

5. Solving the Bellman Equation 5. Solving the Bellman Equation In the next two lectures, we will look at several methods to solve Bellman s Equation (BE) for the stochastic shortest path problem: Value Iteration, Policy Iteration and

More information

1 Jan 28: Overview and Review of Equilibrium

1 Jan 28: Overview and Review of Equilibrium 1 Jan 28: Overview and Review of Equilibrium 1.1 Introduction What is an equilibrium (EQM)? Loosely speaking, an equilibrium is a mapping from environments (preference, technology, information, market

More information

HOMEWORK #3 This homework assignment is due at NOON on Friday, November 17 in Marnix Amand s mailbox.

HOMEWORK #3 This homework assignment is due at NOON on Friday, November 17 in Marnix Amand s mailbox. Econ 50a second half) Yale University Fall 2006 Prof. Tony Smith HOMEWORK #3 This homework assignment is due at NOON on Friday, November 7 in Marnix Amand s mailbox.. This problem introduces wealth inequality

More information

Dynamic Problem Set 1 Solutions

Dynamic Problem Set 1 Solutions Dynamic Problem Set 1 Solutions Jonathan Kreamer July 15, 2011 Question 1 Consider the following multi-period optimal storage problem: An economic agent imizes: c t} T β t u(c t ) (1) subject to the period-by-period

More information

Lecture 9. Dynamic Programming. Randall Romero Aguilar, PhD I Semestre 2017 Last updated: June 2, 2017

Lecture 9. Dynamic Programming. Randall Romero Aguilar, PhD I Semestre 2017 Last updated: June 2, 2017 Lecture 9 Dynamic Programming Randall Romero Aguilar, PhD I Semestre 2017 Last updated: June 2, 2017 Universidad de Costa Rica EC3201 - Teoría Macroeconómica 2 Table of contents 1. Introduction 2. Basics

More information

Lecture 10. Dynamic Programming. Randall Romero Aguilar, PhD II Semestre 2017 Last updated: October 16, 2017

Lecture 10. Dynamic Programming. Randall Romero Aguilar, PhD II Semestre 2017 Last updated: October 16, 2017 Lecture 10 Dynamic Programming Randall Romero Aguilar, PhD II Semestre 2017 Last updated: October 16, 2017 Universidad de Costa Rica EC3201 - Teoría Macroeconómica 2 Table of contents 1. Introduction 2.

More information

Hoover Institution, Stanford University and NBER. University of Alicante and Hoover Institution, Stanford University

Hoover Institution, Stanford University and NBER. University of Alicante and Hoover Institution, Stanford University Supplementary Material Supplement to Numerically stable and accurate stochastic simulation approaches for solving dynamic economic models : Appendices (Quantitative Economics, Vol. 2, No. 2, July 2011,

More information

Lecture 5: The Bellman Equation

Lecture 5: The Bellman Equation Lecture 5: The Bellman Equation Florian Scheuer 1 Plan Prove properties of the Bellman equation (In particular, existence and uniqueness of solution) Use this to prove properties of the solution Think

More information

Reinforcement Learning and Optimal Control. ASU, CSE 691, Winter 2019

Reinforcement Learning and Optimal Control. ASU, CSE 691, Winter 2019 Reinforcement Learning and Optimal Control ASU, CSE 691, Winter 2019 Dimitri P. Bertsekas dimitrib@mit.edu Lecture 8 Bertsekas Reinforcement Learning 1 / 21 Outline 1 Review of Infinite Horizon Problems

More information

Technical appendices: Business cycle accounting for the Japanese economy using the parameterized expectations algorithm

Technical appendices: Business cycle accounting for the Japanese economy using the parameterized expectations algorithm Technical appendices: Business cycle accounting for the Japanese economy using the parameterized expectations algorithm Masaru Inaba November 26, 2007 Introduction. Inaba (2007a) apply the parameterized

More information

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b) Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)

More information

Lecture 2 The Centralized Economy

Lecture 2 The Centralized Economy Lecture 2 The Centralized Economy Economics 5118 Macroeconomic Theory Kam Yu Winter 2013 Outline 1 Introduction 2 The Basic DGE Closed Economy 3 Golden Rule Solution 4 Optimal Solution The Euler Equation

More information

Stochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search

Stochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search Stochastic Problems References: SLP chapters 9, 10, 11; L&S chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology Production function y = Af k where A is random Let A s t

More information

The Market Resources Method for Solving Dynamic Optimization Problems *

The Market Resources Method for Solving Dynamic Optimization Problems * Federal Reserve Bank of Dallas Globalization and Monetary Policy Institute Working Paper No. 274 http://www.dallasfed.org/assets/documents/institute/wpapers/2016/0274.pdf The Market Resources Method for

More information

ECON607 Fall 2010 University of Hawaii Professor Hui He TA: Xiaodong Sun Assignment 2

ECON607 Fall 2010 University of Hawaii Professor Hui He TA: Xiaodong Sun Assignment 2 ECON607 Fall 200 University of Hawaii Professor Hui He TA: Xiaodong Sun Assignment 2 The due date for this assignment is Tuesday, October 2. ( Total points = 50). (Two-sector growth model) Consider the

More information

Exam in TMA4215 December 7th 2012

Exam in TMA4215 December 7th 2012 Norwegian University of Science and Technology Department of Mathematical Sciences Page of 9 Contact during the exam: Elena Celledoni, tlf. 7359354, cell phone 48238584 Exam in TMA425 December 7th 22 Allowed

More information

Solution Methods. Jesús Fernández-Villaverde. University of Pennsylvania. March 16, 2016

Solution Methods. Jesús Fernández-Villaverde. University of Pennsylvania. March 16, 2016 Solution Methods Jesús Fernández-Villaverde University of Pennsylvania March 16, 2016 Jesús Fernández-Villaverde (PENN) Solution Methods March 16, 2016 1 / 36 Functional equations A large class of problems

More information

Uncertainty Per Krusell & D. Krueger Lecture Notes Chapter 6

Uncertainty Per Krusell & D. Krueger Lecture Notes Chapter 6 1 Uncertainty Per Krusell & D. Krueger Lecture Notes Chapter 6 1 A Two-Period Example Suppose the economy lasts only two periods, t =0, 1. The uncertainty arises in the income (wage) of period 1. Not that

More information

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Optimal Control Ömer Özak SMU Macroeconomics II Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Review of the Theory of Optimal Control Section 1 Review of the Theory of Optimal Control Ömer

More information

ECON 2010c Solution to Problem Set 1

ECON 2010c Solution to Problem Set 1 ECON 200c Solution to Problem Set By the Teaching Fellows for ECON 200c Fall 204 Growth Model (a) Defining the constant κ as: κ = ln( αβ) + αβ αβ ln(αβ), the problem asks us to show that the following

More information

Optimization, Part 2 (november to december): mandatory for QEM-IMAEF, and for MMEF or MAEF who have chosen it as an optional course.

Optimization, Part 2 (november to december): mandatory for QEM-IMAEF, and for MMEF or MAEF who have chosen it as an optional course. Paris. Optimization, Part 2 (november to december): mandatory for QEM-IMAEF, and for MMEF or MAEF who have chosen it as an optional course. Philippe Bich (Paris 1 Panthéon-Sorbonne and PSE) Paris, 2016.

More information

DSICE: A Dynamic Stochastic Integrated Model of Climate and Economy

DSICE: A Dynamic Stochastic Integrated Model of Climate and Economy DSICE: A Dynamic Stochastic Integrated Model of Climate and Economy Yongyang Cai Hoover Institution, 424 Galvez Mall, Stanford University, Stanford, CA 94305 yycai@stanford.edu Kenneth L. Judd Hoover Institution,

More information

Notes on Control Theory

Notes on Control Theory Notes on Control Theory max t 1 f t, x t, u t dt # ẋ g t, x t, u t # t 0, t 1, x t 0 x 0 fixed, t 1 can be. x t 1 maybefreeorfixed The choice variable is a function u t which is piecewise continuous, that

More information

Lecture 2 The Centralized Economy: Basic features

Lecture 2 The Centralized Economy: Basic features Lecture 2 The Centralized Economy: Basic features Leopold von Thadden University of Mainz and ECB (on leave) Advanced Macroeconomics, Winter Term 2013 1 / 41 I Motivation This Lecture introduces the basic

More information

Stochastic Dynamic Programming: The One Sector Growth Model

Stochastic Dynamic Programming: The One Sector Growth Model Stochastic Dynamic Programming: The One Sector Growth Model Esteban Rossi-Hansberg Princeton University March 26, 2012 Esteban Rossi-Hansberg () Stochastic Dynamic Programming March 26, 2012 1 / 31 References

More information

Christopher Watkins and Peter Dayan. Noga Zaslavsky. The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015

Christopher Watkins and Peter Dayan. Noga Zaslavsky. The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015 Q-Learning Christopher Watkins and Peter Dayan Noga Zaslavsky The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015 Noga Zaslavsky Q-Learning (Watkins & Dayan, 1992)

More information

Foundation of (virtually) all DSGE models (e.g., RBC model) is Solow growth model

Foundation of (virtually) all DSGE models (e.g., RBC model) is Solow growth model THE BASELINE RBC MODEL: THEORY AND COMPUTATION FEBRUARY, 202 STYLIZED MACRO FACTS Foundation of (virtually all DSGE models (e.g., RBC model is Solow growth model So want/need/desire business-cycle models

More information

Dynamic Discrete Choice Structural Models in Empirical IO

Dynamic Discrete Choice Structural Models in Empirical IO Dynamic Discrete Choice Structural Models in Empirical IO Lecture 4: Euler Equations and Finite Dependence in Dynamic Discrete Choice Models Victor Aguirregabiria (University of Toronto) Carlos III, Madrid

More information

Macroeconomics Qualifying Examination

Macroeconomics Qualifying Examination Macroeconomics Qualifying Examination August 2016 Department of Economics UNC Chapel Hill Instructions: This examination consists of 4 questions. Answer all questions. If you believe a question is ambiguously

More information