Recursive Methods Recursive Methods Nr. 1

Similar documents
Recursive Methods Recursive Methods Nr. 1

Recursive Metho ds. Introduction to Dynamic Optimization Nr

Computing Equilibria of Repeated And Dynamic Games

Chapter 23 Credible Government Policies, I

The Time Consistency Problem - Theory and Applications

Lecture 1. Evolution of Market Concentration

6.254 : Game Theory with Engineering Applications Lecture 13: Extensive Form Games

Copyright (C) 2013 David K. Levine This document is an open textbook; you can redistribute it and/or modify it under the terms of the Creative

Dynamic stochastic game and macroeconomic equilibrium

POLICY GAMES. u t = θ 0 θ 1 π t + ε t. (1) β t (u 2 t + ωπ 2 t ). (3) π t = g t 1 + ν t. (4) g t = θ 1θ 0 ω + θ 2 1. (5)

Political Economy of Institutions and Development: Problem Set 1. Due Date: Thursday, February 23, in class.

NASH IMPLEMENTATION USING SIMPLE MECHANISMS WITHOUT UNDESIRABLE MIXED-STRATEGY EQUILIBRIA

Long-Run versus Short-Run Player

POLICY GAMES. u t = θ 0 θ 1 π t + ε t. (1) β t (u 2 t + ωπ 2 t ). (3)

Simple Consumption / Savings Problems (based on Ljungqvist & Sargent, Ch 16, 17) Jonathan Heathcote. updated, March The household s problem X

Equilibrium Refinements

Ph.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2016

An Application to Growth Theory

3.3.3 Illustration: Infinitely repeated Cournot duopoly.

Eco504, Part II Spring 2008 C. Sims POLICY GAMES

Fiscal Rules and Discretion under Self-Enforcement

A Theory of Financing Constraints and Firm Dynamics by Clementi and Hopenhayn - Quarterly Journal of Economics (2006)

TIME INCONSISTENCY AND REPUTATION IN MONETARY POLICY: A STRATEGIC MODELLING IN CONTINUOUS TIME

Political Economy of Institutions and Development. Lecture 8. Institutional Change and Democratization

Organizational Equilibrium with Capital

Game Theory Fall 2003

Contracts under Asymmetric Information

Basics of Game Theory

Bayes Correlated Equilibrium and Comparing Information Structures

A New and Robust Subgame Perfect Equilibrium in a model of Triadic Power Relations *

Introduction to Numerical Methods

Inflation traps, and rules vs. discretion

Repeated Implementation with Finite Mechanisms and Complexity

Graduate Microeconomics II Lecture 5: Cheap Talk. Patrick Legros

Differentiable Welfare Theorems Existence of a Competitive Equilibrium: Preliminaries

Repeated Games with Perfect Monitoring

14.05 Lecture Notes Crises and Multiple Equilibria

Delegation, Time Inconsistency and Sustainable Equilibrium

Economics 201B Economic Theory (Spring 2017) Bargaining. Topics: the axiomatic approach (OR 15) and the strategic approach (OR 7).

6.207/14.15: Networks Lecture 16: Cooperation and Trust in Networks

Robust Conditional Predictions in Dynamic Games: An Application to Sovereign Debt

Recap Social Choice Functions Fun Game Mechanism Design. Mechanism Design. Lecture 13. Mechanism Design Lecture 13, Slide 1

REPEATED GAMES. Jörgen Weibull. April 13, 2010

: Cryptography and Game Theory Ran Canetti and Alon Rosen. Lecture 8

Microeconomic Theory (501b) Problem Set 10. Auctions and Moral Hazard Suggested Solution: Tibor Heumann

1 Markov decision processes

Correlated Equilibrium in Games with Incomplete Information

Economics 3012 Strategic Behavior Andy McLennan October 20, 2006

6.207/14.15: Networks Lecture 11: Introduction to Game Theory 3

Informed Principal in Private-Value Environments

Avoidance and Mitigation of Public Harm

Sequential Equilibria in a Ramsey Tax Model

The Folk Theorem for Finitely Repeated Games with Mixed Strategies

Time Consistency in Alternating Move Policy Games

Foundations of Modern Macroeconomics Second Edition

The Optimal Degree of Monetary-Discretion in a New Keynesian Model with Private Information

1 Jan 28: Overview and Review of Equilibrium

EC3224 Autumn Lecture #03 Applications of Nash Equilibrium

Pledge-and-Review Bargaining

4: Dynamic games. Concordia February 6, 2017

MS&E338 Reinforcement Learning Lecture 1 - April 2, Introduction

Note on Time Consistency

Economics 385: Suggested Solutions 2

Bayesian-Nash Equilibrium

Real Business Cycle Model (RBC)

Lecture Doraszelski and Judd. Dynamic and Stochastic Model of Industry

Module 16: Signaling

Module 8: Multi-Agent Models of Moral Hazard

NBER WORKING PAPER SERIES POWER FLUCTUATIONS AND POLITICAL ECONOMY. Daron Acemoglu Mikhail Golosov Aleh Tsyvinski

Advanced Microeconomics

Economics 2010c: Lecture 2 Iterative Methods in Dynamic Programming

Game Theory with Information: Introducing the Witsenhausen Intrinsic Model

Solving Extensive Form Games

1 THE GAME. Two players, i=1, 2 U i : concave, strictly increasing f: concave, continuous, f(0) 0 β (0, 1): discount factor, common

Political Economy of Institutions and Development. Lecture 8. Institutional Change and Democratization

Efficient Repeated Implementation

Industrial Organization Lecture 3: Game Theory

Dynamic Common Agency

Organizational Equilibrium with Capital

Submitted to Econometrica SIGNALING AND REPUTATION IN REPEATED GAMES, I: FINITE GAMES CHARLES RODDIE 1

Representation of TU games by coalition production economies

Lecture Notes: Self-enforcing agreements

On Reputation with Imperfect Monitoring

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

Mechanism Design: Basic Concepts

Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Extensive games with perfect information OR6and7,FT3,4and11

Competition relative to Incentive Functions in Common Agency

Lecture Notes on Bargaining

Competing Mechanisms: One-Shot versus Repeated Games

Topic 9. Monetary policy. Notes.

DEPARTMENT OF ECONOMICS IMPLEMENTATION IN MIXED NASH EQUILIBRIUM

First Prev Next Last Go Back Full Screen Close Quit. Game Theory. Giorgio Fagiolo

McCall Model. Prof. Lutz Hendricks. November 22, Econ720

SURPLUS SHARING WITH A TWO-STAGE MECHANISM. By Todd R. Kaplan and David Wettstein 1. Ben-Gurion University of the Negev, Israel. 1.

Lecture notes for Analysis of Algorithms : Markov decision processes

Dynamic Games, II: Bargaining

COMPARISON OF INFORMATION STRUCTURES IN ZERO-SUM GAMES. 1. Introduction

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Microeconomics. 2. Game Theory

Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty

Transcription:

Nr. 1

Outline Today s Lecture Dynamic Programming under Uncertainty notation of sequence problem leave study of dynamics for next week Dynamic Recursive Games: Abreu-Pearce-Stachetti Application: today s Macro seminar Nr. 2

Dynamic Programming with Uncertainty general model of uncertainty: need Measure Theory for simplicity: finite state S Markov process for s (recursive uncertainty) Pr s t+1 s t = p (s t+1 s t ) v (x 0,s 0 ) sup {x t+1 ( )} t=0 ( X X β t F xt s t 1,x t+1 s t Pr s t ) s 0 t s t x t+1 s t Γ x t s t 1 x 0 given Nr. 3

Dynamic Programming Functional Equation (Bellman Equation) ( v (x, s) = sup F (x, y)+ β X ) v (y, s 0 ) p (s 0 s) or simply (or more generally): v (x, s) =sup {F (x, y)+ βe [v (y, s 0 ) s]} where the E [ s] is the conditional expectation operator over s 0 given s basically same: Ppple of Optimality, Contraction Mapping (bounded case), Monotonicity [actually: differentiability sometimes easier!] notational gain is huge! s 0 Nr. 4

Policy Rules Rule more intuitive too! fundamental change in the notion of a solution optimal policy g (x, s) vs. optimal sequence of contingent plan {x t+1 (s t )} t=0 Question: how can we use g to understand the dynamics of the solution? (important for many models) Answer: next week... Nr. 5

Abreu Pearce and Stachetti (APS) Dynamic Programming for Dynamic Games idea: subgame perfect equilibria of repeated games have recursive structure players care about future strategies only through their associated utility values APS study general N person game with non-observable actions we follow Ljungqvist-Sargent: continuum of identical agents vs. benevolent government time consistency problems (credibility through reputation) agent i has preferences u (x i,x,y) where x is average across x i s Nr. 6

One Period competitive equilibria: ½ ¾ C = (x, y) : x arg max u (x i,x,y) x i assume x = h (y) for all (x, y) C 1. Dictatorial allocation: max x,y u (x, x, y) (wishful thinking!) 2. Ramsey commitment allocation: max (x,y) C u (x, x, y) (wishful think - ing?) 3. Nash equilibrium x N,y N : (might be bad outcome) x N arg max x y N arg max y u x u x, x N,y N x N,y N C N,x N,y y N N = H x Nr. 7

Kydland-Prescott / Barro-Gordon v (u, π) = u 2 π 2 u = ū (π π e ) u (π e i,π e,π) = v (ū (π π e ),π) λ (π e i π) 2 then π e i = πe = π = h (π) take λ 0 then First best Ramsey: max π = (ū (π π e )) 2 π 2 λ (π e i π) 2 (ū π + π e ) 2 π 2 n o n o (ū π + h (π)) 2 π 2 = max (ū) 2 π 2 π π =0 Nr. 8

Kydland-Prescott / Barro-Gordon Nash outcome. Gov t optimal reaction: n o max (ū π + π e ) 2 π 2 π this is π = H (π e ) π = ū + πe 2 Nash equilibria is then π = H (h (π)) = H (π) = ū+π 2 which implies π en = π N =ū unemployment stays at ū but positive inflation worse off Andy Atkeson: adds shock θ that is private info of gov t (macro seminar) Nr. 9

Infinitely Repeated Economy Payoff for government: V g = 1 δ δ X δ t r (x t,y t ) t=1 where r (x, y) = u (x, x, y) strategies σ... σ g = σ h = σ g t σ h t x t 1,y t 1 ª t=0 x t 1,y t 1 ª t=0 induce {x t,y t } from which we can write V g (σ). continuation stategies: after history (x t,y t ) we write σ (x t,y t ) Nr. 10

Subgame Perfect Equilibrium Astrategy profile σ = (σ h,σ g ) is a subgame perfect equilibrium of the infinitely repeated economy if for each t 1 and each history (x t 1,y t 1 ) X t 1 Y t 1, 1. The outcome x t = σ h t (x t 1,y t 1 ) is a competitive equilibrium given that y t = σ g t (x t 1,y t 1 ),i.e. (x t,y t ) C 2. For each ŷ Y (1 δ)r(x t,y t )+δv g (σ (x t,y t )) (1 δ)r(x t, ŷ)+δv g (σ (x t ;y t 1,ŷ)) (one shot deviations are not optimal) Nr. 11

Lemma Take σ and let x and y be the associated first period outcome. Then σ is sub-game perfect if and only if: 1. for all (ˆx, ŷ) X Y σ ˆx,ŷ is a sub-game perfect equilibrium 2. (x, y) C 3. ŷ Y (1 δ)r(x t,y t )+ δv g (σ (x,y) ) (1 δ)r(x t, ŷ)+ δv g (σ (ˆx,ŷ) ) note the stellar role of V g (σ (x,y) ) and V g (σ (ˆx,ŷ) ), its all that matters for checking whether it is best to do x or deviate... idea! think about values as fundamental Nr. 12

Values of all SPE Set V of values V = V g (σ) σis a subgame perfect equilibrium Let W R. A 4-tuple (x, y, ω 1,ω 2 ) is said to be admissible with respect to W if (x, y) C, ω 1,ω 2 W W and (1 δ)r(x, y)+ δω 1 (1 δ)r(x, ŷ)+ δω 2, ŷ Y Nr. 13

B(W) operator Definition: For each set W R, let B(W ) be the set of possible values ω = (1 δ)r(x, y)+ δω 1 associated with some admissible tuples (x, y, ω 1,ω 2 ) wrt W : B(W ) ½ w : (x, y) C and ω 1,ω 2 W s.t. (1 δ)r(x, y)+ δω 1 (1 δ)r(x, ŷ)+ δω 2, ŷ Y note that V is a fixed point B (V ) = V we will see that V is the biggest fixed point ¾ Nr. 14

Monotonicity of B. If W W 0 R then B(W ) B(W 0 ) Theorem (self-generation): If W R is bounded and W B(W ) (self-generating ) then B(W ) V Proof Step 1 : for any W B(W ) we can choose and x, y, ω 1, and ω 2 (1 δ)r(x, y)+ δω 1 (1 δ)r(x, ŷ)+ δω 2, ŷ Y Step 2: for ω 1,ω 2 W thus do the same thing for them as in step 1 continue in this way... Nr. 15

Three facts and analgorithm V B(V ) If W B(W ), then B(W ) V (by self-generation) B is monotone and maps compact sets into compact sets Algorithm: start with W 0 such that V B (W 0 ) W 0 then define W n = B n (W 0 ) W n V Proof: since W n are decreasing (and compact) they must converge, the limit must bea fixed point, but V is biggest fixed point Nr. 16

Finding V In this simple case here we can do more... lowest v is self-enforcing highest v is self-rewarding v low = min {(1 δ) r (x, y)+ δv} (x,y) C v V (1 δ)r(x, y)+ δv (1 δ)r(x, ŷ)+ δv low all ŷ Y v low =(1 δ)r(h (y),y)+ δv (1 δ)r(h (y),h (h (y))) + δv low if binds and v > v low then minimize RHS of inequality v low =min y r(h (y),h (h (y))) Nr. 17

Best Value for Best, use Worst to punish and Best as reward solve: max (x,y) C v V = {(1 δ) r (x, y)+ δv high } (1 δ)r(x, y)+ δv high (1 δ)r(x, ŷ)+ δv low all ŷ Y then clearly v high = r (x, y) so max r (h (y),y) subject to r (h (y),y) (1 δ)r(h (y),h (h (y))) + δv low if constraint not binding Ramsey (first best) otherwise value is constrained by v low Nr. 18