Introduction to Mean Field Game Theory Part I

Size: px
Start display at page:

Download "Introduction to Mean Field Game Theory Part I"

Transcription

1 School of Mathematics and Statistics Carleton University Ottawa, Canada Workshop on Game Theory, Institute for Math. Sciences National University of Singapore, June 2018

2 Outline of talk Background and motivation for mean field game (MFG) theory Illustrative modeling examples Fundamental methodologies Applications and references

3 A mean field game is a situation of stochastic (dynamic) decision making where each agent interacts with the aggregate effect of all other agents; agents are non-cooperative. Example: Hybrid electric vehicle recharging control (interacting through aggregate load/price)

4 Noncooperative games: historical background 1. Cournot duopoly Cournot equilibrium, person zero-sum von Neumann s minimax theorem, N-person nonzero-sum Nash equilibrium, 1950 Antoine Cournot John von Neumann John Nash ( ) ( ) ( )

5 Two directions for generalizations: Include dynamics : L. Shapley (1953), MDP model; Rufus Isaacs (1965), Differential games, Wiley (work of 1950s) Consider large populations of players Technical challenges arise... the Curse of Dimensionality with a large number of players Example: 50 players, each having 2 states, 2 actions. Need a system configuration: 4 50 =

6 Early efforts to deal with large populations J. W. Milnor and L.S. Shapley (1978); R. J. Aumann (1975). This is for cooperative games E. J. Green (1984). Non-cooperative games M. Ali Khan and Y. Sun (2002). Survey on Non-cooperative games with many players. Static models B. Jovanovic and R. W. Rosenthal (1988). Anonymous sequential games. It considers distributional strategies for an infinite population. However, individual behaviour is not addressed...

7 The early motivation for mean field games A wireless network power control problem

8 Rate based power control (Huang, Caines, Malhamé 03) dp i = u i dt, u i u max, 1 i N Log-normal channel gain (in db) dx i = a(x i + b)dt + σdw i (Charalambous et al 99) W i : white noise (Brownian motion) Cost function of user i T {[ J i = E p i e xi γ( α 2 N N j i pj e xj + η)] + r(u i ) 2} dt 0 α N N j i pj e xj + η: total interference due to other users and thermal noise at receiver M 1 p1 p2 Base p 3 M 4 p4 Cocktail party model: each speaker deals with bkgrd noise; one against the mass M 3 M 2 Control objective: keep signal to interference ratio around γ Figure : A cell of N users

9 Computational load with N = 2 (Huang, Caines, Malhamé, IFAC 02, IEEE TAC 04): Grid size for the 4 spatial variables in R 4 : 90 4 Discretize time interval [0, 1] with step size 0.1 For the computation, memory allocation of 70 million sizeof(float) in C Very heavy computation via iterations!! For large N, such iteration is no longer feasible This motivated the development of a different approach to tackle dimensionality difficulty See Aziz and Caines (2017) for more recent numerical analysis using MFG

10 Conceptual Modeling of mean field games Mass influence zi ui i Play against mass m(t) Model: Each player interacts with the empirical state distribution of N players via dynamics and/or costs (see examples) δ (N) z Individual state and control (zi, u i ), δ (N) z(t) = 1 N N i=1 δ z i (t) Objective: Overcome dimensionality difficulty

11 Illustrative examples for MFG modeling Discrete time linear quadratic (LQ) model Continuous time LQ model MDP model Stochastic growth with relative performance

12 Example 1: Discrete time linear quadratic (LQ) games X i t+1 = AX i t + Bu i t + GX (N) J i (u i, u i ) = E t=0 t + DW i t+1, ρ t { X i t (Γ X (N) t 1 i N + η) 2 Q + (ui t) T Ru i t } X i t : state; ui t : control; u i : control of all other agents; W i t : noise; Q 0, R > 0, ρ (0, 1); z Q = (z T Qz) 1 2 The coupling term X (N) t = 1 N N k=1 X k t It is possible to consider different parameters such as A i instead of A, etc.

13 Example 1 : Continuous time linear quadratic (LQ) dxt i = (A θi Xt i + But i + GX (N) )dt + DdW i J i (u i, u i ) = E 0 t e ρt { X i t (Γ X (N) t t, 1 i N + η) 2 Q + (ui t) T Ru i t } dt Xt i : state; ui t : control; W t i : white noise as Brownian motion; ρ > 0; z Q = (z T Qz) 1 2 The coupling term X (N) t = 1 N N k=1 X k t θ i : dynamic parameter to indicate heterogeneity Note: This is the continuous time counterpart of Example 1.

14 Example 2: Discrete time Markov decision processes (MDPs) N MDPs (x i t, a i t), 1 i N, where state x i t has transitions affected by action a i t but not by a j t, j i. Player i has cost J i = E T t=0 ρ t l(x i t, x (N) t, a i t), ρ (0, 1].

15 Example 3: Growth with relative performance Dynamics of N agents: dxt i = [ A(Xt i ) α δxt i Ct i ] dt σx i t dwt i, 1 i N Xt i : capital stock, X 0 i > 0, EX 0 i <, C t i : consumption rate Ax α, α (0, 1): Cobb-Douglas production function, 0 < α < 1, A > 0 δdt + σdwt i : stochastic depreciation (see e.g. Wälde 11, Feicht and Stummer 10 for stochastic depreciation modeling) {Wt i, 1 i N} are i.i.d. standard Brownian motions. The i.i.d. initial states {X0 i, 1 i N} are also independent of the N Brownian motions Take the standard choice γ = 1 α, i.e., equalizing the coefficient of the relative risk aversion to capital share See Huang and Nguyen (2016)

16 Example 3 (ctn) The utility functional of agent i: [ ] T J i (C 1,..., C N ) = E e ρt U(Ct i, C (N,γ) t )dt + e ρt S(X T ), where C (N,γ) t = 1 N 0 N i=1 (C t i ) γ, γ (0, 1). Take S(x) = ηx γ, η > 0, and [ U(Ct i, C (N,γ) t ) = 1 γ (C t i ) γ(1 λ) (Ct i ) γ ( = [ 1 γ (C i t ) γ ] 1 λ [ C (N,γ) t 1 γ ] λ (C i t )γ C (N,γ) t γ ] ) λ =: U 1 λ 0 U1 λ. Interpretation of U: as a weighted geometric mean of U 0 (own utility) and U 1 (relative utility).

17 Try conventional methods to solve, for instance, Example 1 : dxt i = (A θi Xt i + But i + GX (N) t J i (u i, u i ) = E 0 )dt + DdW i t, 1 i N e ρt { Xt i (Γ X (N) t + η) 2 Q + (ui t) T Rut}dt i Form large vector X t := [X 1 t,..., X N t ]T. Similarly, u t = [u 1 t,..., un t ]T This gives a standard dynamic game Immediate difficulties (for large N): Solve a very large scale dynamic programming problem In the resulting solution, each u i t will be a function of X t, which is too much information and may be unnecessary Such difficulties motivate the development of MFG theory!

18 The fundamental diagram of MFG theory Problem 3 : player Nash game states: ( 1 ) L strategies: ( L 1 ) costs: L ( ) 1 route 1 A large-scale coupled equation system Example: 1 coupled dynamic programming equations Performance route 2 route 1 Problem 3 : Optimal control of a single player state: control: mean field µ W is fixed and not controlled by L L L route 2 MFG equation system: 1 equation of optimal control; 1 equation of mean field dynamics (for ) µ W Example: HJB PDE; FPK PDE (or McKean-Vlasov SDE) blue: direct approach (Lasry and Lions, 06, 07); red: fixed point approach (Huang, Caines, Malhamé, 03, 07), (Huang, Malhamé, Caines, 06); see overview of the two approaches in (Caines, Huang, Malhamé, 17)

19 Before analyzing MFG, we mention the math toolkit for optimal control Dynamic programming (DP) (in Markov control models) Stochastic maximum principle (SMP)/calculus of variations SMP is useful when the optimal control problem does not have Markov properties (example: random model coefficients) Here we give brief preliminaries on continuous time DP

20 Consider the optimal control problem Dynamics and cost: dx t = (AX t + Bu t + G X t )ds + DdW t, J(u ( ) ) = E T 0 e ρt L(X t, u t )dt, where L = X t (Γ X t + η) 2 Q + ut t Ru t, ρ > 0, Q 0, R > 0. X C[0, T ]. Moreover X t = O(e t(ρ ϵ)/2 ) for some small ϵ > 0 (depending on X ) if T = ; denote X C ρ/2 [0, ). u t is Ft W -adapted (i.e., adapted to the filtration of the BM). Alternatively, one may take u t to be X t dependent (i.e. state feedback). We consider two cases: i) T is finite, ii) T =

21 By the method of DP, we try to solve a family of optimization problems with initial pair (t, x). Consider dx r = (AX r + Bu r + G X r )dr + DdW r, r t, X t = x, T J(t, x, u ( ) ) = E t,x e ρ(r t) L(X r, u r )dr, t where E t,x indicates X t = x, and u r is F W r -adapted. Define the value function V (t, x) = inf u E t,x T t e ρ(r t) L(X r, u r )dr.

22 By DP, for δ > 0, t+δ V (t, x) = inf E[ e ρ(r t) L(X r, u r )dr + e ρδ V (t + δ, X t+δ )], t and we further obtain the HJB equation ρv = V t + V T x (Ax + G X t ) + x (Γ X t + η) 2 Q Tr(DT V xx D) + min[u T Ru + V T u x Bu] = V t 1 4 V T x BR 1 B T V x + V T x (Ax + G X t ) + x (Γ X t + η) 2 Q Tr(DT V xx D). And V (T, x) = 0 when T is finite. V t, V x, V xx : partial derivatives. See e.g. Fleming and Rishel (1975), Yong and Zhou (1999)

23 i) For finite T, denote V (t, x) = x T P(t)x + 2s T (t)x + q(t). We derive differential Riccati equation (DRE) { ρp(t) = Ṗ(t) P(t)BR 1 B T P(t) + P(t)A + A T P(t) + Q, P(T ) = 0. { ρs(t) = ṡ(t) + (A T PBR 1 B T )s(t) + PG X t Q(Γ X t + η), s(t ) = 0. Fact: The DRE has a unique solution (due to Q 0 and R > 0). Use the minimizer in the HJB to give the optimal control u opt t = R 1 B T (P(t)X t + s(t)).

24 ii) For T =, the DRE is replaced by an algebraic Riccati equation (ARE) which has a unique solution Π 0. In this case, the ODE of s has no terminal condition; a growth condition can be imposed to determine a unique solution for s.

25 Solve Example 1 by the fixed point approach We use the LQ model to illustrate this methodology The procedure here is very robust and can be applied to different models

26 Fixed approach to the (as in Ex. 1 ): Problem P 0 dxt i = (A θi Xt i + But i + GX (N) t J i (u i, u i ) = E 0 Assume x (N) 0 = 1 N N i=1 EX i 0 X 0. )dt + DdW i t, 1 i N { e ρt Xt i (Γ X (N) t + η) 2 Q + (ui t) T Rut i Taking approximation: Optimal Control Problem P } dt dxt i = (A θi Xt i + But i + G X t)dt + DdWt i, 1 i N { } J i (u i, u i ) = E e ρt Xt i (Γ X t + η) 2 Q + (ui t )T Rut i dt 0 Step 1. Solve P assuming the deterministic X t were known Step 2. Use consistency condition of MFG to find an equation for X t

27 Step 1. Denote (we consider A θi A just for simplicity) Riccati equation: ρπ = ΠA + A T Π ΠBR 1 B T Π + Q and (for s C ρ/2 ; no given initial condition) ODE: ρs t = ds t dt + (AT ΠBR 1 B T )s t + ΠG X t Q(Γ X t + η) Best response (BR): û i t = R 1 B T (ΠX i t + s t ). Step 2. Closed-loop for P : dx i t = (A BR 1 B T Π)X i t dt BR 1 B T s t dt + G X t dt + DdW i t = d X t dt = (A BR 1 B T Π) X t + G X t BR 1 B T s t, which is obtained from averaging X i t with X 0 determined.

28 Summarizing, the MFG equation system is { d X t dt = (A BR 1 B T Π + G) X t BR 1 B T s t, ρs t = ds t dt + (AT ΠBR 1 B T )s t + ΠG X t Q(Γ X t + η), where X 0 is given. Look for s C ρ/2 [0, ). Analysis and results: This is carried out in the form of a list of Q&A s.

29 Question 1: Does there exist a unique solution ( X, s)? Note: s only has a growth condition Answer: under some conditions, there exists a unique solution. Method: use fixed point theorem (write X as the fixed point of a linear operator), or subspace decomposition of the ODE system

30 Question 2. How is the solution used in the N player model? Answer: provide decentralized strategies The set of decentralized strategies, denoted as û, for the N players: û i t = R 1 B T (ΠX i t + s t ), 1 i N. Key Feature: Player i does not need the state or sample path information of other players.

31 Question 3: What is the performance of the set of decentralized strategies? To analyze performance, we need to introduce strategy spaces. U i centr consists of strategies of the form u i (t, x 1,..., x N ); u i (t, x 1,..., x N ) C(1 + x ); and u i is Lipschitz in x = (x 1,..., x N ). This ensures well defined SDEs. u i is a centralized strategy, here depending on states of all players u i is a Markov feedback strategy Although decentralized strategies are desired, the consideration of U i centr is useful for characterizing performance.

32 To answer Q3 on performance Main assumptions: i) X0 i, i 1, are independent and also independent of the BMs W i, i 1. ii) sup i E X0 i 2 C. iii) The average of initial means x (N) 0 X 0 as N.

33 Question 3 (restated): What is the performance of the set of decentralized strategies? Theorem: If the MFG equation system has a unique solution, the set of decentralized strategies is an ϵ-nash equilibrium, i.e., for each i, inf J i (u i, û i ) J i (û i, û i ) ϵ, u i Ucentr i where ϵ = O( x (N) 0 X 0 + 1/ N). Interpretation: using full information by one deviating player offers little benefit to itself.

34 Question 4: How much efficiency loss occurs in the MFG? In order to answer this question, we need to compare with a social optimization problem. This is doable; we postpone this until introducing the social optimization problem to minimize J (N) soc = i J i.

35 Question 5: Heterogeneous agents? Answer: the MFG equation system will be modified. Assume empirical distribution for the parameter sequence {θ i, i 1} The intuition behind is to average non-uniform particles

36 Nonlinear case Interacting particle modeling for MFG: N agents (players), 1 i N. Dynamics and costs: dx i = 1 N N f (x i, u i, x j )dt + 1 N j=1 J i (u i, u i ) = E T 0 1 N N σ(x i, u i, x j )dw i, j=1 N L(x i, u i, x j )dt, 1 i N. j=1 each x j R n Assume i.i.d. initial states. (u i : controls of all players other than player i) As in the LQ case, a key step is to solve a special single agent optimal control problem. How to construct such a problem?

37 Recall dx i = 1 N N f (x i, u i, x j )dt + 1 N j=1 J i (u i, u i ) = E T 0 1 N N σ(x i, u i, x j )dw i, j=1 N L(x i, u i, x j )dt, 1 i N. j=1 each x j R n Write δ (N) x = 1 N N j=1 δ x j, δ : dirac measure. For φ = f, σ, L, denote 1 N N φ(x i, u i, x j ) = φ(x i, u i, y)δ x (N) R n j=1 Idea: approximate δ (N) x by a deterministic measure µ t (dy) := φ[x i, u i, δ (N) x ]

38 For simplicity, consider constant σ and scalar state. The optimal control problem with dynamics and cost: dx i = f [x i, u i, µ t ]dt + σdw i, J i (u i ) = E T where µ t is called the mean field. 0 L[x i, u i, µ t ]dt, each x j R Again, following the method in the case Assume {µ t, 0 t T } were known; solve the optimal control û i. Consistency condition the closed-loop state process under û i has its marginal equal to µ t. The analysis can be done for a finite classes/types of agents.

39 HJB McKV/FPK for the nonlinear diffusion model: HJB equation (scaler state for simplicity): { V t = inf f [x, u, µ t ] V } u U x + L[x, u, µ t] + σ2 2 V 2 x 2 V (T, x) = 0, (t, x) [0, T ) R. Best Response : u t = φ(t, x µ ), (t, x) [0, T ] R. Closed-loop McK-V equation (which can be written as a Fokker-Planck equation): dx t = f [x t, φ(t, x µ ), µ t ]dt + σdw t, 0 t T. Find a solution (x t, µ t ) in McK-V sense, i.e., Law(x t ) = µ t. Refs. Huang, Caines and Malhame 06, Lasry and Lions (06, 07), Cardaliaguet 12

40 Lasry and Lions (2006, 2007) considered dynamics and costs dx i t = u i dt + σdw i t, x i R d, σ > 0, J i (u 1,..., u N ) = lim inf T where µ N, i t 1 T E T 0 (L(x i, u i ) + V [x i, µ N, i t ])dt, is the empirical distribution of states of other agents. The second argument of V means it is a functional of the distribution for given x i. Assume periodicity of F and V in the cost. Denote Q = [0, 1] d. Solve the N-player game, and show the solution converges to a limiting HJB-FPK equation system σ2 ν(x) = H(x, ν) + λ = V [x, m], 2 σ2 2 m div( H (x, ν)m) = 0, p where Q ν(x)dx = 0, Q m(x) = 1, m 0. H: Hamiltonian, λ: long run average cost, ν(x): differential cost

41 Finite horizon can also be considered for this model. Dynamics of agent i: dx i t = u i tdt + 2dw i t, x i t R d ; cost functional: J N i (u i, u i ) = E T 0 [(ut) i 2 + F (xt, i µ N, i t )]dt + EG(xt, i µ N, i t ), where µ N, i t is the empirical distribution of the states of all other agents. See Lasry and Lions 07, also notes Cardaliaguet 12

42 The MFG equations: t V V DV 2 = F (x, m), (x, t) R d (0, T ) t m m div(mdv ) = 0, m(0) = m 0, V (x, T ) = G(x, m(t )), x R d, where V (t, x) and m(t, x) are the value function and the density of the state distribution, respectively. See an ϵ-nash equilibrium result in Cardaliaguet 12 This relates the MFG solution to finite population games

43 Problem 3 : player Nash game states: ( 1 ) L strategies: ( L 1 ) costs: L ( ) 1 route 1 A large-scale coupled equation system Example: 1 coupled dynamic programming equations Performance route 2 route 1 Problem 3 : Optimal control of a single player state: control: mean field µ W is fixed and not controlled by L L L route 2 MFG equation system: 1 equation of optimal control; 1 equation of mean field dynamics (for ) µ W Example: HJB PDE; FPK PDE (or McKean-Vlasov SDE) So for nonlinear diffusion models, the diagram reduces to the structure on next page

44 The basic framework of MFGs: Math. Aspects P 0 Game with N players; Example dx i = f (x i, u i, δ x (N) )dt + σ( )dw i J i (u i, u i ) = E T 0 l(x i, u i, δ x (N) )dt δ x (N) : empirical distribution of (x j ) N j=1 solution HJBs coupled via densities p N i,t, 1 i N +N Fokker-Planck-Kolmogorov equations u i adapted to σ(w i (s), s t) (i.e., restrict to decentralized info for N players); so giving u N i (t, x i ) construct performance? (subseq. convergence) N P Limiting problem, 1 player dx i = f (x i, u i, µ t )dt + σ( )dw i J i (u i ) = E T 0 l(x i, u i, µ t)dt Freeze µ t, as approx. of δ (N) x solution û i (t, x i ) : optimal response HJB (with v(t, ) given) : v t = inf ui (f T v xi + l Tr[σσT v xi x i ]) Fokker-Planck-Kolmogorov : p t = div(fp) + (( σσt 2 ) jk p) x j i xk i Coupled via µ t (w. density p t; p 0 given) The consistency based approach (red) is more popular; related to ideas in statistical physics (McKean-Vlasov eqn); FPK can be replaced by an MV-SDE

45 Applications: Economic theory Finance Communication networks Social opinion formation Power systems Electric vehicle recharging control Public health (vaccination games)...

46 Some applications of MFGs to economic growth and finance. Guéant, Lasry and Lions (2011): human capital optimization Lucas and Moll (2011): Knowledge growth and allocation of time Carmona and Lacker (2013): Investment of n brokers Huang (2013): capital accumulation with congestion effect Lachapelle et al. (2013): price formation Espinosa and Touzi (2013): Optimal investment with relative 1 performance concern (depending on N 1 j X j ) Jaimungal and Nourian (2014): Optimal execution Chan and Sircar (2015): Bertrand and Cournot models more...

47 References Only some papers are sampled and inserted into the lecture. For further literature, see these papers and books, and references therein. Bensoussan, Frehse, and Yam. Springer Brief Caines. Mean field games, in Encyclopedia of Systems and Control, Springer-Verlag, Caines, Huang, and Malhame, in Springer dynamic game theory handbook, Cardaliaguet. Notes on mean field games, Carmona and Delarue. Probabilistic Theory of MFGs, vol I and II, Cham: Springer, Gomes and Saud. Mean field games models a brief survey Huang, Caines, and Malhame CDC, IEEE TAC 07. Huang, Malhame and Caines, CIS 06. Lasry and Lions. Mean field games, 2006, Weintraub, Benkard, and Van Roy (2008). Econometrica; for oblivious equilibria

48 Thank you!

Production and Relative Consumption

Production and Relative Consumption Mean Field Growth Modeling with Cobb-Douglas Production and Relative Consumption School of Mathematics and Statistics Carleton University Ottawa, Canada Mean Field Games and Related Topics III Henri Poincaré

More information

Mean Field Games: Basic theory and generalizations

Mean Field Games: Basic theory and generalizations Mean Field Games: Basic theory and generalizations School of Mathematics and Statistics Carleton University Ottawa, Canada University of Southern California, Los Angeles, Mar 2018 Outline of talk Mean

More information

Mean Field Control: Selected Topics and Applications

Mean Field Control: Selected Topics and Applications Mean Field Control: Selected Topics and Applications School of Mathematics and Statistics Carleton University Ottawa, Canada University of Michigan, Ann Arbor, Feb 2015 Outline of talk Background and motivation

More information

Large-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed

Large-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed Definition Mean Field Game (MFG) theory studies the existence of Nash equilibria, together with the individual strategies which generate them, in games involving a large number of agents modeled by controlled

More information

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio Mean-Field optimization problems and non-anticipative optimal transport Beatrice Acciaio London School of Economics based on ongoing projects with J. Backhoff, R. Carmona and P. Wang Robust Methods in

More information

arxiv: v1 [math.pr] 18 Oct 2016

arxiv: v1 [math.pr] 18 Oct 2016 An Alternative Approach to Mean Field Game with Major and Minor Players, and Applications to Herders Impacts arxiv:161.544v1 [math.pr 18 Oct 216 Rene Carmona August 7, 218 Abstract Peiqi Wang The goal

More information

Mean Field Consumption-Accumulation Optimization with HAR

Mean Field Consumption-Accumulation Optimization with HAR Mean Field Consumption-Accumulation Optimization with HARA Utility School of Mathematics and Statistics Carleton University, Ottawa Workshop on Mean Field Games and Related Topics 2 Padova, September 4-7,

More information

Mean-Field Games with non-convex Hamiltonian

Mean-Field Games with non-convex Hamiltonian Mean-Field Games with non-convex Hamiltonian Martino Bardi Dipartimento di Matematica "Tullio Levi-Civita" Università di Padova Workshop "Optimal Control and MFGs" Pavia, September 19 21, 2018 Martino

More information

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS René Carmona Department of Operations Research & Financial Engineering PACM Princeton University CEMRACS - Luminy, July 17, 217 MFG WITH MAJOR AND MINOR PLAYERS

More information

Risk-Sensitive and Robust Mean Field Games

Risk-Sensitive and Robust Mean Field Games Risk-Sensitive and Robust Mean Field Games Tamer Başar Coordinated Science Laboratory Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL - 6181 IPAM

More information

Explicit solutions of some Linear-Quadratic Mean Field Games

Explicit solutions of some Linear-Quadratic Mean Field Games Explicit solutions of some Linear-Quadratic Mean Field Games Martino Bardi Department of Pure and Applied Mathematics University of Padua, Italy Men Field Games and Related Topics Rome, May 12-13, 2011

More information

Linear-Quadratic Stochastic Differential Games with General Noise Processes

Linear-Quadratic Stochastic Differential Games with General Noise Processes Linear-Quadratic Stochastic Differential Games with General Noise Processes Tyrone E. Duncan Abstract In this paper a noncooperative, two person, zero sum, stochastic differential game is formulated and

More information

ON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY)

ON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY) Collège de France, Paris (joint project with Jean-Michel LASRY) I INTRODUCTION II A REALLY SIMPLE EXAMPLE III GENERAL STRUCTURE IV TWO PARTICULAR CASES V OVERVIEW AND PERSPECTIVES I. INTRODUCTION New class

More information

From SLLN to MFG and HFT. Xin Guo UC Berkeley. IPAM WFM 04/28/17 Based on joint works with J. Lee, Z. Ruan, and R. Xu (UC Berkeley), and L.

From SLLN to MFG and HFT. Xin Guo UC Berkeley. IPAM WFM 04/28/17 Based on joint works with J. Lee, Z. Ruan, and R. Xu (UC Berkeley), and L. From SLLN to MFG and HFT Xin Guo UC Berkeley IPAM WFM 04/28/17 Based on joint works with J. Lee, Z. Ruan, and R. Xu (UC Berkeley), and L. Zhu (FSU) Outline Starting from SLLN 1 Mean field games 2 High

More information

Variational approach to mean field games with density constraints

Variational approach to mean field games with density constraints 1 / 18 Variational approach to mean field games with density constraints Alpár Richárd Mészáros LMO, Université Paris-Sud (based on ongoing joint works with F. Santambrogio, P. Cardaliaguet and F. J. Silva)

More information

On continuous time contract theory

On continuous time contract theory Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem

More information

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011 Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance

More information

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio Mean-Field optimization problems and non-anticipative optimal transport Beatrice Acciaio London School of Economics based on ongoing projects with J. Backhoff, R. Carmona and P. Wang Thera Stochastics

More information

Mean Field Competitive Binary MDPs and Structured Solutions

Mean Field Competitive Binary MDPs and Structured Solutions Mean Field Competitive Binary MDPs and Structured Solutions Minyi Huang School of Mathematics and Statistics Carleton University Ottawa, Canada MFG217, UCLA, Aug 28 Sept 1, 217 1 / 32 Outline of talk The

More information

Learning in Mean Field Games

Learning in Mean Field Games Learning in Mean Field Games P. Cardaliaguet Paris-Dauphine Joint work with S. Hadikhanloo (Dauphine) Nonlinear PDEs: Optimal Control, Asymptotic Problems and Mean Field Games Padova, February 25-26, 2016.

More information

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem

More information

Numerical methods for Mean Field Games: additional material

Numerical methods for Mean Field Games: additional material Numerical methods for Mean Field Games: additional material Y. Achdou (LJLL, Université Paris-Diderot) July, 2017 Luminy Convergence results Outline 1 Convergence results 2 Variational MFGs 3 A numerical

More information

Mean Field Games on networks

Mean Field Games on networks Mean Field Games on networks Claudio Marchi Università di Padova joint works with: S. Cacace (Rome) and F. Camilli (Rome) C. Marchi (Univ. of Padova) Mean Field Games on networks Roma, June 14 th, 2017

More information

Control of McKean-Vlasov Dynamics versus Mean Field Games

Control of McKean-Vlasov Dynamics versus Mean Field Games Control of McKean-Vlasov Dynamics versus Mean Field Games René Carmona (a),, François Delarue (b), and Aimé Lachapelle (a), (a) ORFE, Bendheim Center for Finance, Princeton University, Princeton, NJ 8544,

More information

Weak solutions to Fokker-Planck equations and Mean Field Games

Weak solutions to Fokker-Planck equations and Mean Field Games Weak solutions to Fokker-Planck equations and Mean Field Games Alessio Porretta Universita di Roma Tor Vergata LJLL, Paris VI March 6, 2015 Outlines of the talk Brief description of the Mean Field Games

More information

Dynamic Bertrand and Cournot Competition

Dynamic Bertrand and Cournot Competition Dynamic Bertrand and Cournot Competition Effects of Product Differentiation Andrew Ledvina Department of Operations Research and Financial Engineering Princeton University Joint with Ronnie Sircar Princeton-Lausanne

More information

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations Lukasz Szpruch School of Mathemtics University of Edinburgh joint work with Shuren Tan and Alvin Tse (Edinburgh) Lukasz Szpruch (University

More information

Stochastic optimal control with rough paths

Stochastic optimal control with rough paths Stochastic optimal control with rough paths Paul Gassiat TU Berlin Stochastic processes and their statistics in Finance, Okinawa, October 28, 2013 Joint work with Joscha Diehl and Peter Friz Introduction

More information

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Conference on the occasion of Giovanni Colombo and Franco Ra... https://events.math.unipd.it/oscgc18/ Optimization,

More information

Continuous Time Finance

Continuous Time Finance Continuous Time Finance Lisbon 2013 Tomas Björk Stockholm School of Economics Tomas Björk, 2013 Contents Stochastic Calculus (Ch 4-5). Black-Scholes (Ch 6-7. Completeness and hedging (Ch 8-9. The martingale

More information

Long time behavior of Mean Field Games

Long time behavior of Mean Field Games Long time behavior of Mean Field Games P. Cardaliaguet (Paris-Dauphine) Joint work with A. Porretta (Roma Tor Vergata) PDE and Probability Methods for Interactions Sophia Antipolis (France) - March 30-31,

More information

Weak solutions of mean-field stochastic differential equations

Weak solutions of mean-field stochastic differential equations Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works

More information

The Nash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and Entry Adaptation

The Nash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and Entry Adaptation Proceedings of the 46th IEEE Conference on Decision and Control ew Orleans, LA, USA, Dec. 12-14, 27 WeA5.1 The ash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and

More information

MEAN FIELD GAMES AND SYSTEMIC RISK

MEAN FIELD GAMES AND SYSTEMIC RISK MEA FIELD GAMES AD SYSTEMIC RISK REÉ CARMOA, JEA-PIERRE FOUQUE, AD LI-HSIE SU Abstract. We propose a simple model of inter-bank borrowing and lending where the evolution of the logmonetary reserves of

More information

Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation

Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation Peter E. Caines McGill University, Montreal, Canada 2 nd Workshop on Mean-Field Games and Related

More information

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION René Carmona Department of Operations Research & Financial Engineering PACM

More information

Solution of Stochastic Optimal Control Problems and Financial Applications

Solution of Stochastic Optimal Control Problems and Financial Applications Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty

More information

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA On Stochastic Adaptive Control & its Applications Bozenna Pasik-Duncan University of Kansas, USA ASEAS Workshop, AFOSR, 23-24 March, 2009 1. Motivation: Work in the 1970's 2. Adaptive Control of Continuous

More information

A numerical approach to the MFG

A numerical approach to the MFG A numerical approach to the MFG Gabriel Turinici CEREMADE, Université Paris Dauphine CIRM Marseille, Summer 2011 Gabriel Turinici (CEREMADE, Université Paris Dauphine A numerical ) approach to the MFG

More information

An Introduction to Noncooperative Games

An Introduction to Noncooperative Games An Introduction to Noncooperative Games Alberto Bressan Department of Mathematics, Penn State University Alberto Bressan (Penn State) Noncooperative Games 1 / 29 introduction to non-cooperative games:

More information

Systemic Risk and Stochastic Games with Delay

Systemic Risk and Stochastic Games with Delay Systemic Risk and Stochastic Games with Delay Jean-Pierre Fouque (with René Carmona, Mostafa Mousavi and Li-Hsien Sun) PDE and Probability Methods for Interactions Sophia Antipolis (France) - March 30-31,

More information

Chapter 2 Event-Triggered Sampling

Chapter 2 Event-Triggered Sampling Chapter Event-Triggered Sampling In this chapter, some general ideas and basic results on event-triggered sampling are introduced. The process considered is described by a first-order stochastic differential

More information

Explicit solutions of some Linear-Quadratic Mean Field Games

Explicit solutions of some Linear-Quadratic Mean Field Games Explicit solutions of some Linear-Quadratic Mean Field Games Martino Bardi To cite this version: Martino Bardi. Explicit solutions of some Linear-Quadratic Mean Field Games. Networks and Heterogeneous

More information

arxiv: v4 [math.oc] 24 Sep 2016

arxiv: v4 [math.oc] 24 Sep 2016 On Social Optima of on-cooperative Mean Field Games Sen Li, Wei Zhang, and Lin Zhao arxiv:607.0482v4 [math.oc] 24 Sep 206 Abstract This paper studies the connections between meanfield games and the social

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

EXPLICIT SOLUTIONS OF SOME LINEAR-QUADRATIC MEAN FIELD GAMES. Martino Bardi

EXPLICIT SOLUTIONS OF SOME LINEAR-QUADRATIC MEAN FIELD GAMES. Martino Bardi NETWORKS AND HETEROGENEOUS MEDIA c American Institute of Mathematical Sciences Volume 7, Number, June 1 doi:1.3934/nhm.1.7.xx pp. X XX EXPLICIT SOLUTIONS OF SOME LINEAR-QUADRATIC MEAN FIELD GAMES Martino

More information

LQG Mean-Field Games with ergodic cost in R d

LQG Mean-Field Games with ergodic cost in R d LQG Mean-Field Games with ergodic cost in R d Fabio S. Priuli University of Roma Tor Vergata joint work with: M. Bardi (Univ. Padova) Padova, September 4th, 2013 Main problems Problems we are interested

More information

Robust control and applications in economic theory

Robust control and applications in economic theory Robust control and applications in economic theory In honour of Professor Emeritus Grigoris Kalogeropoulos on the occasion of his retirement A. N. Yannacopoulos Department of Statistics AUEB 24 May 2013

More information

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009 A new approach for investment performance measurement 3rd WCMF, Santa Barbara November 2009 Thaleia Zariphopoulou University of Oxford, Oxford-Man Institute and The University of Texas at Austin 1 Performance

More information

Chapter Introduction. A. Bensoussan

Chapter Introduction. A. Bensoussan Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

This is a repository copy of Density flow over networks: A mean-field game theoretic approach.

This is a repository copy of Density flow over networks: A mean-field game theoretic approach. This is a repository copy of Density flow over networks: A mean-field game theoretic approach. White Rose Research Online URL for this paper: http://eprints.whiterose.ac.uk/89655/ Version: Accepted Version

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games

Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games Gabriel Y. Weintraub, Lanier Benkard, and Benjamin Van Roy Stanford University {gweintra,lanierb,bvr}@stanford.edu Abstract

More information

Nonlinear Elliptic Systems and Mean-Field Games

Nonlinear Elliptic Systems and Mean-Field Games Nonlinear Elliptic Systems and Mean-Field Games Martino Bardi and Ermal Feleqi Dipartimento di Matematica Università di Padova via Trieste, 63 I-35121 Padova, Italy e-mail: bardi@math.unipd.it and feleqi@math.unipd.it

More information

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct

More information

Uncertainty quantification and systemic risk

Uncertainty quantification and systemic risk Uncertainty quantification and systemic risk Josselin Garnier (Université Paris Diderot) with George Papanicolaou and Tzu-Wei Yang (Stanford University) February 3, 2016 Modeling systemic risk We consider

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

An Uncertain Control Model with Application to. Production-Inventory System

An Uncertain Control Model with Application to. Production-Inventory System An Uncertain Control Model with Application to Production-Inventory System Kai Yao 1, Zhongfeng Qin 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China 2 School of Economics

More information

Mean field games and related models

Mean field games and related models Mean field games and related models Fabio Camilli SBAI-Dipartimento di Scienze di Base e Applicate per l Ingegneria Facoltà di Ingegneria Civile ed Industriale Email: Camilli@dmmm.uniroma1.it Web page:

More information

Uncertainty quantification and systemic risk

Uncertainty quantification and systemic risk Uncertainty quantification and systemic risk Josselin Garnier (Université Paris 7) with George Papanicolaou and Tzu-Wei Yang (Stanford University) April 9, 2013 Modeling Systemic Risk We consider evolving

More information

Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems.

Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems. ES22 Lecture Notes #11 Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems. J = φ(x( ) + L(x,u,t)dt ; x= f(x,u,t)

More information

ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HORIZON MEAN FIELD GAMES

ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HORIZON MEAN FIELD GAMES ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HOIZON MEAN FIELD GAMES MATINO BADI AND MAKUS FISCHE Abstract. This paper presents a class of evolutive Mean Field Games with multiple solutions

More information

ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HORIZON MEAN FIELD GAMES

ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HORIZON MEAN FIELD GAMES ON NON-UNIQUENESS AND UNIQUENESS OF SOLUTIONS IN FINITE-HOIZON MEAN FIELD GAMES MATINO BADI AND MAKUS FISCHE Abstract. This paper presents a class of evolutive Mean Field Games with multiple solutions

More information

Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience

Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience Ulrich Horst 1 Humboldt-Universität zu Berlin Department of Mathematics and School of Business and Economics Vienna, Nov.

More information

Lecture 4: Introduction to stochastic processes and stochastic calculus

Lecture 4: Introduction to stochastic processes and stochastic calculus Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London

More information

Systemic Risk and Stochastic Games with Delay

Systemic Risk and Stochastic Games with Delay Systemic Risk and Stochastic Games with Delay Jean-Pierre Fouque Collaborators: R. Carmona, M. Mousavi, L.-H. Sun, N. Ning and Z. Zhang Mean Field Games 2017 IPAM at UCLA JP Fouque (UC Santa Barbara) Stochastic

More information

On the connection between symmetric N-player games and mean field games

On the connection between symmetric N-player games and mean field games On the connection between symmetric -player games and mean field games Markus Fischer May 14, 214; revised September 7, 215; modified April 15, 216 Abstract Mean field games are limit models for symmetric

More information

Stochastic Modelling in Climate Science

Stochastic Modelling in Climate Science Stochastic Modelling in Climate Science David Kelly Mathematics Department UNC Chapel Hill dtbkelly@gmail.com November 16, 2013 David Kelly (UNC) Stochastic Climate November 16, 2013 1 / 36 Why use stochastic

More information

Some aspects of the mean-field stochastic target problem

Some aspects of the mean-field stochastic target problem Some aspects of the mean-field stochastic target problem Quenched Mass ransport of Particles owards a arget Boualem Djehiche KH Royal Institute of echnology, Stockholm (Joint work with B. Bouchard and

More information

Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm

Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm Gonçalo dos Reis University of Edinburgh (UK) & CMA/FCT/UNL (PT) jointly with: W. Salkeld, U. of

More information

Introduction to Diffusion Processes.

Introduction to Diffusion Processes. Introduction to Diffusion Processes. Arka P. Ghosh Department of Statistics Iowa State University Ames, IA 511-121 apghosh@iastate.edu (515) 294-7851. February 1, 21 Abstract In this section we describe

More information

Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications

Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications Jie Xiong Department of Mathematics University of Macau Macau Based on joint works with Shi and Wang 2015 Barrett

More information

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem

MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear

More information

Mean field equilibria of multiarmed bandit games

Mean field equilibria of multiarmed bandit games Mean field equilibria of multiarmed bandit games Ramesh Johari Stanford University Joint work with Ramki Gummadi (Stanford University) and Jia Yuan Yu (IBM Research) A look back: SN 2000 2 Overview What

More information

Generalized Riccati Equations Arising in Stochastic Games

Generalized Riccati Equations Arising in Stochastic Games Generalized Riccati Equations Arising in Stochastic Games Michael McAsey a a Department of Mathematics Bradley University Peoria IL 61625 USA mcasey@bradley.edu Libin Mou b b Department of Mathematics

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

Mean-field equilibrium: An approximation approach for large dynamic games

Mean-field equilibrium: An approximation approach for large dynamic games Mean-field equilibrium: An approximation approach for large dynamic games Ramesh Johari Stanford University Sachin Adlakha, Gabriel Y. Weintraub, Andrea Goldsmith Single agent dynamic control Two agents:

More information

High-dimensional Problems in Finance and Economics. Thomas M. Mertens

High-dimensional Problems in Finance and Economics. Thomas M. Mertens High-dimensional Problems in Finance and Economics Thomas M. Mertens NYU Stern Risk Economics Lab April 17, 2012 1 / 78 Motivation Many problems in finance and economics are high dimensional. Dynamic Optimization:

More information

Distributed Learning based on Entropy-Driven Game Dynamics

Distributed Learning based on Entropy-Driven Game Dynamics Distributed Learning based on Entropy-Driven Game Dynamics Bruno Gaujal joint work with Pierre Coucheney and Panayotis Mertikopoulos Inria Aug., 2014 Model Shared resource systems (network, processors)

More information

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018 EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal

More information

Systemic risk and uncertainty quantification

Systemic risk and uncertainty quantification Systemic risk and uncertainty quantification Josselin Garnier (Université Paris 7) with George Papanicolaou and Tzu-Wei Yang (Stanford University) October 6, 22 Modeling Systemic Risk We consider evolving

More information

An Introduction to Moral Hazard in Continuous Time

An Introduction to Moral Hazard in Continuous Time An Introduction to Moral Hazard in Continuous Time Columbia University, NY Chairs Days: Insurance, Actuarial Science, Data and Models, June 12th, 2018 Outline 1 2 Intuition and verification 2BSDEs 3 Control

More information

Backward Stochastic Differential Equations with Infinite Time Horizon

Backward Stochastic Differential Equations with Infinite Time Horizon Backward Stochastic Differential Equations with Infinite Time Horizon Holger Metzler PhD advisor: Prof. G. Tessitore Università di Milano-Bicocca Spring School Stochastic Control in Finance Roscoff, March

More information

Minimum average value-at-risk for finite horizon semi-markov decision processes

Minimum average value-at-risk for finite horizon semi-markov decision processes 12th workshop on Markov processes and related topics Minimum average value-at-risk for finite horizon semi-markov decision processes Xianping Guo (with Y.H. HUANG) Sun Yat-Sen University Email: mcsgxp@mail.sysu.edu.cn

More information

Alberto Bressan. Department of Mathematics, Penn State University

Alberto Bressan. Department of Mathematics, Penn State University Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i

More information

Continuous dependence estimates for the ergodic problem with an application to homogenization

Continuous dependence estimates for the ergodic problem with an application to homogenization Continuous dependence estimates for the ergodic problem with an application to homogenization Claudio Marchi Bayreuth, September 12 th, 2013 C. Marchi (Università di Padova) Continuous dependence Bayreuth,

More information

Basics of Game Theory

Basics of Game Theory Basics of Game Theory Giacomo Bacci and Luca Sanguinetti Department of Information Engineering University of Pisa, Pisa, Italy {giacomo.bacci,luca.sanguinetti}@iet.unipi.it April - May, 2010 G. Bacci and

More information

Weakly interacting particle systems on graphs: from dense to sparse

Weakly interacting particle systems on graphs: from dense to sparse Weakly interacting particle systems on graphs: from dense to sparse Ruoyu Wu University of Michigan (Based on joint works with Daniel Lacker and Kavita Ramanan) USC Math Finance Colloquium October 29,

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

On the long time behavior of the master equation in Mean Field Games

On the long time behavior of the master equation in Mean Field Games On the long time behavior of the master equation in Mean Field Games P. Cardaliaguet (Paris-Dauphine) Joint work with A. Porretta (Roma Tor Vergata) Mean Field Games and Related Topics - 4 Rome - June

More information

(1) I. INTRODUCTION (2)

(1) I. INTRODUCTION (2) 920 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 4, APRIL 2012 Synchronization of Coupled Oscillators is a Game Huibing Yin, Member, IEEE, Prashant G. Mehta, Member, IEEE, Sean P. Meyn, Fellow,

More information

Decentralized Convergence to Nash Equilibria in Constrained Deterministic Mean Field Control

Decentralized Convergence to Nash Equilibria in Constrained Deterministic Mean Field Control Decentralized Convergence to Nash Equilibria in Constrained Deterministic Mean Field Control 1 arxiv:1410.4421v2 [cs.sy] 17 May 2015 Sergio Grammatico, Francesca Parise, Marcello Colombino, and John Lygeros

More information

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Mostafa D. Awheda Department of Systems and Computer Engineering Carleton University Ottawa, Canada KS 5B6 Email: mawheda@sce.carleton.ca

More information

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-Field SDE s with Irregular Drift Thilo Meyer-Brandis University of Munich joint with M. Bauer, University of Munich Conference for the 1th Anniversary

More information

On the Solvability of Risk-Sensitive Linear-Quadratic Mean-Field Games

On the Solvability of Risk-Sensitive Linear-Quadratic Mean-Field Games arxiv:141.37v1 [math.oc] 6 Nov 14 On the Solvability of Risk-Sensitive Linear-Quadratic Mean-Field Games Boualem Djehiche and Hamidou Tembine December, 14 Abstract In this paper we formulate and solve

More information

A Parcimonious Long Term Mean Field Model for Mining Industries

A Parcimonious Long Term Mean Field Model for Mining Industries A Parcimonious Long Term Mean Field Model for Mining Industries Yves Achdou Laboratoire J-L. Lions, Université Paris Diderot (joint work with P-N. Giraud, J-M. Lasry and P-L. Lions ) Mining Industries

More information

Short-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility

Short-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility Short-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility José Enrique Figueroa-López 1 1 Department of Statistics Purdue University Statistics, Jump Processes,

More information

Linear Quadratic Zero-Sum Two-Person Differential Games

Linear Quadratic Zero-Sum Two-Person Differential Games Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information