Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation

Size: px
Start display at page:

Download "Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation"

Transcription

1 Mean Field Game Theory for Systems with Major and Minor Agents: Nonlinear Theory and Mean Field Estimation Peter E. Caines McGill University, Montreal, Canada 2 nd Workshop on Mean-Field Games and Related Topics Department of Mathematics University of Padova September 2013 Joint work with Mojtaba Nourian and Arman C. Kizilkale

2 Background Mean Field Game (MFG) Theory Modeling Framework and Central Problem of MFG Theory (Huang, Malhamé, PEC ( 03, 06, 07), Lasry-Lions ( 06, 07)): Framework Games over time with a large number of stochastic dynamical agents such that: Each agent interacts with a mass effect (e.g. average) of other agents via couplings in their individual cost functions and individual dynamics Each agent is minor in the sense that, asymptotically as the population size goes to infinity, it has a negligible influence on the overall system but the mass effect on the agent is significant Problem Establish the existence and uniqueness of equilibria and the corresponding strategies of the agents

3 Background Mean Field Game (MFG) Theory Solution Concepts for MFG Theory: The existence of Nash equilibria between the individual agents and the mass in the infinite population limit where (a) the individual strategy of each agent is a best response to the mass effect, and (b) the set of the strategies collectively replicate that mass effect The ɛ Nash Approximation Property: If agents in a finite population system apply the infinite population equilibrium strategies an approximation to the infinite population equilibrium results

4 Non-linear Major-Minor Mean Field Systems

5 MFG Theory Involving Major-Minor Agents Extension of the LQG MFG model for Major and Minor agents (Huang 2010, Huang-Ngyuan 2011) to the case of nonlinear dynamical systems Dynamic game models will involve nonlinear stochastic systems with (i) a major agent, and (ii) a large population of minor agents Partially observed systems become meaningful and hence estimation of major agent and mean field states becomes a meaningful problem Motivation and Applications: Economic and social models with both minor and massive agents Power markets with large consumers and large utilities together with many domestic consumers and generators using smart meters

6 MFG Nonlinear Major-Minor Agent Formulation Problem Formulation: Notation: Subscript 0 for the major agent A 0 and an integer valued subscript for minor agents {A i : 1 i N}. The states of A 0 and A i are R n valued and denoted z N 0 (t) and z N i Dynamics of the Major and Minor Agents: dz N 0 (t) = 1 N dz N i + 1 N (t) = 1 N + 1 N N f 0(t, z0 N (t), u N 0 (t), zj N (t))dt j=1 (t). N σ 0(t, z0 N (t), zj N (t))dw 0(t), z0 N (0) = z 0(0), 0 t T, j=1 N f(t, zi N (t), u N i (t), zj N (t))dt j=1 N σ(t, zi N (t), zj N (t))dw i(t), zi N (0) = z i(0), 1 i N. j=1

7 MFG Nonlinear Major-Minor Agent Formulation Cost Functions for Major and Minor Agents: The objective of each agent is to minimize its finite time horizon cost function given by T ( J0 N (u N 0 ; u N 1 0) := E 0 N T ( Ji N (u N i ; u N 1 i) := E 0 N N j=1 ) L 0[t, z0 N (t), u N 0 (t), zj N (t)] dt, N L[t, zi N j=1 ) (t), u N i (t), z0 N (t), zj N (t)] dt. Notation The major agent has non-negligible influence on the mean field (mass) behaviour of the minor agents due to presence of z0 N in the cost function of each minor agent. A consequence is that the mean field is no longer a deterministic function of time. (Ω, F, {F t} t 0, P ): a complete filtered probability space F t := σ{z j(s), w j(s) : 0 j N, 0 s t}. F w 0 t := σ{z 0(0), w 0(s) : 0 s t}.

8 Assumptions Assumptions: Let the empirical distribution of N minor agents initial states be defined by F N (x) := 1 N N i=1 1 {z i (0)<x}. (A1) The initial states {z j(0) : 0 j N} are F 0-adapted random variables mutually independent and independent of all Brownian motions, and there exists a constant k independent of N such that sup 0 j N E z j(0) 2 k <. (A2) {F N : N 1} converges weakly to the probability distribution F. (A3) U 0 and U are compact metric spaces. (A4) f 0[t, x, u, y], σ 0[t, x, y], f[t, x, u, y] and σ[t, x, y] are continuous and bounded with respect to all their parameters, and Lipschitz continuous in (x, y, z). In addition, their first order derivatives (w.r.t. x) are all uniformly continuous and bounded with respect to all their parameters, and Lipschitz continuous in y. (A5) f 0[t, x, u, y] and f[t, x, u, y] are Lipschitz continuous in u.

9 Assumptions Assunptions (ctd): (A6) L 0[t, x, u, y] and L[t, x, u, y, z] are continuous and bounded with respect to all their parameters, and Lipschitz continuous in (x, y, z). In addition, their first order derivatives (w.r.t. x) are all uniformly continuous and bounded with respect to all their parameters, and Lipschitz continuous in (y, z). (A7) (Non-degeneracy Assumption) There exists a positive constant α such that σ 0[t, x, y]σ T 0 [t, x, y] αi, σ[t, x, y]σ T (t, x, y) αi, (t, x, y). Otherwise, a notion of viscosity like solutions seems necessary.

10 McKean-Vlasov Approximation for MFG Analysis -Loop Major and Minor Agent Dynamics: Assume ϕ 0(ω, t, x) L 2 F w 0 ([0, T ]; U 0) and ϕ(ω, t, x) L 2 t F w 0 ([0, T ]; U) are t two arbitrary F w 0 t -measurable stochastic processes, Lipschitz continuous in x, constituting the Major and Minor agent control laws. Then: dz N 0 (t) = 1 N dz N i + 1 N (t) = 1 N + 1 N N f 0(t, z0 N (t), ϕ 0(t, z0 N (t)), zj N (t))dt j=1 N σ 0(t, z0 N (t), zj N (t))dw 0(t), z0 N (0) = z 0(0), 0 t T, j=1 N f(t, zi N (t), ϕ(t, zi N (t), z0 N (t)), zj N (t))dt j=1 N σ(t, zi N (t), zj N (t))dw i(t), zi N (0) = z i(0), 1 i N. j=1

11 McKean-Vlasov Approximation for MFG Analysis Major-Minor Agent McKean-Vlasov System: For an arbitrary function g and a probability distribution µ t in R n, set g[t, z, ψ, µ t] = g(t, z, ψ, x)µ t(dx). The pair of infinite population McKean-Vlasov (MV) SDE systems corresponding to the collection of finite population systems above is given by: dz 0(t) = f 0[t, z 0(t), ϕ 0(t, z 0(t)), µ t]dt + σ 0[t, z 0(t), µ t]dw 0(t), dz(t) = f[t, z(t), ϕ(t, z(t), z 0(t)), µ t]dt + σ[t, z(t), µ t]dw(t), with initial conditions (z 0(0), z(0)). 0 t T In using the MV system it is assumed that the (behaviour) of an infinite population of (parameter) uniform minor agents can be modelled by the collection of sample paths of agents with independent initial conditions and independent Brownian sample paths. In the above MV system ( ) ( z 0( ), z( ), µ ( ) is a consistent solution if z0( ), z( ) ) is a solution to the above SDE system, and µ t is the conditional law of z(t) given F w 0 t (i.e., µ t := L ( ) z(t) F w 0 t ).

12 McKean-Vlasov Approximation for MFG Analysis We shall use the notation: dz 0(t) = f 0[t, z 0(t), ϕ 0(t, z 0(t)), µ t]dt + σ 0[t, z 0(t), µ t]dw 0(t), 0 t T, dz i(t) = f[t, z i(t), ϕ(t, z i(t)), z 0(t), µ t]dt + σ[t, z i(t), µ t]dw i(t), 1 i N, with initial conditions z j(0) = z j(0) for 0 j N, to describe N independent samples of the MV SDE system. Theorem (McKean-Vlasov Convergence Result) Assume (A1) and (A3)-(A5) hold. Then unique solutions exist for the finite and MV SDE equation schemes and we have sup sup 0 j N 0 t T E ẑ N j (t) z j(t) = O( 1 N ), where the right hand side may depend upon the terminal time T. The proof is based on the Cauchy-Schwarz inequality, Gronwall s lemma and the conditional independence of minor agents given F w 0 t.

13 Distinct Feature of the Major-Minor MFG Theory The non-standard nature of the SOCPs is due to the fact that the minor agents are optimizing with respect to the future stochastic evolution of the major agent s stochastically evolving state which is partly a result of that agent s future best response control actions. Hence the mean field becomes stochastic. This feature vanishes in the non-game theoretic setting of one controller with one cost function with respect to the trajectories of all the system components (the classical SOCP), moreover it also vanishes in the infinite population limit of the standard MFG models with no major agent. The nonstandard feature of the SOCPs here give rise to the analysis of systems with stochastic parameters and hence BSDEs enter the analysis.

14 An SOCP with Random Coefficients (after Peng 92) Let (W (t)) t 0 and (B(t)) t 0 be mutually independent standard Brownian motions in R m. Denote F W,B t := σ{w (s), B(s) : s t}, F W t := σ{w (s) : s t} U := { u( ) U : u(t) is adapted to σ-field F W,B t and E T 0 u(t) 2 dt < }. Dynamics and optimal control problem for a single agent: dz(t) = f[t, z, u]dt + σ[t, z]dw (t) + ς[t, z]db(t), 0 t T, [ T ] inf J(u) := inf E L[t, z(t), u(t)]dt, u U u U 0 where the coefficients f, σ, ς and L are are F W t -adapted stochastic processes. The value function φ(, ) is defined to be the F W t -adapted process φ(t, x t) = inf u U E F W t T where x t is the initial condition for the process x. t L[s, x(s), u(s)]ds,

15 Solution to the Optimal Control Problem with Random Coefficients Let u o ( ) be the optimal control with corresponding closed-loop solution x( ) By the Principle of Optimality, the process ζ(t) := φ ( t, x(t) ) + t 0 L[s, x(s), u o (s, x(s))]ds, is an {Ft W } 0 t T -martingale. Next, by the martingale representation theorem along the optimal trajectory there exists an Ft W -adapted process ψ (, x( ) ) such that φ ( t, x(t) ) T T = L[s, x(s), u o (s, x(s))]ds ψ T ( s, x(s) ) dw (s) =: t T t Γ ( s, x(s) ) T ds ψ T ( s, x(s) ) dw (s), t [0, T ]. t t

16 Extended Itô-Kunita formula Theorem (Extended Itô-Kunita formula (after Peng 92)) Let φ(t, x) be a stochastic process represented by dφ(t, x) = Γ(t, x)dt + m ψ k (t, x)dw k (t), (t, x) [0, T ] R n, k=1 where Γ(t, x) and ψ k (t, x), 1 k m, are Ft W -adapted stochastic processes. Let x( ) = ( x 1 ( ),, x n ( ) ) be a continuous semimartingale of the form m m dx i (t) = f i(t)dt + σ ik (t)dw k (t) + ς ik (t)db k (t), 1 i n, k=1 where f i, σ i = (σ i1,, σ im) and ς i = (ς i1,, ς im), 1 i n, are F W t adapted stochastic processes. k=1

17 Extended Itô-Kunita formula Then the composition map φ(, x( )) is also a continuous semimartingale which has the form dφ ( t, x(t) ) = Γ ( t, x(t) ) m ( ) n dt + ψ k t, x(t) dwk (t) + xi φ ( t, x(t) ) f i(t)dt i=1 k=1 k=1 n m + xi φ ( t, x(t) ) n m σ ik (t)dw k (t) + xi φ ( t, x(t) ) ς ik (t)db k (t) n i=1 k=1 m ( ) xi ψ k t, x(t) σik (t)dt n i,j=1 k=1 i=1 k=1 n i,j=1 k=1 m x 2 i x j φ ( t, x(t) ) ς ik (t)ς jk (t)dt. i=1 m x 2 i x j φ ( t, x(t) ) σ ik (t)σ jk (t)dt

18 A Stochastic Hamilton-Jacobi-Bellman (SHJB) Equation Using the extended Itô-Kunita formula and the Principle of Optimality, it may be shown that the pair ( φ(s, x), ψ(s, x) ) satisfies the following backward in time SHJB equation: dφ(t, ω, x) = [H[t, ω, x, D xφ(t, ω, x)] + σ[t, ω, x], D xψ(t, ω, x) Tr( a[t, ω, x]d 2 xxφ(t, ω, x) )] dt ψ T (t, ω, x)dw (t, ω), φ(t, x) = 0, in [0, T ] R n, where a[t, ω, x] := σ[t, ω, x]σ T [t, ω, x] + ς[t, ω, x]ς T (t, ω, x), and the stochastic Hamiltonian H is given by { } H[t, ω, z, p] := inf f[t, ω, z, u], p + L[t, ω, z, u]. u U

19 Unique Solution of the SHJB Equation Assumptions: (H1)(Continuity Assumptions 1) f[t, x, u] and L[t, x, u] are a.s. continuous in (x, u) for each t, a.s. continuous in t for each (x, u), f[t, 0, 0] L 2 F t ([0, T ]; R n ) and L[t, 0, 0] L 2 F t ([0, T ]; R +). In addition, they and all their first derivatives (w.r.t. x) are a.s. continuous and bounded. (H2) (Continuity Assumptions 2) σ[t, x] and ς[t, x] are a.s. continuous in x for each t, a.s. continuous in t for each x and σ[t, 0], ς[t, 0] L 2 F t ([0, T ]; R n m ). In addition, they and all their first derivatives (w.r.t. x) are a.s. continuous and bounded. (H3) (Non-degeneracy Assumption) There exist non-negative constants α 1 and α 2, α 1 + α 2 > 0, such that σ[t, ω, x]σ T [t, ω, x] α 1I, ς[t, ω, x]ς T (t, ω, x) α 2I, a.s., (t, ω, x). Theorem (Peng 92) Assume (H1)-(H3) hold. Then the SHJB equation has a unique forward in time Ft W -adapted solution pair (φ(t, ω, x), ψ(t, ω, x)) ( L 2 F t ([0, T ]; R), L 2 F t ([0, T ]; R m ) )

20 Best Response Control Action and Verification Theorem The optimal control process: u o (t, ω, x) := arg inf u U Hu [t, ω, x, D xφ(t, ω, x), u] { = arg inf f[t, ω, x, u], Dxφ(t, ω, x) + L[t, ω, x, u] }. u U is a forward in time F W t -adapted process for any fixed x. By a verification theorem approach, Peng showed that if a unique solution (φ, ψ)(t, x) to the SHJB equation exists, and if it satisfies: (i) for each t, (φ, ψ)(t, x) is a C 2 (R n ) map, (ii) for each x, (φ, ψ)(t, x) and (D xφ, Dxxφ, 2 D xψ)(t, x) are continuous Ft W -adapted stochastic processes, then φ(x, t) coincides with the value function of the optimal control problem.

21 The MFG Consistency Condition The functional dependence loop of observation and control of the major and minor agents yields the following proof iteration loop, initiated with a nominal measure µ t(ω) : µ ( ) (ω) m-smv u o (, ω, x) M-SHJB ( φ 0(, ω, x), ψ 0(, ω, x) ) M-SBR u o 0(, ω, x) m-sbr M-SMV ( φ(, ω, x), ψ(, ω, x) ) m-shjb z o 0(, ω) Closing the Loop: : By substituting u o into the generic minor agent s dynamics we get the SMV dynamics: dz o (t, ω) = f[t, z o (t, ω), u o (t, ω, z), µ t(ω)]dt + σ[t, z o (t, ω), µ t(ω)]dw(t), z o (0) = z(0), where f and σ are random processes via µ, and u o which depend on the Brownian motion of the major agent w 0. Let µ t(ω) be the conditional law of z o ( ) with control u o given F w 0 t. Then: The MFG or Nash certainty equivalence (NCE) consistency condition: The measure and control mapping loop for the MM MFG equation schema is closed if µ t(ω) = µ t(ω) a.s., 0 t T, which consistitutes the measure valued part of the solution.

22 Major-Minor Agent Stochastic MFG System Summary of the Major Agent s Stochastic MFG (SMFG) System: [ MFG-SHJB dφ 0(t, ω, x) = inf H 0[t, ω, x, u, D xφ 0(t, ω, x)] u U 0 + σ 0[t, x, µ t(ω)], D xψ 0(t, ω, x) Tr( a 0[t, ω, x]d 2 xxφ 0(t, ω, x) )] dt ψ T 0 (t, ω, x)dw 0(t, ω), φ 0(T, x) = 0 MFG-SBR u o 0(t, ω, x) = arg inf u U 0 H 0[t, ω, x, u, D xφ 0(t, ω, x)] MFG-SMV dz o 0(t, ω) = f 0[t, z o 0(t, ω), u o 0(t, ω, z 0), µ t(ω)]dt + σ 0[t, z o 0(t, ω), µ t(ω)]dw 0(t, ω), z o 0(0) = z 0(0) where a 0[t, ω, x] := σ 0[t, x, µ t(ω)]σ T 0 [t, x, µ t(ω)], and the stochastic Hamiltonian H 0 is H 0[t, ω, x, u, p] := f 0[t, x, u, µ t(ω)], p + L 0[t, x, u, µ t(ω)].

23 Major-Minor Agent Stochastic MFG System Summary of the Minor Agents SMFG System: [ MFG-SHJB dφ(t, ω, x) = H[t, ω, x, u, Dxφ(t, ω, x)] inf u U Tr( a[t, ω, x]d 2 xxφ(t, ω, x) )] dt ψ T (t, ω, x)dw 0(t, ω), φ(t, x) = 0 MFG-SBR u o (t, ω, x) = arg inf H[t, ω, x, u, Dxφ(t, ω, x)] u U MFG-SMV dz o (t, ω) = f[t, z o (t, ω), u o (t, ω, z), µ t(ω)]dt + σ[t, z o (t, ω), µ t(ω)]dw(t) where a[t, ω, x] := σ[t, x, µ t(ω)]σ T [t, x, µ t(ω)], and the stochastic Hamiltonian H is H[t, ω, x, p] := f[t, x, u, µ t(ω)], p + L[t, x, u, z o 0(t, ω), µ t(ω)].

24 Major-Minor Agent Stochastic MFG System Solution to the Major-Minor Agent SMFG System: The solution of the major-minor SMFG system consists of 8-tuple F w 0 t -adapted random processes ( φ0(t, ω, x), ψ 0(t, ω, x), u o 0(t, ω, x), z0(t, o ω), φ(t, ω, x), ψ(t, ω, x), u o (t, ω, x), z o (t, ω) ) where z o (t, ω) generates the random measure µ t(ω). The solution to the major-minor SMFG system is a stochastic mean field in contrast to the deterministic mean field of the standard MFG problems (HCM 03,HMC 06,LL 06).

25 Analysis of the MM-SMFG System Existence and uniqueness of Solutions to the Major and Minor (MM) Agents SMFG System: A fixed point argument with random parameters in the space of stochastic probability measures. µ ( ) (ω) m-smv u o (, ω, x) M-SHJB ( φ 0(, ω, x), ψ 0(, ω, x) ) M-SBR u o 0(, ω, x) m-sbr M-SMV ( φ(, ω, x), ψ(, ω, x) ) m-shjb z o 0(, ω) Theorem (Existence and Uniqueness of Solutions (Nourian and PEC. SIAM Jnl. Control and Optimization, 2013) Under technical conditions including a contraction gain condition there exists a unique solution for the map Γ, and hence a unique solution to the major and minor agents MM-SMFG system.

26 ɛ-nash Equilibrium of the MFG Control Laws Given ɛ > 0, the set of controls {u o j; 0 j N} generates an ɛ-nash equilibrium w.r.t. the costs J N j, 1 j N} if, for each j, J N j (u 0 j, u 0 j) ɛ inf u j U j J N j (u j, u 0 j) J N j (u 0 j, u 0 j). Theorem (Nourian, PEC, SIAM Jnl Control and Optimization, 2013) Subject to technical conditions, there exists a unique solution to the MM-MFG system such that the set of infinite population MF best response control processes in a finite N population system of minor agents (u o 0,, u o N ) generates an ɛ N -Nash equilibrium where ɛ N = O(1/ N) Agent y is a maximizer Agent x is a minimizer y x 0 1 2

27 Major-Minor Agent LQG -MFG Systems

28 Major-Minor Agent LQG-MFG Systems Application of non-linear theory to MM LQG-MFG systems: Yields retrieval of the MM MFG-LQG equations of [Nguyen-Huang 11] (given here in the uniform agent class case). Dynamics: A 0 : dz N 0 (t) = ( a 0z N 0 (t) + b 0u N 0 (t) + c 0z (N) (t) ) dt + σ 0dw 0(t) A i : dz N i (t) = ( az N i (t) + bu N i (t) + cz (N) (t) ) dt + σdw i(t), 1 i N where z (N) ( ) := (1/N) N i=1 zn i ( ) is the average state of minor agents. Costs: A 0 : J N 0 (u N 0, u N 0) = E A i : J N i (u N i, u N i) = E where r 0, r > 0. T 0 T 0 [( z N 0 (t) ( λ 0z (N) (t) + η 0 ) ) 2 + r0 ( u N 0 (t) ) 2 ] dt [( z N i (t) ( λz (N) (t) + λ 1z N 0 (t) + η )) 2 + r ( u N i (t) ) 2 ] dt

29 Major-Minor Agent LQG-MFG Systems The Major Agent s LQG-MFG System: [ ] [ (a0 Backward SDE : ds0(t) = (b 2 0/r ) 0)Π 0(t) s 0(t) η 0 + ( ) ] c 0Π 0(t) λ 0 z o (t) dt q 0(t)dw 0(t), s 0(T ) = 0 [ ] Best Response : u o 0 (t) := (b ( 0/r 0) Π 0(t)z0(t) o + s ) 0(t) [ Forward SDE ] : dz o 0 (t) = ( a 0z o 0(t) + b 0u o 0(t) + c 0z o (t) ) dt + σ 0dw 0(t), z o 0(0) = z 0(0) where Π 0( ) 0 is the unique solution of the Riccati equation: tπ 0(t) + 2a 0Π 0(t) (b 2 0/r 0)Π 2 0(t) + 1 = 0, Π 0(T ) = 0. The Minor Agents LQG-MFG System: [ ] [ (a Backward SDE : dsi(t) = (b 2 /r)π(t) ) s i(t) η λ 1z0(t) o + ( cπ(t) λ ) ] z o (t) dt q i(t)dw 0(t), s i(t ) = 0 [ ] Best Response : u o i (t) := (b/r) ( Π(t)zi o (t) + s ) i(t) [ Forward SDE ] : dz o i (t) = ( az o i (t) + bu o i (t) + cz o (t) ) dt + σdw i(t), z o i (0) = z i(0) where Π( ) 0 is the unique solution of the Riccati equation: tπ(t) + 2aΠ(t) (b 2 /r)π 2 (t) + 1 = 0, Π(T ) = 0.

30 Major-Minor Agent LQG-MFG Systems The Minor Agents Mean Field z o ( ) Equations: [ ] [ (a Backward SDE : ds(t) = (b 2 /r)π(t) ) s(t) η λ 1z0(t) o + ( cπ(t) λ ) ] z o (t) dt q(t)dw 0(t), s(t ) = 0 [ ] Best Response : u o i (t) := (b/r) ( Π(t)zi o (t) + s(t) ) [ MF FDE ] : dz o (t) = ( (a + c)z o (t) + bu o (t) ) dt. Key assumption for solution existence and uniqueness of MM-MFG system is that all drift and cost functions f and L and their derivatives are bounded which clearly does not hold for the MM-MFG LQG problem (as in classical LQG control), so particular methods must be used to deal with this.

31 The LQG-MFG Solution: The Generalized Four-Step Scheme The Generalized Four-Step Scheme for the SMFG-LQG System: For given z o ( ) we set s 0(t) = θ(t, z0(t)) o where the function θ is to be determined. By Itô s formula: { ds 0(t) = dθ(t, z0(t)) o = θ t(t, z0(t)) o + θ x(t, z0(t)) o [ (a 0 b2 0 )z0(t) o r 0 b2 0 θ(t, z0(t)) o + c 0z o (t) ] + 1 } r 0 2 σ2 0θ xx(t, z0(t)) o dt + σ 0θ x(t, z0(t))dw o 0(t). Comparing this to the Major agent s SDE implies that θ should satisfy the equations: θ t(t, z o 0(t)) + θ x(t, z o 0(t)) [ (a 0 b2 0 r 0 )z o 0(t) b2 0 r 0 θ(t, z o 0(t)) + c 0z o (t) ] σ2 0θ xx(t, z o 0(t)) = [ a 0 b2 0 r 0 Π 0(t) ] θ(t, z o 0(t)) + η 0 [ c 0Π 0(t) λ 0 ] z o (t) σ 0θ x(t, z o 0(t)) = q 0(t). We can get similar equations for the minor agent by setting s(t) = ˇθ(t, z o 0(t), z o (t)).

32 The Four-Step Scheme for the LQG-MFG System Step 1. For given z o ( ) solve the following parabolic PDE for θ(t, x): θ t(t, x) + θ x(t, x) [ (a 0 b2 0 r 0 Π 0(t))x b2 0 r 0 θ(t, x) + c 0z o (t) ] σ2 0θ xx(t, x) = [ a 0 b2 0 r 0 Π 0(t) ] θ(t, x) + η 0 [ c 0Π 0(t) λ 0 ] z o (t), θ(t, x) = 0. Step 2. Use θ in Step 1 to solve the following forward SDE: dz o 0(t) = [ (a 0 b2 0 r 0 Π 0(t))z o 0(t) b2 0 r 0 θ(t, z o 0(t)) + c 0z o (t) ] dt + σ 0dw 0(t), z o 0(0) = z 0(0) Step 3. Set q 0(t) = σ 0θ x(t, z o 0(t)) s 0(t) = θ(t, z o 0(t)). by the use of θ and z o 0(t) obtained in Steps 1 and 2.

33 The Four-Step Scheme for the LQG-MFG System (ctd ) Step 4. Use θ to solve the following PDE for ˇθ(t, x, y): ˇθ t(t, x, y) + ˇθ x(t, x, y) [ (a 0 b2 0 r 0 )x b2 0 r 0 θ(t, x) + c 0y ] σ2 0 ˇθ xx(t, x, y) + ˇθ y(t, x, y) [ (a + c b2 b2 Π(t))y r r ˇθ(t, x, y)] = [ a b2 r Π(t)]ˇθ(t, x, y) + η λ1x ( cπ(t) λ ) y, ˇθ(T, x, y) = 0. Step 5. Use z o 0(t) and ˇθ obtained in Steps 2 and 4 to solve the following forward SDE: dz o (t) = [ (a + c b2 r Π(t))zo (t) b2 r ˇθ(t, z o 0(t), z o (t)) ] dt, z o (0). Step 6. Set q(t) = σ 0 ˇθx(t, z o 0(t), z o (t)) s(t) = ˇθ(t, z o 0(t), z o (t)). by the use of z o 0(t), ˇθ and z o (t) obtained in Steps 2, 4 and 5.

34 Partially Observed Major-Minor Agent Mean Field Systems

35 Infinite Horizon Completely Observed MM MFG Problem Formulation (Huang 2010) Dynamics: Completely Observed Finite Population: Major Agent dx 0 = [A 0 x 0 + B 0 u 0 ]dt + D 0 dw 0 Minor Agents dx i = [A(θ i )x i + B(θ i )u i + Gx 0 ]dt + Ddw i, i N The individual infinite horizon cost for the major agent: J 0 (u 0, u 0 ) = E 0 e ρt { x 0 Φ(x N ) 2 Q 0 + u 0 2 R 0 } dt Φ( ) = H 0 x N + η 0 x N = (1/N) N i=1 x i The individual infinite horizon cost for a minor agent i, i N: { J i (u i, u i ) = E e ρt x i Ψ(x N ) } 2 Q + u i 2 R dt 0 Ψ( ) = H 1 x 0 + H 2 x N + η

36 Minor Agents Types Minor Agents Types: I k = {i : θ i = k, i N}, N k = I k, 1 k K π N = (π1 N,..., πn K ), πn k = N k/n, 1 k K, denotes the empirical distribution of the parameters (θ 1,..., θ N ) of the agents A i N. Assumption: There exists π such that lim N π N = π a.s.

37 Major Agent and Minor Agents Introduce the (auxiliary) state means: x N k = 1 x i, 1 k K N k i I k If it exists, the L 2 limit of the system states means x N = [x N 1,..., xn K ] constitutes the system mean field. Subject to time invariant local state plus mean field plus major agent state feedback control, x N = [x N 1,..., xn K ] satisfies the mean field equation K d x k = Ā k,j x j dt + Ḡkx 0 dt + m k dt, j=1 1 k K i.e., d x(t) = Ā x(t)dt + Ḡx 0(t)dt + m(t)dt where the quantities Ḡk, m k are to be solved for in the tracking solution.

38 Major Agent and Minor Agents LQG - MFG When MF plus x 0 plus local state dependent controls are applied, the MF-dependent extended state closes the system equations into state equations. Major Agent s Extended State: Major agent s state extended by the mean field: [ x0 x Minor Agents Extended States: Minor agent s state extended by major agent s state and the x i mean field: x 0. x ].

39 Major Agent and Minor Agents (Inf. Population) Major Agent s Dynamics (Infinite Population): [ dx0 d x ] [ A0 0 = nk n Ḡ Ā [ + ] [ x0 x ] B 0 u 0 0 dt + nk m ] dt [ 0n 1 m ] [ ] D0 dw dt nk 1 [ ] [ A0 0 A 0 = nk n B Ḡ Ā 0 = [ ] [ 0n 1 M 0 = Q π m 0 = B 0 0 nk m ] Q 0 Q 0 H0 π H0 πt Q 0 H0 πt Q 0 H0 π η 0 = [I n n, H π 0 ] T Q 0 η 0 H π 0 = π H 0 [π 1 H 0 π 2 H 0... π K H 0 ] ]

40 Major Agent and Minor Agents (Inf. Population) Minor Agents Dynamics (Infinite Population): dx i dx 0 d x [ = [ A k = [ B k = A k [G 0 n nk ] 0 (nk+n) n A 0 [ + ] B k 0 (nk+n) m ] x i x 0 x dt [ ] 0n 1 u i dt + M 0 0 n m B 0 0 nk m A k [G 0 n nk ] 0 (nk+n) n A 0 B 0 R 1 ] [ B k M = 0 (nk+n) m dt u 0 dt + 0 BT 0 Π 0 ] 0 n 1 M 0 B 0 R 1 0 BT 0 s 0 η = [I n n, H, H π 2 ] T Qη H π 2 = π H 2 Ddw i D 0 dw 0 0 nk 1 ]

41 Infinite Population Cost Function The individual cost for the major agent: J 0 (u 0, u 0 ) = E 0 e ρt { x 0 Φ( x) 2 Q 0 + u 0 2 R 0 } dt Φ( ) = H0 π x + η 0 The individual cost for a minor agent i, i N: J i (u i, u i ) = E 0 e ρt { x i Ψ( x) 2 Q + u i 2 R Ψ( ) = H 1 x 0 + H π 2 x + η } dt

42 Control Actions (Infinite Population) Major Agent Tracking Problem Solution: ρπ 0 = Π 0 A 0 + A T 0 Π 0 Π 0 B 0 R 1 0 BT 0 Π 0 + Q π 0 ρs 0 = ds 0 dt + (A 0 B 0 R 1 0 BT 0 Π 0 ) T s 0 + Π 0 M 0 η 0 u 0 i = R0 1 [ BT 0 Π0 (x T 0, z T ) T + s 0] Minor Agent Tracking Problem Solution: ρπ k = Π k A k + A T k Π k Π k B k R 1 B T k Π k + Q ρs k = ds k dt + (A k B k R 1 B T k Π k) T s k + Π km η u 0 i = R 1 B T [ k Πk (x T i, x T 0, z T ) T + s k]

43 Major-Minor MF Equations Define: Π k = Π k,11 Π k,12 Π k,13 Π k,21 Π k,22 Π k,23 Π k,31 Π k,32 Π k,33, 1 k K e k = [0 n n,..., 0 n n, I n, 0 n n,..., 0 n n ], where the n n identity matrix I n is at the kth block. Major-Minor MF Equations for Ā, Ḡ, m: Consistency Requirements ρπ 0 = Π 0 A 0 + A T 0 Π 0 Π 0 B 0 R 1 0 BT 0 Π 0 + Q π 0, ρπ k = Π k A k + A T k Π k Π k B k R 1 B T k Π k + Q π, k, Ā k = [A k B k R 1 B T k Π k,11 ]e k B k R 1 B T k Π k,13, k, Ḡ k = B k R 1 B T k Π k,12, k, ρs 0 = ds 0 dt + (A 0 B 0 R 1 0 BT 0 Π 0 ) T s 0 + Π 0 M 0 η 0, ρs k = ds k dt + (A k B k R 1 B T k Π k ) T s k + Π k M η, k, m k = B k R 1 B T k s k, k,

44 Assumptions Define A 1 B 1R 1 B1 T Π 1,11 M 1 =... B 1R 1 B1 T Π 1,13 A K B KR 1 BKΠ T K,11 M 2 =. B KR 1 B KΠ T K,13 A M 3 = Ḡ Ā 0, L 0,H = Q 1/2 0 [I, 0, H0 π ] Ḡ M 2 M 1 H1: There exists a probability vector π such that lim N π N = π. H2: The initial states are independent, Ex i(0) = 0 for each i 1, sup j 0 E x j(0) 2 c. H3: The pair (L 0,H, M 3) is observable. H4: The pair (L a, A 0 (ρ/2)i) is detectable, and for each k = 1,..., K, the pair (L b, A k (ρ/2)i) is detectable, where L a = Q 1/2 0 [I, H0 π ] and L b = Q 1/2 [I, H, Ĥπ ]. The pair (A 0 (ρ/2)i, B 0) is stabilizable and (A k (ρ/2)i, B k ) is stabilizable for each k = 1,..., K.

45 Major - Minor: MF Equilibrium Theorem: (Huang, 2010) Major and Minor Agents: MF Equilibrium: Subject to H1-H4 the MF equations generate a set of stochastic control laws UMF N {u0 i ; 0 i N}, 1 N <, such that (i) All agent systems S(A i ), 0 i N, are second order stable. (ii) {UMF N ; 1 N < } yields an ɛ-nash equilibrium for all ɛ, i.e. for all ɛ > 0, there exists N(ɛ) such that for all N N(ɛ) J N i (u 0 i, u 0 i) ɛ inf u i U g J N i (u i, u 0 i) J N i (u 0 i, u 0 i).

46 Simulation 0.05 Major Agent y x t State trajectories

47 Partially Observed Major-Minor Mean Field Systems

48 Partially Observed Major-Minor Agent Systems Recall: Estimation of major agent and mean field states becomes a meaningful problem in MM case. Major Agent dx 0 = [A 0 x 0 + B 0 u 0 ]dt + D 0 dw 0 Minor Agents dx i = [A(θ i )x i + B(θ i )u i + Gx 0 ]dt + Ddw i, The observation process for minor agent A i : x i dy i (t) = Lx 0, x i dt + dv i (t) L x 0 x where L = [L 1 L 2 0]. dt + dv i (t) Complete observations process for the major agent A 0 : dy o (t) = dx 0 (t) i N

49 Partially Observed Major-Minor Agent Systems The Major agent is assumed to have complete observations of its own state. This permits the Minor agents to form conditional expectations of the Major agent s MFG control action u 0 since it is a (linear) function of the Major agent s state. Such an estimate could not in general be generated by the Minor agents in the case where the Major agent s control action is a (linear) function of the conditional expectation of its state E[x 0 F 0], where F 0 is the Major agent s observation σ-field.

50 Estimation The Riccati equation associated with the Kalman filtering equations for [x i, x 0, x]: x 0, x i V (t) = A k V (t) + V (t)a T k K(t)R v K T (t) + Q w, where Q w = Σ i Σ ζ , A k = A k [G 0 n nk ] [ A0 0 0 (nk n) (nk+n) n Ḡ Ā ] and [ V (0) = E x 0, x i ] [ (0) ˆx 0, x i (0) x 0, x i T (0) ˆx 0, x i (0)].

51 Estimation The innovation process is dν i = dy i L and the Kalman filter gain is given by ˆx i F y i ˆx 0 F y i ˆ x F y i K(t) = V (t)l T R 1 v. H5: The system parameter set Θ satisfies [A k, Q w ] controllable and [L, A k ] observable for 1 k K. The Filtering Equations: dˆx i F y i [ A k ] [ [G 0 n nk ] ] dˆx 0 F y = 0n n A0 0 i nk n dˆ x F y 0 i nk n Ḡ Ā B k 0 n m + 0 n m u i dt + B 0 û 0 F y dt + i 0 nk m 0 nk m, 0 n 1 0 n 1 m ˆx i F y i ˆx 0 F y i ˆ x F y i dt dt + Kdν i,

52 Separation Principle for MM-MFG Systems The control law dependent summand of the individual cost for the major agent A 0 : J N 0 (u 0, u 0 ) = E 0 e ρt { x 0 H 0ˆx N F 0 η 0 2 Q 0 + u 0 2 R 0 } dt ˆx N F 0 = (1/N) N i=1 ˆx i F 0 The control law dependent summand of the individual cost for a minor agent A i, i N: J N i (u i, u i ) = E 0 e ρt { ˆxi F y i H 1ˆx 0 F y i H 2ˆx N F y i η 2 Q + u i 2 R }

53 Separation Principle for PO MM-MFG Systems Key Steps to Main Result: (1) The major agent and individual minor agent state estimation recursive equations schemes are given by the MM KF-MF Equations (for size N finite populations and infinite populations). (2) Apply the Separation Theorem strategy of reducing a partially observed SOC problem to a completely observed SOC problem for the controlled state estimate processes. (3) Step 2 transforms the J 0 and J i MM performance functions into LQG MM tracking performance functions on the controlled state estimate processes in the infinite and finite population cases.

54 Separation Principle for PO MM-MFG Systems (4) The problem in (2) for the state estimate processes is solved using the completely observed LQG MM-MFG methodology which yields the û 0 and û i control laws and the J0 and Ji performance function values. (5) The Major Agent performance value J0 and Minor Agent performance value Ji neccessarily correspond to infinite population Nash equilibria. (6) Approximation Analysis gives ɛ Nash equilibria with respect to J 0 and J i for J N 0 and J N i in finite N populations.

55 Nash Equilibria for Partially Observed MM-MFG Systems Theorem: (PEC, AK, 2013) ɛ-nash Equilibria for PO MM-MF Systems Subject to H1-H5, the KF-MF state estimation scheme plus MM-MFG equations generate the set of control laws ÛMF N {û0 i ; 1 i N}, 1 N <, and u 0 given by u 0 0 = R0 1 [ BT 0 Π0 (x T 0, x T ) T + s ] 0, [ ] û 0 i = R 1 B T k Π k (ˆx T i F y, ˆx T i 0 F y, ˆ x T i F y ) T + s k i such that (i) All agent systems S(A i ), 0 i N, are second order stable. (ii) {Û MF N ; 1 N < } yields an ɛ-nash equilibrium for all ɛ, i.e. for all ɛ > 0, there exists N(ɛ) such that for all N N(ɛ) J N i (û 0 i, û 0 i) ɛ inf u i U g J N i (u i, û 0 i) J N i (û 0 i, û 0 i).

56 Concluding Remarks Future Work Building upon the MM-LQG-MFG theory for partially observed MM systems and the Nonlinear MM-MFG theory, the next step is to generate a MM-MFG theory for partially observed nonlinear MM systems. Investigate applications to power markets via the systematic application of MM-MFG Theory to markets where minor agents (customers and suppliers) receive intermittent observations on active major agents (such as utilities and international energy prices) and on passive major agents (such as wind and ocean behaviour).

57

On continuous time contract theory

On continuous time contract theory Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem

More information

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem

More information

arxiv: v1 [math.pr] 18 Oct 2016

arxiv: v1 [math.pr] 18 Oct 2016 An Alternative Approach to Mean Field Game with Major and Minor Players, and Applications to Herders Impacts arxiv:161.544v1 [math.pr 18 Oct 216 Rene Carmona August 7, 218 Abstract Peiqi Wang The goal

More information

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio Mean-Field optimization problems and non-anticipative optimal transport Beatrice Acciaio London School of Economics based on ongoing projects with J. Backhoff, R. Carmona and P. Wang Robust Methods in

More information

Systemic Risk and Stochastic Games with Delay

Systemic Risk and Stochastic Games with Delay Systemic Risk and Stochastic Games with Delay Jean-Pierre Fouque (with René Carmona, Mostafa Mousavi and Li-Hsien Sun) PDE and Probability Methods for Interactions Sophia Antipolis (France) - March 30-31,

More information

Risk-Sensitive and Robust Mean Field Games

Risk-Sensitive and Robust Mean Field Games Risk-Sensitive and Robust Mean Field Games Tamer Başar Coordinated Science Laboratory Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL - 6181 IPAM

More information

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION René Carmona Department of Operations Research & Financial Engineering PACM

More information

Large-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed

Large-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed Definition Mean Field Game (MFG) theory studies the existence of Nash equilibria, together with the individual strategies which generate them, in games involving a large number of agents modeled by controlled

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011 Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance

More information

Solution of Stochastic Optimal Control Problems and Financial Applications

Solution of Stochastic Optimal Control Problems and Financial Applications Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Weak solutions of mean-field stochastic differential equations

Weak solutions of mean-field stochastic differential equations Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works

More information

Control of McKean-Vlasov Dynamics versus Mean Field Games

Control of McKean-Vlasov Dynamics versus Mean Field Games Control of McKean-Vlasov Dynamics versus Mean Field Games René Carmona (a),, François Delarue (b), and Aimé Lachapelle (a), (a) ORFE, Bendheim Center for Finance, Princeton University, Princeton, NJ 8544,

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Prof. Erhan Bayraktar (University of Michigan)

Prof. Erhan Bayraktar (University of Michigan) September 17, 2012 KAP 414 2:15 PM- 3:15 PM Prof. (University of Michigan) Abstract: We consider a zero-sum stochastic differential controller-and-stopper game in which the state process is a controlled

More information

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009 A new approach for investment performance measurement 3rd WCMF, Santa Barbara November 2009 Thaleia Zariphopoulou University of Oxford, Oxford-Man Institute and The University of Texas at Austin 1 Performance

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS René Carmona Department of Operations Research & Financial Engineering PACM Princeton University CEMRACS - Luminy, July 17, 217 MFG WITH MAJOR AND MINOR PLAYERS

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Backward Stochastic Differential Equations with Infinite Time Horizon

Backward Stochastic Differential Equations with Infinite Time Horizon Backward Stochastic Differential Equations with Infinite Time Horizon Holger Metzler PhD advisor: Prof. G. Tessitore Università di Milano-Bicocca Spring School Stochastic Control in Finance Roscoff, March

More information

Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm

Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm Gonçalo dos Reis University of Edinburgh (UK) & CMA/FCT/UNL (PT) jointly with: W. Salkeld, U. of

More information

Production and Relative Consumption

Production and Relative Consumption Mean Field Growth Modeling with Cobb-Douglas Production and Relative Consumption School of Mathematics and Statistics Carleton University Ottawa, Canada Mean Field Games and Related Topics III Henri Poincaré

More information

Backward martingale representation and endogenous completeness in finance

Backward martingale representation and endogenous completeness in finance Backward martingale representation and endogenous completeness in finance Dmitry Kramkov (with Silviu Predoiu) Carnegie Mellon University 1 / 19 Bibliography Robert M. Anderson and Roberto C. Raimondo.

More information

On the Multi-Dimensional Controller and Stopper Games

On the Multi-Dimensional Controller and Stopper Games On the Multi-Dimensional Controller and Stopper Games Joint work with Yu-Jui Huang University of Michigan, Ann Arbor June 7, 2012 Outline Introduction 1 Introduction 2 3 4 5 Consider a zero-sum controller-and-stopper

More information

Mean Field Consumption-Accumulation Optimization with HAR

Mean Field Consumption-Accumulation Optimization with HAR Mean Field Consumption-Accumulation Optimization with HARA Utility School of Mathematics and Statistics Carleton University, Ottawa Workshop on Mean Field Games and Related Topics 2 Padova, September 4-7,

More information

SDE Coefficients. March 4, 2008

SDE Coefficients. March 4, 2008 SDE Coefficients March 4, 2008 The following is a summary of GARD sections 3.3 and 6., mainly as an overview of the two main approaches to creating a SDE model. Stochastic Differential Equations (SDE)

More information

Weak solutions to Fokker-Planck equations and Mean Field Games

Weak solutions to Fokker-Planck equations and Mean Field Games Weak solutions to Fokker-Planck equations and Mean Field Games Alessio Porretta Universita di Roma Tor Vergata LJLL, Paris VI March 6, 2015 Outlines of the talk Brief description of the Mean Field Games

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Robust control and applications in economic theory

Robust control and applications in economic theory Robust control and applications in economic theory In honour of Professor Emeritus Grigoris Kalogeropoulos on the occasion of his retirement A. N. Yannacopoulos Department of Statistics AUEB 24 May 2013

More information

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Surveys in Mathematics and its Applications ISSN 1842-6298 (electronic), 1843-7265 (print) Volume 5 (2010), 275 284 UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Iuliana Carmen Bărbăcioru Abstract.

More information

Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications

Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications Leader-Follower Stochastic Differential Game with Asymmetric Information and Applications Jie Xiong Department of Mathematics University of Macau Macau Based on joint works with Shi and Wang 2015 Barrett

More information

Variational approach to mean field games with density constraints

Variational approach to mean field games with density constraints 1 / 18 Variational approach to mean field games with density constraints Alpár Richárd Mészáros LMO, Université Paris-Sud (based on ongoing joint works with F. Santambrogio, P. Cardaliaguet and F. J. Silva)

More information

Lecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University

Lecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University Lecture 4: Ito s Stochastic Calculus and SDE Seung Yeal Ha Dept of Mathematical Sciences Seoul National University 1 Preliminaries What is Calculus? Integral, Differentiation. Differentiation 2 Integral

More information

A class of globally solvable systems of BSDE and applications

A class of globally solvable systems of BSDE and applications A class of globally solvable systems of BSDE and applications Gordan Žitković Department of Mathematics The University of Texas at Austin Thera Stochastics - Wednesday, May 31st, 2017 joint work with Hao

More information

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-Field SDE s with Irregular Drift Thilo Meyer-Brandis University of Munich joint with M. Bauer, University of Munich Conference for the 1th Anniversary

More information

I forgot to mention last time: in the Ito formula for two standard processes, putting

I forgot to mention last time: in the Ito formula for two standard processes, putting I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy

More information

Stochastic contraction BACS Workshop Chamonix, January 14, 2008

Stochastic contraction BACS Workshop Chamonix, January 14, 2008 Stochastic contraction BACS Workshop Chamonix, January 14, 2008 Q.-C. Pham N. Tabareau J.-J. Slotine Q.-C. Pham, N. Tabareau, J.-J. Slotine () Stochastic contraction 1 / 19 Why stochastic contraction?

More information

Continuous Time Finance

Continuous Time Finance Continuous Time Finance Lisbon 2013 Tomas Björk Stockholm School of Economics Tomas Björk, 2013 Contents Stochastic Calculus (Ch 4-5). Black-Scholes (Ch 6-7. Completeness and hedging (Ch 8-9. The martingale

More information

Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control

Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control Notes for Quantum Control Summer School, Caltech, August 2005 M.R. James Department of Engineering Australian National

More information

Systemic Risk and Stochastic Games with Delay

Systemic Risk and Stochastic Games with Delay Systemic Risk and Stochastic Games with Delay Jean-Pierre Fouque Collaborators: R. Carmona, M. Mousavi, L.-H. Sun, N. Ning and Z. Zhang Mean Field Games 2017 IPAM at UCLA JP Fouque (UC Santa Barbara) Stochastic

More information

An Introduction to Malliavin calculus and its applications

An Introduction to Malliavin calculus and its applications An Introduction to Malliavin calculus and its applications Lecture 3: Clark-Ocone formula David Nualart Department of Mathematics Kansas University University of Wyoming Summer School 214 David Nualart

More information

Quadratic BSDE systems and applications

Quadratic BSDE systems and applications Quadratic BSDE systems and applications Hao Xing London School of Economics joint work with Gordan Žitković Stochastic Analysis and Mathematical Finance-A Fruitful Partnership 25 May, 2016 1 / 23 Backward

More information

Explicit solutions of some Linear-Quadratic Mean Field Games

Explicit solutions of some Linear-Quadratic Mean Field Games Explicit solutions of some Linear-Quadratic Mean Field Games Martino Bardi Department of Pure and Applied Mathematics University of Padua, Italy Men Field Games and Related Topics Rome, May 12-13, 2011

More information

On the connection between symmetric N-player games and mean field games

On the connection between symmetric N-player games and mean field games On the connection between symmetric -player games and mean field games Markus Fischer May 14, 214; revised September 7, 215; modified April 15, 216 Abstract Mean field games are limit models for symmetric

More information

Chapter Introduction. A. Bensoussan

Chapter Introduction. A. Bensoussan Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688

More information

Uniformly Uniformly-ergodic Markov chains and BSDEs

Uniformly Uniformly-ergodic Markov chains and BSDEs Uniformly Uniformly-ergodic Markov chains and BSDEs Samuel N. Cohen Mathematical Institute, University of Oxford (Based on joint work with Ying Hu, Robert Elliott, Lukas Szpruch) Centre Henri Lebesgue,

More information

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt +

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt + ECE 553, Spring 8 Posted: May nd, 8 Problem Set #7 Solution Solutions: 1. The optimal controller is still the one given in the solution to the Problem 6 in Homework #5: u (x, t) = p(t)x k(t), t. The minimum

More information

Mean-field SDE driven by a fractional BM. A related stochastic control problem

Mean-field SDE driven by a fractional BM. A related stochastic control problem Mean-field SDE driven by a fractional BM. A related stochastic control problem Rainer Buckdahn, Université de Bretagne Occidentale, Brest Durham Symposium on Stochastic Analysis, July 1th to July 2th,

More information

Non-linear wave equations. Hans Ringström. Department of Mathematics, KTH, Stockholm, Sweden

Non-linear wave equations. Hans Ringström. Department of Mathematics, KTH, Stockholm, Sweden Non-linear wave equations Hans Ringström Department of Mathematics, KTH, 144 Stockholm, Sweden Contents Chapter 1. Introduction 5 Chapter 2. Local existence and uniqueness for ODE:s 9 1. Background material

More information

Theoretical Tutorial Session 2

Theoretical Tutorial Session 2 1 / 36 Theoretical Tutorial Session 2 Xiaoming Song Department of Mathematics Drexel University July 27, 216 Outline 2 / 36 Itô s formula Martingale representation theorem Stochastic differential equations

More information

13 The martingale problem

13 The martingale problem 19-3-2012 Notations Ω complete metric space of all continuous functions from [0, + ) to R d endowed with the distance d(ω 1, ω 2 ) = k=1 ω 1 ω 2 C([0,k];H) 2 k (1 + ω 1 ω 2 C([0,k];H) ), ω 1, ω 2 Ω. F

More information

ON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY)

ON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY) Collège de France, Paris (joint project with Jean-Michel LASRY) I INTRODUCTION II A REALLY SIMPLE EXAMPLE III GENERAL STRUCTURE IV TWO PARTICULAR CASES V OVERVIEW AND PERSPECTIVES I. INTRODUCTION New class

More information

Mean Field Control: Selected Topics and Applications

Mean Field Control: Selected Topics and Applications Mean Field Control: Selected Topics and Applications School of Mathematics and Statistics Carleton University Ottawa, Canada University of Michigan, Ann Arbor, Feb 2015 Outline of talk Background and motivation

More information

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations Lukasz Szpruch School of Mathemtics University of Edinburgh joint work with Shuren Tan and Alvin Tse (Edinburgh) Lukasz Szpruch (University

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Mean Field Games on networks

Mean Field Games on networks Mean Field Games on networks Claudio Marchi Università di Padova joint works with: S. Cacace (Rome) and F. Camilli (Rome) C. Marchi (Univ. of Padova) Mean Field Games on networks Roma, June 14 th, 2017

More information

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.

More information

Alberto Bressan. Department of Mathematics, Penn State University

Alberto Bressan. Department of Mathematics, Penn State University Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations.

Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations. Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations. Wenqing Hu. 1 (Joint work with Michael Salins 2, Konstantinos Spiliopoulos 3.) 1. Department of Mathematics

More information

The Convergence of the Minimum Energy Estimator

The Convergence of the Minimum Energy Estimator The Convergence of the Minimum Energy Estimator Arthur J. Krener Department of Mathematics, University of California, Davis, CA 95616-8633, USA, ajkrener@ucdavis.edu Summary. We show that under suitable

More information

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio Mean-Field optimization problems and non-anticipative optimal transport Beatrice Acciaio London School of Economics based on ongoing projects with J. Backhoff, R. Carmona and P. Wang Thera Stochastics

More information

SEPARABLE TERM STRUCTURES AND THE MAXIMAL DEGREE PROBLEM. 1. Introduction This paper discusses arbitrage-free separable term structure (STS) models

SEPARABLE TERM STRUCTURES AND THE MAXIMAL DEGREE PROBLEM. 1. Introduction This paper discusses arbitrage-free separable term structure (STS) models SEPARABLE TERM STRUCTURES AND THE MAXIMAL DEGREE PROBLEM DAMIR FILIPOVIĆ Abstract. This paper discusses separable term structure diffusion models in an arbitrage-free environment. Using general consistency

More information

Lecture 12: Diffusion Processes and Stochastic Differential Equations

Lecture 12: Diffusion Processes and Stochastic Differential Equations Lecture 12: Diffusion Processes and Stochastic Differential Equations 1. Diffusion Processes 1.1 Definition of a diffusion process 1.2 Examples 2. Stochastic Differential Equations SDE) 2.1 Stochastic

More information

Stochastic Calculus. Kevin Sinclair. August 2, 2016

Stochastic Calculus. Kevin Sinclair. August 2, 2016 Stochastic Calculus Kevin Sinclair August, 16 1 Background Suppose we have a Brownian motion W. This is a process, and the value of W at a particular time T (which we write W T ) is a normally distributed

More information

MEAN FIELD GAMES AND SYSTEMIC RISK

MEAN FIELD GAMES AND SYSTEMIC RISK MEA FIELD GAMES AD SYSTEMIC RISK REÉ CARMOA, JEA-PIERRE FOUQUE, AD LI-HSIE SU Abstract. We propose a simple model of inter-bank borrowing and lending where the evolution of the logmonetary reserves of

More information

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p

More information

Introduction Optimality and Asset Pricing

Introduction Optimality and Asset Pricing Introduction Optimality and Asset Pricing Andrea Buraschi Imperial College Business School October 2010 The Euler Equation Take an economy where price is given with respect to the numéraire, which is our

More information

Branching Processes II: Convergence of critical branching to Feller s CSB

Branching Processes II: Convergence of critical branching to Feller s CSB Chapter 4 Branching Processes II: Convergence of critical branching to Feller s CSB Figure 4.1: Feller 4.1 Birth and Death Processes 4.1.1 Linear birth and death processes Branching processes can be studied

More information

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Conference on the occasion of Giovanni Colombo and Franco Ra... https://events.math.unipd.it/oscgc18/ Optimization,

More information

Mean-Field Games with non-convex Hamiltonian

Mean-Field Games with non-convex Hamiltonian Mean-Field Games with non-convex Hamiltonian Martino Bardi Dipartimento di Matematica "Tullio Levi-Civita" Università di Padova Workshop "Optimal Control and MFGs" Pavia, September 19 21, 2018 Martino

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f

More information

Nonlinear Systems Theory

Nonlinear Systems Theory Nonlinear Systems Theory Matthew M. Peet Arizona State University Lecture 2: Nonlinear Systems Theory Overview Our next goal is to extend LMI s and optimization to nonlinear systems analysis. Today we

More information

Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience

Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience Ulrich Horst 1 Humboldt-Universität zu Berlin Department of Mathematics and School of Business and Economics Vienna, Nov.

More information

Random attractors and the preservation of synchronization in the presence of noise

Random attractors and the preservation of synchronization in the presence of noise Random attractors and the preservation of synchronization in the presence of noise Peter Kloeden Institut für Mathematik Johann Wolfgang Goethe Universität Frankfurt am Main Deterministic case Consider

More information

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Joao Meireles joint work with Martino Bardi and Guy Barles University of Padua, Italy Workshop "New Perspectives in Optimal

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455

More information

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA On Stochastic Adaptive Control & its Applications Bozenna Pasik-Duncan University of Kansas, USA ASEAS Workshop, AFOSR, 23-24 March, 2009 1. Motivation: Work in the 1970's 2. Adaptive Control of Continuous

More information

Exercises. T 2T. e ita φ(t)dt.

Exercises. T 2T. e ita φ(t)dt. Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

On a class of stochastic differential equations in a financial network model

On a class of stochastic differential equations in a financial network model 1 On a class of stochastic differential equations in a financial network model Tomoyuki Ichiba Department of Statistics & Applied Probability, Center for Financial Mathematics and Actuarial Research, University

More information

Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Was

Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Was Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Wasserstein Space With many discussions with Yann Brenier and Wilfrid Gangbo Brenierfest, IHP, January 9-13, 2017 ain points of the

More information

Approximation of BSDEs using least-squares regression and Malliavin weights

Approximation of BSDEs using least-squares regression and Malliavin weights Approximation of BSDEs using least-squares regression and Malliavin weights Plamen Turkedjiev (turkedji@math.hu-berlin.de) 3rd July, 2012 Joint work with Prof. Emmanuel Gobet (E cole Polytechnique) Plamen

More information

Quasi-invariant Measures on Path Space. Denis Bell University of North Florida

Quasi-invariant Measures on Path Space. Denis Bell University of North Florida Quasi-invariant Measures on Path Space Denis Bell University of North Florida Transformation of measure under the flow of a vector field Let E be a vector space (or a manifold), equipped with a finite

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

Mean Field Game Systems with Common Noise and Markovian Latent Processes

Mean Field Game Systems with Common Noise and Markovian Latent Processes arxiv:189.7865v2 [math.oc 29 Oct 218 Mean Field Game Systems with Common Noise and Markovian Latent Processes 1 Abstract Dena Firoozi, Peter E. Caines, Sebastian Jaimungal A class of non-cooperative stochastic

More information

Non-Zero-Sum Stochastic Differential Games of Controls and St

Non-Zero-Sum Stochastic Differential Games of Controls and St Non-Zero-Sum Stochastic Differential Games of Controls and Stoppings October 1, 2009 Based on two preprints: to a Non-Zero-Sum Stochastic Differential Game of Controls and Stoppings I. Karatzas, Q. Li,

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

Weak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations

Weak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations Weak Convergence of Numerical Methods for Dynamical Systems and, and a relation with Large Deviations for Stochastic Equations Mattias Sandberg KTH CSC 2010-10-21 Outline The error representation for weak

More information

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N Problem 1. Let f : A R R have the property that for every x A, there exists ɛ > 0 such that f(t) > ɛ if t (x ɛ, x + ɛ) A. If the set A is compact, prove there exists c > 0 such that f(x) > c for all x

More information

Half of Final Exam Name: Practice Problems October 28, 2014

Half of Final Exam Name: Practice Problems October 28, 2014 Math 54. Treibergs Half of Final Exam Name: Practice Problems October 28, 24 Half of the final will be over material since the last midterm exam, such as the practice problems given here. The other half

More information

Introduction to Mean Field Game Theory Part I

Introduction to Mean Field Game Theory Part I School of Mathematics and Statistics Carleton University Ottawa, Canada Workshop on Game Theory, Institute for Math. Sciences National University of Singapore, June 2018 Outline of talk Background and

More information

Continuous dependence estimates for the ergodic problem with an application to homogenization

Continuous dependence estimates for the ergodic problem with an application to homogenization Continuous dependence estimates for the ergodic problem with an application to homogenization Claudio Marchi Bayreuth, September 12 th, 2013 C. Marchi (Università di Padova) Continuous dependence Bayreuth,

More information

Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator

Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator Thi Tuyen Nguyen Ph.D student of University of Rennes 1 Joint work with: Prof. E. Chasseigne(University of

More information

The Nash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and Entry Adaptation

The Nash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and Entry Adaptation Proceedings of the 46th IEEE Conference on Decision and Control ew Orleans, LA, USA, Dec. 12-14, 27 WeA5.1 The ash Certainty Equivalence Principle and McKean-Vlasov Systems: An Invariance Principle and

More information

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME SAUL D. JACKA AND ALEKSANDAR MIJATOVIĆ Abstract. We develop a general approach to the Policy Improvement Algorithm (PIA) for stochastic control problems

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum 9 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June -, 9 FrA.5 An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted

More information