Mean field simulation for Monte Carlo integration. Part I : Intro. nonlinear Markov models (+links integro-diff eq.) P. Del Moral

Similar documents
Concentration inequalities for Feynman-Kac particle models. P. Del Moral. INRIA Bordeaux & IMB & CMAP X. Journées MAS 2012, SMAI Clermond-Ferrand

Mean field simulation for Monte Carlo integration. Part II : Feynman-Kac models. P. Del Moral

A Backward Particle Interpretation of Feynman-Kac Formulae

Advanced Monte Carlo integration methods. P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP

A new class of interacting Markov Chain Monte Carlo methods

An introduction to particle simulation of rare events

On the stability and the applications of interacting particle systems

Contraction properties of Feynman-Kac semigroups

Perfect simulation algorithm of a trajectory under a Feynman-Kac law

Stability of Feynman-Kac Semigroup and SMC Parameters' Tuning

On the Concentration Properties of Interacting Particle Processes. Contents

Sequential Monte Carlo Methods for Bayesian Computation

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations

An Brief Overview of Particle Filtering

On a class of stochastic differential equations in a financial network model

Weak solutions of mean-field stochastic differential equations

1. Stochastic Processes and filtrations

A Backward Particle Interpretation of Feynman-Kac Formulae

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Computer Intensive Methods in Mathematical Statistics

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

On Adaptive Resampling Procedures for Sequential Monte Carlo Methods

University of Toronto Department of Statistics

Weakly interacting particle systems on graphs: from dense to sparse

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS

Monte Carlo methods for sampling-based Stochastic Optimization

Introduction to self-similar growth-fragmentations

Lecture 15 Random variables

Discretization of SDEs: Euler Methods and Beyond

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Branching Processes II: Convergence of critical branching to Feller s CSB

Intertwinings for Markov processes

Interacting and Annealing Particle Filters: Mathematics and a Recipe for Applications. Christoph Schnörr Bodo Rosenhahn Hans-Peter Seidel

Importance splitting for rare event simulation

Lecture 4: Introduction to stochastic processes and stochastic calculus

White noise generalization of the Clark-Ocone formula under change of measure

Weak convergence and large deviation theory

Lycka till!

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION

EE4601 Communication Systems

Multilevel Monte Carlo for Lévy Driven SDEs

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Auxiliary Particle Methods

Brownian Motion and Stochastic Calculus

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

1 Random Variable: Topics

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi

Exact Simulation of Diffusions and Jump Diffusions

Derivation of Itô SDE and Relationship to ODE and CTMC Models

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).

Markov Chain BSDEs and risk averse networks

Full text available at: On the Concentration Properties of Interacting Particle Processes

Normal approximation of Poisson functionals in Kolmogorov distance

Econ 508B: Lecture 5

Formulas for probability theory and linear models SF2941

Optimal Control of Partially Observable Piecewise Deterministic Markov Processes

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

Asymptotic statistics using the Functional Delta Method

STAT 430/510 Probability

On Optimal Stopping Problems with Power Function of Lévy Processes

for all f satisfying E[ f(x) ] <.

Uniformly Uniformly-ergodic Markov chains and BSDEs

Introduction. log p θ (y k y 1:k 1 ), k=1

Stochastic Demography, Coalescents, and Effective Population Size

A Short Introduction to Diffusion Processes and Ito Calculus

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER

Bridging the Gap between Center and Tail for Multiscale Processes

Chapter 4. Chapter 4 sections

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

On Reflecting Brownian Motion with Drift

Lecture 1 Measure concentration

Exercises and Answers to Chapter 1

S. Lototsky and B.L. Rozovskii Center for Applied Mathematical Sciences University of Southern California, Los Angeles, CA

Approximation of BSDEs using least-squares regression and Malliavin weights

Lecture 1a: Basic Concepts and Recaps

Lecture 6: Bayesian Inference in SDE Models

Goodness of fit test for ergodic diffusion processes

Kinetic models of Maxwell type. A brief history Part I

An introduction to stochastic particle integration methods: with applications to risk and insurance

Harnack Inequalities and Applications for Stochastic Equations

On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem

Part II Probability and Measure

MAT 135B Midterm 1 Solutions

Discretization of Stochastic Differential Systems With Singular Coefficients Part II

Consistency of the maximum likelihood estimator for general hidden Markov models

Math Spring Practice for the final Exam.

Exercises. T 2T. e ita φ(t)dt.

Lecture Notes 3 Convergence (Chapter 5)

STABILITY AND UNIFORM APPROXIMATION OF NONLINEAR FILTERS USING THE HILBERT METRIC, AND APPLICATION TO PARTICLE FILTERS 1

Malliavin calculus and central limit theorems

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Learning Session on Genealogies of Interacting Particle Systems

Sequential Monte Carlo Samplers for Applications in High Dimensions

On the concentration properties of Interacting particle processes

Transcription:

Mean field simulation for Monte Carlo integration Part I : Intro. nonlinear Markov models (+links integro-diff eq.) P. Del Moral INRIA Bordeaux & Inst. Maths. Bordeaux & CMAP Polytechnique Lectures, INLN CNRS & Nice Sophia Antipolis Univ. 2012 Some hyper-refs Mean field simulation for Monte Carlo integration. Chapman & Hall - Maths & Stats [600p.] (May 2013). Feynman-Kac formulae, Genealogical & Interacting Particle Systems with appl., Springer [573p.] (2004) Concentration Inequalities for Mean Field Particle Models. Ann. Appl. Probab (2011). (+ Rio). Coalescent tree based functional representations for some Feynman-Kac particle models Annals of Applied Probability (2009) (+ Patras, Rubenthaler) Particle approximations of a class of branching distribution flows arising in multi-target tracking. SIAM Control. & Opt. (2011). (joint work with Caron, Doucet, Pace) On the stability & the approximation of branching distribution flows, with applications to nonlinear multiple target filtering. Stoch. Analysis and Appl. (2011). (joint work with Caron, Pace, Vo). More references on the websitehttp://www.math.u-bordeaux1.fr/ delmoral/index.html [+ Links]

Contents Some basic notation Markov chains Linear evolution equations Finite state space models Monte Carlo methods Links with linear integro-diff. eq. Nonlinear Markov models Nonlinear evolution equations Mean field particle models Graphical illustration Links with nonlinear integro-diff. eq. How & Why it works? Some interpretations A stochastic perturbation analysis Uniform concentration inequalities

Some basic notation Markov chains Nonlinear Markov models How & Why it works?

Lebesgue integral on a measurable state space E (µ, f ) =(measure, function) (M(E) B b (E)) µ(f ) = µ(dx) f (x) Delta-Dirac Measure at a E µ = δ a δ a (f ) = f (x) δ a (dx) = f (a) Normalization of a positive measure µ M + (E) (when µ(1) 0) µ(dx) := µ(dx)/µ(1) = probability P(E)

Lebesgue integral on a measurable state space E (µ, f ) =(measure, function) (M(E) B b (E)) µ(f ) = µ(dx) f (x) Delta-Dirac Measure at a E µ = δ a δ a (f ) = f (x) δ a (dx) = f (a) Normalization of a positive measure µ M + (E) (when µ(1) 0) µ(dx) := µ(dx)/µ(1) = probability P(E) Example µ = 10 Law(X ) µ(1) = 10 & µ = Law(X ) & µ(f ) = E(f (X ))

Examples E={1,...,d} Matrix-Vector notation f (1) µ := [µ(1),..., µ(d)] and f :=. f (d) µ(f ) := d µ(i) f (i) i=1

Examples E={1,...,d} Matrix-Vector notation f (1) µ := [µ(1),..., µ(d)] and f :=. f (d) µ(f ) := d µ(i) f (i) i=1 X = 1 i N δ X i spatial Poisson point process with intensity µ N = Poisson random variable with E(N) = µ(1) and X i iid µ E(X (f )) = E(E(X (f ) N)) = E(N µ(f )) = µ(1) µ(f ) = µ(f )

Q(x, dy) integral operator from E into E Two operator actions : f B b (E ) Q(f ) B b (E) and µ M(E) µq M(E ) with Q(f )(x) = [µq](dx ) = Q(x, dx )f (x ) µ(dx)q(x, dx ) ( [µq](f ) := µ[q(f )] ) and the composition (Q 1 Q 2 )(x, dx ) = Q 1 (x, dx )Q 2 (x, dx )

Q(x, dy) integral operator from E into E Two operator actions : f B b (E ) Q(f ) B b (E) and µ M(E) µq M(E ) with Q(f )(x) = [µq](dx ) = Q(x, dx )f (x ) µ(dx)q(x, dx ) ( [µq](f ) := µ[q(f )] ) and the composition (Q 1 Q 2 )(x, dx ) = Q 1 (x, dx )Q 2 (x, dx ) Notation (used in variance descriptions) Q([f Q(f )] 2 )(x) := Q(x, dy) [f (y) Q(f )(x)] 2 Q Markov operator x Q(x, dy) P(E) notation : M or K (Markov transitions-kernel/stochastic matrices)

Finite state spaces E = {1,..., d} and E = {1,..., d }: Action on the right f (1) f :=. f (d ) Matrix operation B b (E ) Q(f ) = Q(f )(1). Q(f )(d) B b (E) Q(f ) = Q(f )(1). Q(f )(d) = Q(1, 1)... Q(1, d )..... Q(d, 1)... Q(d, d ) f (1). f (d )

Finite state spaces E = {1,..., d} and E = {1,..., d }: Action on the left µ = [µ(1),..., µ(d)] M(E) µq = [(µq)(1),..., (µq)(d )] M(E ) Matrix operation µq = [µ(1),..., µ(d)] Q(1, 1)... Q(1, d )..... Q(d, 1)... Q(d, d )

Boltzmann-Gibbs transformation : G 0 s.t. µ(g) > 0 Ψ G (µ)(dx) = 1 µ(g) G(x) µ(dx)

Boltzmann-Gibbs transformation : G 0 s.t. µ(g) > 0 Ψ G (µ)(dx) = 1 µ(g) Bayes rule (with fixed observation y) G(x) µ(dx) µ(dx) = p(x)dx and G(x) = p(y x) Ψ G (µ)(dx) = 1 p(y) p(y x) p(x)dx = p(x y) dx

Boltzmann-Gibbs transformation : G 0 s.t. µ(g) > 0 Ψ G (µ)(dx) = 1 µ(g) Bayes rule (with fixed observation y) G(x) µ(dx) µ(dx) = p(x)dx and G(x) = p(y x) Restriction Ψ G (µ)(dx) = 1 p(y) p(y x) p(x)dx = p(x y) dx µ(dx) = P(X dx) = p(x)dx and G(x) = 1 A (x) Ψ G (µ)(dx) = 1 P(X A) 1 A(x) p(x)dx = P(X dx X A)

Boltzmann-Gibbs transformation : G 0 s.t. µ(g) > 0 Ψ G (µ)(dx) = 1 µ(g) G(x) µ(dx) Markov transport equation Ψ G (µ)(dy) = µ(dx)s µ (x, dy) Ψ G (µ) = µs µ Example 1 : (G 1) accept/reject/recycling/interacting jumps S µ (x, dy) = G(x)δ x (dy) + (1 G(x)) Ψ G (µ)(dy)

Boltzmann-Gibbs transformation : G 0 s.t. µ(g) > 0 Ψ G (µ)(dx) = 1 µ(g) G(x) µ(dx) Markov transport equation Ψ G (µ)(dy) = µ(dx)s µ (x, dy) Ψ G (µ) = µs µ Example 1 : (G 1) accept/reject/recycling/interacting jumps Note : S µ (x, dy) = G(x)δ x (dy) + (1 G(x)) Ψ G (µ)(dy) S 1 N 1 j N δ (X i, dy) X j = G(X i )δ X i (dy) + (1 G(X i )) 1 j N G(X j ) 1 k N G(X k ) δ X j (dy)

Other examples of transport equations Ψ G (µ) = µs µ Example 2 : ɛ µ s.t. ɛ µ G 1 µ a.e. S µ (x, dy) = ɛ µ G(x) δ x (dy) + (1 ɛ µ G(x)) Ψ G (µ)(dy) Example 3 : a s.t. G > a S µ (x, dy) = Example 4 : G a µ(g) δ x(dy) + ( 1 a ) µ(g) Ψ G a (µ)(dy) S µ (x, dy) = α(x) δ x (dy) + (1 α(x)) Ψ [G G(x)]+ (µ)(dy) with the acceptance rate α(x) = µ[g G(x)]/µ(G)

Some basic notation Markov chains Linear evolution equations Finite state space models Monte Carlo methods Links with linear integro-diff. eq. Nonlinear Markov models How & Why it works?

Example: Markov chain X n+1 = F n (X n, W n ) η 0 = Law(X 0 ) and M n (x n 1, dx n ) = P (X n dx n X n 1 = x n 1 ) η 0 M 1 = Law(X 1 ) and M n (f )(x) = E(f (X n ) X n 1 = x)

Example: Markov chain X n+1 = F n (X n, W n ) η 0 = Law(X 0 ) and M n (x n 1, dx n ) = P (X n dx n X n 1 = x n 1 ) η 0 M 1 = Law(X 1 ) and M n (f )(x) = E(f (X n ) X n 1 = x) as well as (M 1... M n )(x 0, dx n ) = P(X n dx n X 0 = x 0 ) η 0 M 1... M n = Law(X n ) E={1,...,d} Matrix-Vector notation

Markov chain models X n+1 = F n (X n, W n ) Linear evolution equations Law(X n ) := η n η n+1 = η n M n+1 E={1,...,d} simple matrix computations M n+1 (1, 1)... M n+1 (1, d) η n+1 = [η n (1),..., η n (d)]..... M n+1 (d, 1)... M n+1 (d, d)

Markov chain models X n+1 = F n (X n, W n ) Linear evolution equations Law(X n ) := η n η n+1 = η n M n+1 E={1,...,d} simple matrix computations M n+1 (1, 1)... M n+1 (1, d) η n+1 = [η n (1),..., η n (d)]..... M n+1 (d, 1)... M n+1 (d, d) Note : Same equations for X n E n = {1,..., d n } with M n+1 R (dn dn+1) +

(Crude) Monte Carlo methods : Markov chain X n+1 = F n (X n, W n ) E n Linear evolution equations Law(X n ) := η n η n+1 = η n M n+1 N iid copies/samples X i n+1 = F n(x i n, Wi n ), 1 i N ηn N := 1 δ N X i n N η n 1 i N ηn N (f ) = 1 f (X i N n) N η n (f ) = E(f (X n )) 1 i N

(Crude) Monte Carlo methods : Markov chain X n E n with transitions M n Path space models P N n := 1 N Historical process 1 i N X n = (X 0,..., X n ) E n+1 = δ (X i 0,...,X i n ) N P n = Law(X 0,..., X n ) 0 p n E p with Markov transition M n Same model as before and η N n := 1 N P n+1 = η n+1 = η n M n+1 1 i N with N iid copies of the historical process δ X i n = P N n N P n = η n X i n = (X i 0,..., X i n)

Diffusion type & discrete generation models (d=1) Markov chain with Gaussian perturbations X n+1 = F n (X n, W n ) = X n + a n (X n ) + b n (X n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2 Diffusion limits when (t n+1 t n ) = 0 & (a n, b n, X n ) = ( a t n, b t n, X t n ) dx t = a t(x t )dt + b t(x t )dw t

Diffusion type & discrete generation models (d=1) Diffusion model dx t = a t(x t )dt + b t(x t )dw t Taylor expansion (Ito s formula) f (X t + dx t ) f (X t ) = f (X t ) dx t + 1 2 2 f (X t ) dx t dx t +...

Diffusion type & discrete generation models (d=1) Diffusion model Taylor expansion (Ito s formula) dx t = a t(x t )dt + b t(x t )dw t f (X t + dx t ) f (X t ) = f (X t ) dx t + 1 2 2 f (X t ) dx t dx t +... with the infinitesimal generator dw t= dt = L t (f )(X t ) dt + f (X t ) b t(x t )dw t }{{} Martingale (E(.) = 0) L t (f ) = a t f + 1 2 (b t) 2 2 f Weak parabolic PDE η t = Law(X t ) d dt η t(f ) = η t (L t (f ))

Two-steps jump-diffusion type model (d=1) X n X n+1/2 = F n (X n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 ( ) ɛ n+1 e Vn(X n+1/2) δ 1 + 1 e Vn(X n+1/2) δ 0 Y n+1 P n (X n+1/2, dy)

Two-steps jump-diffusion type model (d=1) X n X n+1/2 = F n (X n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 ( ) ɛ n+1 e Vn(X n+1/2) δ 1 + 1 e Vn(X n+1/2) δ 0 Y n+1 P n (X n+1/2, dy) Jump-diffusion limits (a n, b n, V n, P n, X n ) = ( ) a t n, b t n, V t n, P t n, X t n between jumps = diffusion as before, and jumps times T n+1 = inf { t T n : t T n V s (X s ) ds e n } e n iid expo(λ = 1)

Two-steps jump-diffusion type model (d=1) with the diffusion part and the jump part M n+1 J n+1 X n X n+1/2 X n+1 M n+1 (f )(x) f (x) + L diff t (f )(x) J n+1 (f )(x) = e V tn (x) f (x) + (1 e V tn (x) ) P t n (f )(x) = f (x) + V t n (x) [ P t n (f )(x) f (x) ] }{{} =L jump t (f )(x)

Two-steps jump-diffusion type model (d=1) with the diffusion part and the jump part M n+1 J n+1 X n X n+1/2 X n+1 M n+1 (f )(x) f (x) + L diff t (f )(x) J n+1 (f )(x) = e V tn (x) f (x) + (1 e V tn (x) ) P t n (f )(x) = f (x) + V t n (x) [ P t n (f )(x) f (x) ] }{{} =L jump t (f )(x)

Two-steps jump-diffusion type model (d=1) with the diffusion part and the jump part M n+1 J n+1 X n X n+1/2 X n+1 M n+1 (f )(x) f (x) + L diff t (f )(x) J n+1 (f )(x) = e V tn (x) f (x) + (1 e V tn (x) ) P t n (f )(x) = f (x) + V t n (x) [ P t n (f )(x) f (x) ] }{{} =L jump t (f )(x) [ MJ = (I + L diff ) (I + L jump ) I + (L diff + L jump ) ]

Two-steps jump-diffusion type model (d=1) with the diffusion part and the jump part M n+1 J n+1 X n X n+1/2 X n+1 M n+1 (f )(x) f (x) + L diff t (f )(x) J n+1 (f )(x) = e V tn (x) f (x) + (1 e V tn (x) ) P t n (f )(x) = f (x) + V t n (x) [ P t n (f )(x) f (x) ] }{{} =L jump t (f )(x) [ MJ = (I + L diff ) (I + L jump ) I + (L diff + L jump ) ] = Weak integro-differential equation with the infinitesimal generator η t = Law(X t ) d dt η t(f ) = η t (L t (f )) L t (f )(x) = a t(x) f (x)+ 1 2 (b t(x)) 2 2 f (x)+v t (x) (f (y) f (x)) P t(x, dy)

Some basic notation Markov chains Nonlinear Markov models Nonlinear evolution equations Mean field particle models Graphical illustration Links with nonlinear integro-diff. eq. How & Why it works?

Mean field models : Markov chain X n+1 = F n (X n, η n, W n ) Nonlinear evolution equations Law(X n ) := η n η n+1 = Φ n+1 (η n ) := η n K n+1,ηn with Markov transitions K n,η indexed by the symplex η S (d 1) R d + K n+1,ηn (x n, dx n+1 ) = P (X n+1 dx n+1 X n = x n )

Mean field models : Markov chain X n+1 = F n (X n, η n, W n ) Nonlinear evolution equations Law(X n ) := η n η n+1 = Φ n+1 (η n ) := η n K n+1,ηn with Markov transitions K n,η indexed by the symplex η S (d 1) R d + K n+1,ηn (x n, dx n+1 ) = P (X n+1 dx n+1 X n = x n ) When E = {1,..., d} simple matrix computations K n+1,ηn (1, 1)... K n+1,ηn (1, d) η n+1 = [η n (1),..., η n (d)]..... K n+1,ηn (d, 1)... K n+1,ηn (d, d)

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) E := { 1, 1} McKean s two velocities of gases : F n (X n, η n, W n ) = X n 1 Wn η n(1) + ( X n ) 1 Wn>η n(1) X n+1 = ɛ n X n with the Bernoulli random variable P(ɛ n = 1) = 1 P(ɛ n = 1) = η n (1) Nonlinear equation on the symplex S (0) = [0, 1] : η n+1 (1) = η n (1) 2 + η n ( 1) 2 = η n (1) 2 + (1 η n (1)) 2

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) E := { 1, 1} McKean s two velocities of gases : F n (X n, η n, W n ) = X n 1 Wn η n(1) + ( X n ) 1 Wn>η n(1) X n+1 = ɛ n X n with the Bernoulli random variable P(ɛ n = 1) = 1 P(ɛ n = 1) = η n (1) Nonlinear equation on the symplex S (0) = [0, 1] : η n+1 (1) = η n (1) 2 + η n ( 1) 2 = η n (1) 2 + (1 η n (1)) 2 n 1/2 if η 0 (1) {0, 1}

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition K n+1,ηn (i, j) = G n (i) M n+1 (i, j)+(1 G n (i)) k η n (k)g n (k) l η n(l)g n (l) M n+1(k, j)

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition K n+1,ηn (i, j) = G n (i) M n+1 (i, j)+(1 G n (i)) k η n (k)g n (k) l η n(l)g n (l) M n+1(k, j) k η n(k) K n+1,ηn (k, j) = k η n(k)g n (k) M n+1 (k, j) + [1 l η n(l)g n (l)] k = k η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j) η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j)

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition K n+1,ηn (i, j) = G n (i) M n+1 (i, j)+(1 G n (i)) k η n (k)g n (k) l η n(l)g n (l) M n+1(k, j) k η n(k) K n+1,ηn (k, j) = k η n(k)g n (k) M n+1 (k, j) + [1 l η n(l)g n (l)] k = k η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j) k η n(k)g n (k)m n+1 (k, j) η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j)

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition K n+1,ηn (i, j) = G n (i) M n+1 (i, j)+(1 G n (i)) k η n (k)g n (k) l η n(l)g n (l) M n+1(k, j) k η n(k) K n+1,ηn (k, j) = k η n(k)g n (k) M n+1 (k, j) + [1 l η n(l)g n (l)] k = k η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j) k η n(k)g n (k)m n+1 (k, j) η n(k)g n(k) l ηn(l)gn(l) M n+1(k, j) G n (1) 0... 0 0 M n+1 (1, 1)... M n+1 (1, d) η n+1 η n.......... 0 0... 0 G n (d) M n+1 (d, 1)... M n+1 (d, d) }{{}}{{} Multiplication Markov transport

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition η n+1 = [Ψ Gn (η n )(1),..., Ψ Gn (η n )(d)] M n+1 (1, 1)... M n+1 (1, d)..... M n+1 (d, 1)... M n+1 (d, d) with the Boltzmann-Gibbs transformation Ψ Gn : η M(E) Ψ Gn (η) M(E) defined by Ψ Gn (η)(i) = 1 η n (G n ) G n(i) η(i) Nonlinear evolution eq. on the (d 1)-symplex η n S (d 1) R d + η n+1 = Φ n+1 (η n ) := Ψ Gn (η n )M n+1

Example 2 : G n (i)( ex. = e Vn(i) ) ]0, 1] & (M n+1 (i, j)) i,j=1,...,d Markov transition η n+1 = [Ψ Gn (η n )(1),..., Ψ Gn (η n )(d)] M n+1 (1, 1)... M n+1 (1, d)..... M n+1 (d, 1)... M n+1 (d, d) with the Boltzmann-Gibbs transformation Ψ Gn : η M(E) Ψ Gn (η) M(E) defined by Ψ Gn (η)(i) = 1 η n (G n ) G n(i) η(i) Nonlinear evolution eq. on the (d 1)-symplex η n S (d 1) R d + η n+1 = Φ n+1 (η n ) := Ψ Gn (η n )M n+1 Feynman-Kac models

Mean field models : Markov X n+1 = F n (X n, η n, W n ) E = R d Nonlinear evolution equations Law(X n ) := η n η n+1 = Φ n+1 (η n ) := η n K n+1,ηn Mean field particle-simulation models: N almost iid copies/samples X i n+1 = F n(x i n, ηn n, Wi n ), 1 i N with the occupation measure of the system at time n : η N n := 1 N 1 i N δ X i n N η n

Mean field models : Markov X n+1 = F n (X n, η n, W n ) E = R d Nonlinear evolution equations Law(X n ) := η n η n+1 = Φ n+1 (η n ) := η n K n+1,ηn Mean field particle-simulation models: N almost iid copies/samples X i n+1 = F n(x i n, ηn n, Wi n ), 1 i N with the occupation measure of the system at time n : η N n := 1 N 1 i N δ X i n N η n [by induction on n] ηn+1 N = 1 δ N X i n+1 with Xn+1 i F n (Xn, i η n, Wn) i ηn+1 N N η n+1 1 i N

An abstract mathematical model Nonlinear evolution equation η n+1 = Φ n+1 (η n ) := η n K n+1,ηn Nonlinear Markov model interpretation (not unique) K n+1,ηn (x n, dx n+1 ) = P (X n+1 dx n+1 X n = x n ) with Law(X n ) := η n Mean field particle-simulation model N X i n X i n+1 = r.v. with law K n+1,η N n (X i n, dx) where η N n = 1 N i=1 δ X i n

Local sampling errors Nonlinear evolution equation Mean field particle model η n+1 = Φ n+1 (η n ) V N n+1 := N ( η N n+1 Φ n+1 ( η N n )) η N n+1 = Φ n+1 ( η N n ) + 1 N V N n+1 Note f : osc(f ) := sup x,y (f (x) f (y)) 1 E ( ηn+1(f N ) ηn N ) N E ([ ηn+1 N ( )] Φ n+1 η N n (f ) 2 ηn N ) = ηn N ( ) K n+1,η N n (f ) = Φ n+1 η N n (f ) [ = ηn N [ K n+1,η N n f Kn+1,η N n (f ) ] ] 2 1

McKean measures : Markov chain X n E n with transitions K n,ηn 1 P N n := 1 N 1 i N with the McKean measures P n (d(x 0,..., x n )) = η 0 (dx 0 ) δ (X i 0,...,X i n ) N P n = Law(X 0,..., X n ) 1 p n K p,ηp 1 (x p 1, dx p )

McKean measures : Markov chain X n E n with transitions K n,ηn 1 P N n := 1 N 1 i N with the McKean measures P n (d(x 0,..., x n )) = η 0 (dx 0 ) δ (X i 0,...,X i n ) N P n = Law(X 0,..., X n ) 1 p n K p,ηp 1 (x p 1, dx p ) Historical process X n = (X 0,..., X n ) E n+1 = 0 p n E p with Markov transition K n,ηn 1

McKean measures : Markov chain X n E n with transitions K n,ηn 1 P N n := 1 N 1 i N with the McKean measures P n (d(x 0,..., x n )) = η 0 (dx 0 ) δ (X i 0,...,X i n ) N P n = Law(X 0,..., X n ) 1 p n K p,ηp 1 (x p 1, dx p ) Historical process X n = (X 0,..., X n ) E n+1 = 0 p n E p with Markov transition K n,ηn 1 Same mathematical model as before and η N n := 1 N P n+1 = η n+1 = η n K n+1,ηn 1 i N δ X i n = P N n N P n = η n with almost N iid copies of the historical process X i n = (X i 0,..., X i n).

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) E = R d=1 McKean-Vlasov diffusion type & discrete generation models X n+1 = F n (X n, η n, W n ) = a n (X n, η n ) + b n (X n, η n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) E = R d=1 McKean-Vlasov diffusion type & discrete generation models X n+1 = F n (X n, η n, W n ) = a n (X n, η n ) + b n (X n, η n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2 Two-steps jump type model X n X n+1/2 = F n (X n, η n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 with ( ) ɛ n+1 e Vη (Xn+1/2) n+1/2 δ 1 + 1 e Vη (X n+1/2 n+1/2) δ 0 Y n+1 P ηn+1/2 (X n+1/2, dy)

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) R d=1 McKean-Vlasov diffusion type & discrete generation models a n (X n, η n ) = η n (dy) a n(x n, y) & b n (X n, η n ) = η n (dy) b n(x n, y)

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) R d=1 McKean-Vlasov diffusion type & discrete generation models a n (X n, η n ) = η n (dy) a n(x n, y) & b n (X n, η n ) = η n (dy) b n(x n, y) X i n+1 = F n (X i n, η N n, W i n) = 1 N N a n(x n, i Xn) j + 1 N j=1 N b n(x n, i Xn) j Wn i j=1

Example 1 : Markov chain X n+1 = F n (X n, η n, W n ) R d=1 McKean-Vlasov diffusion type & discrete generation models a n (X n, η n ) = η n (dy) a n(x n, y) & b n (X n, η n ) = η n (dy) b n(x n, y) X i n+1 = F n (X i n, η N n, W i n) = 1 N N a n(x n, i Xn) j + 1 N j=1 N b n(x n, i Xn) j Wn i In the same vein : Two-steps mean field jump type model Xn i Xn+1/2 i = F n(xn, i ηn N, Wn) i Xn+1 i = ɛ i n+1 Xn+1/2 i +( 1 ɛ i n+1) Y i n+1 j=1 with ( ɛ i n+1 e V η n+1/2 N (X i n+1/2 ) δ 1 + 1 e V η n+1/2 N (X i )) n+1/2 δ 0 Yn+1 i P η N (X i n+1/2 n+1/2, dy)

Example 2 : Back to the multiplication-transport model η n+1 = Ψ Gn (η n )M n+1 Nonlinear transport equation Ψ Gn (µ) = µs n,µ with S n,µ (x,.) = G n (x)δ x + (1 G n (x)) Ψ Gn (µ)

Example 2 : Back to the multiplication-transport model η n+1 = Ψ Gn (η n )M n+1 Nonlinear transport equation Ψ Gn (µ) = µs n,µ with S n,µ (x,.) = G n (x)δ x + (1 G n (x)) Ψ Gn (µ) Nonlinear Markov chain model η n+1 = Φ n+1 (η n ) := η n K n+1,ηn with K n+1,ηn := S n,ηn M n+1

Example 2 : Back to the multiplication-transport model η n+1 = Ψ Gn (η n )M n+1 Nonlinear transport equation Ψ Gn (µ) = µs n,µ with S n,µ (x,.) = G n (x)δ x + (1 G n (x)) Ψ Gn (µ) Nonlinear Markov chain model η n+1 = Φ n+1 (η n ) := η n K n+1,ηn with K n+1,ηn := S n,ηn M n+1 Mean field particle model = Genetic type interacting jump model X i n S n,η N n X n i M n+1 }{{} K n+1,η N n i Xn+1 with S n,η N n (X i n, dy) = G n (X i n)δ X i n (dy)+(1 G(X i n)) 1 j N G n (X j n) 1 k N G n(x k n ) δ X j n (dy)

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Graphical illustration : η n η N n := 1 N 1 i N δ X i n

Some questions Law(Xn i acceptance) = η n? Convergence of the occupation measures of the genealogical trees? Use of the occupation measures of the complete ancestral tree? Convergence of the product of success proportions? Feynman-Kac integration models on path spaces

Some questions Law(Xn i acceptance) = η n? Convergence of the occupation measures of the genealogical trees? Use of the occupation measures of the complete ancestral tree? Convergence of the product of success proportions? Feynman-Kac integration models on path spaces Part II & Part III (Application areas) & Part V (Theoretical results)

Some questions Law(Xn i acceptance) = η n? Convergence of the occupation measures of the genealogical trees? Use of the occupation measures of the complete ancestral tree? Convergence of the product of success proportions? Feynman-Kac integration models on path spaces Part II & Part III (Application areas) & Part V (Theoretical results) Nonlinear evolution equations on M(E n ) γ n+1 = γ n Q n+1,γn

Some questions Law(Xn i acceptance) = η n? Convergence of the occupation measures of the genealogical trees? Use of the occupation measures of the complete ancestral tree? Convergence of the product of success proportions? Feynman-Kac integration models on path spaces Part II & Part III (Application areas) & Part V (Theoretical results) Nonlinear evolution equations on M(E n ) γ n+1 = γ n Q n+1,γn Part IV (Multiple object filtering)

Some questions Law(Xn i acceptance) = η n? Convergence of the occupation measures of the genealogical trees? Use of the occupation measures of the complete ancestral tree? Convergence of the product of success proportions? Feynman-Kac integration models on path spaces Part II & Part III (Application areas) & Part V (Theoretical results) Nonlinear evolution equations on M(E n ) γ n+1 = γ n Q n+1,γn Part IV (Multiple object filtering) Link with nonlinear integro-differential equations?

Nonlinear diffusion type models (d=1) X n+1 = F n (X n, η n, W n ) = X n + a n (X n, η n ) + b n (X n, η n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2

Nonlinear diffusion type models (d=1) X n+1 = F n (X n, η n, W n ) = X n + a n (X n, η n ) + b n (X n, η n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2 Diffusion limits when (t n+1 t n ) = 0 & (a n, b n, X n, η n ) = ( a t n, b t n, X t n, η t n ) dx t = a t(x t, η t)dt + b t(x t, η t)dw t

Nonlinear diffusion type models (d=1) X n+1 = F n (X n, η n, W n ) = X n + a n (X n, η n ) + b n (X n, η n ) W n with the Gaussian random variable P(W n dw) = 1 exp ( w 2 ) 2π 2 Diffusion limits when (t n+1 t n ) = 0 & (a n, b n, X n, η n ) = ( a t n, b t n, X t n, η t n ) dx t = a t(x t, η t)dt + b t(x t, η t)dw t Weak parabolic nonlinear PDE η t = Law(X t ) d dt η t(f ) = η t(l t,η t (f )) with the infinitesimal generator L t,η t (f )(x) = a t(x, η t) f (x) + 1 2 (b t(x, η t)) 2 2 f (x)

Two-steps jump-diffusion type model (d=1) X n X n+1/2 = F n (X n, η n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 ( ) ɛ n+1 e Vη (X n+1/2 n+1/2) δ 1 + 1 e Vη (X n+1/2 n+1/2) δ 0 Y n+1 P ηn+1/2 (X n+1/2, dy)

Two-steps jump-diffusion type model (d=1) X n X n+1/2 = F n (X n, η n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 ( ) ɛ n+1 e Vη (X n+1/2 n+1/2) δ 1 + 1 e Vη (X n+1/2 n+1/2) δ 0 Y n+1 P ηn+1/2 (X n+1/2, dy) Jump-diffusion limits ) (a n, b n, V n, P n, X n, η n ) = (a tn, b tn, V ηtn, P ηtn, X tn, η tn between jumps = nonlinear diffusion as before, and jumps times { t } T n+1 = inf t T n : V s (X s, η s) ds e n e n iid expo(λ = 1) T n

Two-steps jump-diffusion type model (d=1) X n X n+1/2 = F n (X n, η n, W n ) X n+1 = ɛ n+1 X n+1/2 + (1 ɛ n+1 ) Y n+1 ( ) ɛ n+1 e Vη (X n+1/2 n+1/2) δ 1 + 1 e Vη (X n+1/2 n+1/2) δ 0 Y n+1 P ηn+1/2 (X n+1/2, dy) Jump-diffusion limits ) (a n, b n, V n, P n, X n, η n ) = (a tn, b tn, V ηtn, P ηtn, X tn, η tn between jumps = nonlinear diffusion as before, and jumps times { t } T n+1 = inf t T n : V s (X s, η s) ds e n e n iid expo(λ = 1) T n Weak integro-differential equation d dt η t(f ) = η t(l t,η t (f )) L t,η t (f )(x) = a t(x, η t) f (x) + 1 2 (b t(x, η t)) 2 2 f (x) + V η (x) (f (y) f (x)) P t η (x, dy) t

An abstract mathematical model Nonlinear evolution equation P(E) d dt η t(f ) = η t (L t,ηt (f )) Mean field particle-simulation model= Markov E N with generator L t (ϕ)(x 1,..., x N ) = L (i) (ϕ)(x 1,..., x i,..., x N ) t, 1 N 1 j N δ x j 1 i N

Local sampling errors Nonlinear evolution equation dη t (f ) = η t (L t,ηt (f )) dt

Local sampling errors Nonlinear evolution equation dη t (f ) = η t (L t,ηt (f )) dt Mean field particle model (useless in pratice) dη N t (f ) = η N t (L t,η N t (f )) dt + 1 N dm N t (f ) with a sequence of martingales M N t with predictable increasing process M N (f ) t = t 0 ηs N Γ Ls,η (f, f ) ds N s where Γ L stands for the carré du champ operator Γ L (f, f ) := L([f f (x)] 2 )(x) = L(f 2 )(x) 2f (x)l(f )(x)

Some basic notation Markov chains Nonlinear Markov models How & Why it works? Some interpretations A stochastic perturbation analysis Uniform concentration inequalities

How & Why it works? (Biology) Individual based models (IBM). (Physics) Mean field approximation schemes. (Computer Sciences) Stochastic adaptive grid approximation. (Stats) Universal acceptance-rejection-recycling sampling schemes. (Probab) Stochastic linearization/perturbation technique. η n = Φ n (η n 1 ) η N n = Φ n ( η N n 1 ) + 1 N V N n Theorem: (V N n ) n N (V n ) n independent centered Gaussian fields. Φ p,n (η p ) = η n stable sg No propagation of local errors = Uniform control w.r.t. the time horizon New concentration inequalities for (general) interacting processes

Stochastic perturbation analysis = Taylor expansion Gâteaux derivative of Φ n : P(E n 1 ) P(E n ) ɛ Φ n(η + ɛµ)(f ) ɛ=0 = µ[d η Φ n (f )] with the first order linear/integral operator d η Φ n : f B b (E n ) d η Φ n (f ) B b (E n 1 )

Stochastic perturbation analysis = Taylor expansion Gâteaux derivative of Φ n : P(E n 1 ) P(E n ) ɛ Φ n(η + ɛµ)(f ) ɛ=0 = µ[d η Φ n (f )] with the first order linear/integral operator d η Φ n : f B b (E n ) d η Φ n (f ) B b (E n 1 ) Φ n (µ) Φ n (η) (µ η) d η Φ n

Stochastic perturbation analysis = Taylor expansion Gâteaux derivative of Φ n : P(E n 1 ) P(E n ) ɛ Φ n(η + ɛµ)(f ) ɛ=0 = µ[d η Φ n (f )] with the first order linear/integral operator d η Φ n : f B b (E n ) d η Φ n (f ) B b (E n 1 ) Φ n (µ) Φ n (η) (µ η) d η Φ n Differential rule d η (Φ n+1 Φ n ) = d η Φ n d Φn(η)Φ n+1 Semigroup derivatives Φ p,n (µ) Φ p,n (η) (µ η) d η Φ p,n

Key idea = First order expansions Key telescoping decomposition η N n η n = n p=0 [ Φp,n (η N p ) Φ p,n ( Φp (η N p 1) )] First order expansion N [ Φp,n (η N p ) Φ p,n ( Φp (η N p 1 ))] = N [Φ p,n ( Φ p (η N p 1 ) + 1 N V N p ) ( Np 1 Φ p,n Φp (η ))] V N p D p,n + 1 N R N p,n with a predictable D p,n first order operator }{{} fluctuation term 2nd-order measure R N p,n }{{} bias-term

Stochastic perturbation model Stochastic perturbation model W η,n n := N [ η N n η n ] = 0 p n V N p D p,n + 1 N R N n Under some mixing condition on the limiting semigroups Φ p,n osc (D p,n (f )) Cte e (n p)α and E ( R N n (f ) m) Cte 2 m (2m)!/m! Uniform concentration estimates w.r.t. the time parameter

Uniform concentration inequalities Constants (c 1, c 2 ) related to (bias,variance), c finite constant Test functions/observables f n 1, (x 0, n 0, N 1). When E n = R d : F n (y) := η n ( 1(,y] ) and F N n (y) := η N n ( 1(,y] ) with y R d The probability of any of the following events is greater than 1 e x. η N n η n (fn ) c 1 N ( 1 + x + x ) + c 2 N x sup [ η N ] p η p (fp ) c x log (n + e)/n 0 p n F N n F n c d (x + 1)/N

Coalescent tree based expansions Weak propagation of chaos Taylor s type expansions: P (N,q) n = Law of the first q ancestral lines (q N) = Qn q + ( ) 1 1 N l d l P (q) n + O N m+1 with signed measures d l P (q) n 1 l m expressed in terms of coalescent trees.

Coalescent tree based expansions Weak propagation of chaos Taylor s type expansions: P (N,q) n = Law of the first q ancestral lines (q N) = Qn q + ( ) 1 1 N l d l P (q) n + O N m+1 with signed measures d l P (q) n 1 l m expressed in terms of coalescent trees. Romberg-Richardson interpolation: For any order l 1 1 m l ( 1) l m m! m l (l m)! P(mN,q) n = Q q n + O(1/N l )