Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models

Similar documents
The Invariant Distribution of Wealth and Employment Status in a Small Open Economy with Precautionary Savings

The Invariant Distribution of Wealth and Employment Status in a Small Open Economy with Precautionary Savings

Optimal Saving under Poisson Uncertainty: Corrigendum

Generalized Resolvents and Harris Recurrence of Markov Processes

Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms

Stochastic relations of random variables and processes

A mathematical framework for Exact Milestoning

Walsh Diffusions. Andrey Sarantsev. March 27, University of California, Santa Barbara. Andrey Sarantsev University of Washington, Seattle 1 / 1

Describing the Dynamics of Distributions in Search and Matching Models by Fokker-Planck Equations 1

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

Convergence of Feller Processes

Time-varying Consumption Tax, Productive Government Spending, and Aggregate Instability.

A regeneration proof of the central limit theorem for uniformly ergodic Markov chains

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Non-homogeneous random walks on a semi-infinite strip

Invariance Principle for Variable Speed Random Walks on Trees

Zero-Sum Average Semi-Markov Games: Fixed Point Solutions of the Shapley Equation

Practical conditions on Markov chains for weak convergence of tail empirical processes

Stochastic Processes

On the Central Limit Theorem for an ergodic Markov chain

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

STABILITY OF THE OVERDAMPED LANGEVIN EQUATION IN DOUBLE-WELL POTENTIAL

Introduction to Random Diffusions

Lecture 7: Stochastic Dynamic Programing and Markov Processes

Advanced Computer Networks Lecture 2. Markov Processes

Convergence exponentielle uniforme vers la distribution quasi-stationnaire en dynamique des populations

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Convergence to equilibrium of Markov processes (eventually piecewise deterministic)

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Markov processes and queueing networks

On countably skewed Brownian motion

Macroeconomics Theory II

{σ x >t}p x. (σ x >t)=e at.

Excursion Reflected Brownian Motion and a Loewner Equation

Positive Harris Recurrence and Diffusion Scale Analysis of a Push Pull Queueing Network. Haifa Statistics Seminar May 5, 2008

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Stochastic Process (ENPC) Monday, 22nd of January 2018 (2h30)

Consumption. Consider a consumer with utility. v(c τ )e ρ(τ t) dτ.

Session 1: Probability and Markov chains

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

LONG TIME BEHAVIOUR OF PERIODIC STOCHASTIC FLOWS.

The employment effect of reforming a public employment agency

Poisson Jumps in Credit Risk Modeling: a Partial Integro-differential Equation Formulation

SUBGEOMETRIC RATES OF CONVERGENCE FOR A CLASS OF CONTINUOUS-TIME MARKOV PROCESS

Toulouse School of Economics, M2 Macroeconomics 1 Professor Franck Portier. Exam Solution

Research Article An Optimal Stopping Problem for Jump Diffusion Logistic Population Model

Stability of Cyclic Threshold and Threshold-like Autoregressive Time Series Models

A review of Continuous Time MC STA 624, Spring 2015

Optimal Control of Partially Observable Piecewise Deterministic Markov Processes

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Consistency of the maximum likelihood estimator for general hidden Markov models

Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk

Faithful couplings of Markov chains: now equals forever

Tail inequalities for additive functionals and empirical processes of. Markov chains

University of Toronto Department of Statistics

Data analysis and stochastic modeling

A QUASI-ERGODIC THEOREM FOR EVANESCENT PROCESSES. 1. Introduction

16. Working with the Langevin and Fokker-Planck equations

Continuous Time Approximations to GARCH(1, 1)-Family Models and Their Limiting Properties

SMSTC (2007/08) Probability.

2012 NCTS Workshop on Dynamical Systems

Problemas abiertos en dinámica de operadores

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

A Diffusion Approximation for Stationary Distribution of Many-Server Queueing System In Halfin-Whitt Regime

Some Results on the Ergodicity of Adaptive MCMC Algorithms

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes

Path Decomposition of Markov Processes. Götz Kersting. University of Frankfurt/Main

STEADY STATE AND SCALING LIMIT FOR A TRAFFIC CONGESTION MODEL

STABILITY OF THE INVENTORY-BACKORDER PROCESS IN THE (R, S) INVENTORY/PRODUCTION MODEL

Lecture 4: Introduction to stochastic processes and stochastic calculus

1 Continuous-time chains, finite state space

Some classical results on stationary distributions of continuous time Markov processes

Path integrals for classical Markov processes

Mixing time for a random walk on a ring

Local vs. Nonlocal Diffusions A Tale of Two Laplacians

THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE. By Mogens Bladt National University of Mexico

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

A relative entropy characterization of the growth rate of reward in risk-sensitive control

Lecture 12. F o s, (1.1) F t := s>t

The Mabinogion Sheep Problem

Convergence exponentielle uniforme vers la distribution quasi-stationnaire de processus de Markov absorbés

Stochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search

Department of Econometrics and Business Statistics

arxiv: v1 [q-fin.mf] 8 Jan 2017

Densities for the Navier Stokes equations with noise

Stochastic Volatility and Correction to the Heat Equation

Stochastic Processes (Week 6)

Gradient Estimation for Attractor Networks

Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer

Approximating a Diffusion by a Finite-State Hidden Markov Model

Memory and hypoellipticity in neuronal models

INVARIANT PROBABILITIES FOR

Markov Chain BSDEs and risk averse networks

Uniformly Uniformly-ergodic Markov chains and BSDEs

On Finding Optimal Policies for Markovian Decision Processes Using Simulation

Convergence of price and sensitivities in Carr s randomization approximation globally and near barrier

Regular Variation and Extreme Events for Stochastic Processes

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Transcription:

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Christian Bayer and Klaus Wälde Weierstrass Institute for Applied Analysis and Stochastics and University of Mainz SAET 2012 Conference Brisbane July 1, 2012

Outline 1 Introduction Search and matching models Our model 2 Abstract stability theory in continuous time Setting and methodology Existence of an invariant probability measure Uniqueness of invariant measures Stability 3 Application to the model Existence and stability A sufficient condition for recurrence

Outline 1 Introduction Search and matching models Our model 2 Abstract stability theory in continuous time Setting and methodology Existence of an invariant probability measure Uniqueness of invariant measures Stability 3 Application to the model Existence and stability A sufficient condition for recurrence

Search and matching models Search and matching models Extremely popular these days (and for many years now) Inspired by work of Diamond, Mortensen and Pissarides Apart from some recent examples (e.g. our companion paper and the references therein), these models do not include a saving mechanism Our setup Pissarides textbook model (without Nash-bargaining) extended for a consumption-saving mechanism Process of matching and separation is augmented to allow for self-insurance of workers We describe distributional prediction for labour market status and wealth using Fokker-Planck equations in a companion paper

Search and matching models Search and matching models Extremely popular these days (and for many years now) Inspired by work of Diamond, Mortensen and Pissarides Apart from some recent examples (e.g. our companion paper and the references therein), these models do not include a saving mechanism Our setup Pissarides textbook model (without Nash-bargaining) extended for a consumption-saving mechanism Process of matching and separation is augmented to allow for self-insurance of workers We describe distributional prediction for labour market status and wealth using Fokker-Planck equations in a companion paper

The model Matching on the labour market Transitions between states z {w, b} with (state-dependent) matching rate µ and separation rate s Wage w and benefits b are exogenous (in this stability paper, not in companion paper) Representation for maximisation problem as a stochastic differential equation with two Poisson processes dz (t) = [ dq µ dq s ], w b Corresponds to cont. time Markov chain Budget constraint of an individual da (t) = { ra (t) + z (t) c (t) } dt Interest rate on wealth r, consumption c (t)

The model Matching on the labour market Transitions between states z {w, b} with (state-dependent) matching rate µ and separation rate s Wage w and benefits b are exogenous (in this stability paper, not in companion paper) Representation for maximisation problem as a stochastic differential equation with two Poisson processes dz (t) = [ dq µ dq s ], w b Corresponds to cont. time Markov chain Budget constraint of an individual da (t) = { ra (t) + z (t) c (t) } dt Interest rate on wealth r, consumption c (t)

Optimality Utility functions Intertemporal U (t) = E t e ρ[τ t] u (c (τ)) dτ CRRA instantaneous utility function Optimality condition t u (c (τ)) = c (τ)1 σ 1, σ > 0 1 σ Generalized Keynes-Ramsey rule Represented for this paper by policy function c (a (t), z (t))

Optimality Utility functions Intertemporal U (t) = E t e ρ[τ t] u (c (τ)) dτ CRRA instantaneous utility function Optimality condition t u (c (τ)) = c (τ)1 σ 1, σ > 0 1 σ Generalized Keynes-Ramsey rule Represented for this paper by policy function c (a (t), z (t))

Dynamics System to be understood Frictional labour market equation dz (t) = [ dq µ dq s ], w b Optimal evolution of wealth Different regimes da (t) = { ra (t) + z (t) c (a (t), z (t)) } dt Low interest rate r ρ: bounded state space [ b/r, aw] for wealth High interest rate r ρ + µ: a t increasing to Intermediate case: a t increasing to when larger than a threshold value

Dynamics System to be understood Frictional labour market equation dz (t) = [ dq µ dq s ], w b Optimal evolution of wealth Different regimes da (t) = { ra (t) + z (t) c (a (t), z (t)) } dt Low interest rate r ρ: bounded state space [ b/r, aw] for wealth High interest rate r ρ + µ: a t increasing to Intermediate case: a t increasing to when larger than a threshold value

Outline 1 Introduction Search and matching models Our model 2 Abstract stability theory in continuous time Setting and methodology Existence of an invariant probability measure Uniqueness of invariant measures Stability 3 Application to the model Existence and stability A sufficient condition for recurrence

Setting State space (X, B(X)) locally compact separable metric space (X t ) t [0, [ right-continuous, time-homogeneous strong Markov process Transition kernel P t (x, A) P(X t A X 0 = x) Semi-group P t f(x) E [f(x t ) X 0 = x] = X f(y)pt (x, dy). Example For the wealth-employment process (A t, z t ) in the low-interest-regime, the state space is chosen to be X = [ b/r, a w] {w, b}, a compact, separable metric space.

Setting State space (X, B(X)) locally compact separable metric space (X t ) t [0, [ right-continuous, time-homogeneous strong Markov process Transition kernel P t (x, A) P(X t A X 0 = x) Semi-group P t f(x) E [f(x t ) X 0 = x] = X f(y)pt (x, dy). Example For the wealth-employment process (A t, z t ) in the low-interest-regime, the state space is chosen to be X = [ b/r, a w] {w, b}, a compact, separable metric space.

Methodology There are (at least) two very different approaches: Functional analysis: use the classical theory of strongly continuous semi-groups of linear operators on Banach spaces Probability: analogy to discrete-time Markov chains, i.e., study the recurrence structure We are going to follow the probabilistic road, the semi-group (P t ) t [0, [ and its infinitesimal generator will not be used. Based on a long history of results, ultimate treatment by Meyn and Tweedie and their co-authors in 90 s.

Methodology There are (at least) two very different approaches: Functional analysis: use the classical theory of strongly continuous semi-groups of linear operators on Banach spaces Probability: analogy to discrete-time Markov chains, i.e., study the recurrence structure We are going to follow the probabilistic road, the semi-group (P t ) t [0, [ and its infinitesimal generator will not be used. Based on a long history of results, ultimate treatment by Meyn and Tweedie and their co-authors in 90 s.

Outline Goal Stability is a rather vague concept. Here: ergodicity in the sense that for any initial state x, P t (x, ) t π for some unique probability distribution π. No time-averaging necessary. Definition A measure µ on X is called invariant, iff A B (X), t 0 : Pµ(A) t P t (x, A)µ(dx) = µ(a), i.e., the process X t with law(x 0 ) = µ is stationary. X

Outline Goal Stability is a rather vague concept. Here: ergodicity in the sense that for any initial state x, P t (x, ) t π for some unique probability distribution π. No time-averaging necessary. Definition A measure µ on X is called invariant, iff A B (X), t 0 : Pµ(A) t P t (x, A)µ(dx) = µ(a), i.e., the process X t with law(x 0 ) = µ is stationary. X

Outline 2 Outline of the proof of stability: (1) Existence of an invariant distribution (2) Uniqueness of invariant measures (3) Convergence Remark (1) and (2) are very different in nature and rely on qualitatively different assumptions.

Outline 2 Outline of the proof of stability: (1) Existence of an invariant distribution (2) Uniqueness of invariant measures (3) Convergence Remark (1) and (2) are very different in nature and rely on qualitatively different assumptions.

Outline 2 Outline of the proof of stability: (1) Existence of an invariant distribution (2) Uniqueness of invariant measures (3) Convergence Remark (1) and (2) are very different in nature and rely on qualitatively different assumptions.

Outline 2 Outline of the proof of stability: (1) Existence of an invariant distribution (2) Uniqueness of invariant measures (3) Convergence Remark (1) and (2) are very different in nature and rely on qualitatively different assumptions.

Existence of an invariant probability measure Existence of an invariant probability distribution depends on a growth condition: no mass is allowed to escape to infinity. X t is bounded in probability on average, iff x X, ɛ > 0 there is a compact set C X s.t. lim inf t 1 t t 0 P s (x, C)ds 1 ɛ. Compactness of measures 1 t t 0 Ps (x, C)ds X t has the weak Feller property, iff for any bounded cont. f : X R and t > 0, x X f(y)pt (x, dy) is continuous. Theorem (Beneš 68) If the process X t is bounded in probability on average and has the weak Feller property, then there is an invariant probability measure.

Existence of an invariant probability measure Existence of an invariant probability distribution depends on a growth condition: no mass is allowed to escape to infinity. X t is bounded in probability on average, iff x X, ɛ > 0 there is a compact set C X s.t. lim inf t 1 t t 0 P s (x, C)ds 1 ɛ. Compactness of measures 1 t t 0 Ps (x, C)ds X t has the weak Feller property, iff for any bounded cont. f : X R and t > 0, x X f(y)pt (x, dy) is continuous. Theorem (Beneš 68) If the process X t is bounded in probability on average and has the weak Feller property, then there is an invariant probability measure.

Existence of an invariant probability measure Existence of an invariant probability distribution depends on a growth condition: no mass is allowed to escape to infinity. X t is bounded in probability on average, iff x X, ɛ > 0 there is a compact set C X s.t. lim inf t 1 t t 0 P s (x, C)ds 1 ɛ. Compactness of measures 1 t t 0 Ps (x, C)ds X t has the weak Feller property, iff for any bounded cont. f : X R and t > 0, x X f(y)pt (x, dy) is continuous. Theorem (Beneš 68) If the process X t is bounded in probability on average and has the weak Feller property, then there is an invariant probability measure.

[[SO EXISTENCE FOLLOWS FROM COMPACTNESS, AS THE ABOVE FAMILY OF MEASURES HAS SOME CONVERGENT SUBSEQUENCES. EACH LIMIT ALONG A CONVERGENT SUBSEQUENCE IS AN INVARIANT PROBABILITY MEASURE, BUT THERE CAN BE MORE THAN ONE. CONCEPTUALLY, WHY DO WE EVEN NEED WEAK FELLER? UNDERSTAND THE FOLLOWING PROOF!]]

Uniqueness of the invariant measure X t is recurrent, iff there is a (non-trivial) σ-finite measure µ such that A B(X), µ(a) > 0 x X : P (τ A < X 0 = x) = 1, where τ A inf{t 0 X t A}. Theorem (Azéma, Duflo, Revuz 69) If the process X t is recurrent, then there is a unique σ-finite invariant measure (up to constant multiples). Example Let W t be 1-dimensional Brownian motion. By recurrence, there is a unique invariant measure, whose density satisfies f = 0, implying that f = 1.

Uniqueness of the invariant measure X t is recurrent, iff there is a (non-trivial) σ-finite measure µ such that A B(X), µ(a) > 0 x X : P (τ A < X 0 = x) = 1, where τ A inf{t 0 X t A}. Theorem (Azéma, Duflo, Revuz 69) If the process X t is recurrent, then there is a unique σ-finite invariant measure (up to constant multiples). Example Let W t be 1-dimensional Brownian motion. By recurrence, there is a unique invariant measure, whose density satisfies f = 0, implying that f = 1.

Uniqueness of the invariant measure X t is recurrent, iff there is a (non-trivial) σ-finite measure µ such that A B(X), µ(a) > 0 x X : P (τ A < X 0 = x) = 1, where τ A inf{t 0 X t A}. Theorem (Azéma, Duflo, Revuz 69) If the process X t is recurrent, then there is a unique σ-finite invariant measure (up to constant multiples). Example Let W t be 1-dimensional Brownian motion. By recurrence, there is a unique invariant measure, whose density satisfies f = 0, implying that f = 1.

[[UNDERSTAND WHY STRONG FELLER IMPLIES RECURRENCE/UNIQUENESS]] [[UNDERSTAND WHY A-PERIODICITY NOT NECESSARY]]

Stability Stability for us means convergence P t (x, ) π for any x in total variation, i.e., d TV (P t (x, ), π) sup { P t (x, A) π(a) A B(X) } t 0. Stability holds for a Harris recurrent Markov process X t iff for some > 0, the skeleton chain (X n ) n N is irreducible. Remark Techniques based on Lyapunov functions even allow to specify the speed of convergence. But no general way to construct good Lyapunov functions.

Stability Stability for us means convergence P t (x, ) π for any x in total variation, i.e., d TV (P t (x, ), π) sup { P t (x, A) π(a) A B(X) } t 0. Stability holds for a Harris recurrent Markov process X t iff for some > 0, the skeleton chain (X n ) n N is irreducible. Remark Techniques based on Lyapunov functions even allow to specify the speed of convergence. But no general way to construct good Lyapunov functions.

Outline 1 Introduction Search and matching models Our model 2 Abstract stability theory in continuous time Setting and methodology Existence of an invariant probability measure Uniqueness of invariant measures Stability 3 Application to the model Existence and stability A sufficient condition for recurrence

Existence and stability In the regimes of high and intermediate interest rates, wealth can converge to. We concentrate on the low-interest-regime, where wealth is concentrated in a compact interval [ b/r, a w]. By continuity of solutions of ODEs in the initial value, (a t, z t ) is a continuous function of (a 0, z 0 ), implying the weak Feller property. Existence of invariant probability measures. If z 0 = b or z 0 = w and no jump, then a t b/r or a t a w, respectively, implying the existence of an irreducible skeleton. But how to prove recurrence?

Existence and stability In the regimes of high and intermediate interest rates, wealth can converge to. We concentrate on the low-interest-regime, where wealth is concentrated in a compact interval [ b/r, a w]. By continuity of solutions of ODEs in the initial value, (a t, z t ) is a continuous function of (a 0, z 0 ), implying the weak Feller property. Existence of invariant probability measures. If z 0 = b or z 0 = w and no jump, then a t b/r or a t a w, respectively, implying the existence of an irreducible skeleton. But how to prove recurrence?

Existence and stability In the regimes of high and intermediate interest rates, wealth can converge to. We concentrate on the low-interest-regime, where wealth is concentrated in a compact interval [ b/r, a w]. By continuity of solutions of ODEs in the initial value, (a t, z t ) is a continuous function of (a 0, z 0 ), implying the weak Feller property. Existence of invariant probability measures. If z 0 = b or z 0 = w and no jump, then a t b/r or a t a w, respectively, implying the existence of an irreducible skeleton. But how to prove recurrence?

Sufficient condition for recurrence Recurrence holds when transition kernel is smoothing. Diffusion case: recurrence follows under weak conditions Jump processes cannot smoothen as long as there is a positive probability of no jumps before t Definition & theorem (Meyn and Tweedie 93) X t is called a T-process if there is a Markov kernel T and a prob. measure ν on [0, [ s.t. A B(X) : x T(x, A) is continuous K ν (x, A) 0 Pt (x, A)ν(dt) T(x, A) x : T(x, X) > 0. Any irreducible T-process, which is bounded in probability on average, is recurrent.

Sufficient condition for recurrence Recurrence holds when transition kernel is smoothing. Diffusion case: recurrence follows under weak conditions Jump processes cannot smoothen as long as there is a positive probability of no jumps before t Definition & theorem (Meyn and Tweedie 93) X t is called a T-process if there is a Markov kernel T and a prob. measure ν on [0, [ s.t. A B(X) : x T(x, A) is continuous K ν (x, A) 0 Pt (x, A)ν(dt) T(x, A) x : T(x, X) > 0. Any irreducible T-process, which is bounded in probability on average, is recurrent.

Sufficient condition for recurrence Recurrence holds when transition kernel is smoothing. Diffusion case: recurrence follows under weak conditions Jump processes cannot smoothen as long as there is a positive probability of no jumps before t Definition & theorem (Meyn and Tweedie 93) X t is called a T-process if there is a Markov kernel T and a prob. measure ν on [0, [ s.t. A B(X) : x T(x, A) is continuous K ν (x, A) 0 Pt (x, A)ν(dt) T(x, A) x : T(x, X) > 0. Any irreducible T-process, which is bounded in probability on average, is recurrent.

The wealth-employement process is a T-process Given (a 0, z 0 ), (a t, z t ) is a deterministic function of the jump-times of z t. Conditional on the number of jumps, the jump times have smooth densities. (a t, z t ) is not smoothing, because no jump might occur. If at least one jump occurs, we have smoothing properties. Choose ν = δ τ and T((a 0, z 0 ), A) P ( (a τ, z τ ) A, one jump in [0, τ] a0, z 0 ). Technical condition: c = c(a, z) is C 1. Illustration of how T-property replaces strong Feller condition.

The wealth-employement process is a T-process Given (a 0, z 0 ), (a t, z t ) is a deterministic function of the jump-times of z t. Conditional on the number of jumps, the jump times have smooth densities. (a t, z t ) is not smoothing, because no jump might occur. If at least one jump occurs, we have smoothing properties. Choose ν = δ τ and T((a 0, z 0 ), A) P ( (a τ, z τ ) A, one jump in [0, τ] a0, z 0 ). Technical condition: c = c(a, z) is C 1. Illustration of how T-property replaces strong Feller condition.

The wealth-employement process is a T-process Given (a 0, z 0 ), (a t, z t ) is a deterministic function of the jump-times of z t. Conditional on the number of jumps, the jump times have smooth densities. (a t, z t ) is not smoothing, because no jump might occur. If at least one jump occurs, we have smoothing properties. Choose ν = δ τ and T((a 0, z 0 ), A) P ( (a τ, z τ ) A, one jump in [0, τ] a0, z 0 ). Technical condition: c = c(a, z) is C 1. Illustration of how T-property replaces strong Feller condition.

Conclusions Framework Individual maximization problem inspired by search and matching models Extended for consumption-saving problem Question: Is there a unique long-run distribution to which initial distributions converge? Techniques Markov chain-style ergodicity analysis for general, continuous time Markov processes T-processes by Meyn and Tweedie allow to prove recurrence for a wide class of (degenerate) diffusion and jump models Result Long-run-distribution exists in our matching-saving model

Conclusions Framework Individual maximization problem inspired by search and matching models Extended for consumption-saving problem Question: Is there a unique long-run distribution to which initial distributions converge? Techniques Markov chain-style ergodicity analysis for general, continuous time Markov processes T-processes by Meyn and Tweedie allow to prove recurrence for a wide class of (degenerate) diffusion and jump models Result Long-run-distribution exists in our matching-saving model

Conclusions Framework Individual maximization problem inspired by search and matching models Extended for consumption-saving problem Question: Is there a unique long-run distribution to which initial distributions converge? Techniques Markov chain-style ergodicity analysis for general, continuous time Markov processes T-processes by Meyn and Tweedie allow to prove recurrence for a wide class of (degenerate) diffusion and jump models Result Long-run-distribution exists in our matching-saving model

References Azéma, J,., Duflo, M., Revuz, D.: Propriétés relatives des processus de Markov récurrents, Z. Wahrschenlichkeitsth. 8, 157 181, 1967. Bayer, C., Waelde, K.: Matching and saving in continuous time, preprint 2011. Beneš, V.: Finite regular invariant measures for Feller processes, J. Appl. Prob. 5, 203 209, 1968. Meyn, S., Tweedie, R.: Stability of Markovian processes II: Continuous-time processes and sampled chains, Adv. Appl. Prob. 25, 487 517, 1993.