A regeneration proof of the central limit theorem for uniformly ergodic Markov chains

Size: px
Start display at page:

Download "A regeneration proof of the central limit theorem for uniformly ergodic Markov chains"

Transcription

1 A regeneration proof of the central limit theorem for uniformly ergodic Markov chains By AJAY JASRA Department of Mathematics, Imperial College London, SW7 2AZ, London, UK and CHAO YANG Department of Mathematics, University of Toronto, M5S 2E4, Toronto ON, Canada November 0, 2005 Abstract Let X n be a Markov chain on measurable space E, E with unique stationary distribution π. Let h : E R be a measurable function with finite stationary mean πh := E hxπdx. Ibragimov and Linnik 97 proved that if X n is geometrically ergodic, then a central limit theorem CLT holds for h whenever π h 2+δ <, δ > 0. Cogburn 972 proved that if a Markov chain is uniformly ergodic, with πh 2 < then a CLT holds for h. The first result was re-proved in Roberts and Rosenthal 2004 using a regeneration approach; thus removing many of the technicalities of the original proof. This raised an open problem: to provide a proof of the second result using a regeneration approach. In this paper we provide a solution to this problem. Keywords: Markov chains; Central limit theorems Introduction Let X n be a Markov chain with transition kernel P : E E [0, and a unique stationary distribution π. Let h : E R be a real-valued measurable function. We say that h satisfies a Central Limit Theorem or n CLT if there is some σ 2 < such that the normalized sum n 2 n i= [hx i πh converges weakly to a N0, σ 2 distribution, where N0, σ 2 is a Gaussian distribution with zero mean and variance σ 2 we allow that σ 2 = 0, and e.g. Chan and Geyer 994, see also Bradley 985 and Chen 999 σ 2 = πh E hxp n hxπdx n= with P n hx = E hyp n x, dy and P n x, dy the n step transition law for the Markov chain. To further our discussion we provide the following definitions. Denote the class of probability measures on E, E as PE. The total variation distance between μ, ν PE is: μ ν := sup μa νa. A E We will be concerned with geometrically and uniformly ergodic Markov chains: Corresponding author, chaoyang@math.toronto.edu

2 Definition.. A Markov chain with stationary distribution π PE is geometrically ergodic if n N: P n x, π Mxρ n where ρ < and Mx < π almost everywhere. If M = sup x E Mx is finite then the chain is uniformly ergodic. Theorem.2 Cogburn, 972. If a Markov chain with stationary distribution π PE is uniformly ergodic, then a n CLT holds for h whenever πh 2 <. Ibragimov and Linnik 97 proved a CLT for h when the chain is geometrically ergodic and, for some δ > 0, π h 2+δ <. Roberts and Rosenthal 2004 provided a simpler proof using regeneration arguments. In addition, Roberts and Rosenthal 2004 left an open problem: To provide a proof of Theorem.2 originally proved by Cogburn 972 using regeneration. Many of the recent developments of CLTs for Markov chains are related to the evolution of stochastic simulation algorithms such as Markov chain Monte Carlo MCMC e.g. Robert and Rosenthal For example, Roberts and Rosenthal 2004 posed many open problems, including that considered here, for CLTs; see Häggström 2005 for a solution to another open problem. Additionally, Jones 2004 discusses the link between mixing processes and CLTs, with MCMC algorithms a particular consideration. For an up-to-date review of CLTs for Markov chains see: Bradley 985, Chen 999 and Jones The proof of Theorem.2, using regeneration theory, provides an elegant framework for the proof of CLTs for Markov chains. The approach may also be useful for alternative proofs of CLTs for chains with different ergodicity properties; e.g. polynomial ergodicity see Jarner and Roberts The structure of this paper is as follows. In Section 2 we provide some background knowledge about the small sets and the regeneration construction, we also detail some technical results. In Section 3 we use the results of the previous Section to provide a proof of Theorem.2 using regenerations. 2 Small Sets and Regeneration Construction 2. Small Sets We recall the notion of a small set: Definition 2.. A set C E is small or n 0, ɛ, ν-small if there exists an n 0 N, ɛ > 0 and a non-trivial ν PE such that the following minorization condition holds x C: P n 0 x, ɛν. It is known e.g. Meyn and Tweedie 993 that if P is uniformly ergodic, the whole state space E is small. That is we have the following lemma: Lemma 2.. If X n on E, E with stationary distribution π PE is uniformly ergodic, then E is small. 2

3 2.2 Regeneration Construction and Some related Technical Results Now we consider the regeneration construction for the proof. Since E is small we use the split chain construction Nummelin, 984, for any x E, A E P n 0 x, A = ɛrx, A + ɛνa where Rx, A = ɛ [P n 0 x, A ɛνa. That is, for a single chain X n, with probability ɛ we choose X n+n0 ν, while with probability ɛ we choose X n+n0 RX n,, if n 0 >, we fill in the missing values as X n+ using the appropriate Markov kernel and conditionals. We let T, T 2,... be the regeneration times, i.e. the times such that X Ti ν, clearly T i = in 0. Let T 0 = 0 and rn = sup{i 0 : T i n}, using the regeneration time, we can break up the sum n i=0 [hx i πh into sums over tours as follows: where rn n [hx i πh = i=0 Qn = T j= T j+ [hx j πh + We begin our construction, by noting the following result. Lemma 2.. Under the formulation above, we have that: Proof. Let and [hx i πh + Qn n T rn+ [hx j πh. Qn n /2 p 0. 2 T Q + n = [hx j πh + T Q n = [hx j πh Q + 2 n = n T rn + Q 2 n = n T rn + [hx j πh + [hx j πh where [hx j πh + = max{hx j πh, 0} and [hx j πh = max{ [hx j πh, 0}. 3

4 The strategy of the proof is to show that Q ± i n/n/2 p 0 as n. Consider Q + n, sn 0 Q + n = [hx j πh + w.p ɛ ɛ s 3 where s N. If Q + n/n/2 p 0, i.e. P ɛ, Q + n > ɛn/2, i.o. = for all n, which means that PQ + n =, i.o. =, which is impossible from 3. So Q+ i n/n/2 p 0 as n. Similarly Q i n/n/2 p 0 as n. For Q 2 we have Q + 2 n l n j=rn+ [hx j πh + = Q + 2 n, where ln = inf {i 0 : T i n}. We know that Q + 2 n has the same distribution with Q+ 2 n, so Q + i n/n/2 p 0 as n and therefore, Q + 2 n/n/2 p 0 as n. Similarly Q 2 n/n/2 p 0 as n. From the above discussion, we conclude that Qn/n /2 p 0. The above lemma indicates that our objective is to find the asymptotic distribution of rn j= Tj+ [hx i πh. Given the definition of T i, each random variable s j = T j+ [hx i πh has same distribution. However, we know that T j depends on X Tj +,, X Tj, but does not depend on the value of X Tj. That is, we have the following lemma: Lemma 2.2. For any 0 i <, s i and s i+ are not independent, but the two collections of random variables: {s i : 0 i m 2} and {s i : i m} are independent for any m 2.Therefore the random variable sequence {s i } i=0 is a one-dependent stationary stochastic processes. Proof. Clearly s i+ depends on the distribution T i+, thus: P X Ti + dx,, X Ti +m dy X Ti = x, T i+ T i > m = ɛrx, dy P m P x, dx P x m, dy x, dy and P X Ti + dx,, X Ti +m dy X Ti = x, T i+ T i = m = ɛνdy P m x, dy P x, dx P x m, dy. Note s i depends on T i+. Therefore s i and s i+ are not independent. However, for any 0 i m 2 < m j <, since X Ti ν and X Tj depends X Tj +,, X Tj, but is independent of all the {X k : k T j }. Thus, we have the result. To prove Theorem.2 we follow the strategy: T Step : Prove that I = E ν i=0 [hx i πh = 0 Step 2: Prove that J = E νdxe [ T i=0 [hx i πh 2 X 0 = x <. Step 3: Prove that a n CLT holds for a stationary, one-step dependent stochastic process. 3 Proof of Theorem.2 T Lemma 3.. I = E ν i=0 [hx i πh = 0 4

5 Proof. Denote T = τm and H k = k+m i=km [hx i πh, then we have: I = E ν [ H k I{k < τ} Consider the splitting m skeleton chain { ˇX nm } as in section 5.. of Meyn and Tweedie 993, we know that ˇα = X is an accessible atom. Then we can apply Theorem 0.0. of Meyn and Tweedie 993 to this splitting chain. That is: ˇτ ˇα πb = ˇπB 0 B = ˇπdwE w [ I{ ˇX km ˇB} ˇα = ɛ πdwe w [ X k= ˇτ ˇα k= I{ ˇX km ˇB} Let ˇτˇα = min{n : ˇXnm ˇα}. Since for any w ˇα, ˇP m w, ν, we have ˇτˇα = τ. Following Theorem 5..3 in Meyn and Tweedie 993, we also have P kn 0 x, B = ˇP kn 0 x, ˇB for any B E. Therefore we have: τ πb = ɛe ν [ I{X km B} = ɛe ν [ I{X km B}I{τ > k} So we have: k= k= I = E ν [E H k I{k < τ} X km = = E ν [E H k I{k < τ} X km E ν [E H k X km I{k < τ} The last equation follows since random variables I{τ > k} and X km are independent. In addition, given τ > k and X km, the distribution of H k is equal to H 0 given X 0 ; therefore I = E ν [E H 0 X 0 I{k < τ} = E π [E H 0 X 0 = E π H 0 = 0. Lemma 3.2. We have: [ T 2 J = E ν [hx i πh <. 4 i=0 5

6 Proof. [ τ J = E ν k+m i=km [ 2 E ν I{k < τ} H k 2 [hx i πh [ = E ν H k 2 I{k < τ} + 2 H k = E ν [ = E ν [ = E ν [ H k 2 + 2H k j=i+ E H k H k E H k H k j=k+ H j I{j < τ} I{k < τ} j=k+ j=k+ H j I{j < τ} {k < τ} H j I{j < τ}i{k < τ} X km, I{k < τ} H j I{j < τ} X km I{k < τ}. In the last equation, we have used the fact that random variables I{τ > k} and X km are independent. Since E H i H i j= H j {j < τ} X im = x = E H H 0 define fx = E H H 0 j= H j {j < τ} X 0 = x then we have: J E ν [ fx 0 I{k < τ} k= j= [ = E ν [fx 0 I{0 < τ} + E ν fx 0 I{k < τ} [ E ν [fx 0 + E ν fx 0 E ν [I{k < τ} The last inequality is follows since:. fx 0 I{k < τ} fx 0 ; 2. When k, I{τ > k} is independent with X 0 k= H j {j < τ} X 0 = x Note E ν [I{k < τ = P ν k < τ ɛ k 6

7 and πdy = E P n 0 x, dyπdx ɛνdy therefore we have J ɛ E ν[fx 0 ɛ 2 E π [fx 0 and m E π [fx 0 E π [ hx i πh 2 i=0 From the above arguments we conclude that J <. Finally, we prove Theorem.2: mπh 2 πh 2 < Proof of Theorem.2. Following Lemma 2., we can obtain: n i=0 lim [hx i πh n n /2 = lim n rn j= Tj+ [hx i πh n /2. 5 Define h i = hx i πh, s j = T j+ + h i and η j = s jm+ + +s j+m for an integer m 2. Following Lemma 2.2 we know that two collections of random variables: {s i : 0 j m 2} and {s i : i m} are independent for any m 2; thus n n s j = [n/m η j + [n/m s mj + n n n j= n m[n/m It should be noted that if j i > n 0, then X i and X j are independent, η j are i.i.d random variables and s mj are i.i.d. so we have: [n/m η j d N0, σ2 m n m [n/m s mj d N0, σ2 s n m s j where σ 2 m = m Es 2 + 2m 2Es s 2 and σ 2 s = E[s 2, letting m, we have σ2 m m Es 2 + 2Es s 2 and m σ 2 s 0, so the CLT holds. Let [ n σ 2 2 = lim n n E [hx i πh i= 7

8 then σ 2 = [ n 2 lim n n E [hx i πh i= = [ rn lim n n E s j 2 = lim n n E j= [rns 2 + 2rn 2s s 2 r By the elementary renewal theorem e.g. Feller 968, lim nn n = ET 2 T. Since P[T 2 T = n 0 s = ε ε s, ET 2 T = s= [n 0sε ε s = n 0 ε <. Therefore if we denote σ 2 = E[s 2 + 2s s 2, then As a result, we conclude that lim n as n. rn j= Acknowledgement σ 2 = n 0 ε E[s2 + 2s s 2 = n 0 ε σ2 6 Tj+ [hx i πh n /2 = lim n d n0 ε rn j= = N0, σ 2 /2 N0, σ 2 Tj+ [hx i πh rn /2 r/2 n n /2 Both authors would like to thank Jeffrey Rosenthal for his assistance in writing this paper. The first author was supported by an Engineering and Physical Sciences Research Council Studentship and would like to thank Dave Stephens and Chris Holmes for their advice relating to this paper. REFERENCES Bradley, R. C On the central limit question under absolute regularity. Ann. Prob., 3, Chan, K. S. and Geyer, C. J Discussion of Markov chains for exploring posterior distributions. Ann. Statist., 22, Chen, X Limit theorems for functionals of ergodic Markov chains with general state space. Mem. Amer. Math. Soc., 39. Cogburn, R The central limit theorem for Markov processes. In Le Cam, L. E., Neyman, J. and Scott, E. L. Eds. Proc. Sixth Ann. Berkley Symp. Math. Statist. and Prob., 2,

9 Feller, W An Introduction to Probability Theory and its Applications. 3rd ed, Wiley, Chichester. Häggström, O On the central limit theorem for geometrically ergodic Markov chains. Prob. Theory and Rel. Fields, 32, Ibragimov, I. A. and Linnik, Y. V. 97. Independent and stationary sequences of random variables. Wolter-Noordhoff, Groiningen. Jarner, S. F. and Roberts G. O Polynomial convergence rates of Markov chains. Ann. Appl. Prob., 2, Jones, G. L On the Markov chain central limit theorem. Prob. Surveys,, Meyn, S. P. and Tweedie, R. L Markov chains and stochastic stability. Springer, New York. Nummelin, E General irreducible Markov chains and non-negative operators. Cambridge University Press, Cambridge. Roberts, G. O. and Rosenthal, J. S General state space Markov chains and MCMC algorithms. Prob. Surveys,,

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

Variance Bounding Markov Chains

Variance Bounding Markov Chains Variance Bounding Markov Chains by Gareth O. Roberts * and Jeffrey S. Rosenthal ** (September 2006; revised April 2007.) Abstract. We introduce a new property of Markov chains, called variance bounding.

More information

Faithful couplings of Markov chains: now equals forever

Faithful couplings of Markov chains: now equals forever Faithful couplings of Markov chains: now equals forever by Jeffrey S. Rosenthal* Department of Statistics, University of Toronto, Toronto, Ontario, Canada M5S 1A1 Phone: (416) 978-4594; Internet: jeff@utstat.toronto.edu

More information

Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms

Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms Yan Bai Feb 2009; Revised Nov 2009 Abstract In the paper, we mainly study ergodicity of adaptive MCMC algorithms. Assume that

More information

Tail inequalities for additive functionals and empirical processes of. Markov chains

Tail inequalities for additive functionals and empirical processes of. Markov chains Tail inequalities for additive functionals and empirical processes of geometrically ergodic Markov chains University of Warsaw Banff, June 2009 Geometric ergodicity Definition A Markov chain X = (X n )

More information

ON CONVERGENCE RATES OF GIBBS SAMPLERS FOR UNIFORM DISTRIBUTIONS

ON CONVERGENCE RATES OF GIBBS SAMPLERS FOR UNIFORM DISTRIBUTIONS The Annals of Applied Probability 1998, Vol. 8, No. 4, 1291 1302 ON CONVERGENCE RATES OF GIBBS SAMPLERS FOR UNIFORM DISTRIBUTIONS By Gareth O. Roberts 1 and Jeffrey S. Rosenthal 2 University of Cambridge

More information

Geometric Convergence Rates for Time-Sampled Markov Chains

Geometric Convergence Rates for Time-Sampled Markov Chains Geometric Convergence Rates for Time-Sampled Markov Chains by Jeffrey S. Rosenthal* (April 5, 2002; last revised April 17, 2003) Abstract. We consider time-sampled Markov chain kernels, of the form P µ

More information

On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo

On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo James P. Hobert 1, Galin L. Jones 2, Brett Presnell 1, and Jeffrey S. Rosenthal 3 1 Department of Statistics University of Florida

More information

Quantitative Non-Geometric Convergence Bounds for Independence Samplers

Quantitative Non-Geometric Convergence Bounds for Independence Samplers Quantitative Non-Geometric Convergence Bounds for Independence Samplers by Gareth O. Roberts * and Jeffrey S. Rosenthal ** (September 28; revised July 29.) 1. Introduction. Markov chain Monte Carlo (MCMC)

More information

Introduction to Markov chain Monte Carlo

Introduction to Markov chain Monte Carlo Introduction to Markov chain Monte Carlo Adam M. Johansen 1 February 12, 2018 1 Based closely on slides produced by Anthony Lee in previous years. 1 / 87 Outline Motivation What is a Markov chain? First

More information

Coupling and Ergodicity of Adaptive MCMC

Coupling and Ergodicity of Adaptive MCMC Coupling and Ergodicity of Adaptive MCMC by Gareth O. Roberts* and Jeffrey S. Rosenthal** (March 2005; last revised March 2007.) Abstract. We consider basic ergodicity properties of adaptive MCMC algorithms

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,

More information

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes Front. Math. China 215, 1(4): 933 947 DOI 1.17/s11464-15-488-5 Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes Yuanyuan LIU 1, Yuhui ZHANG 2

More information

Minorization Conditions and Convergence Rates for Markov Chain Monte Carlo. (September, 1993; revised July, 1994.)

Minorization Conditions and Convergence Rates for Markov Chain Monte Carlo. (September, 1993; revised July, 1994.) Minorization Conditions and Convergence Rates for Markov Chain Monte Carlo September, 1993; revised July, 1994. Appeared in Journal of the American Statistical Association 90 1995, 558 566. by Jeffrey

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

The simple slice sampler is a specialised type of MCMC auxiliary variable method (Swendsen and Wang, 1987; Edwards and Sokal, 1988; Besag and Green, 1

The simple slice sampler is a specialised type of MCMC auxiliary variable method (Swendsen and Wang, 1987; Edwards and Sokal, 1988; Besag and Green, 1 Recent progress on computable bounds and the simple slice sampler by Gareth O. Roberts* and Jerey S. Rosenthal** (May, 1999.) This paper discusses general quantitative bounds on the convergence rates of

More information

Convergence Rates for Renewal Sequences

Convergence Rates for Renewal Sequences Convergence Rates for Renewal Sequences M. C. Spruill School of Mathematics Georgia Institute of Technology Atlanta, Ga. USA January 2002 ABSTRACT The precise rate of geometric convergence of nonhomogeneous

More information

On the Central Limit Theorem for an ergodic Markov chain

On the Central Limit Theorem for an ergodic Markov chain Stochastic Processes and their Applications 47 ( 1993) 113-117 North-Holland 113 On the Central Limit Theorem for an ergodic Markov chain K.S. Chan Department of Statistics and Actuarial Science, The University

More information

Practical conditions on Markov chains for weak convergence of tail empirical processes

Practical conditions on Markov chains for weak convergence of tail empirical processes Practical conditions on Markov chains for weak convergence of tail empirical processes Olivier Wintenberger University of Copenhagen and Paris VI Joint work with Rafa l Kulik and Philippe Soulier Toronto,

More information

Chapter 2: Markov Chains and Queues in Discrete Time

Chapter 2: Markov Chains and Queues in Discrete Time Chapter 2: Markov Chains and Queues in Discrete Time L. Breuer University of Kent 1 Definition Let X n with n N 0 denote random variables on a discrete space E. The sequence X = (X n : n N 0 ) is called

More information

Local consistency of Markov chain Monte Carlo methods

Local consistency of Markov chain Monte Carlo methods Ann Inst Stat Math (2014) 66:63 74 DOI 10.1007/s10463-013-0403-3 Local consistency of Markov chain Monte Carlo methods Kengo Kamatani Received: 12 January 2012 / Revised: 8 March 2013 / Published online:

More information

When is a Markov chain regenerative?

When is a Markov chain regenerative? When is a Markov chain regenerative? Krishna B. Athreya and Vivekananda Roy Iowa tate University Ames, Iowa, 50011, UA Abstract A sequence of random variables {X n } n 0 is called regenerative if it can

More information

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes Lectures on Stochastic Stability Sergey FOSS Heriot-Watt University Lecture 4 Coupling and Harris Processes 1 A simple example Consider a Markov chain X n in a countable state space S with transition probabilities

More information

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015 Markov Chains Arnoldo Frigessi Bernd Heidergott November 4, 2015 1 Introduction Markov chains are stochastic models which play an important role in many applications in areas as diverse as biology, finance,

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Christian Bayer and Klaus Wälde Weierstrass Institute for Applied Analysis and Stochastics and University

More information

Strong approximation for additive functionals of geometrically ergodic Markov chains

Strong approximation for additive functionals of geometrically ergodic Markov chains Strong approximation for additive functionals of geometrically ergodic Markov chains Florence Merlevède Joint work with E. Rio Université Paris-Est-Marne-La-Vallée (UPEM) Cincinnati Symposium on Probability

More information

Geometric Ergodicity and Hybrid Markov Chains

Geometric Ergodicity and Hybrid Markov Chains Geometric Ergodicity and Hybrid Markov Chains by Gareth O. Roberts* and Jeffrey S. Rosenthal** (August 1, 1996; revised, April 11, 1997.) Abstract. Various notions of geometric ergodicity for Markov chains

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture

More information

Chapter 7. Markov chain background. 7.1 Finite state space

Chapter 7. Markov chain background. 7.1 Finite state space Chapter 7 Markov chain background A stochastic process is a family of random variables {X t } indexed by a varaible t which we will think of as time. Time can be discrete or continuous. We will only consider

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

On Differentiability of Average Cost in Parameterized Markov Chains

On Differentiability of Average Cost in Parameterized Markov Chains On Differentiability of Average Cost in Parameterized Markov Chains Vijay Konda John N. Tsitsiklis August 30, 2002 1 Overview The purpose of this appendix is to prove Theorem 4.6 in 5 and establish various

More information

Some Results on the Ergodicity of Adaptive MCMC Algorithms

Some Results on the Ergodicity of Adaptive MCMC Algorithms Some Results on the Ergodicity of Adaptive MCMC Algorithms Omar Khalil Supervisor: Jeffrey Rosenthal September 2, 2011 1 Contents 1 Andrieu-Moulines 4 2 Roberts-Rosenthal 7 3 Atchadé and Fort 8 4 Relationship

More information

LARGE DEVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILED DEPENDENT RANDOM VECTORS*

LARGE DEVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILED DEPENDENT RANDOM VECTORS* LARGE EVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILE EPENENT RANOM VECTORS* Adam Jakubowski Alexander V. Nagaev Alexander Zaigraev Nicholas Copernicus University Faculty of Mathematics and Computer Science

More information

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS The Annals of Applied Probability 1999, Vol. 9, No. 3, 719 731 PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS By Elizabeth A. Brooks National Institute of Environmental

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

Weak convergence of Markov chain Monte Carlo II

Weak convergence of Markov chain Monte Carlo II Weak convergence of Markov chain Monte Carlo II KAMATANI, Kengo Mar 2011 at Le Mans Background Markov chain Monte Carlo (MCMC) method is widely used in Statistical Science. It is easy to use, but difficult

More information

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS Statistica Sinica 6(1996), 943-956 THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS H. Z. An and F. C. Huang Academia Sinica Abstract: Consider the nonlinear autoregressive model x t = φ(x

More information

The Codimension of the Zeros of a Stable Process in Random Scenery

The Codimension of the Zeros of a Stable Process in Random Scenery The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar

More information

Stability of Cyclic Threshold and Threshold-like Autoregressive Time Series Models

Stability of Cyclic Threshold and Threshold-like Autoregressive Time Series Models 1 Stability of Cyclic Threshold and Threshold-like Autoregressive Time Series Models Thomas R. Boucher, Daren B.H. Cline Plymouth State University, Texas A&M University Abstract: We investigate the stability,

More information

Random Bernstein-Markov factors

Random Bernstein-Markov factors Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit

More information

1 Kernels Definitions Operations Finite State Space Regular Conditional Probabilities... 4

1 Kernels Definitions Operations Finite State Space Regular Conditional Probabilities... 4 Stat 8501 Lecture Notes Markov Chains Charles J. Geyer April 23, 2014 Contents 1 Kernels 2 1.1 Definitions............................. 2 1.2 Operations............................ 3 1.3 Finite State Space........................

More information

Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC)

Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC) Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC) Cong Ling Imperial College London aaa joint work with Zheng Wang (Huawei Technologies Shanghai) aaa September 20, 2016 Cong Ling (ICL) MCMC

More information

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence

More information

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes

More information

ITERATED FUNCTION SYSTEMS WITH CONTINUOUS PLACE DEPENDENT PROBABILITIES

ITERATED FUNCTION SYSTEMS WITH CONTINUOUS PLACE DEPENDENT PROBABILITIES UNIVERSITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XL 2002 ITERATED FUNCTION SYSTEMS WITH CONTINUOUS PLACE DEPENDENT PROBABILITIES by Joanna Jaroszewska Abstract. We study the asymptotic behaviour

More information

COMPLETE QTH MOMENT CONVERGENCE OF WEIGHTED SUMS FOR ARRAYS OF ROW-WISE EXTENDED NEGATIVELY DEPENDENT RANDOM VARIABLES

COMPLETE QTH MOMENT CONVERGENCE OF WEIGHTED SUMS FOR ARRAYS OF ROW-WISE EXTENDED NEGATIVELY DEPENDENT RANDOM VARIABLES Hacettepe Journal of Mathematics and Statistics Volume 43 2 204, 245 87 COMPLETE QTH MOMENT CONVERGENCE OF WEIGHTED SUMS FOR ARRAYS OF ROW-WISE EXTENDED NEGATIVELY DEPENDENT RANDOM VARIABLES M. L. Guo

More information

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974 LIMITS FOR QUEUES AS THE WAITING ROOM GROWS by Daniel P. Heyman Ward Whitt Bell Communications Research AT&T Bell Laboratories Red Bank, NJ 07701 Murray Hill, NJ 07974 May 11, 1988 ABSTRACT We study the

More information

Geometric Ergodicity of a Random-Walk Metorpolis Algorithm via Variable Transformation and Computer Aided Reasoning in Statistics

Geometric Ergodicity of a Random-Walk Metorpolis Algorithm via Variable Transformation and Computer Aided Reasoning in Statistics Geometric Ergodicity of a Random-Walk Metorpolis Algorithm via Variable Transformation and Computer Aided Reasoning in Statistics a dissertation submitted to the faculty of the graduate school of the university

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

The Monte Carlo Computation Error of Transition Probabilities

The Monte Carlo Computation Error of Transition Probabilities Zuse Institute Berlin Takustr. 7 495 Berlin Germany ADAM NILSN The Monte Carlo Computation rror of Transition Probabilities ZIB Report 6-37 (July 206) Zuse Institute Berlin Takustr. 7 D-495 Berlin Telefon:

More information

Reversible Markov chains

Reversible Markov chains Reversible Markov chains Variational representations and ordering Chris Sherlock Abstract This pedagogical document explains three variational representations that are useful when comparing the efficiencies

More information

FOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1

FOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1 FOURIER TRANSFORMS OF STATIONARY PROCESSES WEI BIAO WU September 8, 003 Abstract. We consider the asymptotic behavior of Fourier transforms of stationary and ergodic sequences. Under sufficiently mild

More information

Central Limit Theorem for Non-stationary Markov Chains

Central Limit Theorem for Non-stationary Markov Chains Central Limit Theorem for Non-stationary Markov Chains Magda Peligrad University of Cincinnati April 2011 (Institute) April 2011 1 / 30 Plan of talk Markov processes with Nonhomogeneous transition probabilities

More information

GENERAL STATE SPACE MARKOV CHAINS AND MCMC ALGORITHMS

GENERAL STATE SPACE MARKOV CHAINS AND MCMC ALGORITHMS GENERAL STATE SPACE MARKOV CHAINS AND MCMC ALGORITHMS by Gareth O. Roberts* and Jeffrey S. Rosenthal** (March 2004; revised August 2004) Abstract. This paper surveys various results about Markov chains

More information

A LeVeque-type lower bound for discrepancy

A LeVeque-type lower bound for discrepancy reprinted from Monte Carlo and Quasi-Monte Carlo Methods 998, H. Niederreiter and J. Spanier, eds., Springer-Verlag, 000, pp. 448-458. A LeVeque-type lower bound for discrepancy Francis Edward Su Department

More information

Mi-Hwa Ko. t=1 Z t is true. j=0

Mi-Hwa Ko. t=1 Z t is true. j=0 Commun. Korean Math. Soc. 21 (2006), No. 4, pp. 779 786 FUNCTIONAL CENTRAL LIMIT THEOREMS FOR MULTIVARIATE LINEAR PROCESSES GENERATED BY DEPENDENT RANDOM VECTORS Mi-Hwa Ko Abstract. Let X t be an m-dimensional

More information

Department of Econometrics and Business Statistics

Department of Econometrics and Business Statistics ISSN 440-77X Australia Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Nonlinear Regression with Harris Recurrent Markov Chains Degui Li, Dag

More information

Adaptive MCMC: Background, Theory, and Simulations

Adaptive MCMC: Background, Theory, and Simulations Adaptive MCMC: Background, Theory, and Simulations Supervisor: Professor Jeffrey S. Rosenthal Student: Shuheng Zheng Summer 2006 Research is supported by the NSERC Undergraduate Student Research Award

More information

The Degree of the Splitting Field of a Random Polynomial over a Finite Field

The Degree of the Splitting Field of a Random Polynomial over a Finite Field The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada fjdixon,danielg@math.carleton.ca

More information

Some functional (Hölderian) limit theorems and their applications (II)

Some functional (Hölderian) limit theorems and their applications (II) Some functional (Hölderian) limit theorems and their applications (II) Alfredas Račkauskas Vilnius University Outils Statistiques et Probabilistes pour la Finance Université de Rouen June 1 5, Rouen (Rouen

More information

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching Communications for Statistical Applications and Methods 2013, Vol 20, No 4, 339 345 DOI: http://dxdoiorg/105351/csam2013204339 A Functional Central Limit Theorem for an ARMAp, q) Process with Markov Switching

More information

VARIABLE TRANSFORMATION TO OBTAIN GEOMETRIC ERGODICITY IN THE RANDOM-WALK METROPOLIS ALGORITHM

VARIABLE TRANSFORMATION TO OBTAIN GEOMETRIC ERGODICITY IN THE RANDOM-WALK METROPOLIS ALGORITHM Submitted to the Annals of Statistics VARIABLE TRANSFORMATION TO OBTAIN GEOMETRIC ERGODICITY IN THE RANDOM-WALK METROPOLIS ALGORITHM By Leif T. Johnson and Charles J. Geyer Google Inc. and University of

More information

A Probabilistic Proof of the Perron-Frobenius Theorem

A Probabilistic Proof of the Perron-Frobenius Theorem A Probabilistic Proof of the Perron-Frobenius Theorem Peter W. Glynn 1 and Paritosh Y. Desai 2 1 Department of Management Science and Engineering, Stanford University, Stanford CA 2 Target Corporation,

More information

The Markov Chain Monte Carlo Method

The Markov Chain Monte Carlo Method The Markov Chain Monte Carlo Method Idea: define an ergodic Markov chain whose stationary distribution is the desired probability distribution. Let X 0, X 1, X 2,..., X n be the run of the chain. The Markov

More information

The Degree of the Splitting Field of a Random Polynomial over a Finite Field

The Degree of the Splitting Field of a Random Polynomial over a Finite Field The Degree of the Splitting Field of a Random Polynomial over a Finite Field John D. Dixon and Daniel Panario School of Mathematics and Statistics Carleton University, Ottawa, Canada {jdixon,daniel}@math.carleton.ca

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

CONVERGENCE RATES AND REGENERATION OF THE BLOCK GIBBS SAMPLER FOR BAYESIAN RANDOM EFFECTS MODELS

CONVERGENCE RATES AND REGENERATION OF THE BLOCK GIBBS SAMPLER FOR BAYESIAN RANDOM EFFECTS MODELS CONVERGENCE RATES AND REGENERATION OF THE BLOCK GIBBS SAMPLER FOR BAYESIAN RANDOM EFFECTS MODELS By AIXIN TAN A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT

More information

MARKOV PARTITIONS FOR HYPERBOLIC SETS

MARKOV PARTITIONS FOR HYPERBOLIC SETS MARKOV PARTITIONS FOR HYPERBOLIC SETS TODD FISHER, HIMAL RATHNAKUMARA Abstract. We show that if f is a diffeomorphism of a manifold to itself, Λ is a mixing (or transitive) hyperbolic set, and V is a neighborhood

More information

Limiting behaviour of moving average processes under ρ-mixing assumption

Limiting behaviour of moving average processes under ρ-mixing assumption Note di Matematica ISSN 1123-2536, e-issn 1590-0932 Note Mat. 30 (2010) no. 1, 17 23. doi:10.1285/i15900932v30n1p17 Limiting behaviour of moving average processes under ρ-mixing assumption Pingyan Chen

More information

Tractability of Multivariate Problems

Tractability of Multivariate Problems Erich Novak University of Jena Chemnitz, Summer School 2010 1 Plan for the talk Example: approximation of C -functions What is tractability? Tractability by smoothness? Tractability by sparsity, finite

More information

Block Gibbs sampling for Bayesian random effects models with improper priors: Convergence and regeneration

Block Gibbs sampling for Bayesian random effects models with improper priors: Convergence and regeneration Block Gibbs sampling for Bayesian random effects models with improper priors: Convergence and regeneration Aixin Tan and James P. Hobert Department of Statistics, University of Florida May 4, 009 Abstract

More information

be the set of complex valued 2π-periodic functions f on R such that

be the set of complex valued 2π-periodic functions f on R such that . Fourier series. Definition.. Given a real number P, we say a complex valued function f on R is P -periodic if f(x + P ) f(x) for all x R. We let be the set of complex valued -periodic functions f on

More information

The Hilbert Transform and Fine Continuity

The Hilbert Transform and Fine Continuity Irish Math. Soc. Bulletin 58 (2006), 8 9 8 The Hilbert Transform and Fine Continuity J. B. TWOMEY Abstract. It is shown that the Hilbert transform of a function having bounded variation in a finite interval

More information

Bayesian Methods with Monte Carlo Markov Chains II

Bayesian Methods with Monte Carlo Markov Chains II Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3

More information

Geometric ergodicity of the Bayesian lasso

Geometric ergodicity of the Bayesian lasso Geometric ergodicity of the Bayesian lasso Kshiti Khare and James P. Hobert Department of Statistics University of Florida June 3 Abstract Consider the standard linear model y = X +, where the components

More information

Markov Chains and De-initialising Processes

Markov Chains and De-initialising Processes Markov Chains and De-initialising Processes by Gareth O. Roberts* and Jeffrey S. Rosenthal** (November 1998; last revised July 2000.) Abstract. We define a notion of de-initialising Markov chains. We prove

More information

Asymptotics and Simulation of Heavy-Tailed Processes

Asymptotics and Simulation of Heavy-Tailed Processes Asymptotics and Simulation of Heavy-Tailed Processes Department of Mathematics Stockholm, Sweden Workshop on Heavy-tailed Distributions and Extreme Value Theory ISI Kolkata January 14-17, 2013 Outline

More information

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University Section 27 The Central Limit Theorem Po-Ning Chen, Professor Institute of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 3000, R.O.C. Identically distributed summands 27- Central

More information

SIMILAR MARKOV CHAINS

SIMILAR MARKOV CHAINS SIMILAR MARKOV CHAINS by Phil Pollett The University of Queensland MAIN REFERENCES Convergence of Markov transition probabilities and their spectral properties 1. Vere-Jones, D. Geometric ergodicity in

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Perturbed Proximal Gradient Algorithm

Perturbed Proximal Gradient Algorithm Perturbed Proximal Gradient Algorithm Gersende FORT LTCI, CNRS, Telecom ParisTech Université Paris-Saclay, 75013, Paris, France Large-scale inverse problems and optimization Applications to image processing

More information

Renewal theory and computable convergence rates for geometrically ergodic Markov chains

Renewal theory and computable convergence rates for geometrically ergodic Markov chains Renewal theory and computable convergence rates for geometrically ergodic Markov chains Peter H. Baxendale July 3, 2002 Revised version July 6, 2003 Abstract We give computable bounds on the rate of convergence

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015 ID NAME SCORE MATH 56/STAT 555 Applied Stochastic Processes Homework 2, September 8, 205 Due September 30, 205 The generating function of a sequence a n n 0 is defined as As : a ns n for all s 0 for which

More information

A central limit theorem for randomly indexed m-dependent random variables

A central limit theorem for randomly indexed m-dependent random variables Filomat 26:4 (2012), 71 717 DOI 10.2298/FIL120471S ublished by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat A central limit theorem for randomly

More information

SMSTC (2007/08) Probability.

SMSTC (2007/08) Probability. SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................

More information

Entropy for zero-temperature limits of Gibbs-equilibrium states for countable-alphabet subshifts of finite type

Entropy for zero-temperature limits of Gibbs-equilibrium states for countable-alphabet subshifts of finite type Entropy for zero-temperature limits of Gibbs-equilibrium states for countable-alphabet subshifts of finite type I. D. Morris August 22, 2006 Abstract Let Σ A be a finitely primitive subshift of finite

More information

LAW OF LARGE NUMBERS FOR THE SIRS EPIDEMIC

LAW OF LARGE NUMBERS FOR THE SIRS EPIDEMIC LAW OF LARGE NUMBERS FOR THE SIRS EPIDEMIC R. G. DOLGOARSHINNYKH Abstract. We establish law of large numbers for SIRS stochastic epidemic processes: as the population size increases the paths of SIRS epidemic

More information

CHAOS AND STABILITY IN SOME RANDOM DYNAMICAL SYSTEMS. 1. Introduction

CHAOS AND STABILITY IN SOME RANDOM DYNAMICAL SYSTEMS. 1. Introduction ØÑ Å Ø Ñ Ø Ð ÈÙ Ð Ø ÓÒ DOI: 10.2478/v10127-012-0008-x Tatra Mt. Math. Publ. 51 (2012), 75 82 CHAOS AND STABILITY IN SOME RANDOM DYNAMICAL SYSTEMS Katarína Janková ABSTRACT. Nonchaotic behavior in the sense

More information

A mathematical framework for Exact Milestoning

A mathematical framework for Exact Milestoning A mathematical framework for Exact Milestoning David Aristoff (joint work with Juan M. Bello-Rivas and Ron Elber) Colorado State University July 2015 D. Aristoff (Colorado State University) July 2015 1

More information

L n = l n (π n ) = length of a longest increasing subsequence of π n.

L n = l n (π n ) = length of a longest increasing subsequence of π n. Longest increasing subsequences π n : permutation of 1,2,...,n. L n = l n (π n ) = length of a longest increasing subsequence of π n. Example: π n = (π n (1),..., π n (n)) = (7, 2, 8, 1, 3, 4, 10, 6, 9,

More information

Complete q-moment Convergence of Moving Average Processes under ϕ-mixing Assumption

Complete q-moment Convergence of Moving Average Processes under ϕ-mixing Assumption Journal of Mathematical Research & Exposition Jul., 211, Vol.31, No.4, pp. 687 697 DOI:1.377/j.issn:1-341X.211.4.14 Http://jmre.dlut.edu.cn Complete q-moment Convergence of Moving Average Processes under

More information

Stochastic Process (ENPC) Monday, 22nd of January 2018 (2h30)

Stochastic Process (ENPC) Monday, 22nd of January 2018 (2h30) Stochastic Process (NPC) Monday, 22nd of January 208 (2h30) Vocabulary (english/français) : distribution distribution, loi ; positive strictement positif ; 0,) 0,. We write N Z,+ and N N {0}. We use the

More information

arxiv: v1 [math.pr] 17 May 2009

arxiv: v1 [math.pr] 17 May 2009 A strong law of large nubers for artingale arrays Yves F. Atchadé arxiv:0905.2761v1 [ath.pr] 17 May 2009 March 2009 Abstract: We prove a artingale triangular array generalization of the Chow-Birnbau- Marshall

More information

MARKOV CHAINS AND HIDDEN MARKOV MODELS

MARKOV CHAINS AND HIDDEN MARKOV MODELS MARKOV CHAINS AND HIDDEN MARKOV MODELS MERYL SEAH Abstract. This is an expository paper outlining the basics of Markov chains. We start the paper by explaining what a finite Markov chain is. Then we describe

More information

Particle Filters: Convergence Results and High Dimensions

Particle Filters: Convergence Results and High Dimensions Particle Filters: Convergence Results and High Dimensions Mark Coates mark.coates@mcgill.ca McGill University Department of Electrical and Computer Engineering Montreal, Quebec, Canada Bellairs 2012 Outline

More information