Weak invariance principles for sums of dependent random functions

Size: px
Start display at page:

Download "Weak invariance principles for sums of dependent random functions"

Transcription

1 Available online at Stochastic Processes and their Applications 13 (013) Weak invariance principles for sums of dependent random functions István Berkes a, Lajos Horváth b,, Gregory Rice b a Institute of Statistics, Graz University of Technology, Kopernikus Gasse 4/3, A 8010 Graz, Austria b Department of Mathematics, University of Utah, Salt Lake City, UT , USA Received 4 February 01; received in revised form 8 September 01; accepted 1 October 01 Available online 6 October 01 Abstract Motivated by problems in functional data analysis, in this paper we prove the weak convergence of normalized partial sums of dependent random functions exhibiting a Bernoulli shift structure. c 01 Published by lsevier B.V. MSC: 60F17; 60F5; 6G10 Keywords: Variables in Hilbert spaces; m approximability; Weak convergence 1. Introduction Functional data analysis in many cases requires central limit theorems and invariance principles for partial sums of random functions. The case of independent summands is much studied and well understood but the theory for the dependent case is less complete. In this paper we study the important class of Bernoulli shift processes which are often used to model econometric and financial data. Let X = {X i (t)} i= be a sequence of random functions, square integrable on [0, 1], and let denote the L [0, 1] norm. To lighten the notation we use f for f (t) when it does not cause confusion. Throughout this paper we assume that X forms a sequence of Bernoulli shifts, i.e. X j (t) = g(ϵ j (t), ϵ j 1 (t),...) for some nonrandom measurable function g : S L and iid random functions ϵ j (t), Research supported by SF grant DMS Corresponding author. -mail address: horvath@math.utah.edu (L. Horváth) /$ - see front matter c 01 Published by lsevier B.V. doi: /j.spa

2 386 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) and < j <, with values in a measurable space S, (1.1) ϵ j (t) = ϵ j (t, ω) is jointly measurable in (t, ω) ( < j < ), (1.) X 0 (t) = 0 for all t, and X 0 +δ < for some 0 < δ < 1, (1.3) the sequence {X n } n= can be approximated by l-dependent sequences {X n,l } n= in the sense that ( X n X n,l +δ ) 1/κ < for some κ > + δ, where X n,l is defined by X n,l = g(ϵ n, ϵ n 1,..., ϵ n l+1, ϵ n,l ), ϵ n,l = (ϵ n,l,n l, ϵ n,l,n l 1,...), where the ϵ n,l,k s are independent copies of ϵ 0, independent of {ϵ i, < i < }. (1.4) We note that assumption (1.1) implies that X n is a stationary and ergodic sequence. Hörmann and Kokoszka [19] call the processes satisfying (1.1) (1.4) L m-decomposable processes. The idea of approximating a stationary sequence with random variables which exhibit finite dependence first appeared in [1] and is used frequently in the literature (cf. [3]). Aue et al. [1] provide several examples when assumption (1.1) (1.4) hold which include autoregressive, moving average and linear processes in Hilbert spaces. Also, the non-linear functional ARCH(1) model (cf. [18]) and bilinear models (cf. [19]) satisfy (1.4). We show in Section (cf. Lemma.) that the series in C(t, s) = [X 0 (t)x 0 (s)] + [X 0 (t)x l (s)] + [X 0 (s)x l (t)] (1.5) are convergent in L. The function C(t, s) is positive definite, and therefore there exist λ 1 λ 0 and orthonormal functions φ i (t), 0 t 1 satisfying λ i φ i (t) = C(t, s)φ i (s)ds, 1 i <, (1.6) where means 1 0. We define Γ (x, t) = λ 1/ i W i (x)φ i (t), where W i are independent and identically distributed Wiener processes (standard Brownian motions). Clearly, Γ (x, t) is Gaussian. We show in Lemma. that λ l <, and therefore Γ (x, t)dt < a.s. sup 0 x 1 Theorem 1.1. If assumption (1.1) (1.4) hold, then for every we can define a Gaussian process Γ (x, t) such that {Γ (x, t), 0 x, t 1} D ={Γ (x, t), 0 x, t 1}

3 and where sup 0 x 1 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) (S (x, t) Γ (x, t)) dt = o P (1), S (x, t) = 1 x 1/ X i (t). The proof of Theorem 1.1 is given in Section. The proof is based on a maximal inequality which is given in Section 3 and is of interest in its own right. There is a wide literature on the central limit theorem for sums of random processes in abstract spaces. For limit theorems for sums of independent Banach space valued random variables we refer to Ledoux and Talagrand (1991). For the central limit theory in the context of functional data analysis we refer to the books of [4,0]. In the real valued case, the martingale approach to weak dependence was developed by Gordin [16], Philipp and Stout [5] and berlein [14], and by using such techniques [3,9] obtained central limit theorems for a large class of dependent variables in Hilbert spaces. For some early influential results on invariance for sums of mixing variables in Banach spaces we refer to [,11,10]. These papers provide very sharp results, but verifying mixing conditions is generally not easy and without additional continuity conditions, even autoregressive (1) processes may fail to be strong mixing (cf. [5]). The weak dependence concept of [13] (cf. also [8]) solves this difficulty, but so far this concept has not been extended to variables in Hilbert spaces. Wu [7,8] proved several limit theorems for one-dimensional stationary processes having a Bernoulli shift representation. Compared to classical mixing conditions, Wu s physical dependence conditions are easier to verify in concrete cases. Condition (1.3) cannot be directly compared to the approximating martingale conditions of [7,8]. For extensions to the Hilbert space case we refer to [19].. Proof of Theorem 1.1 The proof is based on three steps. We recall the definition of X i,m from (1.4). For every fixed m, the sequence {X i,m } is m-dependent. According to our first lemma, the sums of the X i s can be approximated with the sums of m-dependent variables. The second step is the approximation of the infinite dimensional X i,m s with finite dimensional variables (Lemma.4). Then the result in Theorem 1.1 is established for finite dimensional m-dependent random functions (Lemma.6). Lemma.1. If (1.1) (1.4) hold, then for all x > 0 we have 1 k lim lim sup P max (X i X i,m ) m 1 k > x = 0. (.1) Proof. The proof of this lemma requires the maximal inequality of Theorem 3.. Section 3 is devoted to the proof of this result. Using Theorem 3., (.1) is an immediate consequence of Markov s inequality. Define m m C m (t, s) = [X 0,m (t)x 0,m (s)] + [X 0,m (t)x i,m (s)] + [X 0,m (s)x i,m (t)]. (.)

4 388 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) We show in the following lemma that for every m the function C m is square-integrable. Hence there are λ 1,m λ,m 0 and corresponding orthonormal functions φ i,m, i = 1,,... satisfying λ i,m φ i,m (t) = C m (t, s)φ i,m (s)ds, i = 1,,... Lemma.. If (1.1) (1.4) hold, then we have C (t, s)dtds <, and (.3) Cm (t, s)dtds < for all m 1, (.4) lim m (C(t, s) C m (t, s)) dtds = 0, (.5) C(t, t)dt = λ k <, (.6) C m (t, t)dt = lim m k=1 λ k,m < (.7) k=1 C m (t, t)dt = C(t, t)dt. (.8) Proof. Using the Cauchy Schwarz inequality for expected values we get ([X 0 (t)x 0 (s)]) dtds (( X0 (t))1/ ( X0 (s))1/ ) dtds = ( X 0 ) <. Recalling that X 0 and X i,i are independent and both have 0 mean, we conclude first using the triangle inequality and then the Cauchy Schwarz inequality for expected values that 1/ [X 0 (t)x i (s)] dtds = [X 0 (t)(x i (s) X i,i (s))] dtds X 0 (t)(x i (s) X i,i (s)) dtds 1/ 1/ X0 (t)(x i (s) X 1/ i,i(s)) dtds

5 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) = ( X 0 (t))1/ ((X i (s) X i,i (s)) ) 1/ dtds X0 (t)dt = X 0 X 0 X 0,i < (X i (s) X i,i (s)) ds (.9) on account of (1.4). This completes the proof of (.3). Since X 0,m (t)x 0,m (s) = X 0 (t)x 0 (s), in order to establish (.4), it is enough to show that m [X 0,m (t)x i,m (s)] dtds <. It follows from the definition of X i,m that the vectors (X 0,m, X i,m ) and (X 0, X i,m ) have the same distribution for all 1 i m. Also, (X i,m, X i,i ) has the same distribution as (X 0, X 0,i ), 1 i m. Hence following the arguments in (.9) we get 1/ m X 0,m (t)x i,m (s) dtds m = X 0 (t)x i,m (s) dtds X 0 m (X i,m (s) X i,i (s)) ds X 0 X 0 X 0,i <. The proof of (.4) is now complete. The arguments used above also prove (.5). Repeating the previous arguments we have C(t, t)dt = = 1/ X0 (t)dt + [X 0 (t)x i (t)] dt X0 (t)dt + [X 0 (t)(x i (t) X i,i (t))] dt X0 (t)dt + ( X0 (t))1/ ([X i (t) X i,i (t)] ) 1/ dt X 0 + X 0 (t)dt 1/ 1/ [X i (t) X i,i (t)] dt

6 390 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Observing that C(t, t)dt = = X 0 + ( X 0 ) 1/ ( X 0 X 0,i ) 1/ <. λ i φi (t)dt = λ i, the proof of (.6) is complete. The same arguments can be used to establish (.7). The relation in (.8) can be established along the lines of the proof of (.5). By the Karhunen Loéve expansion, we have that Define X i,m (t) = X i,m,k (t) = X i,m, φ l,m φ l,m (t). (.10) K X i,m, φ l,m φ l,m (t) (.11) to be the partial sums of the series in (.10), and X i,m,k (t) = X i,m (t) X i,m,k (t) = X i,m, φ l,m φ l,m (t). (.1) Lemma.3. If {Z i } are independent L valued random variables such that Z 1 (t) = 0 and Z 1 <, (.13) then for all x > 0 we have that P max k Z 1 k i > x 1 x Z i. (.14) Proof. Let F k be the sigma algebra generated by the random variables {Z j } k j=1. By assumption (.13) and the independence of the Z i s we have that k+1 Z i F k k = Z i + Z k+1 k Z i. Therefore k is a non-negative submartingale with respect to the filtration Z i k=1 {F k } k=1. If we define A = ω : max k Z 1 k i > x,

7 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) then it follows from Doob s maximal inequality [6, p. 47] that x P max k Z 1 k i > x Z i I A which completes the proof. Z i, Lemma.4. If (1.1) (1.4) hold, then for all x > 0, 1 k lim lim sup P max X K 1 k i,m,k > x = 0. (.15) Proof. Define Q k ( j) = {i : 1 i k, i = j (mod m)} for j = 0, 1,..., m 1, and all positive integers k. It is then clear that k X i,m,k = m 1 j=0 i Q k ( j) X i,m,k. We thus obtain by the triangle inequality that 1 k P max X 1 k i,m,k > x P m 1 j=0 max 1 k 1 It is therefore sufficient to show that for each fixed j, 1 lim lim sup P max X K 1 k i,m,k > x = 0. i Q k ( j) i Q k ( j) X i,m,k > x By the definition of Q k ( j), { X i,m,k } i Qk ( j) is an iid sequence of random variables. So, by applications of Lemma.3 and assumption (1.3), we have that P max 1 X 1 k i,m,k > x 1 x 1 X i,m,k i Q k ( j) 1 x X 0,m,K = 1 x i Q ( j) λ l,m. (.16) Since the right hand side of (.16) tends to zero as K tends to infinity independently of, (.15) follows. Clearly, with k = x we have 1 k K 1 x X i,m,k (t) = X i,m, φ j,m φ j,m (t). (.17) j=1.

8 39 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Lemma.5. If (1.1) (1.4) hold, then the K dimensional random process 1 x 1 x X i,m, φ 1,m,..., X i,m, φ K,m converges, as, in D K [0, 1] to λ 1/ 1,m W 1(x),..., λ 1/ K,m W K (x), (.18) where {W i } K are independent, identically distributed Wiener processes. 1 x Proof. A similar procedure as in Lemma.4 shows that for each j, X i,m, φ j,m can be written as a sum of sums of independent and identically distributed random variables, and thus, by Billingsley [3], it is tight. This implies that the K dimensional process 1 x 1 x X i,m, φ 1,m,..., X i,m, φ K,m is tight, since it is tight in each coordinate. Furthermore, the Cramér Wold device and the central limit theorem for m-dependent random variables (cf. [7, p. 119]) shows that the finite dimensional distributions of the vector process converge to the finite dimensional distributions of the process in (.18). The lemma follows. In light of the Skorokhod Dudley Wichura theorem (cf. [6, p. 47]), we may reformulate Lemma.5 as follows. Corollary.1. If (1.1) (1.4) hold, then for each positive integer, there exists K independent, identically distributed Wiener processes {W i, } K such that for each j, 1 x sup X i,m, φ j,m λ 1/ j,m W P j, (x) 0, 0 x 1 as. Lemma.6. If (1.1) (1.4) hold, then for {W i, } K defined in Corollary.1, we have that 1 x sup 0 x 1 as. X i,m,k (t) Proof. By using (.17), we get that K 1 x K X i,m,k (t) λ 1/ l,m W l, (x)φ l,m (t) = K 1 x X i,m, φ l,m λ 1/ l,m W l, (x) φ l,m (t). λ 1/ l,m W P l, (x)φ l,m (t) dt 0, (.19)

9 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) The substitution of this into the expression in (.19) along with a simple calculation shows that 1 x sup X i,m,k (t) 0 x 1 K 1 x = sup 0 x 1 K sup 0 x 1 as, by Corollary.1. K λ 1/ l,m W l, (x)φ l,m (t) dt X i,m, φ l,m λ 1/ l,m W l, (x) 1 x X i,m, φ l,m λ 1/ l,m W l, (x) Lemma.7. If (1.1) (1.4) hold, sup 0 x 1 λ 1/ l,m W l(x)φ l,m (t) dt P 0, P 0, (.0) as K, where W 1, W,... are independent and identically distributed Wiener processes. Proof. Since the functions {φ l,m } are orthonormal, we have that sup λ 1/ l,m W l(x)φ l,m (t) dt = sup 0 x 1 0 x 1 as K. Therefore (.0) follows from Markov s inequality. λ l,m W l (x) λ l,m sup Wl (x) 0, 0 x 1 Lemma.8. If (1.1) (1.4) hold, then for each we can define independent identically distributed Wiener processes {W i, } K such that 1 x sup 0 x 1 as. X i,m (t) Proof. It follows from Lemmas.4.7. λ 1/ l,m W P l, (x)φ l,m (t) dt 0, Since the distribution of W l,, 1 l < does not depend on, it is enough to consider the asymptotics for λ 1/ l,m W l(x)φ l,m (t), where W l are independent Wiener processes. Lemma.9. If (1.1) (1.4) hold, then for each m we can define independent and identically distributed Wiener processes W l,m (x), 1 l < such that sup 0 x 1 as m. λ 1/ l,m W l(x)φ l,m (t) λ 1/ l W l,m (x)φ l (t) dt P 0, (.1)

10 394 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Proof. Let m (x, t) = λ 1/ l,m W l(x)φ l,m (t). Let M be a positive integer and define x i = i/m, 0 i M. It is easy to see that max sup ( m (x i + h, t) m (x i, t)) dt 0 i<m 0 h 1/M = λ l,m max max 0 i<m sup 0 i<m 0 h 1/M (W l (x i + h) W l (x i )) 0 h 1/M(W sup 1 (x i + h) W 1 (x i )) λ l,m. Using Lemma. we get that λ l,m = m (1, t)dt = C m (t, t)dt C(t, t)dt = λ l. So by the modulus of continuity of the Wiener process (cf. [15]) we get that lim lim sup max sup ( m (x i + h, t) m (x i, t)) dt = 0. (.) M m 0 i<m 0 h 1/M By the Karhunen Loéve expansion we can also write m as m (x, t) = m (x, ), φ l φ l (t) and m (x, t)dt = ( m (x, ), φ l ). So by Lemma. we have ( m (x, ), φ l ) x λ l. Also, for any positive integer l, ( m (x, ), φ l ) = C m (t, s)φ l (t)φ l (s)dtds C(t, s)φ l (t)φ l (s)dtds = λ l, as m. Hence for every z > 0 we have lim sup lim sup P m (x, ), φ l φ l (t) dt > z = 0. (.3) K m The joint distribution of (x i, ), φ l, 1 i M, 1 l K is multivariate normal with zero mean. Hence they converge jointly to a multivariate normal distribution. To show their joint

11 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) convergence in distribution, we need to show the convergence of the covariance matrix. Using again Lemma. we get that (x i, ), φ l (x j, ), φ k = min(x i, x j ) C m (t, s)φ l (t)φ k (s)dtds min(x i, x j ) C(t, s)φ l (t)φ k (s)dtds = min(x i, x j )λ l I {k = l}. Due to this covariance structure and the Skorokhod Dudley Wichura theorem (cf. [6, p. 47]) we can find independent Wiener processes W l,m (x), 1 l < such that max 1 i M max (x i, ), φ l λ 1/ l W l,m (x i ) = o P (1), as m. 1 l K Clearly, for all 0 x 1 W l,m (x)φ l (t) dt = x λ 1/ l and therefore similarly to (.3) lim sup lim sup P K m λ 1/ l for all z > 0. Similarly to (.) one can show that max sup 0 i<m 0 h 1/M max 0 i<m λ l 0, as m, W l,m (x)φ l (t) dt > z = 0 ( W l,m (x i + h) W l,m (x i ))φ l (t) dt 0 h 1/M(W sup (x i + h) W (x i )) λ l 0, as M, where W is a Wiener process. This also completes the proof of Lemma.9. Proof of Theorem 1.1. First we approximate S (x,t) with m-dependent processes (Lemma.1). The second step of the proof is the approximation of the sums of m-dependent processes with a Gaussian process with covariance function min(x, x )C m (t, s), where C m is defined in (.) (Lemma.8). The last step of the proof is the convergence of Gaussian processes with covariance functions min(x, x )C m (t, s) to a Gaussian process with covariance function min(x, x )C(t, s) (Lemma.9). 3. Some moment and maximal inequalities In this section we give the proof of the maximal inequality used in Lemma.1 which is a crucial ingredient of the proof of Theorem 1.1. Actually, we will prove below some moment and maximal inequalities for partial sums of function valued Bernoulli shift sequences which have their own interest and can be used in various related problems. Our first lemma is a Hilbert space version of Doob s [1, p. 6] inequality.

12 396 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Lemma 3.1. If Z 1 and Z are independent mean zero Hilbert space valued random variables, and if 0 < δ 1, then Z 1 + Z +δ Z 1 +δ + Z +δ + Z 1 ( Z ) δ/ + Z ( Z 1 ) δ/. Proof. Since 0 < δ 1, for any A, B 0 we have that (A + B) δ A δ + B δ (cf. [17, p. 3]). An application of this inequality along with Minkowski s inequality gives that Z 1 + Z δ ( Z 1 + Z ) δ Z 1 δ + Z δ. We also have by Hölder s inequality that Z 1 δ ( Z 1 ) δ/. This yields that Z 1 + Z +δ = Z 1 + Z Z 1 + Z δ which proves the lemma. Z 1 + Z ( Z 1 δ + Z δ ) = [ Z 1 + Z + Z 1, Z ]( Z 1 δ + Z δ ) = Z 1 +δ + Z +δ + Z 1 Z δ + Z Z 1 δ Z 1 +δ + Z +δ + Z 1 ( Z ) δ/ + Z ( Z 1 ) δ/, Remark 3.1. If Z 1 and Z are independent and identically distributed, then the result of Lemma 3.1 can be written as Let Z 1 + Z +δ Z 1 +δ + ( Z 1 ) 1+δ/. I (r) = ( X 0 X 0,l r ) 1/r. (3.1) We note that by (1.4), I (r) < for all r + δ. Lemma 3.. If (1.1) (1.4) hold, then we have n (X i X i,m ) n A, where A = (X 0 (t) X 0,m (t)) dt + 5/ (X 0 (t) X 0,m (t)) dt 1/ I (). (3.) Proof. Let i = X i X i,m. By Fubini s theorem and the fact that the random variables are identically distributed, we conclude

13 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) n n i = i (t) dt n 1 = n 0 (t)dt + (n i)[ 0 (t) i (t)]dt n n 0 n 1 (t)dt + n [ 0 (t) i (t)] dt 0 (t)dt + n [ 0 (t) i (t)] dt. (3.3) We recall X i,i from (1.4). Under this definition, the random variables 0 and X i,i are independent for all i 1. Let Z i = X i,m, if i > m and Z i = g(ϵ i,..., ϵ 1, δ i ), if 1 i m, where δ i = (δ i,0, δ i, 1,...) and δ i, j are iid copies of ϵ 0, independent of the ϵ l s and ϵ k,l s. Clearly, Z i and 0 are independent and thus with i,i = X i,i Z i we have [ 0 (t) i (t)] = [ 0 (t)( i (t) i,i (t))]. Furthermore, by first applying the Cauchy Schwarz inequality for expected values and then by the Cauchy Schwarz inequality for functions in L, we get that 1/ [ 0 (t)( i (t) i,i (t))] dt 0 (t) i (t) i,i (t) 1/ dt 1/ 0 (t)dt i (t) i,i (t) 1/ dt. Also, i (t) i,i (t) dt X i (t) X i,i (t) dt + X i,m (t) Z i (t) dt. The substitution of this expression into (3.3) gives that n 1/ i n 0 (t)dt + 3/ n 0 (t)dt X i (t) X 1/ i,i(t) dt + n X i,m (t) Z i (t) dt 1/ 0 (t)dt + 5/ 0 (t)dt 1/ I (), which completes the proof.

14 398 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Theorem 3.1. If (1.1) (1.4) hold, then for all 1 +δ (X i X i,m ) 1+δ/ B, where B = X 0 X 0,m +δ + c +δ δ [A 1+δ/ + J +δ m + J m A (1+δ)/ + A (1+δ/)δ J m ] + (c δ J m )1/(1 δ) with A defined in (3.), c δ = δ/ (3.5) and J m = ( X 0 X 0,m +δ ) (κ δ)/(κ(+δ)) ( X 0 X 0,l +δ ) 1/κ. Proof. We prove Theorem 3.1 using mathematical induction. By the definition of B, the inequality is obvious when = 1. Assume that it holds for all k which are less than or equal to 1. We assume that is even, i.e. = n. The case when is odd can be done in the same way with minor modifications. Let i = X i X i,m. For all i satisfying n + 1 i n, we define X i,n = g(ϵ i, ϵ i 1,..., ϵ n+1, ϵ n, ϵ n 1,...) where the ϵ j s denote iid copies of ϵ 0, independent of {ϵ i, < i < } and {ϵ k,l, < k, l < }. We define Z i,n = X i,m, if m + n + 1 i n and Z i,n = g(ϵ i,..., ϵ n+1, ϵ n,... ϵ i m+1, δ i) with δ i = (δ i,n, δ i,n 1,...), if n + 1 i n + m, where the δ k,l s are iid copies of ϵ 0, independent of the ϵ k s and ϵ k,l s. Let i,n = X i,n Z i,n, if n + 1 i n. Under this definition, the sequences { i, 1 i n} and {i,n, n + 1 i n} are independent and have the same distribution. Let n n Θ = i + i,n n i,n and Ψ = i. i=n+1 i=n+1 By applying the triangle inequality for L and expected values, we get n +δ n n n i = i + i,n + i i,n +δ i=n+1 i=n+1 (3.4) (Θ + Ψ) +δ (Θ +δ ) 1/(+δ) + (Ψ +δ ) 1/(+δ) +δ. (3.6) The two-term Taylor expansion gives for all a, b 0 and r > that (a + b) r a r + ra r 1 b + r(r 1) (a + b) r b. (3.7)

15 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) Since both of the expected values in the last line of the inequality in (3.6) are positive, we obtain by (3.7) that n +δ i Θ +δ + ( + δ)(θ +δ ) (1+δ)/(+δ) (Ψ +δ ) 1/(+δ) + ( + δ)(1 + δ) (Θ +δ ) 1/(+δ) + (Ψ +δ ) 1/(+δ) δ (Ψ +δ ) /(+δ). (3.8) We proceed by bounding the terms (Ψ +δ ) 1/(+δ), and Θ +δ individually. Applications of both the triangle inequality for L and for expected values yield that (Ψ +δ ) 1/(+δ) n = i +δ 1/(+δ) i,n n i=n+1 i=n+1 +δ i i,n i=n+1 n ( i i,n +δ ) 1/(+δ). By Hölder s inequality we have, with κ in (1.4), 1/(+δ) ( i i,n +δ ) 1/(+δ) = ([ i i,n (+δ) /κ i i,n (+δ) (+δ) /κ ]) 1/(+δ) ( i i,n +δ ) 1/κ ( i i,n +δ ) (κ δ)/(κ(+δ)). It follows from the definition of i, i,n and the convexity of x+δ that and Thus we get i i,n +δ 1+δ ( X i X i,n +δ + X i,m Z i,n +δ ) +δ X 0 X 0,i n +δ i i,n +δ 1+δ ( X i X i,m +δ + X i,n Z i,n +δ ) +δ X 0 X 0,m +δ. (Ψ +δ ) 1/(+δ) ( X 0 X 0,m +δ ) (κ δ)/(κ(+δ)) ( X 0 X 0,l +δ ) 1/κ = J m. To bound Θ +δ, since n i and n i=n+1 i,n are independent and have the same distribution, we have by Lemma 3., Remark 3.1 and the inductive assumption that Θ +δ n n +δ = i + i,n i=n+1

16 400 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) n +δ n i + i n 1+δ/ B + (n A) 1+δ/. The substitution of these two bounds into (3.8) give that n +δ i n 1+δ/ B + (n A) 1+δ/ 1+δ/ + ( + δ)[n 1+δ/ B + (n A) 1+δ/ ] (1+δ)/(+δ) J m δ + ( + δ)(1 + δ) n 1+δ/ B + (n A) 1+δ/ + J m J m. (3.9) Furthermore, by the definition of B, we may further bound each summand on the right hand side of (3.9). We obtain for the first two terms that n 1+δ/ B + (n A) 1+δ/ (n) 1+δ/ B δ/ + A1+δ/ B (n) 1+δ/ B δ/ + 6cδ 1. A similar factoring procedure applied to the expression in the second line of (3.9) yields that ( + δ) n 1+δ/ B + (n A) 1+δ/ (1+δ)/(+δ) Jm 6 (n 1+δ/ B) (1+δ)/(+δ) + (n A) (1+δ/)[(1+δ)/(+δ)] J m (n) 1+δ/ 6J m B B 1/(+δ) + 6J m A (1+δ/)[(1+δ)/(+δ)] B (n) 1+δ/ B. 1c 1 δ Since 0 < δ < 1, the expression in the third line of (3.9) may be broken into three separate terms: δ ( + δ)(1 + δ) n 1+δ/ B + (n A) 1+δ/ + J m J m 6(n 1+δ/ B) δ J m + 6((n A)(1+δ/) ) δ J m + 6J +δ m. Furthermore by again applying the definition of B we have that 6(n 6(n 1+δ/ B) δ Jm 1+δ/ = B) δ J (n)1+δ/ m B (n) 1+δ/ B 6J (n) 1+δ/ B m B 1 δ (n) 1+δ/ B[6c 1 δ ], 6((n A) (1+δ/) ) δ J m = (n)1+δ/ B 6((n A) (1+δ/) ) δ J m (n) 1+δ/ B

17 and 6J +δ m I. Berkes et al. / Stochastic Processes and their Applications 13 (013) (n) 1+δ/ 6A (1+δ/)δ Jm B B (n) 1+δ/ B[6c 1 δ ], = (n) 1+δ/ 6J +δ m B (n) 1+δ/ B (n) 1+δ/ B The application of these bounds to the right hand side of (3.9) give that n +δ i (n) 1+δ/ B δ/ + 36cδ 1 = (n) 1+δ/ B, which concludes the induction step and thus the proof. 6J +δ m (n) 1+δ/ B[6cδ 1 ]. B Theorem 3.. If (1.1) (1.4) hold, then we have k +δ (X i X i,m ) a m 1+δ/ (3.10) max 1 k with some sequence a m satisfying a m 0 as m. Proof. By examining the proofs, it is evident that Theorem 3.1 in [4] holds for L valued random variables. Furthermore, by the stationarity of the sequence {X i X i,m } and Theorem 3.1, the conditions of Theorem 3.1 in Móricz are satisfied and therefore k +δ (X i X i,m ) c δ 1+δ/ B, max 1 k with some constant c δ, depending only on δ and B is defined in (3.4). Observing that B = B m 0 as m, the result is proven. Theorem 3.1 provides inequality for the moments of the norm of partial sums of X i X i,m which are not Bernoulli shifts. However, checking the proof of Theorem 3.1, we get the following result for Bernoulli shifts. Theorem 3.3. If (1.1), (1.3) are satisfied and X is a Bernoulli shift satisfying I ( + δ) = ( X 0 X 0,l +δ ) 1/(+δ) < with some 0 < δ < 1, where X 0,l is defined by (1.4), then for all 1 +δ X i 1+δ/ B, where B = X 0 +δ + cδ +δ [A 1+δ/ + I ( + δ)a (1+δ)/ + I +δ ( + δ) + A (1+δ/)δ I ( + δ)] + (c δ I ()) 1/(1 δ),

18 40 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) A = X 0 (t)dt + and c δ is defined in (3.5) and I () in (3.1). X 0 (t)dt 1/ I () Remark 3.. The inequality in Theorem 3.1 is an extension of Proposition 4 in [] to random variables in Hilbert spaces; we have computed how B depends on the distribution of X explicitly. Acknowledgment We are grateful to Professor Herold Dehling for pointing out several important references. References [1] A. Aue, S. Hörmann, L. Horváth, M. Hušková, J. Steinebach, Sequential testing for the stability of portfolio betas, conometric Theory 8 (01) [] I. Berkes, S. Hörmann, J. Schauer, Split invariance principles for stationary processes, Annals of Probability 39 (011) [3] P. Billingsley, Convergence of Probability Measures, Wiley, ew ork, [4] D. Bosq, Linear Processes in Function Spaces, Theory and Applications, in: Lecture otes in Statistics, vol. 149, Springer-Verlag, ew ork, 000. [5] R.C. Bradley, Introduction to Strong Mixing Conditions I III, Kendrick Press, Heber City UT, 007. [6]. Chow, H. Teicher, Probability Theory: Independence, Interchangeability, Martingales, Springer-Verlag, Heidelberg, [7] A. DasGupta, Asymptotic Theory of Statistics and Probability, Springer, ew ork, 008. [8] J. Dedecker, P. Doukhan, G. Lang, R.J.R. León, S. Louhichi, C. Prieur, Weak Dependence: With xamples and Applications, in: Lecture otes in Statistics, vol. 190, Springer, ew ork, 007, p. xiv+318. [9] J. Dedecker, F. Merlevède, The conditional central limit theorem in Hilbert spaces, Stochastic Processes and Their Applications 108 (003) 9 6. [10] H. Dehling, Limit theorems for sums of weakly dependent Banach space valued random variables, Zeitschrift fur Wahrscheinlichkeitstheorie und Verwandte Gebiete 63 (1983) [11] H. Dehling, W. Philip, Almost sure invariance principles for weakly dependent vector-valued random variables, Annals of Probability 10 (198) [1] J. Doob, Stochastic Processes, Wiley, ew ork, [13] P. Doukhan, S. Louhichi, A new weak dependence condition and applications to moment inequalities, Stochastic Processes and Their Applications 84 (1999) [14]. berlein, An invariance principle for lattices of dependent random variables, Zeitschrift fur Wahrscheinlichkeitstheorie und Verwandte Gebiete 50 (1979) [15] A.M. Garsia, Continuity properties of Gaussian processes with multidimensional time parameter, in: Proceedings of the 6th Berkeley Symposium Mathematical Statistics and Probability, vol., University of California, Berkeley, 1970, pp [16] M.I. Gordin, The central limit theorem for stationary processes, Doklady Akademii auk SSSR 188 (1969) (in Russian). [17] G.H. Hardy, J.. Littlewood, G. Pólya, Inequalities, second ed., Cambridge University Press, [18] S. Hörmann, L. Horváth, R. Reeder, A functional version of the ARCH model, conometric Theory 01 (in press). [19] S. Hörmann, P. Kokoszka, Weakly dependent functional data, Annals of Statistics 38 (010) [0] L. Horváth, P. Kokoszka, Inference for Functional Data with Applications, Springer, ew ork, 01. [1] I.A. Ibragimov, Some limit theorems for stationary processes, Theory of Probabability and Its Applications 7 (196) [] J. Kuelbs, W. Philipp, Almost sure invariance principles for partial sums of mixing B-valued random variables, Annals of Probability 8 (1980) [3] F. Merlevède, Central limit theorem for linear processes with values in a Hilbert space, Stochastic Processes and Their Applications 65 (1996) [4] F. Móricz, R. Serfling, W. Stout, Moment and probability bounds with quasi-superadditive structure for the maximal partial sum, Annals of Probability 10 (198)

19 I. Berkes et al. / Stochastic Processes and their Applications 13 (013) [5] W. Philipp, W. Stout, Almost sure invariance principles for partial sums of weakly dependent random variables, Memoirs of the American Mathematical Society 161 (1975) iv+140. [6] G. Shorack, J. Wellner, mpirical Processes With Applications To Statistics, Wiley, ew ork, [7] W. Wu, onlinear system theory: another look at dependence, Proceedings of the ational Academy of Sciences of the USA 10 (005) [8] W. Wu, Strong invariance principles for dependent random variables, Annals of Probability 35 (007)

An almost sure invariance principle for additive functionals of Markov chains

An almost sure invariance principle for additive functionals of Markov chains Statistics and Probability Letters 78 2008 854 860 www.elsevier.com/locate/stapro An almost sure invariance principle for additive functionals of Markov chains F. Rassoul-Agha a, T. Seppäläinen b, a Department

More information

ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR

ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR Periodica Mathematica Hungarica Vol. 51 1, 2005, pp. 11 25 ALMOST SURE CONVERGENCE OF THE BARTLETT ESTIMATOR István Berkes Graz, Budapest, LajosHorváth Salt Lake City, Piotr Kokoszka Logan Qi-man Shao

More information

Mi-Hwa Ko. t=1 Z t is true. j=0

Mi-Hwa Ko. t=1 Z t is true. j=0 Commun. Korean Math. Soc. 21 (2006), No. 4, pp. 779 786 FUNCTIONAL CENTRAL LIMIT THEOREMS FOR MULTIVARIATE LINEAR PROCESSES GENERATED BY DEPENDENT RANDOM VECTORS Mi-Hwa Ko Abstract. Let X t be an m-dimensional

More information

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching Communications for Statistical Applications and Methods 2013, Vol 20, No 4, 339 345 DOI: http://dxdoiorg/105351/csam2013204339 A Functional Central Limit Theorem for an ARMAp, q) Process with Markov Switching

More information

GARCH( 1, 1) processes are near epoch dependent

GARCH( 1, 1) processes are near epoch dependent Economics Letters 36 (1991) 181-186 North-Holland 181 GARCH( 1, 1) processes are near epoch dependent Bruce E. Hansen 1Jnrrwsity of Rochester, Rochester, NY 14627, USA Keceived 22 October 1990 Accepted

More information

An empirical Central Limit Theorem in L 1 for stationary sequences

An empirical Central Limit Theorem in L 1 for stationary sequences Stochastic rocesses and their Applications 9 (9) 3494 355 www.elsevier.com/locate/spa An empirical Central Limit heorem in L for stationary sequences Sophie Dede LMA, UMC Université aris 6, Case courrier

More information

Robustní monitorování stability v modelu CAPM

Robustní monitorování stability v modelu CAPM Robustní monitorování stability v modelu CAPM Ondřej Chochola, Marie Hušková, Zuzana Prášková (MFF UK) Josef Steinebach (University of Cologne) ROBUST 2012, Němčičky, 10.-14.9. 2012 Contents Introduction

More information

STAT 331. Martingale Central Limit Theorem and Related Results

STAT 331. Martingale Central Limit Theorem and Related Results STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal

More information

Convergence at first and second order of some approximations of stochastic integrals

Convergence at first and second order of some approximations of stochastic integrals Convergence at first and second order of some approximations of stochastic integrals Bérard Bergery Blandine, Vallois Pierre IECN, Nancy-Université, CNRS, INRIA, Boulevard des Aiguillettes B.P. 239 F-5456

More information

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS SOME CONVERSE LIMIT THEOREMS OR EXCHANGEABLE BOOTSTRAPS Jon A. Wellner University of Washington The bootstrap Glivenko-Cantelli and bootstrap Donsker theorems of Giné and Zinn (990) contain both necessary

More information

On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables

On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables On the Set of Limit Points of Normed Sums of Geometrically Weighted I.I.D. Bounded Random Variables Deli Li 1, Yongcheng Qi, and Andrew Rosalsky 3 1 Department of Mathematical Sciences, Lakehead University,

More information

Almost sure limit theorems for U-statistics

Almost sure limit theorems for U-statistics Almost sure limit theorems for U-statistics Hajo Holzmann, Susanne Koch and Alesey Min 3 Institut für Mathematische Stochasti Georg-August-Universität Göttingen Maschmühlenweg 8 0 37073 Göttingen Germany

More information

Change detection problems in branching processes

Change detection problems in branching processes Change detection problems in branching processes Outline of Ph.D. thesis by Tamás T. Szabó Thesis advisor: Professor Gyula Pap Doctoral School of Mathematics and Computer Science Bolyai Institute, University

More information

1. A remark to the law of the iterated logarithm. Studia Sci. Math. Hung. 7 (1972)

1. A remark to the law of the iterated logarithm. Studia Sci. Math. Hung. 7 (1972) 1 PUBLICATION LIST OF ISTVÁN BERKES 1. A remark to the law of the iterated logarithm. Studia Sci. Math. Hung. 7 (1972) 189-197. 2. Functional limit theorems for lacunary trigonometric and Walsh series.

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

Sequential block bootstrap in a Hilbert space with application to change point analysis

Sequential block bootstrap in a Hilbert space with application to change point analysis Sequential block bootstrap in a Hilbert space with application to change point analysis Olimjon Sharipov, Johannes Tewes and Martin Wendler arxiv:42.0446v [math.st] Dec 204 December 2, 204 Abstract A new

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Estimation of the mean of functional time series and a two-sample problem

Estimation of the mean of functional time series and a two-sample problem J. R. Statist. Soc. B (01) 74, Part 5, pp. Estimation of the mean of functional time series and a two-sample problem Lajos Horváth, University of Utah, Salt Lake City, USA Piotr okoszka Colorado State

More information

FOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1

FOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1 FOURIER TRANSFORMS OF STATIONARY PROCESSES WEI BIAO WU September 8, 003 Abstract. We consider the asymptotic behavior of Fourier transforms of stationary and ergodic sequences. Under sufficiently mild

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

Packing-Dimension Profiles and Fractional Brownian Motion

Packing-Dimension Profiles and Fractional Brownian Motion Under consideration for publication in Math. Proc. Camb. Phil. Soc. 1 Packing-Dimension Profiles and Fractional Brownian Motion By DAVAR KHOSHNEVISAN Department of Mathematics, 155 S. 1400 E., JWB 233,

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

Gaussian Processes. 1. Basic Notions

Gaussian Processes. 1. Basic Notions Gaussian Processes 1. Basic Notions Let T be a set, and X : {X } T a stochastic process, defined on a suitable probability space (Ω P), that is indexed by T. Definition 1.1. We say that X is a Gaussian

More information

QED. Queen s Economics Department Working Paper No. 1244

QED. Queen s Economics Department Working Paper No. 1244 QED Queen s Economics Department Working Paper No. 1244 A necessary moment condition for the fractional functional central limit theorem Søren Johansen University of Copenhagen and CREATES Morten Ørregaard

More information

Martingale approximations for random fields

Martingale approximations for random fields Electron. Commun. Probab. 23 (208), no. 28, 9. https://doi.org/0.24/8-ecp28 ISSN: 083-589X ELECTRONIC COMMUNICATIONS in PROBABILITY Martingale approximations for random fields Peligrad Magda * Na Zhang

More information

PACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION

PACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION PACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION DAVAR KHOSHNEVISAN AND YIMIN XIAO Abstract. In order to compute the packing dimension of orthogonal projections Falconer and Howroyd 997) introduced

More information

Some functional (Hölderian) limit theorems and their applications (II)

Some functional (Hölderian) limit theorems and their applications (II) Some functional (Hölderian) limit theorems and their applications (II) Alfredas Račkauskas Vilnius University Outils Statistiques et Probabilistes pour la Finance Université de Rouen June 1 5, Rouen (Rouen

More information

A Note on the Central Limit Theorem for a Class of Linear Systems 1

A Note on the Central Limit Theorem for a Class of Linear Systems 1 A Note on the Central Limit Theorem for a Class of Linear Systems 1 Contents Yukio Nagahata Department of Mathematics, Graduate School of Engineering Science Osaka University, Toyonaka 560-8531, Japan.

More information

ASYMPTOTIC NORMALITY OF THE QMLE ESTIMATOR OF ARCH IN THE NONSTATIONARY CASE

ASYMPTOTIC NORMALITY OF THE QMLE ESTIMATOR OF ARCH IN THE NONSTATIONARY CASE Econometrica, Vol. 7, No. March, 004), 4 4 ASYMPOIC NORMALIY OF HE QMLE ESIMAOR OF ARCH IN HE NONSAIONARY CASE BY SØREN OLVER JENSEN AND ANDERS RAHBEK We establish consistency and asymptotic normality

More information

COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE

COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE Communications on Stochastic Analysis Vol. 4, No. 3 (21) 299-39 Serials Publications www.serialspublications.com COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE NICOLAS PRIVAULT

More information

Optimal series representations of continuous Gaussian random fields

Optimal series representations of continuous Gaussian random fields Optimal series representations of continuous Gaussian random fields Antoine AYACHE Université Lille 1 - Laboratoire Paul Painlevé A. Ayache (Lille 1) Optimality of continuous Gaussian series 04/25/2012

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)

More information

NOTES AND PROBLEMS IMPULSE RESPONSES OF FRACTIONALLY INTEGRATED PROCESSES WITH LONG MEMORY

NOTES AND PROBLEMS IMPULSE RESPONSES OF FRACTIONALLY INTEGRATED PROCESSES WITH LONG MEMORY Econometric Theory, 26, 2010, 1855 1861. doi:10.1017/s0266466610000216 NOTES AND PROBLEMS IMPULSE RESPONSES OF FRACTIONALLY INTEGRATED PROCESSES WITH LONG MEMORY UWE HASSLER Goethe-Universität Frankfurt

More information

Chapter 13: Functional Autoregressive Models

Chapter 13: Functional Autoregressive Models Chapter 13: Functional Autoregressive Models Jakub Černý Department of Probability and Mathematical Statistics Stochastic Modelling in Economics and Finance December 9, 2013 1 / 25 Contents 1 Introduction

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Location Multiplicative Error Model. Asymptotic Inference and Empirical Analysis

Location Multiplicative Error Model. Asymptotic Inference and Empirical Analysis : Asymptotic Inference and Empirical Analysis Qian Li Department of Mathematics and Statistics University of Missouri-Kansas City ql35d@mail.umkc.edu October 29, 2015 Outline of Topics Introduction GARCH

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Strong approximation for additive functionals of geometrically ergodic Markov chains

Strong approximation for additive functionals of geometrically ergodic Markov chains Strong approximation for additive functionals of geometrically ergodic Markov chains Florence Merlevède Joint work with E. Rio Université Paris-Est-Marne-La-Vallée (UPEM) Cincinnati Symposium on Probability

More information

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA On Stochastic Adaptive Control & its Applications Bozenna Pasik-Duncan University of Kansas, USA ASEAS Workshop, AFOSR, 23-24 March, 2009 1. Motivation: Work in the 1970's 2. Adaptive Control of Continuous

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and

More information

Module 9: Stationary Processes

Module 9: Stationary Processes Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.

More information

A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand

A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand Carl Mueller 1 and Zhixin Wu Abstract We give a new representation of fractional

More information

Conditional independence, conditional mixing and conditional association

Conditional independence, conditional mixing and conditional association Ann Inst Stat Math (2009) 61:441 460 DOI 10.1007/s10463-007-0152-2 Conditional independence, conditional mixing and conditional association B. L. S. Prakasa Rao Received: 25 July 2006 / Revised: 14 May

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Logarithmic scaling of planar random walk s local times

Logarithmic scaling of planar random walk s local times Logarithmic scaling of planar random walk s local times Péter Nándori * and Zeyu Shen ** * Department of Mathematics, University of Maryland ** Courant Institute, New York University October 9, 2015 Abstract

More information

UNIT ROOT TESTING FOR FUNCTIONALS OF LINEAR PROCESSES

UNIT ROOT TESTING FOR FUNCTIONALS OF LINEAR PROCESSES Econometric Theory, 22, 2006, 1 14+ Printed in the United States of America+ DOI: 10+10170S0266466606060014 UNIT ROOT TESTING FOR FUNCTIONALS OF LINEAR PROCESSES WEI BIAO WU University of Chicago We consider

More information

The Kadec-Pe lczynski theorem in L p, 1 p < 2

The Kadec-Pe lczynski theorem in L p, 1 p < 2 The Kadec-Pe lczynski theorem in L p, 1 p < 2 I. Berkes and R. Tichy Abstract By a classical result of Kadec and Pe lczynski (1962), every normalized weakly null sequence in L p, p > 2 contains a subsequence

More information

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation Statistics 62: L p spaces, metrics on spaces of probabilites, and connections to estimation Moulinath Banerjee December 6, 2006 L p spaces and Hilbert spaces We first formally define L p spaces. Consider

More information

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,

More information

Random Bernstein-Markov factors

Random Bernstein-Markov factors Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit

More information

A Note on Jackknife Based Estimates of Sampling Distributions. Abstract

A Note on Jackknife Based Estimates of Sampling Distributions. Abstract A Note on Jackknife Based Estimates of Sampling Distributions C. Houdré Abstract Tail estimates of statistics are given; they depend on jackknife estimates of variance of the statistics of interest. 1

More information

Strong Mixing Conditions. Richard C. Bradley Department of Mathematics, Indiana University, Bloomington, Indiana, USA

Strong Mixing Conditions. Richard C. Bradley Department of Mathematics, Indiana University, Bloomington, Indiana, USA Strong Mixing Conditions Richard C. Bradley Department of Mathematics, Indiana University, Bloomington, Indiana, USA There has been much research on stochastic models that have a well defined, specific

More information

The Codimension of the Zeros of a Stable Process in Random Scenery

The Codimension of the Zeros of a Stable Process in Random Scenery The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

arxiv: v2 [math.pr] 27 Oct 2015

arxiv: v2 [math.pr] 27 Oct 2015 A brief note on the Karhunen-Loève expansion Alen Alexanderian arxiv:1509.07526v2 [math.pr] 27 Oct 2015 October 28, 2015 Abstract We provide a detailed derivation of the Karhunen Loève expansion of a stochastic

More information

An elementary proof of the weak convergence of empirical processes

An elementary proof of the weak convergence of empirical processes An elementary proof of the weak convergence of empirical processes Dragan Radulović Department of Mathematics, Florida Atlantic University Marten Wegkamp Department of Mathematics & Department of Statistical

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

UND INFORMATIK. Maximal φ-inequalities for Nonnegative Submartingales. G. Alsmeyer und U. Rösler

UND INFORMATIK. Maximal φ-inequalities for Nonnegative Submartingales. G. Alsmeyer und U. Rösler ANGEWANDTE MATHEMATIK UND INFORMATIK Maximal φ-inequalities for Nonnegative Submartingales G. Alsmeyer und U. Rösler Universität Münster und Universität Kiel e-mail: gerolda@math.uni-muenster.de, roesler@math.uni-kiel.de

More information

LAN property for sde s with additive fractional noise and continuous time observation

LAN property for sde s with additive fractional noise and continuous time observation LAN property for sde s with additive fractional noise and continuous time observation Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Samy Tindel (Purdue University) Vlad s 6th birthday,

More information

Random Locations of Periodic Stationary Processes

Random Locations of Periodic Stationary Processes Random Locations of Periodic Stationary Processes Yi Shen Department of Statistics and Actuarial Science University of Waterloo Joint work with Jie Shen and Ruodu Wang May 3, 2016 The Fields Institute

More information

A central limit theorem for randomly indexed m-dependent random variables

A central limit theorem for randomly indexed m-dependent random variables Filomat 26:4 (2012), 71 717 DOI 10.2298/FIL120471S ublished by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat A central limit theorem for randomly

More information

Spectral representations and ergodic theorems for stationary stochastic processes

Spectral representations and ergodic theorems for stationary stochastic processes AMS 263 Stochastic Processes (Fall 2005) Instructor: Athanasios Kottas Spectral representations and ergodic theorems for stationary stochastic processes Stationary stochastic processes Theory and methods

More information

New Versions of Some Classical Stochastic Inequalities

New Versions of Some Classical Stochastic Inequalities New Versions of Some Classical Stochastic Inequalities Deli Li Lakehead University, Canada 19 July 2018 at The 14th WMPRT in Chengdu This is a joint work with Professor Andrew ROSALSKY at University of

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Itô s formula. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Itô s formula. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Itô s formula Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Itô s formula Probability Theory

More information

Detecting instants of jumps and estimating intensity of jumps from continuous or discrete data

Detecting instants of jumps and estimating intensity of jumps from continuous or discrete data Detecting instants of jumps and estimating intensity of jumps from continuous or discrete data Denis Bosq 1 Delphine Blanke 2 1 LSTA, Université Pierre et Marie Curie - Paris 6 2 LMA, Université d'avignon

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Austrian Journal of Statistics April 27, Volume 46, 67 78. AJS http://www.ajs.or.at/ doi:.773/ajs.v46i3-4.672 Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Yuliya

More information

Evgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields

Evgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields Evgeny Spodarev 23.01.2013 WIAS, Berlin Limit theorems for excursion sets of stationary random fields page 2 LT for excursion sets of stationary random fields Overview 23.01.2013 Overview Motivation Excursion

More information

First passage time for Brownian motion and piecewise linear boundaries

First passage time for Brownian motion and piecewise linear boundaries To appear in Methodology and Computing in Applied Probability, (2017) 19: 237-253. doi 10.1007/s11009-015-9475-2 First passage time for Brownian motion and piecewise linear boundaries Zhiyong Jin 1 and

More information

On Least Squares Linear Regression Without Second Moment

On Least Squares Linear Regression Without Second Moment On Least Squares Linear Regression Without Second Moment BY RAJESHWARI MAJUMDAR University of Connecticut If \ and ] are real valued random variables such that the first moments of \, ], and \] exist and

More information

STAT 7032 Probability. Wlodek Bryc

STAT 7032 Probability. Wlodek Bryc STAT 7032 Probability Wlodek Bryc Revised for Spring 2019 Printed: January 14, 2019 File: Grad-Prob-2019.TEX Department of Mathematical Sciences, University of Cincinnati, Cincinnati, OH 45221 E-mail address:

More information

Combinatorial Dimension in Fractional Cartesian Products

Combinatorial Dimension in Fractional Cartesian Products Combinatorial Dimension in Fractional Cartesian Products Ron Blei, 1 Fuchang Gao 1 Department of Mathematics, University of Connecticut, Storrs, Connecticut 0668; e-mail: blei@math.uconn.edu Department

More information

On detection of unit roots generalizing the classic Dickey-Fuller approach

On detection of unit roots generalizing the classic Dickey-Fuller approach On detection of unit roots generalizing the classic Dickey-Fuller approach A. Steland Ruhr-Universität Bochum Fakultät für Mathematik Building NA 3/71 D-4478 Bochum, Germany February 18, 25 1 Abstract

More information

A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN. Dedicated to the memory of Mikhail Gordin

A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN. Dedicated to the memory of Mikhail Gordin A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN Dedicated to the memory of Mikhail Gordin Abstract. We prove a central limit theorem for a square-integrable ergodic

More information

SAMPLE CHAPTERS UNESCO EOLSS STATIONARY PROCESSES. E ξ t (1) K.Grill Institut für Statistik und Wahrscheinlichkeitstheorie, TU Wien, Austria

SAMPLE CHAPTERS UNESCO EOLSS STATIONARY PROCESSES. E ξ t (1) K.Grill Institut für Statistik und Wahrscheinlichkeitstheorie, TU Wien, Austria STATIONARY PROCESSES K.Grill Institut für Statistik und Wahrscheinlichkeitstheorie, TU Wien, Austria Keywords: Stationary process, weak sense stationary process, strict sense stationary process, correlation

More information

Testing the Equality of Covariance Operators in Functional Samples

Testing the Equality of Covariance Operators in Functional Samples Scandinavian Journal of Statistics, Vol. 4: 38 5, 3 doi:./j.467-9469..796.x Board of the Foundation of the Scandinavian Journal of Statistics. Published by Blackwell Publishing Ltd. Testing the Equality

More information

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze

More information

Convergence of the Ensemble Kalman Filter in Hilbert Space

Convergence of the Ensemble Kalman Filter in Hilbert Space Convergence of the Ensemble Kalman Filter in Hilbert Space Jan Mandel Center for Computational Mathematics Department of Mathematical and Statistical Sciences University of Colorado Denver Parts based

More information

LARGE DEVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILED DEPENDENT RANDOM VECTORS*

LARGE DEVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILED DEPENDENT RANDOM VECTORS* LARGE EVIATION PROBABILITIES FOR SUMS OF HEAVY-TAILE EPENENT RANOM VECTORS* Adam Jakubowski Alexander V. Nagaev Alexander Zaigraev Nicholas Copernicus University Faculty of Mathematics and Computer Science

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

Uniform and non-uniform quantization of Gaussian processes

Uniform and non-uniform quantization of Gaussian processes Uniform and non-uniform quantization of Gaussian processes Oleg Seleznjev a,b,, Mykola Shykula a a Department of Mathematics and Mathematical Statistics, Umeå University, SE-91 87 Umeå, Sweden b Faculty

More information

STIT Tessellations via Martingales

STIT Tessellations via Martingales STIT Tessellations via Martingales Christoph Thäle Department of Mathematics, Osnabrück University (Germany) Stochastic Geometry Days, Lille, 30.3.-1.4.2011 Christoph Thäle (Uni Osnabrück) STIT Tessellations

More information

Self-normalized laws of the iterated logarithm

Self-normalized laws of the iterated logarithm Journal of Statistical and Econometric Methods, vol.3, no.3, 2014, 145-151 ISSN: 1792-6602 print), 1792-6939 online) Scienpress Ltd, 2014 Self-normalized laws of the iterated logarithm Igor Zhdanov 1 Abstract

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Sequential block bootstrap in a Hilbert space with application to change point analysis

Sequential block bootstrap in a Hilbert space with application to change point analysis Sequential block bootstrap in a Hilbert space with application to change point analysis Olimjon Sharipov, Johannes Tewes and Martin Wendler arxiv:42.0446v2 [math.st] 5 Sep 205 September 6, 208 Abstract

More information

Convex Feasibility Problems

Convex Feasibility Problems Laureate Prof. Jonathan Borwein with Matthew Tam http://carma.newcastle.edu.au/drmethods/paseky.html Spring School on Variational Analysis VI Paseky nad Jizerou, April 19 25, 2015 Last Revised: May 6,

More information

On an uniqueness theorem for characteristic functions

On an uniqueness theorem for characteristic functions ISSN 392-53 Nonlinear Analysis: Modelling and Control, 207, Vol. 22, No. 3, 42 420 https://doi.org/0.5388/na.207.3.9 On an uniqueness theorem for characteristic functions Saulius Norvidas Institute of

More information

The Wiener Itô Chaos Expansion

The Wiener Itô Chaos Expansion 1 The Wiener Itô Chaos Expansion The celebrated Wiener Itô chaos expansion is fundamental in stochastic analysis. In particular, it plays a crucial role in the Malliavin calculus as it is presented in

More information

Functional Central Limit Theorem for the Measure of Level Sets Generated by a Gaussian Random Field

Functional Central Limit Theorem for the Measure of Level Sets Generated by a Gaussian Random Field Functional Central Limit Theorem for the Measure of Level Sets Generated by a Gaussian Random Field Daniel Meschenmoser Institute of Stochastics Ulm University 89069 Ulm Germany daniel.meschenmoser@uni-ulm.de

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information