On Differentiability of Average Cost in Parameterized Markov Chains

Size: px
Start display at page:

Download "On Differentiability of Average Cost in Parameterized Markov Chains"

Transcription

1 On Differentiability of Average Cost in Parameterized Markov Chains Vijay Konda John N. Tsitsiklis August 30, Overview The purpose of this appendix is to prove Theorem 4.6 in 5 and establish various facts used in verifying some of the assumptions of Theorem A.1 therein. We use the same notation and assumptions as in 5, and to simplify the analysis, we assume that the parameter is scalar. The setting used here is similar (but not identical) to the one in 4. The major difference is that the setting of 4 allows only RSPs of the form U k = f(x k,w k ), where the parameter enters only through the distribution of W k which does not depend on x. Incontrast, our setting allows for any RSPs of the form U k = f (X k,w k ), as long as the distribution µ (u x) off (x, U) satisfies the required assumptions. The differentiability results presented in this report are based on the following key lemma (see 3, Corollary 2.2.1). Lemma 1.1. Consider a parameterized family of random variables {F (ω); R} on a Polish space Ω. Suppose the following hold: 1. For each 0, there exists some d>1 and K d > 0 such that, E F F 0 d <K d 0 d, for all sufficiently close to There is a family of random variables {f } such that where F +h F h P denotes convergence in probability. P f, as h 0,, 1

2 Then the map E F is differentiable with d d E F =E f. Our proof of the differentiability results involves the following steps. 1. Obtain a representation of the steady-state expected cost and of solutions to the Poisson equation in terms of suitably defined expectations E,x. 2. Using likelihood ratios, represent the above expectations using only the expectations E 0,x corresponding to some fixed 0. This new representation has the advantage that the probability law w.r.t. which expectations are taken does not depend on the parameter. Only the random variables inside the expectation depend on the parameter, making our key result Lemma 1.1 applicable to this representation. 3. Verify the hypotheses of Lemma 1.1. The next section carries out these steps for expectations over finite horizon. In Section 3, we recall the results on the regenerative representations for the average cost and solutions to the Poisson equation. In Section 4, we introduce the likelihood ratio representations and show that they satisfy the assumptions of Lemma 1.1. In Sections 5 and 6, we use all the previous results to prove differentiability of expectations over infinite horizon and verify some of the assumptions of Theorem A.1 in 5. 2 Finite Horizon In this section, we prove the differentiability of expectations of the form E,x W where W is a random variable that depends only on a finite number of state-decision pairs (X k,u k ). For each and k, let Λ,k = k µ (U l X l ) µ 0 (U l X l ). It is easy to see that Λ,k is the Radon-Nikodym derivative of the distribution of (X l,u l ) k under policy with respect to the distribution under policy 0.Wewill use the following basic lemma to derive some bounds on Λ,k. Lemma 2.1. If W 0,...,W k 1 are positive random variables satisfying E W l 1 d ɛ d C(d), d 1, l =0,...,k 1, for some monotonic function C( ) 1, and if ɛ<1/k2 k, then k 1 d E W l 1 k 1 ɛ d (2k) d C(kd), E 2 d C(kd), for any d 1. W d l 2

3 Proof. For any d 1 and z 0, we have z d 2 d 1 ( 1+ z 1 d), (1) we have E W d l 2 d 1 (1 + ɛ d C(d)) 2 d C(d), l =0,...,k 1. To prove the first part, note that and therefore z 0 z k 1 1 z z 0 z z 0 z k 2 z k 1 1 z 0 z k 1 1 d k d 1 z 0 1 d + z d 0 z 1 1 d + + z d 0 z d k 2 z k 1 1 d. We now use Holder s inequality to see that k 1 d E k 1 l 1 W l 1 k d 1 E Wj d W l 1 d k 1 l 1 k d 1 j=0 j=0 E W (l+1)d j k 1 k d 1 2 ld ɛ d C((l +1)d) k d ɛ d 2 kd C(kd). 1/(l+1) E W l 1 (l+1)d 1/(l+1) To prove the second part, use (1) and the fact that ɛ<1/k2 k to simplify the bound. The following are two immediate consequences of the above lemma. In the following lemma, note that the constant K d might depend on k. This dependence can be safely ignored as the next lemma is intended to verify the first assumption of Lemma 1.1 for a fixed k. Lemma For every d 1, there exists K d > 0 such that Λ,k 1 d K d 0 d L(x),, x. E 0,x 2. For every d/ (1, 1), there exists K d > 0 such that Λ,k d K d L(x), x, E 0,x and for all sufficiently close to 0. 3

4 Proof. Use the mean-value theorem to see that µ (u x) µ 0 (u x) 1 0 sup µ (u x) µ 0 (u x). (2) Since µ (u x)/µ 0 (u x) belongs to D, wehave E,x sup µ (X k,u k ) K d L(x), 0 µ 0 (X k,u k ) for some constant K d.part 1, and Part 2 for d 1, are implied by Lemma 2.1. To prove Part 2 for d 1, note that the inequality (2) holds even if and 0 are interchanged. The above lemma verifies the first assumption of Lemma 1.1 for expectations of random variables that are functions of state-decision pairs up to time k. The next lemma verifies the second assumption and derives a formula for the gradient. Lemma 2.3. Consider a family of random variables {W } such that the map W (ω) is differentiable for each ω, W is a function of the state-decision pairs up to time k, and, 0 E,x W d <, E,x sup W d <, x, for some d>1. Then, the map E,x W is differentiable with where E,x W =E,x W +E,x Λ,k W, Λ,k = k ψ (X l,u l )= k µ (X l,u l ) µ (X l,u l ). Proof. Note that E,x W =E 0,x Λ,k W. Using the bounds of the previous lemma, the fact that d-moments of W and sup W are finite, and the mean-value theorem, verify the first assumption of Lemma 1.1. Using the fact that µ (u x) and W are differentiable in, the second assumption of Lemma 1.1 follows easily. The gradient formula now follows from interchanging the order of differentiation and expectation. The following is an immediate consequence of the previous result. It uses the formula in the previous result to derive bounds on the rate at which the gradient grows with k. E,x f (X k,u k ) 4

5 Corollary 2.4. Consider a family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D. Then, the map E,x f (X k,u k ) is differentiable for each x, and there exists some K>0 such that E,x f (X k,u k ) (k +1)KL(x), x. Furthermore, the families of functions belong to D. {E,x f (X k,u k )}, { E,x f (X k,u k )} Before we move on, we would like to point out that the above result holds for expectations of the form E,x U 0 = u aswell, if we redefine Λ,k as 3 Splitting k Λ,0 =0, Λ,k = ψ (X l,u l ), k >0. l=1 In this section, we recall the splitting technique of Athreya and Ney 2, and Nummelin 7 to obtain a regenerative representation for the average cost function ᾱ() and the solutions Q of the Poisson equation. Let δ, N, X 0 and ϑ be as in Assumption 4.2 of 5. Consider the {0, 1}-valued process (B k ) constructed as follows: 1. If k is not divisible by N, then B k =0. 2. If k is divisible by N, then P(B k =1 X l,u l,l =0, 1,...,B l,l k) = P(B k =1 X k,x k+n ) = (X k,x k+n ) where = δ 2 I ϑ(dy) X 0 (x) P,x (X N dy). (Assumption 4.2(a) ensures existence of.) Let It is easy to see that for each, x, f (0) = 1,, x, y, Q,x (dy) = P,x(X N dy) δ 2 I X 0 (x)ϑ(dy) 1 δ 2 I. X 0 (x) f (0) (x, ) = ( 1 δ 2 I X 0 (x) 5 ) Q,x (dy) P,x (X N dy)

6 where the existence of the Radon-Nikodym derivative is guaranteed by the fact that 2 Q,x (dy) 2 δ P,x(X N dy). With this construction it is not difficult to see that the process (X kn,b kn )isamarkov chain for which X {1} is an atom. Consider the first time this atom is hit, i.e., τ = min{k B k =1}. The time τ is well-defined and finite w.p.1. as the set X 0 is hit infinitely often from any initial state (cf. Assumption 4.2 in 5) and the probability that B k is 1 given that X k = x is δ/2 whenever x X 0. Using standard results 6, Theorems and on Markov chains with atoms, the average cost ᾱ() can be written as (τ/n) 1 E,ϑ c,n (X kn ) ᾱ() =, E,ϑ τ where c,n (x) = N 1 Furthermore, the functions V (x) given by V (x) =E,x satisfy the Poisson equation (τ/n) 1 E,x c(x k,u k ). ( c,n (X kn ) Nᾱ()) (3) V (x) =E,x c(x, U 0 )+V (X 1 ) for the Markov chain {X k }.Tosee this, note that V (x) isasolution to the Poisson equation for the cost function c,n and the Markov chain {X kn } (see 6, p.441 ). Furthermore, if ˆV is a solution to the Poisson equation for the cost function c,1 and the Markov chain {X k }, then ˆV is a solution to the Poisson equation for the cost function c,n and the Markov chain {X kn }. Since ˆV and V are two solutions to the Poisson equation for a positive Harris chain {X kn }, they must differ by a constant (cf. Proposition of 6) and therefore, V is a solution to the Poisson equation for the cost function c,1 and the Markov chain {X k }. 4 Likelihood Ratio For each, k, let where Λ,k =Λ,kN ˆΛ,k 6

7 k 1 f ˆΛ (0) (X ln,x ln+n ),k = f (0) 0 (X ln,x ln+n ), and 0/0 isinterpreted as 1. It is easy to see that Λ,k is the Radon-Nikodym derivative of the joint distribution of (X l,u l ;0 l kn, B l ;0 l (k 1)N) under P with respect to that under P 0 on the set of outcomes for which τ>kn. Therefore for any W that is a function of only these random variables, we have E,x WI{τ >kn} =E 0,x WI{τ >kn} Λ,k. To show that the map E,x WI{τ >kn} is differentiable, we need to verify the hypothesis of Lemma 1.1 for the likelihood ratio Λ,k.Wenow derive several bounds to verify the first assumption of Lemma 1.1. Lemma For each d 1, there exists K d > 0 such that E 0 f (i),x (x, X d N) f (i) 0 (x, X N ) 1 K d 0 d L(x), x, i =0, For each d 1, there exists K d > 0 such that E 0 f (i),x (x, X d N) f (i) K 0 (x, X N ) d L(x) x, i =0, 1, and for all sufficiently close to For each d 1, there exists K d > 0 such that E 0,x ˆΛ,k 1 d K d 0 d L(x), x. 4. For each d 1, there exists K d > 0 such that E 0,x ˆΛ,k d K d L(x) x, and for all sufficiently close to 0. 7

8 Proof. To prove the first part, note that for x/ X 0, 0 = 1, and for x X 0, 0 = P 0,x(X N dy) P,x (X N dy) = {E 0,x Λ,N X N = y} 1. Therefore, we have E 0 (x, X N ) d,x 0 (x, X N ) 1 E 0,x E 0,x E 0,x E0,x Λ,N X N 1 1 d E 0,x Λ,N X N 2d E 0,x E 0,x Λ,N 1 X N 2d Λ 2d,N E 0,x Λ,N 1 2d, where the last two inequalities follow from Holder s and Jensen s inequalities respectively. For i =0,note that f (0) f (0) 0 1 = ( (1) 0 1 f f (0) 0 since 1/2 for all. The proof of the second part is similar. The last two parts follow from the first two and Lemma 2.1. The next lemma proves that the first of the two hypotheses for Lemma 1.1 hold for Λ. Lemma For each d 1, there exists K d > 0 such that E 0,x Λ,k 1 d K d 0 d L(x), for sufficiently close to For each d 1, there exists K d > 0 such that E 0,x Λ,k d K d L(x). ) 8

9 Proof. Follows from Lemmas 2.2, 4.1 and 2.1. We are now ready to state the following result on differentiability of finite horizon expectations. Lemma 4.3. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map E,x f (X kn )I{τ >kn}, where f (x) = N 1 E,x f (X k,u k ), is differentiable for each x, with E,x f (X kn )I{τ >kn} (k +1)ρ k 0KL(x), for some ρ 0 < 1. Proof. Since E,x f (X kn )I{τ >kn} = E 0,x Λ,k f (X kn )I{τ >kn}, we can use Lemma 2.1 to prove its differentiability and to calculate the derivative. To verify the first assumption of Lemma 2.1, use Holder s inequality to see that E 0,x f (X kn )I{τ >kn} Λ,k f (X kn )I{τ >kn} Λ 0,k d = E 0,x f (X kn ) d Λ,k 1 d E 0,x f (X kn ) 2d 1/2 E0,x Λ,k 1 2d 1/2. The first assumption now follows from the fact that f (x, u) (and therefore f (x)) belongs to D, and the first part of the previous lemma. To verify the second assumption of Lemma 1.1, use Assumption 4.4 of 5 to conclude that the map Λ,k (ω) is continuously differentiable for all ω, and therefore it is enough to prove that ˆΛ 0 +h,k 1 converges in probability. This in turn is verified if the functions f (i) h (i) /f 0,i=0, 1, can be shown to be differentiable w.r.t. for all x, y. Since 0 / 9

10 is equal to E 0,x Λ,N X N = y, for all x X 0 and is equal to 1 otherwise, use part 3 of Assumption 4.4 in 5 and the mean-value theorem to see that Λ 0 +h,n 1 h is bounded above by an integrable random variable. Therefore, we invoke dominated convergence theorem for conditional expectations to show that there exists a version of / 0 that is differentiable at 0 for all x, y. Furthermore, Λ 0 +h 1 h converges in probability to Λ 0,k where kn k 1 jn+n 1 Λ,k = ψ (X k,u k ) I X0 (X jn )E,x ψ (X l,u l ) j=0 l=jn X jn,x jn+n. Therefore, we have E,x f (X kn )I{τ >kn} = E,x Λ,k f (X kn )I{τ >kn} +E,x Λ,k f (X kn )I{τ >kn} To derive the bound on the derivative, use Holder s inequality to obtain E,x f (X kn )I{τ >k} 3 1/3 E,x Λ,k E,x f (X kn ) 3 1/3 P,x (τ>kn) 1/3 +E,x f (X kn ) 3 1/3 P,x (τ>kn) 1/3. To calculate E,x Λ,k 3 1/3, note that Λ,k is a sum of 2kN +1 terms. Since ψ (x, u) belongs to D, the 3-norm of each of these terms is bounded by KL(x) 1/3 for some K. Similarly, since f (x, u) and f (x, u) (therefore f (x) and f (x)) belong to D, the terms E,x f (X kn ) 3 1/3 and E,x f (X kn ) 3 1/3 are also bounded by KL(x) 1/3 for some K. Finally, due to uniform geometric ergodicity, we have P,x (τ>kn) Kρ k 0L(x), for some ρ 0 < 1. Combining all these terms it is easy to establish the stated bound on the derivative. 10

11 5 Infinite Horizon The above lemma concerns the differentiability of expectation w.r.t. finite dimensional marginals of P,x. We use the following result from advanced calculus (e.g., see Apostol 1) to extend the above result to the infinite horizon case. Theorem 5.1. Assume that each f k is a real-valued function defined on a neighborhood Θ of 0 such that the derivative f k () exists for each in Θ. Assume that 1. f k ( 0 ) converges, 2. there exists a g() such that f k () =g() uniformly on Θ. Then: 1. There exists a function f() such that f k () =f() uniformly on Θ. 2. If Θ, the derivative f() exists and equals f k (). By combining the previous two results we have the following theorem. Recall that η is the steady state expectation of the state-decision process {(X k,u k )}. Theorem 5.2. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map (τ/n) 1 E,x f (X kn ) is differentiable for each x, and the families of functions (τ/n) 1 E,x f (X kn ), E,x (τ/n) 1 f (X kn ), belong to D. Proof. Note that E,x (τ/n) 1 f (X kn ) = E,x f (X kn )I{τ >kn}. The differentiability now follows from the previous two lemmas. To show that the functions (τ/n) 1 E (τ/n) 1,x f (X kn ), E,x f (X kn ) 11

12 are in D, consider E,x sup E,x E,x (τ/n) 1 (τ/n) 1 d f (X kn ) d sup f (X kn ) sup f (X d kn ) (τ/n) 1 (τ/n) d 1 E,x (τ/n) d 1 sup f (X kn ) E,x (τ/n) 3(d 1) 1/3 E,x sup P,x (τ>kn) 1/3. d I{τ >kn} f (X kn ) 3d 1/3 It is now easy to see that the right hand side is bounded by K d L(x) for some K d > 0. The following corollaries are immediate consequences of this result. Corollary 5.3. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map η (f ) is bounded and differentiable with bounded derivatives. Corollary 5.4. (Lemma 5.1 in 5) If X 0 is singleton (i.e., X 0 is an atom and therefore splitting in unnecessary) and τ is the first hitting time of X 0 then the functions defined as τ 1 Q (x, u) = E,x c(x k,u k ) U 0 = u, T (x, u) = E,x τ U 0 = u, belong to D. Furthermore, for each (x, u), the maps Q (x, u), T (x, u) are differentiable, and the derivatives Q (x, u), T (x, u) also belong to D. Corollary 5.5. The average expected cost function ᾱ() is bounded and differentiable with bounded derivatives. Furthermore, the solutions V (x) to the Poisson equation given by (3) belong to D and the map V (x) is differentiable for each x and the derivatives V (x) also belong to D. Furthermore, the family of functions Q = c ᾱ()1 + P V 12

13 belongs to D, with the map Q (x, u) being differentiable for all (x, u) and Q (x, u) belonging to D. This corollary is used in 5 to establish Theorem 4.6 in 5, which is again proved here for completeness. Theorem 5.6. For any solution Q to the Poisson equation we have ᾱ() = ψ,q. Furthermore, ᾱ() has bounded derivatives. Proof. Note that the Q defined in the previous corollary satisfies the Poisson equation: Q (x, u) =c(x, u) ᾱ()+e,x Q (X 1,U 1 ) U 0 = u. Since Q belongs to D, Q is differentiable in, and Q also belongs to D, wecan differentiate both sides of the above equation to obtain Q (x, u) = ᾱ()+e,x Q (X 1,U 1 ) U 0 = u +E,x ψ (X 1,U 1 )Q (X 1,U 1 ) U 0 = u. Taking expectation with respect to the steady state distribution of (X k,u k )weobtain the desired formula with Q defined as in the previous corollary. To see that Q can be replaced by any other solution ˆQ to the Poisson equation, note that Q ˆQ is some function C() (Proposition of 6) and that E,x ψ (x, U 0 ) = 0. To prove that ᾱ is twice differentiable with bounded derivatives, note that ᾱ = η (ψ Q ) and apply Corollary 5.3. The above result leads to the following theorem on differentiability of a common representation for solutions to Poisson equation. Theorem 5.7. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map E,x f (X k,u k ) η (f ) is differentiable for each x, and the families of functions { } { } E,x f (X k,u k ) η (f ), E,x f (X k,u k ) η (f ), belong to D. 13

14 Proof. To simplify the proof, assume that f = c. Then the results on differentiability for the finite horizon case, imply that the function is differentiable with E,x c(x k,u k ) ᾱ() E,x c(x k,u k ) ᾱ() k = E,x ψ (X l,u l )(c(x k,u k ) ᾱ()) ᾱ() k/2 + E,x ψ (X l,u l )E,Xl+1 c(x k,u k ) ᾱ() k l=k/2+1 + l=k/2 E,x ψ (X l,u l )(c(x k,u k ) ᾱ()) ψ,p k l ψ,p l (c ᾱ()1), (c ᾱ()1) where k/2 represents the floor of k/2. Using geometric ergodicity, the fact that V 1/d also satisfies the geometric Foster Lyapunov criterion with the decay factor ρ 1/d, and Schwartz inequality, we can see that each of these terms can be bounded by Kρ k 0L(x) for some ρ 0 < 1. The result now follows from an easy application of Theorem Verification of Assumption A.3 parts (d) and (e) In this section, we verify parts (d) and (e) of Assumption A.3 of 5 for the two cases of TD(1) and of TD(λ), λ<1 separately. Since φ,q,t belong to D, part (d) in both cases follows from Corollary 5.3. Therefore, we will verify only part (e). We start with the TD(1) case. 6.1 TD(1) We will now verify part (e) for ĥ( ). Note that τ 1 ĥ (y) = E, x (h (Y k ) h()) Y 0 = y, = E, x (h (Y k ) h())i{τ >k} Y 0 = y. Since h (Y k )isafunction of the state-decision pairs up to time k and satisfies the assumptions of Lemma 2.3, a formula can be derived for the derivative of the expectation E, x (h (Y k ) h())i{τ >k} Y 0 = y. 14

15 Using this formula, a bound on this derivative of the form K(k +1) d ρ k L(x, 0 u)z,d 1,K >0,ρ 0 < 1, L D. can be obtained. This bound and Lemma 5.1 can be used to conclude that ĥ is differentiable with respect to and the derivative is bounded by K L(x, u)z for some other K>0. It is now easy to verify that E,x ĥ (Y 1 ) Y 0 = y is differentiable with the derivative bounded appropriately. The verification of part (e) for Ĝ is similar. 6.2 TD(λ), λ<1 Since the verification of part (e) for ĥ is simpler and the verification for Ĝ is similar, we will consider ĥ only. Note that the first component of the vector ĥ does not depend on z and is equal to L E,x (c(x k,u k ) ᾱ()) U 0 = u. Therefore, part (e) for this component follows from Lemma 5.7. For the second component, consider the sum E,x c(xk,u k )Z k h 1 () ᾱ() Z() Y0 = y = z + = z + λ k E,x c(x k,u k ) P l c, φ Y0 = y k 1 λ l E,x c(xk,u k )φ (X k l,u k l ) P l c, φ Y0 = y + l=k+1 λ l P l c, φ λ k E,x c(x k,u k ) P l c, φ Y0 = y + λ l k=l+1 λ l l 1 E,x c(xk,u k ))φ (X k l,u k l ) P l c ᾱ()1,φ Y 0 = y P l c, φ. It is now easy to see that the second sums of the first two terms are differentiable with the derivatives uniformly bounded in l as the second sum of the second term is an expectation of the solution to a Poisson equation for the Markov chain (X k l,u k l,x k,u k ) (cf. Lemma 5.7). The second sum of the third term is also differentiable with the bound on the derivative linear in k. Itnow follows from Theorem 5.1 that ĥ(y) isdifferentiable with the derivative bounded above by K L(x, u)z for some L D. 15

16 References 1 T. M. Apostol. Mathematical Analysis: A Modern Approach to Advanced Calculus. Addison-Wesley Pub Co, 2 edition, January K. B. Athreya and P. Ney. A new approach to the limit theory of recurrent Markov chains. Trans. Amer. Math. Soc., 245: , V. S. Borkar. Probability theory: an advanced course. Springer-Verlag, New York, P. W. Glynn and P. L Ecuyer. Likelihood ratio gradient estimation for stochastic recursions. Advances in applied probability, 27: , V. R. Konda and J. N. Tsitsiklis. Actor-critic algorithms. Submitted to the SIAM Journal on Control and Optimization, February S. P. Meyn and R. L. Tweedie. Markov Chains and Stochastic Stability. Springer- Verlag, E. Nummelin. A splitting technique for Harris recurrent chains. Z. Wahrscheinlichkeitstheorie and Verw. Geb., 43: ,

A System Theoretic Perspective of Learning and Optimization

A System Theoretic Perspective of Learning and Optimization A System Theoretic Perspective of Learning and Optimization Xi-Ren Cao* Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong eecao@ee.ust.hk Abstract Learning and optimization

More information

Convergence Rates for Renewal Sequences

Convergence Rates for Renewal Sequences Convergence Rates for Renewal Sequences M. C. Spruill School of Mathematics Georgia Institute of Technology Atlanta, Ga. USA January 2002 ABSTRACT The precise rate of geometric convergence of nonhomogeneous

More information

Verifying Regularity Conditions for Logit-Normal GLMM

Verifying Regularity Conditions for Logit-Normal GLMM Verifying Regularity Conditions for Logit-Normal GLMM Yun Ju Sung Charles J. Geyer January 10, 2006 In this note we verify the conditions of the theorems in Sung and Geyer (submitted) for the Logit-Normal

More information

A regeneration proof of the central limit theorem for uniformly ergodic Markov chains

A regeneration proof of the central limit theorem for uniformly ergodic Markov chains A regeneration proof of the central limit theorem for uniformly ergodic Markov chains By AJAY JASRA Department of Mathematics, Imperial College London, SW7 2AZ, London, UK and CHAO YANG Department of Mathematics,

More information

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past. 1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if

More information

Faithful couplings of Markov chains: now equals forever

Faithful couplings of Markov chains: now equals forever Faithful couplings of Markov chains: now equals forever by Jeffrey S. Rosenthal* Department of Statistics, University of Toronto, Toronto, Ontario, Canada M5S 1A1 Phone: (416) 978-4594; Internet: jeff@utstat.toronto.edu

More information

Linear stochastic approximation driven by slowly varying Markov chains

Linear stochastic approximation driven by slowly varying Markov chains Available online at www.sciencedirect.com Systems & Control Letters 50 2003 95 102 www.elsevier.com/locate/sysconle Linear stochastic approximation driven by slowly varying Marov chains Viay R. Konda,

More information

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm Chapter 13 Radon Measures Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm (13.1) f = sup x X f(x). We want to identify

More information

Approximate Dynamic Programming

Approximate Dynamic Programming Master MVA: Reinforcement Learning Lecture: 5 Approximate Dynamic Programming Lecturer: Alessandro Lazaric http://researchers.lille.inria.fr/ lazaric/webpage/teaching.html Objectives of the lecture 1.

More information

On Finding Optimal Policies for Markovian Decision Processes Using Simulation

On Finding Optimal Policies for Markovian Decision Processes Using Simulation On Finding Optimal Policies for Markovian Decision Processes Using Simulation Apostolos N. Burnetas Case Western Reserve University Michael N. Katehakis Rutgers University February 1995 Abstract A simulation

More information

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes

More information

Nonparametric inference for ergodic, stationary time series.

Nonparametric inference for ergodic, stationary time series. G. Morvai, S. Yakowitz, and L. Györfi: Nonparametric inference for ergodic, stationary time series. Ann. Statist. 24 (1996), no. 1, 370 379. Abstract The setting is a stationary, ergodic time series. The

More information

x log x, which is strictly convex, and use Jensen s Inequality:

x log x, which is strictly convex, and use Jensen s Inequality: 2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M. Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although

More information

Dynamic Programming with Hermite Interpolation

Dynamic Programming with Hermite Interpolation Dynamic Programming with Hermite Interpolation Yongyang Cai Hoover Institution, 424 Galvez Mall, Stanford University, Stanford, CA, 94305 Kenneth L. Judd Hoover Institution, 424 Galvez Mall, Stanford University,

More information

When is a Markov chain regenerative?

When is a Markov chain regenerative? When is a Markov chain regenerative? Krishna B. Athreya and Vivekananda Roy Iowa tate University Ames, Iowa, 50011, UA Abstract A sequence of random variables {X n } n 0 is called regenerative if it can

More information

Harmonic Functions and Brownian motion

Harmonic Functions and Brownian motion Harmonic Functions and Brownian motion Steven P. Lalley April 25, 211 1 Dynkin s Formula Denote by W t = (W 1 t, W 2 t,..., W d t ) a standard d dimensional Wiener process on (Ω, F, P ), and let F = (F

More information

Chapter 4: Asymptotic Properties of the MLE

Chapter 4: Asymptotic Properties of the MLE Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider

More information

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes Lectures on Stochastic Stability Sergey FOSS Heriot-Watt University Lecture 4 Coupling and Harris Processes 1 A simple example Consider a Markov chain X n in a countable state space S with transition probabilities

More information

An almost sure invariance principle for additive functionals of Markov chains

An almost sure invariance principle for additive functionals of Markov chains Statistics and Probability Letters 78 2008 854 860 www.elsevier.com/locate/stapro An almost sure invariance principle for additive functionals of Markov chains F. Rassoul-Agha a, T. Seppäläinen b, a Department

More information

Independence of some multiple Poisson stochastic integrals with variable-sign kernels

Independence of some multiple Poisson stochastic integrals with variable-sign kernels Independence of some multiple Poisson stochastic integrals with variable-sign kernels Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological

More information

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by L p Functions Given a measure space (, µ) and a real number p [, ), recall that the L p -norm of a measurable function f : R is defined by f p = ( ) /p f p dµ Note that the L p -norm of a function f may

More information

Local semiconvexity of Kantorovich potentials on non-compact manifolds

Local semiconvexity of Kantorovich potentials on non-compact manifolds Local semiconvexity of Kantorovich potentials on non-compact manifolds Alessio Figalli, Nicola Gigli Abstract We prove that any Kantorovich potential for the cost function c = d / on a Riemannian manifold

More information

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,

More information

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes Front. Math. China 215, 1(4): 933 947 DOI 1.17/s11464-15-488-5 Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes Yuanyuan LIU 1, Yuhui ZHANG 2

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence 1 Overview We started by stating the two principal laws of large numbers: the strong

More information

AW -Convergence and Well-Posedness of Non Convex Functions

AW -Convergence and Well-Posedness of Non Convex Functions Journal of Convex Analysis Volume 10 (2003), No. 2, 351 364 AW -Convergence Well-Posedness of Non Convex Functions Silvia Villa DIMA, Università di Genova, Via Dodecaneso 35, 16146 Genova, Italy villa@dima.unige.it

More information

Central-limit approach to risk-aware Markov decision processes

Central-limit approach to risk-aware Markov decision processes Central-limit approach to risk-aware Markov decision processes Jia Yuan Yu Concordia University November 27, 2015 Joint work with Pengqian Yu and Huan Xu. Inventory Management 1 1 Look at current inventory

More information

Almost Sure Convergence of Two Time-Scale Stochastic Approximation Algorithms

Almost Sure Convergence of Two Time-Scale Stochastic Approximation Algorithms Almost Sure Convergence of Two Time-Scale Stochastic Approximation Algorithms Vladislav B. Tadić Abstract The almost sure convergence of two time-scale stochastic approximation algorithms is analyzed under

More information

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL First published in Journal of Applied Probability 43(1) c 2006 Applied Probability Trust STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL LASSE LESKELÄ, Helsinki

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming

University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming University of Warwick, EC9A0 Maths for Economists 1 of 63 University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming Peter J. Hammond Autumn 2013, revised 2014 University of

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information

2 light traffic derivatives for the GI/G/ queue. We shall see that our proof of analyticity is mainly based on some recursive formulas very similar to

2 light traffic derivatives for the GI/G/ queue. We shall see that our proof of analyticity is mainly based on some recursive formulas very similar to Analyticity of Single-Server Queues in Light Traffic Jian-Qiang Hu Manufacturing Engineering Department Boston University Cummington Street Boston, MA 0225 September 993; Revised May 99 Abstract Recently,

More information

Tail inequalities for additive functionals and empirical processes of. Markov chains

Tail inequalities for additive functionals and empirical processes of. Markov chains Tail inequalities for additive functionals and empirical processes of geometrically ergodic Markov chains University of Warsaw Banff, June 2009 Geometric ergodicity Definition A Markov chain X = (X n )

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

A Probabilistic Proof of the Perron-Frobenius Theorem

A Probabilistic Proof of the Perron-Frobenius Theorem A Probabilistic Proof of the Perron-Frobenius Theorem Peter W. Glynn 1 and Paritosh Y. Desai 2 1 Department of Management Science and Engineering, Stanford University, Stanford CA 2 Target Corporation,

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Limit theorems for dependent regularly varying functions of Markov chains

Limit theorems for dependent regularly varying functions of Markov chains Limit theorems for functions of with extremal linear behavior Limit theorems for dependent regularly varying functions of In collaboration with T. Mikosch Olivier Wintenberger wintenberger@ceremade.dauphine.fr

More information

Chapter 7. Markov chain background. 7.1 Finite state space

Chapter 7. Markov chain background. 7.1 Finite state space Chapter 7 Markov chain background A stochastic process is a family of random variables {X t } indexed by a varaible t which we will think of as time. Time can be discrete or continuous. We will only consider

More information

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS Statistica Sinica 6(1996), 943-956 THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS H. Z. An and F. C. Huang Academia Sinica Abstract: Consider the nonlinear autoregressive model x t = φ(x

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

Chapter 2: Markov Chains and Queues in Discrete Time

Chapter 2: Markov Chains and Queues in Discrete Time Chapter 2: Markov Chains and Queues in Discrete Time L. Breuer University of Kent 1 Definition Let X n with n N 0 denote random variables on a discrete space E. The sequence X = (X n : n N 0 ) is called

More information

On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo

On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo James P. Hobert 1, Galin L. Jones 2, Brett Presnell 1, and Jeffrey S. Rosenthal 3 1 Department of Statistics University of Florida

More information

1 Problem Formulation

1 Problem Formulation Book Review Self-Learning Control of Finite Markov Chains by A. S. Poznyak, K. Najim, and E. Gómez-Ramírez Review by Benjamin Van Roy This book presents a collection of work on algorithms for learning

More information

Properties of an infinite dimensional EDS system : the Muller s ratchet

Properties of an infinite dimensional EDS system : the Muller s ratchet Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population

More information

Approximate Dynamic Programming

Approximate Dynamic Programming Approximate Dynamic Programming A. LAZARIC (SequeL Team @INRIA-Lille) Ecole Centrale - Option DAD SequeL INRIA Lille EC-RL Course Value Iteration: the Idea 1. Let V 0 be any vector in R N A. LAZARIC Reinforcement

More information

STAT STOCHASTIC PROCESSES. Contents

STAT STOCHASTIC PROCESSES. Contents STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5

More information

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads Operations Research Letters 37 (2009) 312 316 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl Instability of FIFO in a simple queueing

More information

arxiv: v2 [cs.sy] 27 Sep 2016

arxiv: v2 [cs.sy] 27 Sep 2016 Analysis of gradient descent methods with non-diminishing, bounded errors Arunselvan Ramaswamy 1 and Shalabh Bhatnagar 2 arxiv:1604.00151v2 [cs.sy] 27 Sep 2016 1 arunselvan@csa.iisc.ernet.in 2 shalabh@csa.iisc.ernet.in

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

0.1 Uniform integrability

0.1 Uniform integrability Copyright c 2009 by Karl Sigman 0.1 Uniform integrability Given a sequence of rvs {X n } for which it is known apriori that X n X, n, wp1. for some r.v. X, it is of great importance in many applications

More information

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson

More information

A generalization of Strassen s functional LIL

A generalization of Strassen s functional LIL A generalization of Strassen s functional LIL Uwe Einmahl Departement Wiskunde Vrije Universiteit Brussel Pleinlaan 2 B-1050 Brussel, Belgium E-mail: ueinmahl@vub.ac.be Abstract Let X 1, X 2,... be a sequence

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Symmetric polynomials and symmetric mean inequalities

Symmetric polynomials and symmetric mean inequalities Symmetric polynomials and symmetric mean inequalities Karl Mahlburg Department of Mathematics Louisiana State University Baton Rouge, LA 70803, U.S.A. mahlburg@math.lsu.edu Clifford Smyth Department of

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence

More information

Convergence of generalized entropy minimizers in sequences of convex problems

Convergence of generalized entropy minimizers in sequences of convex problems Proceedings IEEE ISIT 206, Barcelona, Spain, 2609 263 Convergence of generalized entropy minimizers in sequences of convex problems Imre Csiszár A Rényi Institute of Mathematics Hungarian Academy of Sciences

More information

A mathematical framework for Exact Milestoning

A mathematical framework for Exact Milestoning A mathematical framework for Exact Milestoning David Aristoff (joint work with Juan M. Bello-Rivas and Ron Elber) Colorado State University July 2015 D. Aristoff (Colorado State University) July 2015 1

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

A Note on the Central Limit Theorem for a Class of Linear Systems 1

A Note on the Central Limit Theorem for a Class of Linear Systems 1 A Note on the Central Limit Theorem for a Class of Linear Systems 1 Contents Yukio Nagahata Department of Mathematics, Graduate School of Engineering Science Osaka University, Toyonaka 560-8531, Japan.

More information

MTH 404: Measure and Integration

MTH 404: Measure and Integration MTH 404: Measure and Integration Semester 2, 2012-2013 Dr. Prahlad Vaidyanathan Contents I. Introduction....................................... 3 1. Motivation................................... 3 2. The

More information

arxiv: v1 [cs.lg] 23 Oct 2017

arxiv: v1 [cs.lg] 23 Oct 2017 Accelerated Reinforcement Learning K. Lakshmanan Department of Computer Science and Engineering Indian Institute of Technology (BHU), Varanasi, India Email: lakshmanank.cse@itbhu.ac.in arxiv:1710.08070v1

More information

A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition

A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition Junfeng Wu, Karl Henrik Johansson and Ling Shi E-mail: jfwu@ust.hk Stockholm, 9th, January 2014 1 / 19 Outline Outline

More information

Continuous Functions on Metric Spaces

Continuous Functions on Metric Spaces Continuous Functions on Metric Spaces Math 201A, Fall 2016 1 Continuous functions Definition 1. Let (X, d X ) and (Y, d Y ) be metric spaces. A function f : X Y is continuous at a X if for every ɛ > 0

More information

Weak-Star Convergence of Convex Sets

Weak-Star Convergence of Convex Sets Journal of Convex Analysis Volume 13 (2006), No. 3+4, 711 719 Weak-Star Convergence of Convex Sets S. Fitzpatrick A. S. Lewis ORIE, Cornell University, Ithaca, NY 14853, USA aslewis@orie.cornell.edu www.orie.cornell.edu/

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms

Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms Yan Bai Feb 2009; Revised Nov 2009 Abstract In the paper, we mainly study ergodicity of adaptive MCMC algorithms. Assume that

More information

Research Article Almost Periodic Solutions of Prey-Predator Discrete Models with Delay

Research Article Almost Periodic Solutions of Prey-Predator Discrete Models with Delay Hindawi Publishing Corporation Advances in Difference Equations Volume 009, Article ID 976865, 19 pages doi:10.1155/009/976865 Research Article Almost Periodic Solutions of Prey-Predator Discrete Models

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

An Empirical Algorithm for Relative Value Iteration for Average-cost MDPs

An Empirical Algorithm for Relative Value Iteration for Average-cost MDPs 2015 IEEE 54th Annual Conference on Decision and Control CDC December 15-18, 2015. Osaka, Japan An Empirical Algorithm for Relative Value Iteration for Average-cost MDPs Abhishek Gupta Rahul Jain Peter

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

NUMERICAL RADIUS OF A HOLOMORPHIC MAPPING

NUMERICAL RADIUS OF A HOLOMORPHIC MAPPING Geometric Complex Analysis edited by Junjiro Noguchi et al. World Scientific, Singapore, 1995 pp.1 7 NUMERICAL RADIUS OF A HOLOMORPHIC MAPPING YUN SUNG CHOI Department of Mathematics Pohang University

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

Stability of optimization problems with stochastic dominance constraints

Stability of optimization problems with stochastic dominance constraints Stability of optimization problems with stochastic dominance constraints D. Dentcheva and W. Römisch Stevens Institute of Technology, Hoboken Humboldt-University Berlin www.math.hu-berlin.de/~romisch SIAM

More information

Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer

Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer 2009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 10-12, 2009 ThA06.6 Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer Serdar Yüksel

More information

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk Instructor: Victor F. Araman December 4, 2003 Theory and Applications of Stochastic Systems Lecture 0 B60.432.0 Exponential Martingale for Random Walk Let (S n : n 0) be a random walk with i.i.d. increments

More information

TECHNIQUES FOR ESTABLISHING DOMINATED SPLITTINGS

TECHNIQUES FOR ESTABLISHING DOMINATED SPLITTINGS TECHNIQUES FOR ESTABLISHING DOMINATED SPLITTINGS ANDY HAMMERLINDL ABSTRACT. We give theorems which establish the existence of a dominated splitting and further properties, such as partial hyperbolicity.

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

RATIO ERGODIC THEOREMS: FROM HOPF TO BIRKHOFF AND KINGMAN

RATIO ERGODIC THEOREMS: FROM HOPF TO BIRKHOFF AND KINGMAN RTIO ERGODIC THEOREMS: FROM HOPF TO BIRKHOFF ND KINGMN HNS HENRIK RUGH ND DMIEN THOMINE bstract. Hopf s ratio ergodic theorem has an inherent symmetry which we exploit to provide a simplification of standard

More information

The Art of Sequential Optimization via Simulations

The Art of Sequential Optimization via Simulations The Art of Sequential Optimization via Simulations Stochastic Systems and Learning Laboratory EE, CS* & ISE* Departments Viterbi School of Engineering University of Southern California (Based on joint

More information

A generalization of Dobrushin coefficient

A generalization of Dobrushin coefficient A generalization of Dobrushin coefficient Ü µ ŒÆ.êÆ ÆÆ 202.5 . Introduction and main results We generalize the well-known Dobrushin coefficient δ in total variation to weighted total variation δ V, which

More information

for all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true

for all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true 3 ohn Nirenberg inequality, Part I A function ϕ L () belongs to the space BMO() if sup ϕ(s) ϕ I I I < for all subintervals I If the same is true for the dyadic subintervals I D only, we will write ϕ BMO

More information

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem 56 Chapter 7 Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem Recall that C(X) is not a normed linear space when X is not compact. On the other hand we could use semi

More information

On the Central Limit Theorem for an ergodic Markov chain

On the Central Limit Theorem for an ergodic Markov chain Stochastic Processes and their Applications 47 ( 1993) 113-117 North-Holland 113 On the Central Limit Theorem for an ergodic Markov chain K.S. Chan Department of Statistics and Actuarial Science, The University

More information

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω

More information

A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1.

A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1. A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE THOMAS CHEN AND NATAŠA PAVLOVIĆ Abstract. We prove a Beale-Kato-Majda criterion

More information

Computations of Critical Groups at a Degenerate Critical Point for Strongly Indefinite Functionals

Computations of Critical Groups at a Degenerate Critical Point for Strongly Indefinite Functionals Journal of Mathematical Analysis and Applications 256, 462 477 (2001) doi:10.1006/jmaa.2000.7292, available online at http://www.idealibrary.com on Computations of Critical Groups at a Degenerate Critical

More information

AN EFFECTIVE METRIC ON C(H, K) WITH NORMAL STRUCTURE. Mona Nabiei (Received 23 June, 2015)

AN EFFECTIVE METRIC ON C(H, K) WITH NORMAL STRUCTURE. Mona Nabiei (Received 23 June, 2015) NEW ZEALAND JOURNAL OF MATHEMATICS Volume 46 (2016), 53-64 AN EFFECTIVE METRIC ON C(H, K) WITH NORMAL STRUCTURE Mona Nabiei (Received 23 June, 2015) Abstract. This study first defines a new metric with

More information

On the fast convergence of random perturbations of the gradient flow.

On the fast convergence of random perturbations of the gradient flow. On the fast convergence of random perturbations of the gradient flow. Wenqing Hu. 1 (Joint work with Chris Junchi Li 2.) 1. Department of Mathematics and Statistics, Missouri S&T. 2. Department of Operations

More information

Lecture Notes 10: Dynamic Programming

Lecture Notes 10: Dynamic Programming University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 81 Lecture Notes 10: Dynamic Programming Peter J. Hammond 2018 September 28th University of Warwick, EC9A0 Maths for Economists Peter

More information

Wald for non-stopping times: The rewards of impatient prophets

Wald for non-stopping times: The rewards of impatient prophets Electron. Commun. Probab. 19 (2014), no. 78, 1 9. DOI: 10.1214/ECP.v19-3609 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS in PROBABILITY Wald for non-stopping times: The rewards of impatient prophets Alexander

More information

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University February 7, 2007 2 Contents 1 Metric Spaces 1 1.1 Basic definitions...........................

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information