On Differentiability of Average Cost in Parameterized Markov Chains
|
|
- Junior Allen Harrell
- 5 years ago
- Views:
Transcription
1 On Differentiability of Average Cost in Parameterized Markov Chains Vijay Konda John N. Tsitsiklis August 30, Overview The purpose of this appendix is to prove Theorem 4.6 in 5 and establish various facts used in verifying some of the assumptions of Theorem A.1 therein. We use the same notation and assumptions as in 5, and to simplify the analysis, we assume that the parameter is scalar. The setting used here is similar (but not identical) to the one in 4. The major difference is that the setting of 4 allows only RSPs of the form U k = f(x k,w k ), where the parameter enters only through the distribution of W k which does not depend on x. Incontrast, our setting allows for any RSPs of the form U k = f (X k,w k ), as long as the distribution µ (u x) off (x, U) satisfies the required assumptions. The differentiability results presented in this report are based on the following key lemma (see 3, Corollary 2.2.1). Lemma 1.1. Consider a parameterized family of random variables {F (ω); R} on a Polish space Ω. Suppose the following hold: 1. For each 0, there exists some d>1 and K d > 0 such that, E F F 0 d <K d 0 d, for all sufficiently close to There is a family of random variables {f } such that where F +h F h P denotes convergence in probability. P f, as h 0,, 1
2 Then the map E F is differentiable with d d E F =E f. Our proof of the differentiability results involves the following steps. 1. Obtain a representation of the steady-state expected cost and of solutions to the Poisson equation in terms of suitably defined expectations E,x. 2. Using likelihood ratios, represent the above expectations using only the expectations E 0,x corresponding to some fixed 0. This new representation has the advantage that the probability law w.r.t. which expectations are taken does not depend on the parameter. Only the random variables inside the expectation depend on the parameter, making our key result Lemma 1.1 applicable to this representation. 3. Verify the hypotheses of Lemma 1.1. The next section carries out these steps for expectations over finite horizon. In Section 3, we recall the results on the regenerative representations for the average cost and solutions to the Poisson equation. In Section 4, we introduce the likelihood ratio representations and show that they satisfy the assumptions of Lemma 1.1. In Sections 5 and 6, we use all the previous results to prove differentiability of expectations over infinite horizon and verify some of the assumptions of Theorem A.1 in 5. 2 Finite Horizon In this section, we prove the differentiability of expectations of the form E,x W where W is a random variable that depends only on a finite number of state-decision pairs (X k,u k ). For each and k, let Λ,k = k µ (U l X l ) µ 0 (U l X l ). It is easy to see that Λ,k is the Radon-Nikodym derivative of the distribution of (X l,u l ) k under policy with respect to the distribution under policy 0.Wewill use the following basic lemma to derive some bounds on Λ,k. Lemma 2.1. If W 0,...,W k 1 are positive random variables satisfying E W l 1 d ɛ d C(d), d 1, l =0,...,k 1, for some monotonic function C( ) 1, and if ɛ<1/k2 k, then k 1 d E W l 1 k 1 ɛ d (2k) d C(kd), E 2 d C(kd), for any d 1. W d l 2
3 Proof. For any d 1 and z 0, we have z d 2 d 1 ( 1+ z 1 d), (1) we have E W d l 2 d 1 (1 + ɛ d C(d)) 2 d C(d), l =0,...,k 1. To prove the first part, note that and therefore z 0 z k 1 1 z z 0 z z 0 z k 2 z k 1 1 z 0 z k 1 1 d k d 1 z 0 1 d + z d 0 z 1 1 d + + z d 0 z d k 2 z k 1 1 d. We now use Holder s inequality to see that k 1 d E k 1 l 1 W l 1 k d 1 E Wj d W l 1 d k 1 l 1 k d 1 j=0 j=0 E W (l+1)d j k 1 k d 1 2 ld ɛ d C((l +1)d) k d ɛ d 2 kd C(kd). 1/(l+1) E W l 1 (l+1)d 1/(l+1) To prove the second part, use (1) and the fact that ɛ<1/k2 k to simplify the bound. The following are two immediate consequences of the above lemma. In the following lemma, note that the constant K d might depend on k. This dependence can be safely ignored as the next lemma is intended to verify the first assumption of Lemma 1.1 for a fixed k. Lemma For every d 1, there exists K d > 0 such that Λ,k 1 d K d 0 d L(x),, x. E 0,x 2. For every d/ (1, 1), there exists K d > 0 such that Λ,k d K d L(x), x, E 0,x and for all sufficiently close to 0. 3
4 Proof. Use the mean-value theorem to see that µ (u x) µ 0 (u x) 1 0 sup µ (u x) µ 0 (u x). (2) Since µ (u x)/µ 0 (u x) belongs to D, wehave E,x sup µ (X k,u k ) K d L(x), 0 µ 0 (X k,u k ) for some constant K d.part 1, and Part 2 for d 1, are implied by Lemma 2.1. To prove Part 2 for d 1, note that the inequality (2) holds even if and 0 are interchanged. The above lemma verifies the first assumption of Lemma 1.1 for expectations of random variables that are functions of state-decision pairs up to time k. The next lemma verifies the second assumption and derives a formula for the gradient. Lemma 2.3. Consider a family of random variables {W } such that the map W (ω) is differentiable for each ω, W is a function of the state-decision pairs up to time k, and, 0 E,x W d <, E,x sup W d <, x, for some d>1. Then, the map E,x W is differentiable with where E,x W =E,x W +E,x Λ,k W, Λ,k = k ψ (X l,u l )= k µ (X l,u l ) µ (X l,u l ). Proof. Note that E,x W =E 0,x Λ,k W. Using the bounds of the previous lemma, the fact that d-moments of W and sup W are finite, and the mean-value theorem, verify the first assumption of Lemma 1.1. Using the fact that µ (u x) and W are differentiable in, the second assumption of Lemma 1.1 follows easily. The gradient formula now follows from interchanging the order of differentiation and expectation. The following is an immediate consequence of the previous result. It uses the formula in the previous result to derive bounds on the rate at which the gradient grows with k. E,x f (X k,u k ) 4
5 Corollary 2.4. Consider a family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D. Then, the map E,x f (X k,u k ) is differentiable for each x, and there exists some K>0 such that E,x f (X k,u k ) (k +1)KL(x), x. Furthermore, the families of functions belong to D. {E,x f (X k,u k )}, { E,x f (X k,u k )} Before we move on, we would like to point out that the above result holds for expectations of the form E,x U 0 = u aswell, if we redefine Λ,k as 3 Splitting k Λ,0 =0, Λ,k = ψ (X l,u l ), k >0. l=1 In this section, we recall the splitting technique of Athreya and Ney 2, and Nummelin 7 to obtain a regenerative representation for the average cost function ᾱ() and the solutions Q of the Poisson equation. Let δ, N, X 0 and ϑ be as in Assumption 4.2 of 5. Consider the {0, 1}-valued process (B k ) constructed as follows: 1. If k is not divisible by N, then B k =0. 2. If k is divisible by N, then P(B k =1 X l,u l,l =0, 1,...,B l,l k) = P(B k =1 X k,x k+n ) = (X k,x k+n ) where = δ 2 I ϑ(dy) X 0 (x) P,x (X N dy). (Assumption 4.2(a) ensures existence of.) Let It is easy to see that for each, x, f (0) = 1,, x, y, Q,x (dy) = P,x(X N dy) δ 2 I X 0 (x)ϑ(dy) 1 δ 2 I. X 0 (x) f (0) (x, ) = ( 1 δ 2 I X 0 (x) 5 ) Q,x (dy) P,x (X N dy)
6 where the existence of the Radon-Nikodym derivative is guaranteed by the fact that 2 Q,x (dy) 2 δ P,x(X N dy). With this construction it is not difficult to see that the process (X kn,b kn )isamarkov chain for which X {1} is an atom. Consider the first time this atom is hit, i.e., τ = min{k B k =1}. The time τ is well-defined and finite w.p.1. as the set X 0 is hit infinitely often from any initial state (cf. Assumption 4.2 in 5) and the probability that B k is 1 given that X k = x is δ/2 whenever x X 0. Using standard results 6, Theorems and on Markov chains with atoms, the average cost ᾱ() can be written as (τ/n) 1 E,ϑ c,n (X kn ) ᾱ() =, E,ϑ τ where c,n (x) = N 1 Furthermore, the functions V (x) given by V (x) =E,x satisfy the Poisson equation (τ/n) 1 E,x c(x k,u k ). ( c,n (X kn ) Nᾱ()) (3) V (x) =E,x c(x, U 0 )+V (X 1 ) for the Markov chain {X k }.Tosee this, note that V (x) isasolution to the Poisson equation for the cost function c,n and the Markov chain {X kn } (see 6, p.441 ). Furthermore, if ˆV is a solution to the Poisson equation for the cost function c,1 and the Markov chain {X k }, then ˆV is a solution to the Poisson equation for the cost function c,n and the Markov chain {X kn }. Since ˆV and V are two solutions to the Poisson equation for a positive Harris chain {X kn }, they must differ by a constant (cf. Proposition of 6) and therefore, V is a solution to the Poisson equation for the cost function c,1 and the Markov chain {X k }. 4 Likelihood Ratio For each, k, let where Λ,k =Λ,kN ˆΛ,k 6
7 k 1 f ˆΛ (0) (X ln,x ln+n ),k = f (0) 0 (X ln,x ln+n ), and 0/0 isinterpreted as 1. It is easy to see that Λ,k is the Radon-Nikodym derivative of the joint distribution of (X l,u l ;0 l kn, B l ;0 l (k 1)N) under P with respect to that under P 0 on the set of outcomes for which τ>kn. Therefore for any W that is a function of only these random variables, we have E,x WI{τ >kn} =E 0,x WI{τ >kn} Λ,k. To show that the map E,x WI{τ >kn} is differentiable, we need to verify the hypothesis of Lemma 1.1 for the likelihood ratio Λ,k.Wenow derive several bounds to verify the first assumption of Lemma 1.1. Lemma For each d 1, there exists K d > 0 such that E 0 f (i),x (x, X d N) f (i) 0 (x, X N ) 1 K d 0 d L(x), x, i =0, For each d 1, there exists K d > 0 such that E 0 f (i),x (x, X d N) f (i) K 0 (x, X N ) d L(x) x, i =0, 1, and for all sufficiently close to For each d 1, there exists K d > 0 such that E 0,x ˆΛ,k 1 d K d 0 d L(x), x. 4. For each d 1, there exists K d > 0 such that E 0,x ˆΛ,k d K d L(x) x, and for all sufficiently close to 0. 7
8 Proof. To prove the first part, note that for x/ X 0, 0 = 1, and for x X 0, 0 = P 0,x(X N dy) P,x (X N dy) = {E 0,x Λ,N X N = y} 1. Therefore, we have E 0 (x, X N ) d,x 0 (x, X N ) 1 E 0,x E 0,x E 0,x E0,x Λ,N X N 1 1 d E 0,x Λ,N X N 2d E 0,x E 0,x Λ,N 1 X N 2d Λ 2d,N E 0,x Λ,N 1 2d, where the last two inequalities follow from Holder s and Jensen s inequalities respectively. For i =0,note that f (0) f (0) 0 1 = ( (1) 0 1 f f (0) 0 since 1/2 for all. The proof of the second part is similar. The last two parts follow from the first two and Lemma 2.1. The next lemma proves that the first of the two hypotheses for Lemma 1.1 hold for Λ. Lemma For each d 1, there exists K d > 0 such that E 0,x Λ,k 1 d K d 0 d L(x), for sufficiently close to For each d 1, there exists K d > 0 such that E 0,x Λ,k d K d L(x). ) 8
9 Proof. Follows from Lemmas 2.2, 4.1 and 2.1. We are now ready to state the following result on differentiability of finite horizon expectations. Lemma 4.3. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map E,x f (X kn )I{τ >kn}, where f (x) = N 1 E,x f (X k,u k ), is differentiable for each x, with E,x f (X kn )I{τ >kn} (k +1)ρ k 0KL(x), for some ρ 0 < 1. Proof. Since E,x f (X kn )I{τ >kn} = E 0,x Λ,k f (X kn )I{τ >kn}, we can use Lemma 2.1 to prove its differentiability and to calculate the derivative. To verify the first assumption of Lemma 2.1, use Holder s inequality to see that E 0,x f (X kn )I{τ >kn} Λ,k f (X kn )I{τ >kn} Λ 0,k d = E 0,x f (X kn ) d Λ,k 1 d E 0,x f (X kn ) 2d 1/2 E0,x Λ,k 1 2d 1/2. The first assumption now follows from the fact that f (x, u) (and therefore f (x)) belongs to D, and the first part of the previous lemma. To verify the second assumption of Lemma 1.1, use Assumption 4.4 of 5 to conclude that the map Λ,k (ω) is continuously differentiable for all ω, and therefore it is enough to prove that ˆΛ 0 +h,k 1 converges in probability. This in turn is verified if the functions f (i) h (i) /f 0,i=0, 1, can be shown to be differentiable w.r.t. for all x, y. Since 0 / 9
10 is equal to E 0,x Λ,N X N = y, for all x X 0 and is equal to 1 otherwise, use part 3 of Assumption 4.4 in 5 and the mean-value theorem to see that Λ 0 +h,n 1 h is bounded above by an integrable random variable. Therefore, we invoke dominated convergence theorem for conditional expectations to show that there exists a version of / 0 that is differentiable at 0 for all x, y. Furthermore, Λ 0 +h 1 h converges in probability to Λ 0,k where kn k 1 jn+n 1 Λ,k = ψ (X k,u k ) I X0 (X jn )E,x ψ (X l,u l ) j=0 l=jn X jn,x jn+n. Therefore, we have E,x f (X kn )I{τ >kn} = E,x Λ,k f (X kn )I{τ >kn} +E,x Λ,k f (X kn )I{τ >kn} To derive the bound on the derivative, use Holder s inequality to obtain E,x f (X kn )I{τ >k} 3 1/3 E,x Λ,k E,x f (X kn ) 3 1/3 P,x (τ>kn) 1/3 +E,x f (X kn ) 3 1/3 P,x (τ>kn) 1/3. To calculate E,x Λ,k 3 1/3, note that Λ,k is a sum of 2kN +1 terms. Since ψ (x, u) belongs to D, the 3-norm of each of these terms is bounded by KL(x) 1/3 for some K. Similarly, since f (x, u) and f (x, u) (therefore f (x) and f (x)) belong to D, the terms E,x f (X kn ) 3 1/3 and E,x f (X kn ) 3 1/3 are also bounded by KL(x) 1/3 for some K. Finally, due to uniform geometric ergodicity, we have P,x (τ>kn) Kρ k 0L(x), for some ρ 0 < 1. Combining all these terms it is easy to establish the stated bound on the derivative. 10
11 5 Infinite Horizon The above lemma concerns the differentiability of expectation w.r.t. finite dimensional marginals of P,x. We use the following result from advanced calculus (e.g., see Apostol 1) to extend the above result to the infinite horizon case. Theorem 5.1. Assume that each f k is a real-valued function defined on a neighborhood Θ of 0 such that the derivative f k () exists for each in Θ. Assume that 1. f k ( 0 ) converges, 2. there exists a g() such that f k () =g() uniformly on Θ. Then: 1. There exists a function f() such that f k () =f() uniformly on Θ. 2. If Θ, the derivative f() exists and equals f k (). By combining the previous two results we have the following theorem. Recall that η is the steady state expectation of the state-decision process {(X k,u k )}. Theorem 5.2. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map (τ/n) 1 E,x f (X kn ) is differentiable for each x, and the families of functions (τ/n) 1 E,x f (X kn ), E,x (τ/n) 1 f (X kn ), belong to D. Proof. Note that E,x (τ/n) 1 f (X kn ) = E,x f (X kn )I{τ >kn}. The differentiability now follows from the previous two lemmas. To show that the functions (τ/n) 1 E (τ/n) 1,x f (X kn ), E,x f (X kn ) 11
12 are in D, consider E,x sup E,x E,x (τ/n) 1 (τ/n) 1 d f (X kn ) d sup f (X kn ) sup f (X d kn ) (τ/n) 1 (τ/n) d 1 E,x (τ/n) d 1 sup f (X kn ) E,x (τ/n) 3(d 1) 1/3 E,x sup P,x (τ>kn) 1/3. d I{τ >kn} f (X kn ) 3d 1/3 It is now easy to see that the right hand side is bounded by K d L(x) for some K d > 0. The following corollaries are immediate consequences of this result. Corollary 5.3. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map η (f ) is bounded and differentiable with bounded derivatives. Corollary 5.4. (Lemma 5.1 in 5) If X 0 is singleton (i.e., X 0 is an atom and therefore splitting in unnecessary) and τ is the first hitting time of X 0 then the functions defined as τ 1 Q (x, u) = E,x c(x k,u k ) U 0 = u, T (x, u) = E,x τ U 0 = u, belong to D. Furthermore, for each (x, u), the maps Q (x, u), T (x, u) are differentiable, and the derivatives Q (x, u), T (x, u) also belong to D. Corollary 5.5. The average expected cost function ᾱ() is bounded and differentiable with bounded derivatives. Furthermore, the solutions V (x) to the Poisson equation given by (3) belong to D and the map V (x) is differentiable for each x and the derivatives V (x) also belong to D. Furthermore, the family of functions Q = c ᾱ()1 + P V 12
13 belongs to D, with the map Q (x, u) being differentiable for all (x, u) and Q (x, u) belonging to D. This corollary is used in 5 to establish Theorem 4.6 in 5, which is again proved here for completeness. Theorem 5.6. For any solution Q to the Poisson equation we have ᾱ() = ψ,q. Furthermore, ᾱ() has bounded derivatives. Proof. Note that the Q defined in the previous corollary satisfies the Poisson equation: Q (x, u) =c(x, u) ᾱ()+e,x Q (X 1,U 1 ) U 0 = u. Since Q belongs to D, Q is differentiable in, and Q also belongs to D, wecan differentiate both sides of the above equation to obtain Q (x, u) = ᾱ()+e,x Q (X 1,U 1 ) U 0 = u +E,x ψ (X 1,U 1 )Q (X 1,U 1 ) U 0 = u. Taking expectation with respect to the steady state distribution of (X k,u k )weobtain the desired formula with Q defined as in the previous corollary. To see that Q can be replaced by any other solution ˆQ to the Poisson equation, note that Q ˆQ is some function C() (Proposition of 6) and that E,x ψ (x, U 0 ) = 0. To prove that ᾱ is twice differentiable with bounded derivatives, note that ᾱ = η (ψ Q ) and apply Corollary 5.3. The above result leads to the following theorem on differentiability of a common representation for solutions to Poisson equation. Theorem 5.7. For any family of functions {f (x, u)} in D such that the family { f (x, u)} is also in D, the map E,x f (X k,u k ) η (f ) is differentiable for each x, and the families of functions { } { } E,x f (X k,u k ) η (f ), E,x f (X k,u k ) η (f ), belong to D. 13
14 Proof. To simplify the proof, assume that f = c. Then the results on differentiability for the finite horizon case, imply that the function is differentiable with E,x c(x k,u k ) ᾱ() E,x c(x k,u k ) ᾱ() k = E,x ψ (X l,u l )(c(x k,u k ) ᾱ()) ᾱ() k/2 + E,x ψ (X l,u l )E,Xl+1 c(x k,u k ) ᾱ() k l=k/2+1 + l=k/2 E,x ψ (X l,u l )(c(x k,u k ) ᾱ()) ψ,p k l ψ,p l (c ᾱ()1), (c ᾱ()1) where k/2 represents the floor of k/2. Using geometric ergodicity, the fact that V 1/d also satisfies the geometric Foster Lyapunov criterion with the decay factor ρ 1/d, and Schwartz inequality, we can see that each of these terms can be bounded by Kρ k 0L(x) for some ρ 0 < 1. The result now follows from an easy application of Theorem Verification of Assumption A.3 parts (d) and (e) In this section, we verify parts (d) and (e) of Assumption A.3 of 5 for the two cases of TD(1) and of TD(λ), λ<1 separately. Since φ,q,t belong to D, part (d) in both cases follows from Corollary 5.3. Therefore, we will verify only part (e). We start with the TD(1) case. 6.1 TD(1) We will now verify part (e) for ĥ( ). Note that τ 1 ĥ (y) = E, x (h (Y k ) h()) Y 0 = y, = E, x (h (Y k ) h())i{τ >k} Y 0 = y. Since h (Y k )isafunction of the state-decision pairs up to time k and satisfies the assumptions of Lemma 2.3, a formula can be derived for the derivative of the expectation E, x (h (Y k ) h())i{τ >k} Y 0 = y. 14
15 Using this formula, a bound on this derivative of the form K(k +1) d ρ k L(x, 0 u)z,d 1,K >0,ρ 0 < 1, L D. can be obtained. This bound and Lemma 5.1 can be used to conclude that ĥ is differentiable with respect to and the derivative is bounded by K L(x, u)z for some other K>0. It is now easy to verify that E,x ĥ (Y 1 ) Y 0 = y is differentiable with the derivative bounded appropriately. The verification of part (e) for Ĝ is similar. 6.2 TD(λ), λ<1 Since the verification of part (e) for ĥ is simpler and the verification for Ĝ is similar, we will consider ĥ only. Note that the first component of the vector ĥ does not depend on z and is equal to L E,x (c(x k,u k ) ᾱ()) U 0 = u. Therefore, part (e) for this component follows from Lemma 5.7. For the second component, consider the sum E,x c(xk,u k )Z k h 1 () ᾱ() Z() Y0 = y = z + = z + λ k E,x c(x k,u k ) P l c, φ Y0 = y k 1 λ l E,x c(xk,u k )φ (X k l,u k l ) P l c, φ Y0 = y + l=k+1 λ l P l c, φ λ k E,x c(x k,u k ) P l c, φ Y0 = y + λ l k=l+1 λ l l 1 E,x c(xk,u k ))φ (X k l,u k l ) P l c ᾱ()1,φ Y 0 = y P l c, φ. It is now easy to see that the second sums of the first two terms are differentiable with the derivatives uniformly bounded in l as the second sum of the second term is an expectation of the solution to a Poisson equation for the Markov chain (X k l,u k l,x k,u k ) (cf. Lemma 5.7). The second sum of the third term is also differentiable with the bound on the derivative linear in k. Itnow follows from Theorem 5.1 that ĥ(y) isdifferentiable with the derivative bounded above by K L(x, u)z for some L D. 15
16 References 1 T. M. Apostol. Mathematical Analysis: A Modern Approach to Advanced Calculus. Addison-Wesley Pub Co, 2 edition, January K. B. Athreya and P. Ney. A new approach to the limit theory of recurrent Markov chains. Trans. Amer. Math. Soc., 245: , V. S. Borkar. Probability theory: an advanced course. Springer-Verlag, New York, P. W. Glynn and P. L Ecuyer. Likelihood ratio gradient estimation for stochastic recursions. Advances in applied probability, 27: , V. R. Konda and J. N. Tsitsiklis. Actor-critic algorithms. Submitted to the SIAM Journal on Control and Optimization, February S. P. Meyn and R. L. Tweedie. Markov Chains and Stochastic Stability. Springer- Verlag, E. Nummelin. A splitting technique for Harris recurrent chains. Z. Wahrscheinlichkeitstheorie and Verw. Geb., 43: ,
A System Theoretic Perspective of Learning and Optimization
A System Theoretic Perspective of Learning and Optimization Xi-Ren Cao* Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong eecao@ee.ust.hk Abstract Learning and optimization
More informationConvergence Rates for Renewal Sequences
Convergence Rates for Renewal Sequences M. C. Spruill School of Mathematics Georgia Institute of Technology Atlanta, Ga. USA January 2002 ABSTRACT The precise rate of geometric convergence of nonhomogeneous
More informationVerifying Regularity Conditions for Logit-Normal GLMM
Verifying Regularity Conditions for Logit-Normal GLMM Yun Ju Sung Charles J. Geyer January 10, 2006 In this note we verify the conditions of the theorems in Sung and Geyer (submitted) for the Logit-Normal
More informationA regeneration proof of the central limit theorem for uniformly ergodic Markov chains
A regeneration proof of the central limit theorem for uniformly ergodic Markov chains By AJAY JASRA Department of Mathematics, Imperial College London, SW7 2AZ, London, UK and CHAO YANG Department of Mathematics,
More informationLecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.
1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if
More informationFaithful couplings of Markov chains: now equals forever
Faithful couplings of Markov chains: now equals forever by Jeffrey S. Rosenthal* Department of Statistics, University of Toronto, Toronto, Ontario, Canada M5S 1A1 Phone: (416) 978-4594; Internet: jeff@utstat.toronto.edu
More informationLinear stochastic approximation driven by slowly varying Markov chains
Available online at www.sciencedirect.com Systems & Control Letters 50 2003 95 102 www.elsevier.com/locate/sysconle Linear stochastic approximation driven by slowly varying Marov chains Viay R. Konda,
More informationRecall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm
Chapter 13 Radon Measures Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm (13.1) f = sup x X f(x). We want to identify
More informationApproximate Dynamic Programming
Master MVA: Reinforcement Learning Lecture: 5 Approximate Dynamic Programming Lecturer: Alessandro Lazaric http://researchers.lille.inria.fr/ lazaric/webpage/teaching.html Objectives of the lecture 1.
More informationOn Finding Optimal Policies for Markovian Decision Processes Using Simulation
On Finding Optimal Policies for Markovian Decision Processes Using Simulation Apostolos N. Burnetas Case Western Reserve University Michael N. Katehakis Rutgers University February 1995 Abstract A simulation
More informationWeighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing
Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes
More informationNonparametric inference for ergodic, stationary time series.
G. Morvai, S. Yakowitz, and L. Györfi: Nonparametric inference for ergodic, stationary time series. Ann. Statist. 24 (1996), no. 1, 370 379. Abstract The setting is a stationary, ergodic time series. The
More informationx log x, which is strictly convex, and use Jensen s Inequality:
2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and
More informationA CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE
Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received
More informationLecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.
Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although
More informationDynamic Programming with Hermite Interpolation
Dynamic Programming with Hermite Interpolation Yongyang Cai Hoover Institution, 424 Galvez Mall, Stanford University, Stanford, CA, 94305 Kenneth L. Judd Hoover Institution, 424 Galvez Mall, Stanford University,
More informationWhen is a Markov chain regenerative?
When is a Markov chain regenerative? Krishna B. Athreya and Vivekananda Roy Iowa tate University Ames, Iowa, 50011, UA Abstract A sequence of random variables {X n } n 0 is called regenerative if it can
More informationHarmonic Functions and Brownian motion
Harmonic Functions and Brownian motion Steven P. Lalley April 25, 211 1 Dynkin s Formula Denote by W t = (W 1 t, W 2 t,..., W d t ) a standard d dimensional Wiener process on (Ω, F, P ), and let F = (F
More informationChapter 4: Asymptotic Properties of the MLE
Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider
More informationLectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes
Lectures on Stochastic Stability Sergey FOSS Heriot-Watt University Lecture 4 Coupling and Harris Processes 1 A simple example Consider a Markov chain X n in a countable state space S with transition probabilities
More informationAn almost sure invariance principle for additive functionals of Markov chains
Statistics and Probability Letters 78 2008 854 860 www.elsevier.com/locate/stapro An almost sure invariance principle for additive functionals of Markov chains F. Rassoul-Agha a, T. Seppäläinen b, a Department
More informationIndependence of some multiple Poisson stochastic integrals with variable-sign kernels
Independence of some multiple Poisson stochastic integrals with variable-sign kernels Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological
More informationL p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by
L p Functions Given a measure space (, µ) and a real number p [, ), recall that the L p -norm of a measurable function f : R is defined by f p = ( ) /p f p dµ Note that the L p -norm of a function f may
More informationLocal semiconvexity of Kantorovich potentials on non-compact manifolds
Local semiconvexity of Kantorovich potentials on non-compact manifolds Alessio Figalli, Nicola Gigli Abstract We prove that any Kantorovich potential for the cost function c = d / on a Riemannian manifold
More informationSUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS
Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,
More informationCentral limit theorems for ergodic continuous-time Markov chains with applications to single birth processes
Front. Math. China 215, 1(4): 933 947 DOI 1.17/s11464-15-488-5 Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes Yuanyuan LIU 1, Yuhui ZHANG 2
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationIEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence
IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence 1 Overview We started by stating the two principal laws of large numbers: the strong
More informationAW -Convergence and Well-Posedness of Non Convex Functions
Journal of Convex Analysis Volume 10 (2003), No. 2, 351 364 AW -Convergence Well-Posedness of Non Convex Functions Silvia Villa DIMA, Università di Genova, Via Dodecaneso 35, 16146 Genova, Italy villa@dima.unige.it
More informationCentral-limit approach to risk-aware Markov decision processes
Central-limit approach to risk-aware Markov decision processes Jia Yuan Yu Concordia University November 27, 2015 Joint work with Pengqian Yu and Huan Xu. Inventory Management 1 1 Look at current inventory
More informationAlmost Sure Convergence of Two Time-Scale Stochastic Approximation Algorithms
Almost Sure Convergence of Two Time-Scale Stochastic Approximation Algorithms Vladislav B. Tadić Abstract The almost sure convergence of two time-scale stochastic approximation algorithms is analyzed under
More informationSTABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL
First published in Journal of Applied Probability 43(1) c 2006 Applied Probability Trust STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL LASSE LESKELÄ, Helsinki
More informationis a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.
Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable
More informationUniversity of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming
University of Warwick, EC9A0 Maths for Economists 1 of 63 University of Warwick, EC9A0 Maths for Economists Lecture Notes 10: Dynamic Programming Peter J. Hammond Autumn 2013, revised 2014 University of
More informationLaplace s Equation. Chapter Mean Value Formulas
Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic
More informationMeasure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond
Measure Theory on Topological Spaces Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond May 22, 2011 Contents 1 Introduction 2 1.1 The Riemann Integral........................................ 2 1.2 Measurable..............................................
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationOn Reparametrization and the Gibbs Sampler
On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department
More information2 light traffic derivatives for the GI/G/ queue. We shall see that our proof of analyticity is mainly based on some recursive formulas very similar to
Analyticity of Single-Server Queues in Light Traffic Jian-Qiang Hu Manufacturing Engineering Department Boston University Cummington Street Boston, MA 0225 September 993; Revised May 99 Abstract Recently,
More informationTail inequalities for additive functionals and empirical processes of. Markov chains
Tail inequalities for additive functionals and empirical processes of geometrically ergodic Markov chains University of Warsaw Banff, June 2009 Geometric ergodicity Definition A Markov chain X = (X n )
More informationBrownian Motion and Conditional Probability
Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical
More informationA Probabilistic Proof of the Perron-Frobenius Theorem
A Probabilistic Proof of the Perron-Frobenius Theorem Peter W. Glynn 1 and Paritosh Y. Desai 2 1 Department of Management Science and Engineering, Stanford University, Stanford CA 2 Target Corporation,
More informationLecture 22: Variance and Covariance
EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce
More informationLimit theorems for dependent regularly varying functions of Markov chains
Limit theorems for functions of with extremal linear behavior Limit theorems for dependent regularly varying functions of In collaboration with T. Mikosch Olivier Wintenberger wintenberger@ceremade.dauphine.fr
More informationChapter 7. Markov chain background. 7.1 Finite state space
Chapter 7 Markov chain background A stochastic process is a family of random variables {X t } indexed by a varaible t which we will think of as time. Time can be discrete or continuous. We will only consider
More informationTHE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS
Statistica Sinica 6(1996), 943-956 THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS H. Z. An and F. C. Huang Academia Sinica Abstract: Consider the nonlinear autoregressive model x t = φ(x
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationChapter 2: Markov Chains and Queues in Discrete Time
Chapter 2: Markov Chains and Queues in Discrete Time L. Breuer University of Kent 1 Definition Let X n with n N 0 denote random variables on a discrete space E. The sequence X = (X n : n N 0 ) is called
More informationOn the Applicability of Regenerative Simulation in Markov Chain Monte Carlo
On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo James P. Hobert 1, Galin L. Jones 2, Brett Presnell 1, and Jeffrey S. Rosenthal 3 1 Department of Statistics University of Florida
More information1 Problem Formulation
Book Review Self-Learning Control of Finite Markov Chains by A. S. Poznyak, K. Najim, and E. Gómez-Ramírez Review by Benjamin Van Roy This book presents a collection of work on algorithms for learning
More informationProperties of an infinite dimensional EDS system : the Muller s ratchet
Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population
More informationApproximate Dynamic Programming
Approximate Dynamic Programming A. LAZARIC (SequeL Team @INRIA-Lille) Ecole Centrale - Option DAD SequeL INRIA Lille EC-RL Course Value Iteration: the Idea 1. Let V 0 be any vector in R N A. LAZARIC Reinforcement
More informationSTAT STOCHASTIC PROCESSES. Contents
STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5
More informationOperations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads
Operations Research Letters 37 (2009) 312 316 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl Instability of FIFO in a simple queueing
More informationarxiv: v2 [cs.sy] 27 Sep 2016
Analysis of gradient descent methods with non-diminishing, bounded errors Arunselvan Ramaswamy 1 and Shalabh Bhatnagar 2 arxiv:1604.00151v2 [cs.sy] 27 Sep 2016 1 arunselvan@csa.iisc.ernet.in 2 shalabh@csa.iisc.ernet.in
More informationSTAT 7032 Probability Spring Wlodek Bryc
STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,
More information0.1 Uniform integrability
Copyright c 2009 by Karl Sigman 0.1 Uniform integrability Given a sequence of rvs {X n } for which it is known apriori that X n X, n, wp1. for some r.v. X, it is of great importance in many applications
More information40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology
Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson
More informationA generalization of Strassen s functional LIL
A generalization of Strassen s functional LIL Uwe Einmahl Departement Wiskunde Vrije Universiteit Brussel Pleinlaan 2 B-1050 Brussel, Belgium E-mail: ueinmahl@vub.ac.be Abstract Let X 1, X 2,... be a sequence
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationSymmetric polynomials and symmetric mean inequalities
Symmetric polynomials and symmetric mean inequalities Karl Mahlburg Department of Mathematics Louisiana State University Baton Rouge, LA 70803, U.S.A. mahlburg@math.lsu.edu Clifford Smyth Department of
More informationExercises with solutions (Set D)
Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where
More informationMGMT 69000: Topics in High-dimensional Data Analysis Falll 2016
MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence
More informationConvergence of generalized entropy minimizers in sequences of convex problems
Proceedings IEEE ISIT 206, Barcelona, Spain, 2609 263 Convergence of generalized entropy minimizers in sequences of convex problems Imre Csiszár A Rényi Institute of Mathematics Hungarian Academy of Sciences
More informationA mathematical framework for Exact Milestoning
A mathematical framework for Exact Milestoning David Aristoff (joint work with Juan M. Bello-Rivas and Ron Elber) Colorado State University July 2015 D. Aristoff (Colorado State University) July 2015 1
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More informationA Note on the Central Limit Theorem for a Class of Linear Systems 1
A Note on the Central Limit Theorem for a Class of Linear Systems 1 Contents Yukio Nagahata Department of Mathematics, Graduate School of Engineering Science Osaka University, Toyonaka 560-8531, Japan.
More informationMTH 404: Measure and Integration
MTH 404: Measure and Integration Semester 2, 2012-2013 Dr. Prahlad Vaidyanathan Contents I. Introduction....................................... 3 1. Motivation................................... 3 2. The
More informationarxiv: v1 [cs.lg] 23 Oct 2017
Accelerated Reinforcement Learning K. Lakshmanan Department of Computer Science and Engineering Indian Institute of Technology (BHU), Varanasi, India Email: lakshmanank.cse@itbhu.ac.in arxiv:1710.08070v1
More informationA Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition
A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition Junfeng Wu, Karl Henrik Johansson and Ling Shi E-mail: jfwu@ust.hk Stockholm, 9th, January 2014 1 / 19 Outline Outline
More informationContinuous Functions on Metric Spaces
Continuous Functions on Metric Spaces Math 201A, Fall 2016 1 Continuous functions Definition 1. Let (X, d X ) and (Y, d Y ) be metric spaces. A function f : X Y is continuous at a X if for every ɛ > 0
More informationWeak-Star Convergence of Convex Sets
Journal of Convex Analysis Volume 13 (2006), No. 3+4, 711 719 Weak-Star Convergence of Convex Sets S. Fitzpatrick A. S. Lewis ORIE, Cornell University, Ithaca, NY 14853, USA aslewis@orie.cornell.edu www.orie.cornell.edu/
More informationConsistency of the maximum likelihood estimator for general hidden Markov models
Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models
More informationSimultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms
Simultaneous drift conditions for Adaptive Markov Chain Monte Carlo algorithms Yan Bai Feb 2009; Revised Nov 2009 Abstract In the paper, we mainly study ergodicity of adaptive MCMC algorithms. Assume that
More informationResearch Article Almost Periodic Solutions of Prey-Predator Discrete Models with Delay
Hindawi Publishing Corporation Advances in Difference Equations Volume 009, Article ID 976865, 19 pages doi:10.1155/009/976865 Research Article Almost Periodic Solutions of Prey-Predator Discrete Models
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationAn Empirical Algorithm for Relative Value Iteration for Average-cost MDPs
2015 IEEE 54th Annual Conference on Decision and Control CDC December 15-18, 2015. Osaka, Japan An Empirical Algorithm for Relative Value Iteration for Average-cost MDPs Abhishek Gupta Rahul Jain Peter
More informationEstimation of Dynamic Regression Models
University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.
More informationNUMERICAL RADIUS OF A HOLOMORPHIC MAPPING
Geometric Complex Analysis edited by Junjiro Noguchi et al. World Scientific, Singapore, 1995 pp.1 7 NUMERICAL RADIUS OF A HOLOMORPHIC MAPPING YUN SUNG CHOI Department of Mathematics Pohang University
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationStability of optimization problems with stochastic dominance constraints
Stability of optimization problems with stochastic dominance constraints D. Dentcheva and W. Römisch Stevens Institute of Technology, Hoboken Humboldt-University Berlin www.math.hu-berlin.de/~romisch SIAM
More informationStochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer
2009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 10-12, 2009 ThA06.6 Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer Serdar Yüksel
More informationTheory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk
Instructor: Victor F. Araman December 4, 2003 Theory and Applications of Stochastic Systems Lecture 0 B60.432.0 Exponential Martingale for Random Walk Let (S n : n 0) be a random walk with i.i.d. increments
More informationTECHNIQUES FOR ESTABLISHING DOMINATED SPLITTINGS
TECHNIQUES FOR ESTABLISHING DOMINATED SPLITTINGS ANDY HAMMERLINDL ABSTRACT. We give theorems which establish the existence of a dominated splitting and further properties, such as partial hyperbolicity.
More informationTheorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1
Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be
More informationRATIO ERGODIC THEOREMS: FROM HOPF TO BIRKHOFF AND KINGMAN
RTIO ERGODIC THEOREMS: FROM HOPF TO BIRKHOFF ND KINGMN HNS HENRIK RUGH ND DMIEN THOMINE bstract. Hopf s ratio ergodic theorem has an inherent symmetry which we exploit to provide a simplification of standard
More informationThe Art of Sequential Optimization via Simulations
The Art of Sequential Optimization via Simulations Stochastic Systems and Learning Laboratory EE, CS* & ISE* Departments Viterbi School of Engineering University of Southern California (Based on joint
More informationA generalization of Dobrushin coefficient
A generalization of Dobrushin coefficient Ü µ ŒÆ.êÆ ÆÆ 202.5 . Introduction and main results We generalize the well-known Dobrushin coefficient δ in total variation to weighted total variation δ V, which
More informationfor all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true
3 ohn Nirenberg inequality, Part I A function ϕ L () belongs to the space BMO() if sup ϕ(s) ϕ I I I < for all subintervals I If the same is true for the dyadic subintervals I D only, we will write ϕ BMO
More informationLocally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem
56 Chapter 7 Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem Recall that C(X) is not a normed linear space when X is not compact. On the other hand we could use semi
More informationOn the Central Limit Theorem for an ergodic Markov chain
Stochastic Processes and their Applications 47 ( 1993) 113-117 North-Holland 113 On the Central Limit Theorem for an ergodic Markov chain K.S. Chan Department of Statistics and Actuarial Science, The University
More informationHigh Dimensional Inverse Covariate Matrix Estimation via Linear Programming
High Dimensional Inverse Covariate Matrix Estimation via Linear Programming Ming Yuan October 24, 2011 Gaussian Graphical Model X = (X 1,..., X p ) indep. N(µ, Σ) Inverse covariance matrix Σ 1 = Ω = (ω
More informationA LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE. 1.
A LOWER BOUND ON BLOWUP RATES FOR THE 3D INCOMPRESSIBLE EULER EQUATION AND A SINGLE EXPONENTIAL BEALE-KATO-MAJDA ESTIMATE THOMAS CHEN AND NATAŠA PAVLOVIĆ Abstract. We prove a Beale-Kato-Majda criterion
More informationComputations of Critical Groups at a Degenerate Critical Point for Strongly Indefinite Functionals
Journal of Mathematical Analysis and Applications 256, 462 477 (2001) doi:10.1006/jmaa.2000.7292, available online at http://www.idealibrary.com on Computations of Critical Groups at a Degenerate Critical
More informationAN EFFECTIVE METRIC ON C(H, K) WITH NORMAL STRUCTURE. Mona Nabiei (Received 23 June, 2015)
NEW ZEALAND JOURNAL OF MATHEMATICS Volume 46 (2016), 53-64 AN EFFECTIVE METRIC ON C(H, K) WITH NORMAL STRUCTURE Mona Nabiei (Received 23 June, 2015) Abstract. This study first defines a new metric with
More informationOn the fast convergence of random perturbations of the gradient flow.
On the fast convergence of random perturbations of the gradient flow. Wenqing Hu. 1 (Joint work with Chris Junchi Li 2.) 1. Department of Mathematics and Statistics, Missouri S&T. 2. Department of Operations
More informationLecture Notes 10: Dynamic Programming
University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 81 Lecture Notes 10: Dynamic Programming Peter J. Hammond 2018 September 28th University of Warwick, EC9A0 Maths for Economists Peter
More informationWald for non-stopping times: The rewards of impatient prophets
Electron. Commun. Probab. 19 (2014), no. 78, 1 9. DOI: 10.1214/ECP.v19-3609 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS in PROBABILITY Wald for non-stopping times: The rewards of impatient prophets Alexander
More informationLecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University
Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University February 7, 2007 2 Contents 1 Metric Spaces 1 1.1 Basic definitions...........................
More informationAdditive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535
Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June
More information