Bounds on the Constant in the Mean Central Limit Theorem
|
|
- Ralph Watson
- 5 years ago
- Views:
Transcription
1 Bounds on the Constant in the Mean Central Limit Theorem Larry Goldstein University of Southern California, Ann. Prob. 2010
2 Classical Berry Esseen Theorem Let X, X 1, X 2,... be i.i.d. with distribution G having mean zero, variance σ 2 and finite third moment. Then there exists C such that F n Φ CE X 3 σ 3 n where F n is the distribution function of S n = 1 n σ X i, n i=1 where for distribution functions F and G F G = for n N sup F (x) G(x). <x<
3 Different Metrics L, type of worse case error: F G = L 1, type of average case error: F G 1 = sup F (x) G(x) <x< F (x) G(x) dx
4 L p Berry Esseen Theorem For p 1 there exists a constant C such that F n Φ p CE X 3 σ 3 n for n N. (1) Let F σ be the collection of all distributions with mean zero, positive variance σ 2, and finite third moment. The L p Berry-Esseen constant c p is given by c p = inf{c : nσ 3 F n Φ p E X 3 C, n N, G F σ }. Each C in (1) is an upper bound on c p.
5 Upper Bounds in the Classical Case Classical case p =, /7.59 (Berry/Esseen, 1941/1942) (P. van Beeck, 1972) (I. S. Shiganov in 1986) (I.G. Shevtsova in 2006)
6 Asymptotic Refinements Let c p,m = inf{c : nσ 3 F n Φ p E X 3 C, n m, G F σ } Clearly c p,m decreases in m, so we have existence of the limit lim m c p,m = c p,.
7 Asymptotically Correct Constant: p = 1 For G F σ Esseen explicitly calculates the limit lim n n1/2 F n Φ 1 = A 1 (G). Zolotarev (1964), using characteristic function techniques, shows that σ 3 A 1 (G) sup G F σ E X 3 = 1 2, so c 1, = 1 2.
8 Stein Functional A bound on the (non-asymptotic) L 1 constant can be obtained by considering the extremum of a Stein functional. Extrema of Stein functionals are considered by Utev and Lefévre, 2003, who computed some exact norms of Stein operators.
9 Bound using Zero Bias Let W be a mean zero random variable with finite positive variance σ 2. We say W has the W zero bias distribution if E[W f(w )] = σ 2 E[f (W )] for all smooth f. If the distribution F of W has variance 1 and W and W are on the same space with W having the W zero bias distribution, then F Φ 1 2E W W.
10 Exchange One Zero Bias Coupling If W = X X n, independent mean zero variables with variances σ 2 1,..., σ 2 n, and I is an independent index with distribution P (I = i) = σi 2 n, j=1 σ2 j then W = j I X j + X I has the W -zero bias distribution, when for each i, Xi has the X i -zero bias distribution independent of X j, j i.
11 Functional B(X) For X F σ let B(X) = 2σ2 L(X ) L(X) 1 E X 3. For R (more generally on any Polish space) valued random variables, given distributions F and G, one can construct X F and Y G such that E X Y = F G 1 In fact, let X = F 1 (U), Y = G 1 (U) for U U[0, 1].
12 Exchange One Zero Bias Coupling Let X 1,..., X n be independent random variables with distributions G i F σi, i = 1,..., n and let F n be the distribution function of W = (X X n )/σ with σ 2 = σ σ 2 n. Then with E X i X i = G i G i 1, E W W = 1 σ E X I X I = 1 σ n i=1 = 1 n σi 2E X i X i σ 3 E X i=1 i 3 E Xi 3 1 n = 2σ 3 B(X i )E Xi 3. i=1 σ 2 i σ 2 E X i X i
13 Exchange One Zero Bias Coupling If X 1,..., X n are independent mean zero random variables with distributions G 1,..., G n having finite variances σ 2 1,..., σ 2 n and finite third moments, then the distribution function F n of (X X n )/σ with σ 2 = σ σ 2 n obeys F n Φ 1 1 σ 3 n i=1 where the functional B(G) is given by B(G i )E X i 3 B(G) = 2σ2 G G 1 E X 3 when X has distribution G F σ.
14 Distribution Specific Constants In the i.i.d. case, and e.g., F n Φ 1 B(G)E X3, nσ 3 1. B(G) = 1 for mean zero two point distributions 2. B(G) = 1/3 for mean zero uniform distributions 3. B(G) = 0 for mean zero normal distributions
15 Universal Bound Recall that for G F σ B(G) = 2σ2 G G 1 E X 3. For a collection of distributions F σ>0 F σ, let B(F) = sup B(G). G F Then for X 1,..., X n i.i.d. with distribution in F σ, F n Φ 1 B(F σ)e X 3. nσ 3
16 Bounds on B(F σ ) Mean zero two point distributions give B(F σ ) 1 for all σ > 0. Using essentially only E X X E X + E X gives B(F σ ) 3 for all σ > 0. By coupling X and X together we improve the value of the constant from 3.
17 Value of Supremum Theorem 1 For all σ (0, ), B(F σ ) = 1. Hence, when X 1,..., X n are independent with distributions in F σi, i = 1,..., n and n i=1 σ2 i = σ2, F n Φ 1 1 σ 3 n i=1 E X i 3, and when these variables are identically distributed with variances σ 2, F n Φ 1 E X i 3 nσ 3.
18 Bounds on the Constant c 1 We can also prove the lower bound c 1 2 π(2φ(1) 1) ( π + 2) + 2e 1/2 2 π
19 Supremum of B(F σ ) Want to compute sup B(G) where B(G) = 2σ2 G G 1 G F σ E X 3. Successively reduce, in four steps, the computation of the supreumum of B(G) on F σ to computations over smaller collections of distributions.
20 First Reduction: σ = 1 Recall B(G) = 2σ2 G G 1 E X 3. By the scaling property it suffices to consider F 1. B(aX) = B(X) for all a 0
21 Second Reduction: compact support For X F 1, show that there exists X n, n = 1, 2,..., each in F 1 and having compact support, such that B(X n ) B(X). Hence it suffices to consider the class of distributions M F 1 with compact support.
22 Third Reduction: finite support For X M show that there exists X n, n = 1, 2,... in M, finitely supported, such that B(X n ) B(X). Hence it suffices to consider m 3 D m, where D m are mean zero variance one distributions supported on at most m points.
23 Fourth Reduction: three point support Use a convexity type property of B(G), which depends on the behavior of the zero bias transformation on a mixture, to obtain B(D 3 ) = B( m 3 D m ). Hence it suffices to consider D 3.
24 Lastly Show B(D 3 ) = 1.
25 Finding Extremes of Expectations Arguments along these lines were first considered by Hoeffding for the calculation of the extremes of EK(X 1,..., X n ) where X 1,..., X n are independent. Though B(G) is not of this form, the reasoning of Hoeffding applies. In some cases the final result obtained is not in closed form.
26 Reduction to Compact Support and Finite Support Continuity of the zero bias transformation: If then X n d X, and lim n EX2 n = EX 2 X n d X as n. Leads to continuity of B(G): If X n d X, then lim n EX2 n = EX 2 and lim n E X3 n = E X 3 B(X n ) B(X) as n.
27 From m 3 D m to D 3 If X µ be the µ mixture of a collection {X s, s S} of mean zero, variance 1 random variables satisfying E X 3 µ <. Then B(X µ ) sup B(X s ). s S In particular, if C is a collection of mean zero, variance 1 random variables with finite absolute third moments and C D such that every distribution in C can be represented as a mixture of distributions in D, then B(C) = B(D).
28 Zero Biasing a Mixture Theorem 2 Let {m s, s S} be a collection of mean zero distributions on R and µ a probability measure on S such that the variance σ 2 µ of the mixture distribution is positive and finite. Then m µ, the m µ zero bias distribution exists and is given by the mixture m µ = m sdν where dν dµ = σ2 s σµ 2. In particular, ν = µ if and only if σ 2 s is a constant µ a.s.
29 Mixture of Constant Variance: ν = µ L(Xµ) L(X µ ) 1 = sup Ef(Xµ) Ef(X µ ) f L = sup = f L sup f L sup f L S S S S Ef(Xs )dµ Ef(X s )dµ S Ef(X s ) Ef(X s ) dµ L(X s ) L(X s ) 1 dµ L(X s ) L(X s ) 1 dµ.
30 B(X µ ) sup s B(X s ) The relation dτ dµ = E X3 s E Xµ 3. (2) defines a probability measure, as E X 3 µ = E X 3 s ds.
31 B(X µ ) sup s B(X s ) Then B(X µ ) = 2 L(X µ) L(X µ ) 1 E Xµ 3 S 2 L(X s ) L(X s ) 1 dµ E Xµ 3 S = B(X s)e Xs 3 dµ E Xµ 3 = B(X s )dτ S sup s S B(X s )
32 Reduction of D m, m > 3 The distribution of any X D m is determined by the values a 1 < < a m and probabilities p = (p 1,..., p m ), all positive. The vector p must satisfy Ap = c where A = a 1 a 2... a m a 2 1 a a 2 m and c = For m > 3 there exists vector v 0 satisfying Av = 0. Hence X D m can be represented of a mixture of two distributions in D m 1..
33 Reduction to D 3 For every m > 3, every G D m can be represented as a finite mixture of distributions in D m 1. Hence B(D 3 ) = B( m 3 D m ). Every distribution D 3 with support points, say x < y < 0 < z, can be written as m α = αm 1 + (1 α)m 0, a mixture of the (unequal variance) mean zero distributions m 1 and m 0 supported on {x, z} and {y, z}, respectively.
34 Mixture with Unequal Variance For α [0, 1], with we have m α = αm 1 + (1 α)m 0 m α = βm 1 + (1 β)m 0 where β = Since x < y < 0, αx αx + (1 α)y. β 1 β = α x 1 α y > α 1 α and therefore β > α.
35 Calculating G(D 3 ) Write m D 3 as m α = αm 1 + (1 α)m 0 where m 1 and m 0 are mean zero two point distributions on {x, z} and {y, z}, respectively, x < y < 0 < z. Need to bound m α m α 1. (3) Any coupling of variables Y α and Y α with distributions m α and m α, respectively, gives an upper bound to (3). Let F 0, F 1, F 0, F 1 be the distribution functions of m 0, m 1, m 0 and m 1, respectively.
36 Bound by Coupling Set (Y 1, Y 0, Y 1, Y 0 ) equal to (F 1 1 (U), F 1 0 (U), (F 1 ) 1 (U), (F 0 ) 1 (U)) and let L(Y α, Y α ) be αl(y 1, Y 1 ) + (1 β)l(y 0, Y 0 ) + (β α)l(y 0, Y 1 ). Then (Y α, Y α ) has marginals Y α = d X α and Y α = d X α, and therefore m α m α 1 is upper bounded by α m 1 m (1 β) m 0 m (β α) m 1 m 0 1.
37 Bound on D 3 Goal is to have m α m α 1, or its upper bound α m 1 m (1 β) m 0 m (β α) m 1 m 0 1, bounded by E X 3 α /(2EX 2 α) = β m 1 m (1 β) m 0 m 0 1. Hence it suffices to show m 1 m 0 1 m 1 m 1 1. Reduces to computation of L 1 distances between uniform distribution on [x, z] and two point distributions on {y, z} and {x, z}.
38 m 1 m 0 1 m 1 m 1 1 m 0 on {y, z}, m 1 on {x, z} with x < y < 0 < z. Right hand side is m 1 m 1 1 = z2 + x 2 2(z x). Left hand side, under case where F 1 (y) F 0 (y), is [2(z x)(z y) 2 ] 1 ( z 4 2yz 3 + x 2 z 2 2x 2 yz +5y 2 z 2 + 3x 2 y 2 4xy 3 + 4xy 2 z 4xyz 2 + 2y 4 4y 3 z ).
39 Using Mathematica Taking the difference, after much cancelation m 1 m 1 1 m 1 m 0 1 is seen to equal 4y 2 z 2 2x 2 y 2 + 4xy 3 4xy 2 z + 4xyz 2 2y 4 + 4y 3 z 2(z x)(z y) 2, which factors as y(y x)(y 2 + 2z 2 y(x + 2z)) (z x)(z y) 2 and is positive, due to being in case F 1 (y) F 0 (y).
40 Bound over D 3 Since m 1 m 0 1 m 1 m 1 1 we have and therefore B(D 3 ) 1. m α m α 1 E X 3 α /(2EX 2 α),
41 Bound over D 3 Since m 1 m 0 1 m 1 m 1 1 we have m α m α 1 E Xα /(2EX 3 α), 2 and therefore B(D 3 ) 1. Hence 1 B(D 3 ) = B( D m ) = B(M) = B(F 1 ) 1. m 3
42 The Anti-Normal Distributions G F 1 is normal if and only if B(G) = 0; small B(G) close to normal. G, a mean zero two point distribution on x < 0 < y achieves sup G F1 B(G), the worst case for B(G), so anti-normal.
43 Lower Bound For L(X) = G F 1, F n Φ 1 c 1E X 3 n for all n N, and in particular for n = 1 c 1 F 1 Φ 1 E X 3 = G Φ 1 E X 3.
44 Lower Bound: For B B(p) for p (0, 1) let G p be the distribution function of X = (B p)/ pq. Then G p Φ 1 equals p q and letting q p Φ(x)dx + Φ(x) q dx + Φ(x) 1 dx, p q q p ψ(p) = pq p 2 + q 2 G p Φ 1 for p (0, 1) ψ(1/2) = 2 π(2φ(1) 1) ( π + 2) + 2e 1/2 2 π.
45 Upper Bounds in the Classical Case Classical case p =, /7.59 (Berry/Esseen, 1941/1942) (P. van Beeck, 1972) (I. S. Shiganov in 1986) (I.G. Shevtsova in 2006) (I. Tyurin in 2010)
46 Higher Order Hermite Functionals Letting H k (x) be the k th Hermite Polynomial, if the moments of X match those of the standard normal up to order 2k, then there exists X (k) such that EH k (X)f(X) = Ef (k) (X (k) ). Can one compute extreme values of the natural generalizations of B(G) such as B k (G) = σ2k X (k) X 1 E X 2k+1 which might be the values of like constants when higher moments match the normal.
arxiv: v2 [math.pr] 19 Oct 2010
The Annals of Probability 2010, Vol. 38, No. 4, 1672 1689 DOI: 10.1214/10-AOP527 c Institute of Mathematical tatistics, 2010 arxiv:0906.5145v2 [math.pr] 19 Oct 2010 BOUND ON THE CONTANT IN THE MEAN CENTRAL
More informationBOUNDS ON THE CONSTANT IN THE MEAN CENTRAL LIMIT THEOREM. BY LARRY GOLDSTEIN University of Southern California
The Annals of Probability 2010, Vol. 38, No. 4, 1672 1689 DOI: 10.1214/10-AOP527 Institute of Mathematical Statistics, 2010 BOUNDS ON THE CONSTANT IN THE MEAN CENTRAL LIMIT THEOREM BY LARRY GOLDSTEIN University
More informationA Gentle Introduction to Stein s Method for Normal Approximation I
A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein
More informationKolmogorov Berry-Esseen bounds for binomial functionals
Kolmogorov Berry-Esseen bounds for binomial functionals Raphaël Lachièze-Rey, Univ. South California, Univ. Paris 5 René Descartes, Joint work with Giovanni Peccati, University of Luxembourg, Singapore
More informationStein s Method: Distributional Approximation and Concentration of Measure
Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Stein s method for Distributional
More informationA square bias transformation: properties and applications
A square bias transformation: properties and applications Irina Shevtsova arxiv:11.6775v4 [math.pr] 13 Dec 13 Abstract The properties of the square bias transformation are studied, in particular, the precise
More informationStein s Method: Distributional Approximation and Concentration of Measure
Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Concentration of Measure Distributional
More informationConcentration of Measures by Bounded Size Bias Couplings
Concentration of Measures by Bounded Size Bias Couplings Subhankar Ghosh, Larry Goldstein University of Southern California [arxiv:0906.3886] January 10 th, 2013 Concentration of Measure Distributional
More informationConcentration of Measures by Bounded Couplings
Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional
More informationStein s Method and the Zero Bias Transformation with Application to Simple Random Sampling
Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling Larry Goldstein and Gesine Reinert November 8, 001 Abstract Let W be a random variable with mean zero and variance
More informationA Short Introduction to Stein s Method
A Short Introduction to Stein s Method Gesine Reinert Department of Statistics University of Oxford 1 Overview Lecture 1: focusses mainly on normal approximation Lecture 2: other approximations 2 1. The
More informationGaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula
Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula Larry Goldstein, University of Southern California Nourdin GIoVAnNi Peccati Luxembourg University University British
More informationStatistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it
Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students
More informationLectures 22-23: Conditional Expectations
Lectures 22-23: Conditional Expectations 1.) Definitions Let X be an integrable random variable defined on a probability space (Ω, F 0, P ) and let F be a sub-σ-algebra of F 0. Then the conditional expectation
More informationFerromagnets and the classical Heisenberg model. Kay Kirkpatrick, UIUC
Ferromagnets and the classical Heisenberg model Kay Kirkpatrick, UIUC Ferromagnets and the classical Heisenberg model: asymptotics for a mean-field phase transition Kay Kirkpatrick, Urbana-Champaign June
More informationLecture 1 Measure concentration
CSE 29: Learning Theory Fall 2006 Lecture Measure concentration Lecturer: Sanjoy Dasgupta Scribe: Nakul Verma, Aaron Arvey, and Paul Ruvolo. Concentration of measure: examples We start with some examples
More informationNormal Approximation for Hierarchical Structures
Normal Approximation for Hierarchical Structures Larry Goldstein University of Southern California July 1, 2004 Abstract Given F : [a, b] k [a, b] and a non-constant X 0 with P (X 0 [a, b]) = 1, define
More informationl(y j ) = 0 for all y j (1)
Problem 1. The closed linear span of a subset {y j } of a normed vector space is defined as the intersection of all closed subspaces containing all y j and thus the smallest such subspace. 1 Show that
More informationThe Contraction Method on C([0, 1]) and Donsker s Theorem
The Contraction Method on C([0, 1]) and Donsker s Theorem Henning Sulzbach J. W. Goethe-Universität Frankfurt a. M. YEP VII Probability, random trees and algorithms Eindhoven, March 12, 2010 joint work
More informationNEW FUNCTIONAL INEQUALITIES
1 / 29 NEW FUNCTIONAL INEQUALITIES VIA STEIN S METHOD Giovanni Peccati (Luxembourg University) IMA, Minneapolis: April 28, 2015 2 / 29 INTRODUCTION Based on two joint works: (1) Nourdin, Peccati and Swan
More informationMath 205b Homework 2 Solutions
Math 5b Homework Solutions January 5, 5 Problem (R-S, II.) () For the R case, we just expand the right hand side and use the symmetry of the inner product: ( x y x y ) = = ((x, x) (y, y) (x, y) (y, x)
More information9. Banach algebras and C -algebras
matkt@imf.au.dk Institut for Matematiske Fag Det Naturvidenskabelige Fakultet Aarhus Universitet September 2005 We read in W. Rudin: Functional Analysis Based on parts of Chapter 10 and parts of Chapter
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationFinal Exam Practice Problems Math 428, Spring 2017
Final xam Practice Problems Math 428, Spring 2017 Name: Directions: Throughout, (X,M,µ) is a measure space, unless stated otherwise. Since this is not to be turned in, I highly recommend that you work
More informationTHE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION
THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence
More informationL p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by
L p Functions Given a measure space (, µ) and a real number p [, ), recall that the L p -norm of a measurable function f : R is defined by f p = ( ) /p f p dµ Note that the L p -norm of a function f may
More informationOn the absolute constants in the Berry Esseen type inequalities for identically distributed summands
On the absolute constants in the Berry Esseen type inequalities for identically distributed summands Irina Shevtsova arxiv:1111.6554v1 [math.pr] 28 Nov 2011 Abstract By a modification of the method that
More informationGeneralization theory
Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1
More informationFunctional Analysis. Martin Brokate. 1 Normed Spaces 2. 2 Hilbert Spaces The Principle of Uniform Boundedness 32
Functional Analysis Martin Brokate Contents 1 Normed Spaces 2 2 Hilbert Spaces 2 3 The Principle of Uniform Boundedness 32 4 Extension, Reflexivity, Separation 37 5 Compact subsets of C and L p 46 6 Weak
More informationStein Couplings for Concentration of Measure
Stein Couplings for Concentration of Measure Jay Bartroff, Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] [arxiv:1402.6769] Borchard
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationMultivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More informationConvergence rates in weighted L 1 spaces of kernel density estimators for linear processes
Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,
More informationKernel Density Estimation
EECS 598: Statistical Learning Theory, Winter 2014 Topic 19 Kernel Density Estimation Lecturer: Clayton Scott Scribe: Yun Wei, Yanzhen Deng Disclaimer: These notes have not been subjected to the usual
More informationMODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT
Statistica Sinica 23 (2013), 1523-1540 doi:http://dx.doi.org/10.5705/ss.2012.203s MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT Louis H. Y. Chen 1, Xiao Fang 1,2 and Qi-Man Shao 3 1 National
More informationreview To find the coefficient of all the terms in 15ab + 60bc 17ca: Coefficient of ab = 15 Coefficient of bc = 60 Coefficient of ca = -17
1. Revision Recall basic terms of algebraic expressions like Variable, Constant, Term, Coefficient, Polynomial etc. The coefficients of the terms in 4x 2 5xy + 6y 2 are Coefficient of 4x 2 is 4 Coefficient
More informationLecture I: Asymptotics for large GUE random matrices
Lecture I: Asymptotics for large GUE random matrices Steen Thorbjørnsen, University of Aarhus andom Matrices Definition. Let (Ω, F, P) be a probability space and let n be a positive integer. Then a random
More informationStat 5101 Notes: Algorithms (thru 2nd midterm)
Stat 5101 Notes: Algorithms (thru 2nd midterm) Charles J. Geyer October 18, 2012 Contents 1 Calculating an Expectation or a Probability 2 1.1 From a PMF........................... 2 1.2 From a PDF...........................
More informationConcentration inequalities and the entropy method
Concentration inequalities and the entropy method Gábor Lugosi ICREA and Pompeu Fabra University Barcelona what is concentration? We are interested in bounding random fluctuations of functions of many
More informationStat 5101 Notes: Algorithms
Stat 5101 Notes: Algorithms Charles J. Geyer January 22, 2016 Contents 1 Calculating an Expectation or a Probability 3 1.1 From a PMF........................... 3 1.2 From a PDF...........................
More informationMath 5588 Final Exam Solutions
Math 5588 Final Exam Solutions Prof. Jeff Calder May 9, 2017 1. Find the function u : [0, 1] R that minimizes I(u) = subject to u(0) = 0 and u(1) = 1. 1 0 e u(x) u (x) + u (x) 2 dx, Solution. Since the
More informationFunctional Analysis Exercise Class
Functional Analysis Exercise Class Week 9 November 13 November Deadline to hand in the homeworks: your exercise class on week 16 November 20 November Exercises (1) Show that if T B(X, Y ) and S B(Y, Z)
More informationMATH 425, FINAL EXAM SOLUTIONS
MATH 425, FINAL EXAM SOLUTIONS Each exercise is worth 50 points. Exercise. a The operator L is defined on smooth functions of (x, y by: Is the operator L linear? Prove your answer. L (u := arctan(xy u
More informationCentral limit theorem for random multiplicative functions
Central limit theorem for random multiplicative functions (Courant Institute) joint work with Kannan Soundararajan (Stanford) Multiplicative functions Many of the functions of interest to number theorists
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationIntroduction to Self-normalized Limit Theory
Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized
More informationWasserstein-2 bounds in normal approximation under local dependence
Wasserstein- bounds in normal approximation under local dependence arxiv:1807.05741v1 [math.pr] 16 Jul 018 Xiao Fang The Chinese University of Hong Kong Abstract: We obtain a general bound for the Wasserstein-
More informationMATH 19520/51 Class 5
MATH 19520/51 Class 5 Minh-Tam Trinh University of Chicago 2017-10-04 1 Definition of partial derivatives. 2 Geometry of partial derivatives. 3 Higher derivatives. 4 Definition of a partial differential
More informationUnit 3 Factors & Products
1 Unit 3 Factors & Products General Outcome: Develop algebraic reasoning and number sense. Specific Outcomes: 3.1 Demonstrate an understanding of factors of whole number by determining the: o prime factors
More informationStatistics 3657 : Moment Approximations
Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the
More information4 Invariant Statistical Decision Problems
4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More informationHomework 2: Solution
0-704: Information Processing and Learning Sring 0 Lecturer: Aarti Singh Homework : Solution Acknowledgement: The TA graciously thanks Rafael Stern for roviding most of these solutions.. Problem Hence,
More informationWEYL S LEMMA, ONE OF MANY. Daniel W. Stroock
WEYL S LEMMA, ONE OF MANY Daniel W Stroock Abstract This note is a brief, and somewhat biased, account of the evolution of what people working in PDE s call Weyl s Lemma about the regularity of solutions
More information4 Expectation & the Lebesgue Theorems
STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does
More informationPartial Derivatives. w = f(x, y, z).
Partial Derivatives 1 Functions of Several Variables So far we have focused our attention of functions of one variable. These functions model situations in which a variable depends on another independent
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationMATH 426, TOPOLOGY. p 1.
MATH 426, TOPOLOGY THE p-norms In this document we assume an extended real line, where is an element greater than all real numbers; the interval notation [1, ] will be used to mean [1, ) { }. 1. THE p
More informationSelf-normalized Cramér-Type Large Deviations for Independent Random Variables
Self-normalized Cramér-Type Large Deviations for Independent Random Variables Qi-Man Shao National University of Singapore and University of Oregon qmshao@darkwing.uoregon.edu 1. Introduction Let X, X
More informationGrade 8 Factorisation
ID : ae-8-factorisation [1] Grade 8 Factorisation For more such worksheets visit www.edugain.com Answer the questions (1) Find factors of following polynomial A) y 2-2xy + 3y - 6x B) 3y 2-12xy - 2y + 8x
More informationOn the Error Bound in the Normal Approximation for Jack Measures (Joint work with Le Van Thanh)
On the Error Bound in the Normal Approximation for Jack Measures (Joint work with Le Van Thanh) Louis H. Y. Chen National University of Singapore International Colloquium on Stein s Method, Concentration
More informationLecture No 1 Introduction to Diffusion equations The heat equat
Lecture No 1 Introduction to Diffusion equations The heat equation Columbia University IAS summer program June, 2009 Outline of the lectures We will discuss some basic models of diffusion equations and
More informationRandom Fermionic Systems
Random Fermionic Systems Fabio Cunden Anna Maltsev Francesco Mezzadri University of Bristol December 9, 2016 Maltsev (University of Bristol) Random Fermionic Systems December 9, 2016 1 / 27 Background
More informationStein s method, logarithmic Sobolev and transport inequalities
Stein s method, logarithmic Sobolev and transport inequalities M. Ledoux University of Toulouse, France and Institut Universitaire de France Stein s method, logarithmic Sobolev and transport inequalities
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 13: Entropy Calculations
Introduction to Empirical Processes and Semiparametric Inference Lecture 13: Entropy Calculations Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More information7 Influence Functions
7 Influence Functions The influence function is used to approximate the standard error of a plug-in estimator. The formal definition is as follows. 7.1 Definition. The Gâteaux derivative of T at F in the
More informationNonparametric estimation under Shape Restrictions
Nonparametric estimation under Shape Restrictions Jon A. Wellner University of Washington, Seattle Statistical Seminar, Frejus, France August 30 - September 3, 2010 Outline: Five Lectures on Shape Restrictions
More informationMath 240 Calculus III
DE Higher Order Calculus III Summer 2015, Session II Tuesday, July 28, 2015 Agenda DE 1. of order n An example 2. constant-coefficient linear Introduction DE We now turn our attention to solving linear
More informationMidterm 1 Solutions Thursday, February 26
Math 59 Dr. DeTurck Midterm 1 Solutions Thursday, February 26 1. First note that since f() = f( + ) = f()f(), we have either f() = (in which case f(x) = f(x + ) = f(x)f() = for all x, so c = ) or else
More informationFunctional Analysis Winter 2018/2019
Functional Analysis Winter 2018/2019 Peer Christian Kunstmann Karlsruher Institut für Technologie (KIT) Institut für Analysis Englerstr. 2, 76131 Karlsruhe e-mail: peer.kunstmann@kit.edu These lecture
More informationMath 5520 Homework 2 Solutions
Math 552 Homework 2 Solutions March, 26. Consider the function fx) = 2x ) 3 if x, 3x ) 2 if < x 2. Determine for which k there holds f H k, 2). Find D α f for α k. Solution. We show that k = 2. The formulas
More informationTWO MAPPINGS RELATED TO SEMI-INNER PRODUCTS AND THEIR APPLICATIONS IN GEOMETRY OF NORMED LINEAR SPACES. S.S. Dragomir and J.J.
RGMIA Research Report Collection, Vol. 2, No. 1, 1999 http://sci.vu.edu.au/ rgmia TWO MAPPINGS RELATED TO SEMI-INNER PRODUCTS AND THEIR APPLICATIONS IN GEOMETRY OF NORMED LINEAR SPACES S.S. Dragomir and
More informationLecture 4 Lebesgue spaces and inequalities
Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how
More informationAlgebraic Expressions
Algebraic Expressions 1. Expressions are formed from variables and constants. 2. Terms are added to form expressions. Terms themselves are formed as product of factors. 3. Expressions that contain exactly
More informationOn the Bennett-Hoeffding inequality
On the Bennett-Hoeffding inequality of Iosif 1,2,3 1 Department of Mathematical Sciences Michigan Technological University 2 Supported by NSF grant DMS-0805946 3 Paper available at http://arxiv.org/abs/0902.4058
More informationRonalda Benjamin. Definition A normed space X is complete if every Cauchy sequence in X converges in X.
Group inverses in a Banach algebra Ronalda Benjamin Talk given in mathematics postgraduate seminar at Stellenbosch University on 27th February 2012 Abstract Let A be a Banach algebra. An element a A is
More informationExample continued. Math 425 Intro to Probability Lecture 37. Example continued. Example
continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with
More informationOverview of normed linear spaces
20 Chapter 2 Overview of normed linear spaces Starting from this chapter, we begin examining linear spaces with at least one extra structure (topology or geometry). We assume linearity; this is a natural
More informationComparison Method in Random Matrix Theory
Comparison Method in Random Matrix Theory Jun Yin UW-Madison Valparaíso, Chile, July - 2015 Joint work with A. Knowles. 1 Some random matrices Wigner Matrix: H is N N square matrix, H : H ij = H ji, EH
More informationMetric Spaces. Exercises Fall 2017 Lecturer: Viveka Erlandsson. Written by M.van den Berg
Metric Spaces Exercises Fall 2017 Lecturer: Viveka Erlandsson Written by M.van den Berg School of Mathematics University of Bristol BS8 1TW Bristol, UK 1 Exercises. 1. Let X be a non-empty set, and suppose
More informationIf Y and Y 0 satisfy (1-2), then Y = Y 0 a.s.
20 6. CONDITIONAL EXPECTATION Having discussed at length the limit theory for sums of independent random variables we will now move on to deal with dependent random variables. An important tool in this
More informationarxiv: v2 [math.st] 20 Feb 2013
Exact bounds on the closeness between the Student and standard normal distributions arxiv:1101.3328v2 [math.st] 20 Feb 2013 Contents Iosif Pinelis Department of Mathematical Sciences Michigan Technological
More informationCHAPTER 7 DIV, GRAD, AND CURL
CHAPTER 7 DIV, GRAD, AND CURL 1 The operator and the gradient: Recall that the gradient of a differentiable scalar field ϕ on an open set D in R n is given by the formula: (1 ϕ = ( ϕ, ϕ,, ϕ x 1 x 2 x n
More informationOn Ergodic Impulse Control with Constraint
On Ergodic Impulse Control with Constraint Maurice Robin Based on joint papers with J.L. Menaldi University Paris-Sanclay 9119 Saint-Aubin, France (e-mail: maurice.robin@polytechnique.edu) IMA, Minneapolis,
More information19. From a resolution of the identity to operators - and bac
19. From a resolution of the identity to operators - and back matkt@imf.au.dk Institut for Matematiske Fag Det Naturvidenskabelige Fakultet Aarhus Universitet December 2005 We read Sections 12.21 in Rudins
More informationSelçuk Demir WS 2017 Functional Analysis Homework Sheet
Selçuk Demir WS 2017 Functional Analysis Homework Sheet 1. Let M be a metric space. If A M is non-empty, we say that A is bounded iff diam(a) = sup{d(x, y) : x.y A} exists. Show that A is bounded iff there
More informationThe sum x 1 + x 2 + x 3 is (A): 4 (B): 6 (C): 8 (D): 14 (E): None of the above. How many pairs of positive integers (x, y) are there, those satisfy
Important: Answer to all 15 questions. Write your answers on the answer sheets provided. For the multiple choice questions, stick only the letters (A, B, C, D or E) of your choice. No calculator is allowed.
More informationClass 2 & 3 Overfitting & Regularization
Class 2 & 3 Overfitting & Regularization Carlo Ciliberto Department of Computer Science, UCL October 18, 2017 Last Class The goal of Statistical Learning Theory is to find a good estimator f n : X Y, approximating
More information1 Solutions to selected problems
1 Solutions to selected problems 1. Let A B R n. Show that int A int B but in general bd A bd B. Solution. Let x int A. Then there is ɛ > 0 such that B ɛ (x) A B. This shows x int B. If A = [0, 1] and
More informationRANDOM FIELDS AND GEOMETRY. Robert Adler and Jonathan Taylor
RANDOM FIELDS AND GEOMETRY from the book of the same name by Robert Adler and Jonathan Taylor IE&M, Technion, Israel, Statistics, Stanford, US. ie.technion.ac.il/adler.phtml www-stat.stanford.edu/ jtaylor
More informationOn large deviations for combinatorial sums
arxiv:1901.0444v1 [math.pr] 14 Jan 019 On large deviations for combinatorial sums Andrei N. Frolov Dept. of Mathematics and Mechanics St. Petersburg State University St. Petersburg, Russia E-mail address:
More informationStochastic relations of random variables and processes
Stochastic relations of random variables and processes Lasse Leskelä Helsinki University of Technology 7th World Congress in Probability and Statistics Singapore, 18 July 2008 Fundamental problem of applied
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationLocally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem
56 Chapter 7 Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem Recall that C(X) is not a normed linear space when X is not compact. On the other hand we could use semi
More informationPractice problems for Exam 1. a b = (2) 2 + (4) 2 + ( 3) 2 = 29
Practice problems for Exam.. Given a = and b =. Find the area of the parallelogram with adjacent sides a and b. A = a b a ı j k b = = ı j + k = ı + 4 j 3 k Thus, A = 9. a b = () + (4) + ( 3)
More information1 Hanoi Open Mathematical Competition 2017
1 Hanoi Open Mathematical Competition 017 1.1 Junior Section Question 1. Suppose x 1, x, x 3 are the roots of polynomial P (x) = x 3 6x + 5x + 1. The sum x 1 + x + x 3 is (A): 4 (B): 6 (C): 8 (D): 14 (E):
More informationCalculus 2502A - Advanced Calculus I Fall : Local minima and maxima
Calculus 50A - Advanced Calculus I Fall 014 14.7: Local minima and maxima Martin Frankland November 17, 014 In these notes, we discuss the problem of finding the local minima and maxima of a function.
More informationHamburger Beiträge zur Angewandten Mathematik
Hamburger Beiträge zur Angewandten Mathematik Numerical analysis of a control and state constrained elliptic control problem with piecewise constant control approximations Klaus Deckelnick and Michael
More information