SDS : Theoretical Statistics
|
|
- Nathaniel Hodge
- 5 years ago
- Views:
Transcription
1 SDS : Theoretical Statistics Lecture 1: Introduction Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
2 Manegerial Stuff Instructor- Purnamrita Sarkar Course material and homeworks will be posted under Office hours: Tuesdays 2-3 pm. GDC TA: Jack Liu Homeworks are due Biweekly after class on Wednesdays Grading - 5 homeworks (70% ), class participation (5%) Final Exam (25% ) Books Asymptotic Statistics, Aad van der Vaart. Cambridge Martin Wainwright s High dimensional statistics: A non-asymptotic view point 1
3 Why do theory? Say you have estimated ˆθn from data X 1,..., Xn. How do we know we have a good estimation method? Does ˆθ n θ? This brings us to Stochastic Convergence. How do I know if one estimation method is better than another? Does the estimate from one converge faster than the other? Does one algorithm work under broader parameter regimes, or weaker assumptions? What is the optimal rate for a given estimation problem? 2
4 This class Your instructor hopes to cover : Consistency of parameter estimates Stochastic Convergence Concentration inequalities Asymptotic normality of estimators Empirical processes, VC classes, covering numbers Asymptotic testing Examples of network clustering with a bit of random matrix theory Bootstrap, Nonparametric regression and density estimation 3
5 Stochastic Convergence Assume that X n, n 1 and X are elements of a separable metric space (S, d). Definition (Weak Convergence) A sequence of random variable s converge in law or in distribution to a random variable X, i.e. Xn d X if P(Xn x) P(X x) x at which P(X x) is continuous. Definition ( Convergence in Probability) A sequence of random variables converge in probability to a random variable X, i.e. Xn P X if ɛ > 0, P(d(Xn, X ) ɛ) 0. 4
6 Stochastic Convergence Assume that X n, n 1 and X are elements of a separable metric space (S, d). Definition (Almost Sure Convergence) A sequence of random variables converges almost surely to a random variable X, i.e. Xn a.s. ( ) X if P lim n d(xn, X ) = 0 = 1. If you think about a (scalar) random variable as a function that maps events to a real number, almost sure convergence means P(ω Ω : lim Xn(ω) = X (ω)) = 1 n Definition (Convergence in quadratic mean) A sequence of random variables converges in quadratic mean to a random variable X, i.e. Xn q.m [ X if E d(xn, X ) 2] 0. 5
7 Stochastic Convergence Theorem Xn a.s. X, Xn q.m. X Xn P X Xn d X Xn d c Xn P c 6
8 Converses:X n d X Xn P X Convergence in law needs no knowledge of the joint distribution of Xn and the limiting random variable X. Convergence in probability does. Example Consider X N(0, 1), X n = X. Xn d X. But how about Xn P X? 7
9 Converses:X n d X Xn P X Convergence in law needs no knowledge of the joint distribution of Xn and the limiting random variable X. Convergence in probability does. Example Consider X N(0, 1), X n = X. Xn d X. But how about Xn P X? P( Xn X ɛ) = P(2 X ɛ) 0 ɛ > 0. So Xn does not converge in probability to X. 7
10 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). 8
11 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. 8
12 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. Xn are a sequence of bernoulli s with probabilities pn = 1/2 k where k = log n. 8
13 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. Xn are a sequence of bernoulli s with probabilities pn = 1/2 k where k = log n. So X n P 0 and Xn qm 0 8
14 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? 9
15 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n 9
16 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n E Xn 2 = 2 2n /n. So Xn qm 0 9
17 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n E Xn 2 = 2 2n /n. So Xn qm 0 P( X n ɛ) = P(Xn = 2 n ) = P(Z [0, 1/n)) = 1/n 0 9
18 Continuous Mapping Theorem Theorem Let g be continuous on a set C where P(X C) = 1. Then, Xn d X g(xn) d g(x ) Xn P X g(xn) P g(x ) Xn a.s. X g(xn) a.s. g(x ) 10
19 Example Let X n d X where X N(0, 1). Then X 2 n d? 11
20 Example Let X n d X where X N(0, 1). Then X 2 n d? Use g(x) = x 2. 11
21 Example Let X n d X where X N(0, 1). Then X 2 n d? Use g(x) = x 2. Use X 2 χ
22 Example d Let X n X where X N(0, 1). Then X 2 d n? Use g(x) = x 2. Use X 2 χ 2 1. So Xn 2 d χ
23 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? 12
24 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 12
25 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 Can we use Continuous Mapping Theorem to claim that g( Xn µ) 2 d 0? 12
26 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 Can we use Continuous Mapping Theorem to claim that g( Xn µ) 2 d 0? NO. Because, 0 is a random variable whose mass is at 0, where g is discontinuous. 12
27 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? 13
28 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. 13
29 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. E[ f (X n) f (X ) 2 ] L 2 E[ Xn X 2 ] 0. So for Lipschitz functions quadratic mean convergence goes through. 13
30 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. E[ f (X n) f (X ) 2 ] L 2 E[ Xn X 2 ] 0. So for Lipschitz functions quadratic mean convergence goes through. Can you come up with a non-lipschitz function and a sequence {X n} where f (Xn) qm 0? 13
31 Portmanteau Theorem Theorem The following are equivalent. Xn d X E[f (Xn)] E[f (X )] for all bounded and continuous f. E[f (Xn)] E[f (X )] for all bounded and Lipschitz f. E[e itt Xn ] E[e it T Xn ], t R k. (Levy s continuity theorem) t T X n d t T X t R k. (Cramer-Wold device) lim inf E[f (Xn)] E[f (X )] for all non-negative continuous f n lim sup P(Xn F ) P(X F ) for all closed F n lim inf P(Xn F ) P(X F ) for all open F n P(X n B) P(X B) for all continuity sets B (P(X B) = 0) 14
32 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n 15
33 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? 15
34 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? E[Xn] = 1. What went wrong? 15
35 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? E[Xn] = 1. What went wrong? f (x) = x is not bounded. 15
36 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) 16
37 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. 16
38 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. Example: Xn N(0, 1), Yn = Xn. X Y and X, Y are independent standard normal random variables. 16
39 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. Example: Xn N(0, 1), Yn = Xn. X Y and X, Y are independent standard normal random variables. Then X n d X and Yn d Y. But (Xn, Yn) d (X, X ), not (Xn, Yn) d (X, Y ). 16
40 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c 17
41 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c Does X n d X and Yn d Y imply Xn + Yn d X + Y? 17
42 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c Does X n d X and Yn d Y imply Xn + Yn d X + Y? Take Yn = Xn, and X, Y as independent standard normal random variables. Xn d X and Yn d Y but Xn + Yn d 0. 17
43 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn 18
44 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 i X 2 i X 2 n = n n 1 i X 2 i n X 2 n 18
45 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n 18
46 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. 18
47 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. Finally, n( Xn µ) d N(0, σ 2 ) using CLT. 18
48 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. Finally, n( Xn µ) d N(0, σ 2 ) using CLT. Now using Slutsky s lemma, n( Xn µ)/sn d N(0, 1) using CLT. 18
49 Uniformly tight Definition X is defined to be tight if ɛ > 0 M for which, P( X > M) < ɛ {Xn} is defined to uniformly tight if ɛ > 0 M for which, sup n P( Xn > M) < ɛ 19
50 Prohorov s theorem Theorem Xn d X {Xn} is UT. {Xn} is UT implies that, there exists a subsequence {n j } such that d Xn j X. 20
51 Notation for rates, small oh-pee and big oh-pee Definition The small o P : X n = op (1) Xn P 0 Xn = o P (Rn) Xn = YnRn and Yn = o P (1) Xn is vanishing in probability The big O P : Xn = O P (1) {Xn} is UT Xn = O P (Rn) Xn = YnRn and Yn = O P (1) Xn lies within a ball of finite radius with high probability 21
52 How do they interact o P (an) + o P (bn) = o P (max(an, bn)). o P (an) + O P (bn) = O P (bn). O P (a n)op (b n) = op (a nbn). 1 + O P (1) = O P (1). (1 + o P (1)) 1 = 1 + o P (1). o P (O P (1)) = o P (1). Xn = o P (an), Xn r = o P (an) r Be careful: e o P (1) o P (1) O P (1) + O P (1) Can actually be o P (1) because of cancellation. 22
53 23
Theoretical Statistics. Lecture 1.
1. Organizational issues. 2. Overview. 3. Stochastic convergence. Theoretical Statistics. Lecture 1. eter Bartlett 1 Organizational Issues Lectures: Tue/Thu 11am 12:30pm, 332 Evans. eter Bartlett. bartlett@stat.
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence
Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and
More informationChapter 5. Weak convergence
Chapter 5 Weak convergence We will see later that if the X i are i.i.d. with mean zero and variance one, then S n / p n converges in the sense P(S n / p n 2 [a, b])! P(Z 2 [a, b]), where Z is a standard
More informationStochastic Convergence, Delta Method & Moment Estimators
Stochastic Convergence, Delta Method & Moment Estimators Seminar on Asymptotic Statistics Daniel Hoffmann University of Kaiserslautern Department of Mathematics February 13, 2015 Daniel Hoffmann (TU KL)
More informationEconomics 620, Lecture 8: Asymptotics I
Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators
More informationLecture 2: Convergence of Random Variables
Lecture 2: Convergence of Random Variables Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Introduction to Stochastic Processes, Fall 2013 1 / 9 Convergence of Random Variables
More informationConvergence in Distribution
Convergence in Distribution Undergraduate version of central limit theorem: if X 1,..., X n are iid from a population with mean µ and standard deviation σ then n 1/2 ( X µ)/σ has approximately a normal
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationConvergence Concepts of Random Variables and Functions
Convergence Concepts of Random Variables and Functions c 2002 2007, Professor Seppo Pynnonen, Department of Mathematics and Statistics, University of Vaasa Version: January 5, 2007 Convergence Modes Convergence
More informationEconomics 241B Review of Limit Theorems for Sequences of Random Variables
Economics 241B Review of Limit Theorems for Sequences of Random Variables Convergence in Distribution The previous de nitions of convergence focus on the outcome sequences of a random variable. Convergence
More informationWeak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria
Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationLecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationRegression #4: Properties of OLS Estimator (Part 2)
Regression #4: Properties of OLS Estimator (Part 2) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #4 1 / 24 Introduction In this lecture, we continue investigating properties associated
More informationLecture 1: Review of Basic Asymptotic Theory
Lecture 1: Instructor: Department of Economics Stanfor University Prepare by Wenbo Zhou, Renmin University Basic Probability Theory Takeshi Amemiya, Avance Econometrics, 1985, Harvar University Press.
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationElements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley
Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationThe Moment Method; Convex Duality; and Large/Medium/Small Deviations
Stat 928: Statistical Learning Theory Lecture: 5 The Moment Method; Convex Duality; and Large/Medium/Small Deviations Instructor: Sham Kakade The Exponential Inequality and Convex Duality The exponential
More informationUniversity of California San Diego and Stanford University and
First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 13: Expectation and Variance and joint distributions Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More informationLecture Notes on Asymptotic Statistics. Changliang Zou
Lecture Notes on Asymptotic Statistics Changliang Zou Prologue Why asymptotic statistics? The use of asymptotic approximation is two-fold. First, they enable us to find approximate tests and confidence
More informationLecture 5: Asymptotic Equipartition Property
Lecture 5: Asymptotic Equipartition Property Law of large number for product of random variables AEP and consequences Dr. Yao Xie, ECE587, Information Theory, Duke University Stock market Initial investment
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationLarge Deviations for Small-Noise Stochastic Differential Equations
Chapter 22 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large
More informationLarge Deviations for Small-Noise Stochastic Differential Equations
Chapter 21 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large
More informationLarge Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationIntroduction to Self-normalized Limit Theory
Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized
More informationGraduate Econometrics I: Asymptotic Theory
Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory
More informationPart II Probability and Measure
Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationSEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE
SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE FRANÇOIS BOLLEY Abstract. In this note we prove in an elementary way that the Wasserstein distances, which play a basic role in optimal transportation
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results
Introduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics
More informationStat 8112 Lecture Notes Weak Convergence in Metric Spaces Charles J. Geyer January 23, Metric Spaces
Stat 8112 Lecture Notes Weak Convergence in Metric Spaces Charles J. Geyer January 23, 2013 1 Metric Spaces Let X be an arbitrary set. A function d : X X R is called a metric if it satisfies the folloing
More informationLECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)]
LECTURE 7 NOTES 1. Convergence of random variables. Before delving into the large samle roerties of the MLE, we review some concets from large samle theory. 1. Convergence in robability: x n x if, for
More informationBrownian Motion and Stochastic Calculus
ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationProblem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1
Problem Sheet 1 1. Let Ω = {1, 2, 3}. Let F = {, {1}, {2, 3}, {1, 2, 3}}, F = {, {2}, {1, 3}, {1, 2, 3}}. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X :
More informationStatistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it
Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students
More informationLecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf
Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationBootstrap - theoretical problems
Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationAdvanced Signal Processing Introduction to Estimation Theory
Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,
More informationHomework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018
Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Topics: consistent estimators; sub-σ-fields and partial observations; Doob s theorem about sub-σ-field measurability;
More informationRESEARCH REPORT. A note on gaps in proofs of central limit theorems. Christophe A.N. Biscio, Arnaud Poinas and Rasmus Waagepetersen
CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING 2017 www.csgb.dk RESEARCH REPORT Christophe A.N. Biscio, Arnaud Poinas and Rasmus Waagepetersen A note on gaps in proofs of central limit theorems
More informationCHAPTER 3: LARGE SAMPLE THEORY
CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually
More informationStochastic Models (Lecture #4)
Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence
More informationLecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1]
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture 7: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 3, 06 [Ed. Apr ] In last lecture, we studied the minimax
More informationMath Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT
Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set
More informationMath 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson
Math 832 Fall 2013 University of Wisconsin at Madison Instructor: David F. Anderson Pertinent information Instructor: David Anderson Office: Van Vleck 617 email: anderson@math.wisc.edu Office hours: Mondays
More informationMidterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016
Midterm Exam Time: 09:10 12:10 11/23, 2016 Name: Student ID: Policy: (Read before You Start to Work) The exam is closed book. However, you are allowed to bring TWO A4-size cheat sheet (single-sheet, two-sided).
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More information7 Influence Functions
7 Influence Functions The influence function is used to approximate the standard error of a plug-in estimator. The formal definition is as follows. 7.1 Definition. The Gâteaux derivative of T at F in the
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More information1 Probability theory. 2 Random variables and probability theory.
Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationGood luck! Problem 1. (b) The random variables X 1 and X 2 are independent and N(0, 1)-distributed. Show that the random variables X 1
Avd. Matematisk statistik TETAME I SF2940 SAOLIKHETSTEORI/EXAM I SF2940 PROBABILITY THE- ORY, WEDESDAY OCTOBER 26, 2016, 08.00-13.00. Examinator : Boualem Djehiche, tel. 08-7907875, email: boualem@kth.se
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationMAT 135B Midterm 1 Solutions
MAT 35B Midterm Solutions Last Name (PRINT): First Name (PRINT): Student ID #: Section: Instructions:. Do not open your test until you are told to begin. 2. Use a pen to print your name in the spaces above.
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationIntroduction to Probability
LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute
More informationDerivatives and the Product Rule
Derivatives and the Product Rule James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University January 28, 2014 Outline 1 Differentiability 2 Simple Derivatives
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationTheoretical Statistics. Lecture 17.
Theoretical Statistics. Lecture 17. Peter Bartlett 1. Asymptotic normality of Z-estimators: classical conditions. 2. Asymptotic equicontinuity. 1 Recall: Delta method Theorem: Supposeφ : R k R m is differentiable
More informationIntroduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued
Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research
More informationChapter 6: Large Random Samples Sections
Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in
More informationThe Multivariate Gaussian Distribution [DRAFT]
The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,
More informationExercise 1. Let f be a nonnegative measurable function. Show that. where ϕ is taken over all simple functions with ϕ f. k 1.
Real Variables, Fall 2014 Problem set 3 Solution suggestions xercise 1. Let f be a nonnegative measurable function. Show that f = sup ϕ, where ϕ is taken over all simple functions with ϕ f. For each n
More informationUses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationMath 320: Real Analysis MWF 1pm, Campion Hall 302 Homework 8 Solutions Please write neatly, and in complete sentences when possible.
Math 320: Real Analysis MWF pm, Campion Hall 302 Homework 8 Solutions Please write neatly, and in complete sentences when possible. Do the following problems from the book: 4.3.5, 4.3.7, 4.3.8, 4.3.9,
More informationMcGill University Math 354: Honors Analysis 3
Practice problems McGill University Math 354: Honors Analysis 3 not for credit Problem 1. Determine whether the family of F = {f n } functions f n (x) = x n is uniformly equicontinuous. 1st Solution: The
More informationStat 710: Mathematical Statistics Lecture 31
Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:
More information17. Convergence of Random Variables
7. Convergence of Random Variables In elementary mathematics courses (such as Calculus) one speaks of the convergence of functions: f n : R R, then lim f n = f if lim f n (x) = f(x) for all x in R. This
More informationExample continued. Math 425 Intro to Probability Lecture 37. Example continued. Example
continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with
More informationδ -method and M-estimation
Econ 2110, fall 2016, Part IVb Asymptotic Theory: δ -method and M-estimation Maximilian Kasy Department of Economics, Harvard University 1 / 40 Example Suppose we estimate the average effect of class size
More informationBennett-type Generalization Bounds: Large-deviation Case and Faster Rate of Convergence
Bennett-type Generalization Bounds: Large-deviation Case and Faster Rate of Convergence Chao Zhang The Biodesign Institute Arizona State University Tempe, AZ 8587, USA Abstract In this paper, we present
More informationAustin Mohr Math 704 Homework 6
Austin Mohr Math 704 Homework 6 Problem 1 Integrability of f on R does not necessarily imply the convergence of f(x) to 0 as x. a. There exists a positive continuous function f on R so that f is integrable
More informationOn the convergence of sequences of random variables: A primer
BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :
More informationlim F n(x) = F(x) will not use either of these. In particular, I m keeping reserved for implies. ) Note:
APPM/MATH 4/5520, Fall 2013 Notes 9: Convergence in Distribution and the Central Limit Theorem Definition: Let {X n } be a sequence of random variables with cdfs F n (x) = P(X n x). Let X be a random variable
More informationSOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS
SOME CONVERSE LIMIT THEOREMS OR EXCHANGEABLE BOOTSTRAPS Jon A. Wellner University of Washington The bootstrap Glivenko-Cantelli and bootstrap Donsker theorems of Giné and Zinn (990) contain both necessary
More informationMidterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example
Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:
More informationECO Class 6 Nonparametric Econometrics
ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................
More informationBahadur representations for bootstrap quantiles 1
Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by
More informationSTAT 331. Martingale Central Limit Theorem and Related Results
STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 10: Expectation and Variance Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/ psarkar/teaching
More informationMidterm 1. Every element of the set of functions is continuous
Econ 200 Mathematics for Economists Midterm Question.- Consider the set of functions F C(0, ) dened by { } F = f C(0, ) f(x) = ax b, a A R and b B R That is, F is a subset of the set of continuous functions
More informationThus f is continuous at x 0. Matthew Straughn Math 402 Homework 6
Matthew Straughn Math 402 Homework 6 Homework 6 (p. 452) 14.3.3, 14.3.4, 14.3.5, 14.3.8 (p. 455) 14.4.3* (p. 458) 14.5.3 (p. 460) 14.6.1 (p. 472) 14.7.2* Lemma 1. If (f (n) ) converges uniformly to some
More informationLecture I: Asymptotics for large GUE random matrices
Lecture I: Asymptotics for large GUE random matrices Steen Thorbjørnsen, University of Aarhus andom Matrices Definition. Let (Ω, F, P) be a probability space and let n be a positive integer. Then a random
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationMachine Learning. Instructor: Pranjal Awasthi
Machine Learning Instructor: Pranjal Awasthi Course Info Requested an SPN and emailed me Wait for Carol Difrancesco to give them out. Not registered and need SPN Email me after class No promises It s a
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction
More information