SDS : Theoretical Statistics

Size: px
Start display at page:

Download "SDS : Theoretical Statistics"

Transcription

1 SDS : Theoretical Statistics Lecture 1: Introduction Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin

2 Manegerial Stuff Instructor- Purnamrita Sarkar Course material and homeworks will be posted under Office hours: Tuesdays 2-3 pm. GDC TA: Jack Liu Homeworks are due Biweekly after class on Wednesdays Grading - 5 homeworks (70% ), class participation (5%) Final Exam (25% ) Books Asymptotic Statistics, Aad van der Vaart. Cambridge Martin Wainwright s High dimensional statistics: A non-asymptotic view point 1

3 Why do theory? Say you have estimated ˆθn from data X 1,..., Xn. How do we know we have a good estimation method? Does ˆθ n θ? This brings us to Stochastic Convergence. How do I know if one estimation method is better than another? Does the estimate from one converge faster than the other? Does one algorithm work under broader parameter regimes, or weaker assumptions? What is the optimal rate for a given estimation problem? 2

4 This class Your instructor hopes to cover : Consistency of parameter estimates Stochastic Convergence Concentration inequalities Asymptotic normality of estimators Empirical processes, VC classes, covering numbers Asymptotic testing Examples of network clustering with a bit of random matrix theory Bootstrap, Nonparametric regression and density estimation 3

5 Stochastic Convergence Assume that X n, n 1 and X are elements of a separable metric space (S, d). Definition (Weak Convergence) A sequence of random variable s converge in law or in distribution to a random variable X, i.e. Xn d X if P(Xn x) P(X x) x at which P(X x) is continuous. Definition ( Convergence in Probability) A sequence of random variables converge in probability to a random variable X, i.e. Xn P X if ɛ > 0, P(d(Xn, X ) ɛ) 0. 4

6 Stochastic Convergence Assume that X n, n 1 and X are elements of a separable metric space (S, d). Definition (Almost Sure Convergence) A sequence of random variables converges almost surely to a random variable X, i.e. Xn a.s. ( ) X if P lim n d(xn, X ) = 0 = 1. If you think about a (scalar) random variable as a function that maps events to a real number, almost sure convergence means P(ω Ω : lim Xn(ω) = X (ω)) = 1 n Definition (Convergence in quadratic mean) A sequence of random variables converges in quadratic mean to a random variable X, i.e. Xn q.m [ X if E d(xn, X ) 2] 0. 5

7 Stochastic Convergence Theorem Xn a.s. X, Xn q.m. X Xn P X Xn d X Xn d c Xn P c 6

8 Converses:X n d X Xn P X Convergence in law needs no knowledge of the joint distribution of Xn and the limiting random variable X. Convergence in probability does. Example Consider X N(0, 1), X n = X. Xn d X. But how about Xn P X? 7

9 Converses:X n d X Xn P X Convergence in law needs no knowledge of the joint distribution of Xn and the limiting random variable X. Convergence in probability does. Example Consider X N(0, 1), X n = X. Xn d X. But how about Xn P X? P( Xn X ɛ) = P(2 X ɛ) 0 ɛ > 0. So Xn does not converge in probability to X. 7

10 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). 8

11 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. 8

12 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. Xn are a sequence of bernoulli s with probabilities pn = 1/2 k where k = log n. 8

13 Example Example Let Z U(0, 1) and for n = 2 k + m for k 0, 0 m < 2 k X n = 1(Z [m2 k, (m + 1)2 k ]), i.e. X1 = 1, X 2 = 1(Z [0, 1/2)), X 3 = 1(Z [1/2, 1)), X 4 = 1(Z [0, 1/4)), X 5 = 1(Z [1/4, 1/2)). For any Z (0, 1), the sequence {X n(z)} does not converge. So Xn a.s. 0. Xn are a sequence of bernoulli s with probabilities pn = 1/2 k where k = log n. So X n P 0 and Xn qm 0 8

14 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? 9

15 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n 9

16 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n E Xn 2 = 2 2n /n. So Xn qm 0 9

17 Example Example Let Z U([0, 1]) and Xn = 2 n 1(Z [0, 1/n)). Does Xn converge to X in quadratic mean, almost surely or in probability? a.s. P( lim Xn = X ) = P(Z > 0) = 1. So Xn X. n E Xn 2 = 2 2n /n. So Xn qm 0 P( X n ɛ) = P(Xn = 2 n ) = P(Z [0, 1/n)) = 1/n 0 9

18 Continuous Mapping Theorem Theorem Let g be continuous on a set C where P(X C) = 1. Then, Xn d X g(xn) d g(x ) Xn P X g(xn) P g(x ) Xn a.s. X g(xn) a.s. g(x ) 10

19 Example Let X n d X where X N(0, 1). Then X 2 n d? 11

20 Example Let X n d X where X N(0, 1). Then X 2 n d? Use g(x) = x 2. 11

21 Example Let X n d X where X N(0, 1). Then X 2 n d? Use g(x) = x 2. Use X 2 χ

22 Example d Let X n X where X N(0, 1). Then X 2 d n? Use g(x) = x 2. Use X 2 χ 2 1. So Xn 2 d χ

23 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? 12

24 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 12

25 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 Can we use Continuous Mapping Theorem to claim that g( Xn µ) 2 d 0? 12

26 Example-continuity points Let X 1,..., Xn be i.i.d. with mean µ and variance σ 2. We have X n µ d 0. Consider g(x) = 1x>0. Then g(( X n µ) 2 ) d? Using Continuous Mapping Theorem, ( Xn µ) 2 d 0 Can we use Continuous Mapping Theorem to claim that g( Xn µ) 2 d 0? NO. Because, 0 is a random variable whose mass is at 0, where g is discontinuous. 12

27 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? 13

28 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. 13

29 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. E[ f (X n) f (X ) 2 ] L 2 E[ Xn X 2 ] 0. So for Lipschitz functions quadratic mean convergence goes through. 13

30 How about convergence in q.m.? If Xn qm X, then is it true that for continuous f (discontinuous only at a measure zero set), f (Xn) qm f (X )? Consider an L- Lipschitz function f (X ). f (x) f (y) L x y. E[ f (X n) f (X ) 2 ] L 2 E[ Xn X 2 ] 0. So for Lipschitz functions quadratic mean convergence goes through. Can you come up with a non-lipschitz function and a sequence {X n} where f (Xn) qm 0? 13

31 Portmanteau Theorem Theorem The following are equivalent. Xn d X E[f (Xn)] E[f (X )] for all bounded and continuous f. E[f (Xn)] E[f (X )] for all bounded and Lipschitz f. E[e itt Xn ] E[e it T Xn ], t R k. (Levy s continuity theorem) t T X n d t T X t R k. (Cramer-Wold device) lim inf E[f (Xn)] E[f (X )] for all non-negative continuous f n lim sup P(Xn F ) P(X F ) for all closed F n lim inf P(Xn F ) P(X F ) for all open F n P(X n B) P(X B) for all continuity sets B (P(X B) = 0) 14

32 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n 15

33 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? 15

34 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? E[Xn] = 1. What went wrong? 15

35 Example-bounded Consider f (x) = x and n w.p. 1/n Xn = 0 w.p. 1 1/n X n d 0, but E[Xn]? E[Xn] = 1. What went wrong? f (x) = x is not bounded. 15

36 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) 16

37 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. 16

38 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. Example: Xn N(0, 1), Yn = Xn. X Y and X, Y are independent standard normal random variables. 16

39 Putting everything together Theorem X n d X and d(xn, Yn) P 0 Yn d X (1) X n d X and Yn d c (Xn, Yn) d (X, c) (2) X n P X and Yn P Y (Xn, Yn) P (X, Y ) (3) Eq 3 does not hold if we replace convergence in probability by convergence in distribution. Example: Xn N(0, 1), Yn = Xn. X Y and X, Y are independent standard normal random variables. Then X n d X and Yn d Y. But (Xn, Yn) d (X, X ), not (Xn, Yn) d (X, Y ). 16

40 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c 17

41 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c Does X n d X and Yn d Y imply Xn + Yn d X + Y? 17

42 Putting everything together Theorem (Slutsky s theorem) Xn d X and Yn d c imply that Xn + Yn d X + c XnYn d cx Xn/Yn d X /c Does X n d X and Yn d Y imply Xn + Yn d X + Y? Take Yn = Xn, and X, Y as independent standard normal random variables. Xn d X and Yn d Y but Xn + Yn d 0. 17

43 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn 18

44 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 i X 2 i X 2 n = n n 1 i X 2 i n X 2 n 18

45 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n 18

46 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. 18

47 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. Finally, n( Xn µ) d N(0, σ 2 ) using CLT. 18

48 Using all this If X 1,... Xn are i.i.d. random variables with mean µ and variance σ 2, prove that n Xn µ d N(0, 1). Sn First note that Sn = 1 n 1 Law of large numbers give So ( X 2 i X 2 n = i i X i 2 n n n 1 i X 2 i n P E[X 2 ] and Xn P µ. X 2 n i X i 2, Xn) P (E[X 2 ], µ) and now using the continuous mapping n theorem, Sn 2 P σ 2. Finally, n( Xn µ) d N(0, σ 2 ) using CLT. Now using Slutsky s lemma, n( Xn µ)/sn d N(0, 1) using CLT. 18

49 Uniformly tight Definition X is defined to be tight if ɛ > 0 M for which, P( X > M) < ɛ {Xn} is defined to uniformly tight if ɛ > 0 M for which, sup n P( Xn > M) < ɛ 19

50 Prohorov s theorem Theorem Xn d X {Xn} is UT. {Xn} is UT implies that, there exists a subsequence {n j } such that d Xn j X. 20

51 Notation for rates, small oh-pee and big oh-pee Definition The small o P : X n = op (1) Xn P 0 Xn = o P (Rn) Xn = YnRn and Yn = o P (1) Xn is vanishing in probability The big O P : Xn = O P (1) {Xn} is UT Xn = O P (Rn) Xn = YnRn and Yn = O P (1) Xn lies within a ball of finite radius with high probability 21

52 How do they interact o P (an) + o P (bn) = o P (max(an, bn)). o P (an) + O P (bn) = O P (bn). O P (a n)op (b n) = op (a nbn). 1 + O P (1) = O P (1). (1 + o P (1)) 1 = 1 + o P (1). o P (O P (1)) = o P (1). Xn = o P (an), Xn r = o P (an) r Be careful: e o P (1) o P (1) O P (1) + O P (1) Can actually be o P (1) because of cancellation. 22

53 23

Theoretical Statistics. Lecture 1.

Theoretical Statistics. Lecture 1. 1. Organizational issues. 2. Overview. 3. Stochastic convergence. Theoretical Statistics. Lecture 1. eter Bartlett 1 Organizational Issues Lectures: Tue/Thu 11am 12:30pm, 332 Evans. eter Bartlett. bartlett@stat.

More information

Lecture 4: September Reminder: convergence of sequences

Lecture 4: September Reminder: convergence of sequences 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and

More information

Chapter 5. Weak convergence

Chapter 5. Weak convergence Chapter 5 Weak convergence We will see later that if the X i are i.i.d. with mean zero and variance one, then S n / p n converges in the sense P(S n / p n 2 [a, b])! P(Z 2 [a, b]), where Z is a standard

More information

Stochastic Convergence, Delta Method & Moment Estimators

Stochastic Convergence, Delta Method & Moment Estimators Stochastic Convergence, Delta Method & Moment Estimators Seminar on Asymptotic Statistics Daniel Hoffmann University of Kaiserslautern Department of Mathematics February 13, 2015 Daniel Hoffmann (TU KL)

More information

Economics 620, Lecture 8: Asymptotics I

Economics 620, Lecture 8: Asymptotics I Economics 620, Lecture 8: Asymptotics I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 8: Asymptotics I 1 / 17 We are interested in the properties of estimators

More information

Lecture 2: Convergence of Random Variables

Lecture 2: Convergence of Random Variables Lecture 2: Convergence of Random Variables Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 2 Introduction to Stochastic Processes, Fall 2013 1 / 9 Convergence of Random Variables

More information

Convergence in Distribution

Convergence in Distribution Convergence in Distribution Undergraduate version of central limit theorem: if X 1,..., X n are iid from a population with mean µ and standard deviation σ then n 1/2 ( X µ)/σ has approximately a normal

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

Convergence Concepts of Random Variables and Functions

Convergence Concepts of Random Variables and Functions Convergence Concepts of Random Variables and Functions c 2002 2007, Professor Seppo Pynnonen, Department of Mathematics and Statistics, University of Vaasa Version: January 5, 2007 Convergence Modes Convergence

More information

Economics 241B Review of Limit Theorems for Sequences of Random Variables

Economics 241B Review of Limit Theorems for Sequences of Random Variables Economics 241B Review of Limit Theorems for Sequences of Random Variables Convergence in Distribution The previous de nitions of convergence focus on the outcome sequences of a random variable. Convergence

More information

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 3 Convergence (Chapter 5) Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let

More information

Regression #4: Properties of OLS Estimator (Part 2)

Regression #4: Properties of OLS Estimator (Part 2) Regression #4: Properties of OLS Estimator (Part 2) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #4 1 / 24 Introduction In this lecture, we continue investigating properties associated

More information

Lecture 1: Review of Basic Asymptotic Theory

Lecture 1: Review of Basic Asymptotic Theory Lecture 1: Instructor: Department of Economics Stanfor University Prepare by Wenbo Zhou, Renmin University Basic Probability Theory Takeshi Amemiya, Avance Econometrics, 1985, Harvar University Press.

More information

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/

More information

The Moment Method; Convex Duality; and Large/Medium/Small Deviations

The Moment Method; Convex Duality; and Large/Medium/Small Deviations Stat 928: Statistical Learning Theory Lecture: 5 The Moment Method; Convex Duality; and Large/Medium/Small Deviations Instructor: Sham Kakade The Exponential Inequality and Convex Duality The exponential

More information

University of California San Diego and Stanford University and

University of California San Diego and Stanford University and First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 13: Expectation and Variance and joint distributions Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin

More information

Lecture Notes on Asymptotic Statistics. Changliang Zou

Lecture Notes on Asymptotic Statistics. Changliang Zou Lecture Notes on Asymptotic Statistics Changliang Zou Prologue Why asymptotic statistics? The use of asymptotic approximation is two-fold. First, they enable us to find approximate tests and confidence

More information

Lecture 5: Asymptotic Equipartition Property

Lecture 5: Asymptotic Equipartition Property Lecture 5: Asymptotic Equipartition Property Law of large number for product of random variables AEP and consequences Dr. Yao Xie, ECE587, Information Theory, Duke University Stock market Initial investment

More information

Selected Exercises on Expectations and Some Probability Inequalities

Selected Exercises on Expectations and Some Probability Inequalities Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ

More information

Large Deviations for Small-Noise Stochastic Differential Equations

Large Deviations for Small-Noise Stochastic Differential Equations Chapter 22 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large

More information

Large Deviations for Small-Noise Stochastic Differential Equations

Large Deviations for Small-Noise Stochastic Differential Equations Chapter 21 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large

More information

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

Introduction to Self-normalized Limit Theory

Introduction to Self-normalized Limit Theory Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized

More information

Graduate Econometrics I: Asymptotic Theory

Graduate Econometrics I: Asymptotic Theory Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE FRANÇOIS BOLLEY Abstract. In this note we prove in an elementary way that the Wasserstein distances, which play a basic role in optimal transportation

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results

Introduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results Introduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics

More information

Stat 8112 Lecture Notes Weak Convergence in Metric Spaces Charles J. Geyer January 23, Metric Spaces

Stat 8112 Lecture Notes Weak Convergence in Metric Spaces Charles J. Geyer January 23, Metric Spaces Stat 8112 Lecture Notes Weak Convergence in Metric Spaces Charles J. Geyer January 23, 2013 1 Metric Spaces Let X be an arbitrary set. A function d : X X R is called a metric if it satisfies the folloing

More information

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)]

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)] LECTURE 7 NOTES 1. Convergence of random variables. Before delving into the large samle roerties of the MLE, we review some concets from large samle theory. 1. Convergence in robability: x n x if, for

More information

Brownian Motion and Stochastic Calculus

Brownian Motion and Stochastic Calculus ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1 Problem Sheet 1 1. Let Ω = {1, 2, 3}. Let F = {, {1}, {2, 3}, {1, 2, 3}}, F = {, {2}, {1, 3}, {1, 2, 3}}. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X :

More information

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students

More information

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

Bootstrap - theoretical problems

Bootstrap - theoretical problems Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem

More information

8 Laws of large numbers

8 Laws of large numbers 8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable

More information

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2 Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,

More information

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018

Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Homework Assignment #2 for Prob-Stats, Fall 2018 Due date: Monday, October 22, 2018 Topics: consistent estimators; sub-σ-fields and partial observations; Doob s theorem about sub-σ-field measurability;

More information

RESEARCH REPORT. A note on gaps in proofs of central limit theorems. Christophe A.N. Biscio, Arnaud Poinas and Rasmus Waagepetersen

RESEARCH REPORT. A note on gaps in proofs of central limit theorems.   Christophe A.N. Biscio, Arnaud Poinas and Rasmus Waagepetersen CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING 2017 www.csgb.dk RESEARCH REPORT Christophe A.N. Biscio, Arnaud Poinas and Rasmus Waagepetersen A note on gaps in proofs of central limit theorems

More information

CHAPTER 3: LARGE SAMPLE THEORY

CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 1 CHAPTER 3: LARGE SAMPLE THEORY CHAPTER 3 LARGE SAMPLE THEORY 2 Introduction CHAPTER 3 LARGE SAMPLE THEORY 3 Why large sample theory studying small sample property is usually

More information

Stochastic Models (Lecture #4)

Stochastic Models (Lecture #4) Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence

More information

Lecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1]

Lecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1] ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture 7: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 3, 06 [Ed. Apr ] In last lecture, we studied the minimax

More information

Math Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT

Math Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set

More information

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson Math 832 Fall 2013 University of Wisconsin at Madison Instructor: David F. Anderson Pertinent information Instructor: David Anderson Office: Van Vleck 617 email: anderson@math.wisc.edu Office hours: Mondays

More information

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016 Midterm Exam Time: 09:10 12:10 11/23, 2016 Name: Student ID: Policy: (Read before You Start to Work) The exam is closed book. However, you are allowed to bring TWO A4-size cheat sheet (single-sheet, two-sided).

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

7 Influence Functions

7 Influence Functions 7 Influence Functions The influence function is used to approximate the standard error of a plug-in estimator. The formal definition is as follows. 7.1 Definition. The Gâteaux derivative of T at F in the

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

1 Probability theory. 2 Random variables and probability theory.

1 Probability theory. 2 Random variables and probability theory. Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Good luck! Problem 1. (b) The random variables X 1 and X 2 are independent and N(0, 1)-distributed. Show that the random variables X 1

Good luck! Problem 1. (b) The random variables X 1 and X 2 are independent and N(0, 1)-distributed. Show that the random variables X 1 Avd. Matematisk statistik TETAME I SF2940 SAOLIKHETSTEORI/EXAM I SF2940 PROBABILITY THE- ORY, WEDESDAY OCTOBER 26, 2016, 08.00-13.00. Examinator : Boualem Djehiche, tel. 08-7907875, email: boualem@kth.se

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

MAT 135B Midterm 1 Solutions

MAT 135B Midterm 1 Solutions MAT 35B Midterm Solutions Last Name (PRINT): First Name (PRINT): Student ID #: Section: Instructions:. Do not open your test until you are told to begin. 2. Use a pen to print your name in the spaces above.

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

Derivatives and the Product Rule

Derivatives and the Product Rule Derivatives and the Product Rule James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University January 28, 2014 Outline 1 Differentiability 2 Simple Derivatives

More information

6.1 Moment Generating and Characteristic Functions

6.1 Moment Generating and Characteristic Functions Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Theoretical Statistics. Lecture 17.

Theoretical Statistics. Lecture 17. Theoretical Statistics. Lecture 17. Peter Bartlett 1. Asymptotic normality of Z-estimators: classical conditions. 2. Asymptotic equicontinuity. 1 Recall: Delta method Theorem: Supposeφ : R k R m is differentiable

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research

More information

Chapter 6: Large Random Samples Sections

Chapter 6: Large Random Samples Sections Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

Exercise 1. Let f be a nonnegative measurable function. Show that. where ϕ is taken over all simple functions with ϕ f. k 1.

Exercise 1. Let f be a nonnegative measurable function. Show that. where ϕ is taken over all simple functions with ϕ f. k 1. Real Variables, Fall 2014 Problem set 3 Solution suggestions xercise 1. Let f be a nonnegative measurable function. Show that f = sup ϕ, where ϕ is taken over all simple functions with ϕ f. For each n

More information

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ). 1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need

More information

Math 320: Real Analysis MWF 1pm, Campion Hall 302 Homework 8 Solutions Please write neatly, and in complete sentences when possible.

Math 320: Real Analysis MWF 1pm, Campion Hall 302 Homework 8 Solutions Please write neatly, and in complete sentences when possible. Math 320: Real Analysis MWF pm, Campion Hall 302 Homework 8 Solutions Please write neatly, and in complete sentences when possible. Do the following problems from the book: 4.3.5, 4.3.7, 4.3.8, 4.3.9,

More information

McGill University Math 354: Honors Analysis 3

McGill University Math 354: Honors Analysis 3 Practice problems McGill University Math 354: Honors Analysis 3 not for credit Problem 1. Determine whether the family of F = {f n } functions f n (x) = x n is uniformly equicontinuous. 1st Solution: The

More information

Stat 710: Mathematical Statistics Lecture 31

Stat 710: Mathematical Statistics Lecture 31 Stat 710: Mathematical Statistics Lecture 31 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 31 April 13, 2009 1 / 13 Lecture 31:

More information

17. Convergence of Random Variables

17. Convergence of Random Variables 7. Convergence of Random Variables In elementary mathematics courses (such as Calculus) one speaks of the convergence of functions: f n : R R, then lim f n = f if lim f n (x) = f(x) for all x in R. This

More information

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with

More information

δ -method and M-estimation

δ -method and M-estimation Econ 2110, fall 2016, Part IVb Asymptotic Theory: δ -method and M-estimation Maximilian Kasy Department of Economics, Harvard University 1 / 40 Example Suppose we estimate the average effect of class size

More information

Bennett-type Generalization Bounds: Large-deviation Case and Faster Rate of Convergence

Bennett-type Generalization Bounds: Large-deviation Case and Faster Rate of Convergence Bennett-type Generalization Bounds: Large-deviation Case and Faster Rate of Convergence Chao Zhang The Biodesign Institute Arizona State University Tempe, AZ 8587, USA Abstract In this paper, we present

More information

Austin Mohr Math 704 Homework 6

Austin Mohr Math 704 Homework 6 Austin Mohr Math 704 Homework 6 Problem 1 Integrability of f on R does not necessarily imply the convergence of f(x) to 0 as x. a. There exists a positive continuous function f on R so that f is integrable

More information

On the convergence of sequences of random variables: A primer

On the convergence of sequences of random variables: A primer BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :

More information

lim F n(x) = F(x) will not use either of these. In particular, I m keeping reserved for implies. ) Note:

lim F n(x) = F(x) will not use either of these. In particular, I m keeping reserved for implies. ) Note: APPM/MATH 4/5520, Fall 2013 Notes 9: Convergence in Distribution and the Central Limit Theorem Definition: Let {X n } be a sequence of random variables with cdfs F n (x) = P(X n x). Let X be a random variable

More information

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS

SOME CONVERSE LIMIT THEOREMS FOR EXCHANGEABLE BOOTSTRAPS SOME CONVERSE LIMIT THEOREMS OR EXCHANGEABLE BOOTSTRAPS Jon A. Wellner University of Washington The bootstrap Glivenko-Cantelli and bootstrap Donsker theorems of Giné and Zinn (990) contain both necessary

More information

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

Bahadur representations for bootstrap quantiles 1

Bahadur representations for bootstrap quantiles 1 Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by

More information

STAT 331. Martingale Central Limit Theorem and Related Results

STAT 331. Martingale Central Limit Theorem and Related Results STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 10: Expectation and Variance Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/ psarkar/teaching

More information

Midterm 1. Every element of the set of functions is continuous

Midterm 1. Every element of the set of functions is continuous Econ 200 Mathematics for Economists Midterm Question.- Consider the set of functions F C(0, ) dened by { } F = f C(0, ) f(x) = ax b, a A R and b B R That is, F is a subset of the set of continuous functions

More information

Thus f is continuous at x 0. Matthew Straughn Math 402 Homework 6

Thus f is continuous at x 0. Matthew Straughn Math 402 Homework 6 Matthew Straughn Math 402 Homework 6 Homework 6 (p. 452) 14.3.3, 14.3.4, 14.3.5, 14.3.8 (p. 455) 14.4.3* (p. 458) 14.5.3 (p. 460) 14.6.1 (p. 472) 14.7.2* Lemma 1. If (f (n) ) converges uniformly to some

More information

Lecture I: Asymptotics for large GUE random matrices

Lecture I: Asymptotics for large GUE random matrices Lecture I: Asymptotics for large GUE random matrices Steen Thorbjørnsen, University of Aarhus andom Matrices Definition. Let (Ω, F, P) be a probability space and let n be a positive integer. Then a random

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Machine Learning. Instructor: Pranjal Awasthi

Machine Learning. Instructor: Pranjal Awasthi Machine Learning Instructor: Pranjal Awasthi Course Info Requested an SPN and emailed me Wait for Carol Difrancesco to give them out. Not registered and need SPN Email me after class No promises It s a

More information

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2 Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction

More information