Convergence Concepts of Random Variables and Functions

Similar documents
Convergence of Random Variables

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

17. Convergence of Random Variables

Lecture Notes 3 Convergence (Chapter 5)

MATH 140B - HW 5 SOLUTIONS

On the convergence of sequences of random variables: A primer

Math 328 Course Notes

Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain.

Stochastic Models (Lecture #4)

Thus f is continuous at x 0. Matthew Straughn Math 402 Homework 6

At t = T the investors learn the true state.

Summer Jump-Start Program for Analysis, 2012 Song-Ying Li

The Theory of Statistics and Its Applications

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence

e x = 1 + x + x2 2! + x3 If the function f(x) can be written as a power series on an interval I, then the power series is of the form

1. Is the set {f a,b (x) = ax + b a Q and b Q} of all linear functions with rational coefficients countable or uncountable?

Math 651 Introduction to Numerical Analysis I Fall SOLUTIONS: Homework Set 1

CLASSICAL PROBABILITY MODES OF CONVERGENCE AND INEQUALITIES

Lecture 2: Convergence of Random Variables

Homework 11. Solutions

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Analysis Qualifying Exam

Immerse Metric Space Homework

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Convergence of Random Variables

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote

Summer Jump-Start Program for Analysis, 2012 Song-Ying Li. 1 Lecture 7: Equicontinuity and Series of functions

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

McGill University Math 354: Honors Analysis 3

Real Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi

Economics 620, Lecture 8: Asymptotics I

1 Probability theory. 2 Random variables and probability theory.

Notes on uniform convergence

The Heine-Borel and Arzela-Ascoli Theorems

Probability and Measure

Math Bootcamp 2012 Miscellaneous

Problem set 1, Real Analysis I, Spring, 2015.

Math 118B Solutions. Charles Martin. March 6, d i (x i, y i ) + d i (y i, z i ) = d(x, y) + d(y, z). i=1

1 Probability space and random variables

Part II Probability and Measure

Riesz Representation Theorems

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Math 104: Homework 7 solutions

Definitions & Theorems

Hence, (f(x) f(x 0 )) 2 + (g(x) g(x 0 )) 2 < ɛ

1 Measurable Functions

3 Integration and Expectation

1 Sequences of events and their limits

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University

Probability and Measure

Advanced Calculus: MATH 410 Professor David Levermore 28 November 2006

STAT331 Lebesgue-Stieltjes Integrals, Martingales, Counting Processes

Induction, sequences, limits and continuity

The Borel-Cantelli Group

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Metric Spaces and Topology

Brownian Motion and Stochastic Calculus

Lecture 4: September Reminder: convergence of sequences

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Review of Probability. CS1538: Introduction to Simulations

LECTURE 3 RANDOM VARIABLES, CUMULATIVE DISTRIBUTION FUNCTIONS (CDFs)

MATH 202B - Problem Set 5

Chapter 5. Measurable Functions

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

Math Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT

MATH 1372, SECTION 33, MIDTERM 3 REVIEW ANSWERS

SDS : Theoretical Statistics

1. Let A R be a nonempty set that is bounded from above, and let a be the least upper bound of A. Show that there exists a sequence {a n } n N

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

8 Laws of large numbers

converges as well if x < 1. 1 x n x n 1 1 = 2 a nx n

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

9 Sequences of Functions

Review of Probability Theory II

0.1 Uniform integrability

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Math 140A - Fall Final Exam

Probability A exam solutions

Real Analysis Notes. Thomas Goller

Probability and Measure

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

REAL AND COMPLEX ANALYSIS

Entrance Exam, Real Analysis September 1, 2017 Solve exactly 6 out of the 8 problems

Estimates for probabilities of independent events and infinite series

Topic 7: Convergence of Random Variables

Constructing Approximations to Functions

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

MS 2001: Test 1 B Solutions

Definition 6.1. A metric space (X, d) is complete if every Cauchy sequence tends to a limit in X.

Introduction to Empirical Processes and Semiparametric Inference Lecture 13: Entropy Calculations

Spectral representations and ergodic theorems for stationary stochastic processes

Lecture Notes for MA 623 Stochastic Processes. Ionut Florescu. Stevens Institute of Technology address:

Continuous distributions

Math 209B Homework 2

From now on, we will represent a metric space with (X, d). Here are some examples: i=1 (x i y i ) p ) 1 p, p 1.

Math 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Transcription:

Convergence Concepts of Random Variables and Functions c 2002 2007, Professor Seppo Pynnonen, Department of Mathematics and Statistics, University of Vaasa Version: January 5, 2007

Convergence Modes Convergence of real numbers and real valued functions (a) Limit of a sequence of real numbers: Definition 1 The real number a IR is the limit of a sequence {x n }, denoted as lim n x n = a if for any ɛ > 0 there is an integer n(ɛ) such that x n a < ɛ whenever n n(ɛ), where n(ɛ) indicates that it may depend on ɛ. (b) Limit of a sequence of functions: Definition 2 A sequence of real-valued functions {f n } defined on a set S IR converge point wise to a function f : S IR, denoted as lim f n (x) = f(x), if for any ɛ > 0 and x S there is an integer n(ɛ, x) such that f n (x) f(x) < ɛ for each x S, whenever n n(ɛ, x), where n(ɛ, x) indicates that the integer may depend on ɛ and x. If n(ɛ, x) = n(ɛ), and for any ɛ > 0 f n (x) f(x) < ɛ for all x S, whenever n n(ɛ), then {f n } is converging uniformly to f. Note. In the latter case whenever n n(ɛ) the graphs of f n (x) become indistinguishable from f(x). Note. The uniform convergence implies the pointwise convergence, but not vice versa. Example. Let f n (x) = x n if 0 x 1. Then f(x) = lim n f n (x) = { 0 if 0 x < 1, 1 if x = 1. Thus {f n } converge pointwise but not uniformly to the limit function f(x) = 0 for 0 x < 1 and f(x) = 1 for x = 1. 2

Definition 3 (Limit of a function) A point c is the limit of a function f(x) as x p lim x p f(x) = c if for any ɛ > 0 there exists a δ > 0 such that f(x) c < ɛ, whenever 0 < x p < δ. Note. f need not be defined at p. Example. It is well known that f(x) = sin x x 1 as x 0. Definition 4 A function f : S IR is continuous at a point p S if f is defined at p, and for all ɛ > 0 there exist a δ > 0 such that f(x) f(p) < ɛ, whenever x p < δ Properties of lim: Theorem 1 Let lim x n = a and f(x) is continuous at a then lim f(x n ) = f(lim x n ) = f(a). Proof. For an arbitrary ɛ > 0 select δ > 0 (which may depend on ɛ) such that f(x) f(a) < ɛ, whenever x a < δ. Because x n a then there exists an n δ such that x n a < δ when n n δ, and hence f(x n ) f(a) < ɛ i.e., lim n f(x n ) = f(a), and the proof is complete. 3

Examples: If lim n x n = a 0 then lim 1 x n = 1 lim x n = 1 a lim x n = lim x n = a. Given that lim x n and lim y n exist then for an b, c IR lim(bx n + cy n ) = b lim x n + c lim y n lim x n y n = (lim x n )(lim y n ) provided that lim y n 0. lim x n y n = lim x n lim y n 4

Convergence of random variables Suppose next that the sequence X n is a sequence of random variables defined on a probability space (Ω, F, P ). Thus X n : Ω IR are just real valued functions with domain Ω and values X n (ω) IR), and all the convergence concepts defined so far apply as such. These are, however, pretty uninteresting, because they are tied to particular outcome ω. This would be just equivalent to investigate sequences {f n (x)} of real valued functions, and we would loose the new component randomness. Introducing randomness, then the interest in the convergence modes is the probability of those events where the sequence does not converge. If this diminishes to zero it is reasonable to say that a sequence of random variables converges. Definition 5 (Convergence with probability one, w.p.1-convergence or almost sure (a.s.) convergence). A sequence {X n } of random variables converges almost surely to a random variable X, denoted as if for any ɛ > 0 X n a.s. X (or X n X w.p.1) lim P ({ω Ω : X n(ω) X(ω) < ɛ, whenever n k}) = 1. k That is ( ) P lim X n = X = 1. n Note. The sets S k = {ω Ω : X n X < ɛ whenever n k} are an increasing sequence of sets, i.e., S k S k+1, and in set notations can be written as S k = {ω Ω : X k (ω) X(ω) < ɛ and X k+1 (ω) X(ω) < ɛ and } = n=k{ω Ω : X n (ω) X(ω) < ɛ}. Thus lim k P (S k ) = 1 and hence lim k P (S c k) = 0, i.e., for the sets S c k = Ω \ S k = n=k{ω Ω : X n X ɛ}, where the convergence does not occur, the probability diminishes to zero. Note. Confirming almost sure convergence of a sequence is, unfortunately, not an easy exercise. There are many sufficient conditions. One is the so 5

called Borel-Cantelli Lemma: Let {X n } be as sequence of random variables. If for every ɛ > 0 P ( X n X ɛ) < then n=1 X n a.s. X. Proof. Now P ( X n X ɛ) 0, and by assumption the sum above converges, which together imply that But then lim P ( X n X ɛ) = 0. k n=k lim k P ( n=k{ X n X < ɛ}) = 1 lim k P ( n=k{ X n X ɛ}) and the proof is complete. 1 lim k n=k P ( X n X ɛ) = 1, Definition 6 A sequence of random variables {X n } converges in probability to random variable X, denoted as plim X n = X or X n P X if for all ɛ > 0 lim P ( X n X ɛ) = 0. n Assume X n a.s X. Obviously, because { X n X ɛ} k=n{ X k X ɛ}, so that lim P ( X n X ɛ) lim P ( n n k=n{ X k X ɛ}) = 0. Thus a.s. convergence implies convergence in probability. The converse does not hold. Example. Let {X n } be a sequence of independent random variables, such that P (X n = 0) = 1 1 n, and P (X n = 1) = 1 n. 6

Then for any 1 > ɛ > 0 P ( X n < ɛ) = P (X n = 0) = 1 1 n 1, as n. So X n P 0. Nevertheless, lim k P ( n=k{ X n ɛ}) = 1 lim k P ( n=k{ X n < ɛ}) since n=k (1 1 ) = 0 for all k.1 n Thus X n 0 in probability but not a.s. = 1 lim k n=k P ( X n < ɛ) (by independence) = 1 lim k n=k (1 1 n ) = 1 Definition 7 (Convergence in distribution) A sequence of random variables {X n } in distribution to a random X, denoted as X n D X if lim F n(x) = F (x) n in all continuity points x of F (x), where F n (x) = P (X n x) is the cumulative distribution function (cdf) of X n and F (x) = P (X x) is the cdf of X. If g( ) is a continuous monotonically increasing function. Then if Y n = g(x n ), we have F Yn (y) = P (Y n y) = P (g(x n ) y) = P (X n g 1 (y)) = F Xn (g 1 (y))0. Similarly, if Y = g(x) then 1 n ( k 1 F Y (y) = P (Y y) = P (g(x) y) = P (X g 1 (y)) = F X (g 1 (y)). m=2 m=2 (1 1 m = n ( 1 1 m) ) 1 n m=2 m=2 ( m 1 m ( 1 1 m ) = (n 1)! n! = 1 n 0 as n. Generally n m=k (1 1 m ) ( ) ) = 1 = 1 1 k 1 n = (k 1) 1 n 0 as n. 7

If X n D X, then F Yn (y) = F Xn (g 1 (y)) F X (g 1 (y)) = F Y (y) as n. That is g(x n ) D g(x). More generally: D X and g( ) is con- Theorem 2 (Continuous Mapping Theorem) If X n tinuous, then g(x n ) D X. Definition 8 (Convergence in quadratic mean) A sequence of random variables {X n } with E(Xn) 2 < converges in quadratic mean to X (with E(X 2 ) < ), denoted as X n X if lim E(X n X) 2 = 0. n Using the Chebysev inequality we obtain directly that if X n P X. Generally X n X then [X n X] [X n a.s. X] [X n P X] [X n D X] Figure. Relations between convergence modes. Furthermore Theorem 3 Let {X n } and {Y n } be sequences of random variables. (a) If c is a constant then X n c X n P c 8

(b) If c is constant then X n c lim E[X n ] = c and lim Var[X n ] = lim E (X n E[X n ]) 2 = 0. (c) If plim X n and plim Y n exist and a and b are constants, then (1) plim (ax n ± by n ) = a plim X n ± b plim Y n, (2) plim (X n Y n ) = (plim X n )(plim Y n ) (3) if plim Y n 0 then plim Xn Y n = plim X n plim Y n. (4) if X n D X and plim X n = plim Y n, then Y n D X (d) If X n D X and plim Y n = c, a constant, then (1) X n ± Y n D X ± c. (2) X n Y n D cx given that c 0. (3) If c = 0, then plim X n Y n = 0. We observe that plim shares the same properties as the usual lim, but the other stochastic convergence concepts do not. Especially (and unfortunately) limit shares very few of these properties. For example X n X and Y n Y does not imply X n + Y n X + Y! (If it did, then by defining Z n = X n Y n one gets Z n 0, and property (c) would make equivalent to plim.) 9