The Central Limit Theorem: More of the Story
|
|
- Cameron Newman
- 5 years ago
- Views:
Transcription
1 The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
2 Central Limit Theorem Theorem (Central Limit Theorem) Let X 1, X 2,... be a sequence of independent and identically distributed random variables, each with expectation µ and variance σ 2. Then the distribution of Z n = X 1 + X X n nµ σ n converges to the distribution of a standard normal random variable. lim P(Z n x) = 1 x e y2 2 dy n 2π Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
3 Central Limit Theorem Applications The sampling distribution of the mean is approximately normal. The distribution of experimental errors is approximately normal. >> But why the normal distribution? Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
4 Benford s Law In an arbitrary table of data, such as populations or lake areas, P[Leading digit is d] = log 10 (1 + 1 d ) Data: List of 60 Tallest Buildings Lead Digit Meters Feet Benford Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
5 Benford Justification Simon Newcomb 1881 Frank Benford 1938 Proof arguments: Positional number system Densities Scale invariance Scale and base unbiased (Hill 1995) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
6 Central Limit Theorem Theorem (Central Limit Theorem) Let X 1, X 2,... be a sequence of independent and identically distributed random variables, each with expectation µ and variance σ 2. Then the distribution of Z n = X 1 + X X n nµ σ n converges to the distribution of a standard normal random variable. lim P(Z n x) = 1 x e y2 2 dy n 2π Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
7 Central Limit Theorem Proof Proof Sketch: Let Y i = X i µ Moment Generating Function of Y i is M Yi (t) = Ee ty i MGF of Z n is M Zn (t) = [M Y1 ( t σ n ]n lim n ln M Zn (t) = t2 2 The MGF of the standard normal is e t2 2 Since the MGF s converge, the distributions converge. (Lévy Continuity Theorem). Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
8 Counter-Examples Moment problem: Lognormal R.V. not determined by moments. No first moment: Cauchy R.V. has no MGF, EX = so CLT does not hold. No second moment: f (x) = 1 x 3 for x 1. Pairwise independence is not sufficient for CLT. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
9 Demoivre s Theorem 1733 Each X i is Bernoulli ( 0 or 1 ). b(k) = P[S n = k] = ( n k) 1 2 n n! ( 2π)n n ne n (Stirling s formula) b( n 2 ) 2 πn log( b( n 2 +d b( n 2 ) ) 2d2 n b( n 2 + d) 2 πn e 2d2 n lim n P[a Sn n/2 n/2 b] = 1 b x2 2 a e 2 dx Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
10 Laplace 1810 Dealt with independent and identically distributed case. Started with discrete variables: Consider X i where p k = P[X i = k m ] for k = m, m + 1,, m 1, m Generating function: T (t) = m k= m p kt k q j = P[ X i = j m ] is coefficient of tj in T (t) n Substitute e ix 1 π for t and recall 2π π e itx e isx dx = δ ts Then, q j = 1 π 2π π e ijx [ m m p ke ikx ] n dx Now, expand e ikx in a power series around 0 and use the fact that the mean of X i is zero. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
11 Why Normal? Normal Characteristic Function f (u) = 1 2π e iux e x2 2 dx = e u2 2 f Sn σ n (u) = E[e iu(sn/σ n) u ] = (f ( σ n ))n = (1 σ2 2σ 2 n u2 + o( σ2 σ 2 n u2 )) n = (1 u2 2n + o(u2 n ))n e u 2 2 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
12 Levy Continuity Theorem If distribution functions F n converge to F, then the corresponding ch.f. f n converge to f. Conversely, if f n converges to g continuous at 0, then F n converges to F. Proof Sketch: First direction is the Helly-Bray theorem. The set {e iux } is a separating set for distribution functions. In both directions, continuity points and mass of F n are critical. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
13 History Laplace never presented a general CLT statement. (Concerned with limiting probabilities for particular problems). Concern over convergence led Poisson to improvements (not identically distributed case). Dirichlet and Cauchy changed conception of analysis (epsilon/delta). Counter-examples uncovered limitations. Chebychev proved CLT using convergence of moments. (Markov and Liapounov were students). First rigorous proof (Liapounov 1900). CLT holds with independent (but not necessarily i.i.d.) X i if E Xj 3 [ Xj 2 = E Xj 3 ]3/2 sn 3 0 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
14 Liapounov proof Assume: E Xj 3 [ E Xj = 3 Xj 2]3/2 sn 3 0 g n (u) = Π n 1 f k( u s n ) = Π n 1 [1 + (f k( u s n ) 1)] f k ( u s n ) = 1 u2 (σ 2 2sn 2 k + δ k( u s n )) f k ( u s n ) 1 2u 2 σ2 k s 2 n = k 1 f k( u s n ) 1 2u 2 (E X 2 k ) 3 2 < E X k 3 σ sup k k n s n 0 = sup f k ( u s n ) 1 0 Use log(1 + z) = z(1 + θz) where θ 1 for z 1 2 log g n (u) = n 1 (f k( u s n ) 1) + θ n 1 (f k( u s n ) 1) 2 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
15 Liapounov proof continued log g n (u) = n 1 (f k( u s n ) 1) + θ n 1 (f k( u s n ) 1) 2 θ n 1 (f k( u s n ) 1) 2 sup f k ( u s n ) 1 n 1 (f k( u s n ) 1) 0 σk 2 2 sn 2 u + θ 3 k E X k 3 f k ( u s n ) 1 = u2 sn 3 n 1 (f k( u s n ) 1) = u2 2 + θ u3 sn 3 n 1 E X k 3 u2 2 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
16 Lindeberg 1922 Theorem (Central Limit Theorem) Let the variables X i be independent with EX i = 0 and EXi 2 = σi 2. Let s be the standard deviation of the sum S and let F be the distribution of S s. With Φ(x) the normal distribution, then if 1 sn 2 x ɛs n x 2 df k 0, we have sup F (x) Φ(x) 5ɛ x Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
17 Lindeberg Proof Pick auxiliary function f. Arbitrary distribution V, define F (x) = f (x t)dv (x) With φ(x) the normal density, define Ψ(x) = (f (x t)φ(x)dx Taylor expansion of f to third power gives F (x) Ψ(x) < k x 3 dv (x) With U i the distribution of X i, F 1 (x) = f (x t)du 1 (x)... F n (x) = F n 1 (x t)du n (x) Note U(x) = U(x t 1 t 2 t n )du 1 (t 1 ) du n (t n ) By selecting f carefully, U(x) Φ(x) < 3( n i x 3 du i (x)) 1 4 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
18 Still Why Normal? Let X = X 1 + X 2 where X is N (0, 1) and X 1 independent of X 2 f (u) = f 1 (u)f 2 (u) = e u2 2 e u2 2 is an entire, non-vanishing function with f 1 (z) e c u 2, Hadamard factorization theorem = log f 1 (u) is a polynomial in u of at most degree 2. f is a characteristic function = f (0) = 1, f (u) = f ( u), and it is bounded. Hence, log f (u) = iua + bu 2. This is the general form of the normal characteristic function. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
19 Feller - Levy 1935 Theorem (Final Central Limit Theorem) Let the variables X i be independent with EX i = 0 and EXi 2 = σi 2. Let S n = n 1 X i and sn 2 = n 1 σ2 k. Φ is the normal distribution and F k is the distribution of X k. Then as n, if and only if for every ɛ > 0 P[S n /s n x] Φ(x) and max k n σ k s n 0 1 sn 2 x 2 df k 0 x ɛs n Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
20 Levy Following appeared in Levy s monograph in Theorem (Levy s version) In order that a sum S = j X j of independent variables have a distribution close to Gaussian, it is necessary and sufficient that, after reducing medians to zero, the following conditions be satisfied. Each summand that is not negligible compared to the dispersion of the entire sum has a distribution close to Gaussian. The maximum of the absolute value of the negligible summands is itself negligible compared to the dispersion of the sum. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
21 Normed sums and Stable Laws Definition X is said to have a stable law if whenever X 1, X 2,, X k are independent with the same distribution as X, we have X 1 + X X k D = ax + b Both the Normal (σ 2 < 0) and Cauchy (σ 2 = ) Laws are stable. Theorem (Limit of Normed Sums) Suppose that then X has a stable law. S n = X 1 + X X n A n B n D X Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
22 Entropy Definition (Discrete Entropy) Let X be a discrete random variable taking values x i with probability p i. The entropy of X is H(X ) = p i log p i H(X ) > 0 unless X is constant. H(aX + b) = H(X ) If X is the result of flipping a fair coin, H(X ) = 1. If X takes n values, H(X ) is maximized when p i = 1 n. Extend to joint entropy H(X, Y ). Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
23 Entropy Axioms Symmetric in p i Continuous in p i Normalized: H(X ) = 1 for fair coin. X and Y independent gives H(X, Y ) = H(X ) + H(Y ) Decomposition: H(r 1,, r m, q 1,, q n ) = αh(r 1,, r m ) + (1 α)h(q 1,, q n ) Axioms = H(X ) = p i log p i Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
24 Entropy Definition (Differential Entropy) Let X be a continuous random variable with density p. The differential entropy of X is H(X ) = p(t) log p(t)dt X is uniform on [0, c] gives H(X ) = log(c) X is N (0, σ 2 ) gives H(X ) = 1 2 log(2πeσ2 ) H(aX + b) = H(X ) + log(a) H(X ) can be negative. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
25 Entropy Definition (Relative Entropy) Let p and q be densities. The relative entropy distance from p to q is D(p q) = p(x) log( p(x) q(x) )dx If supp(p) supp(q), then D(p q) =. D is not a metric (not symmetric and no triangle inequality). D(p q) 0 with equality if and only if p = q a.e. Convergence in D is stronger than convergence in L 1. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
26 Entropy Lemma (Maximum Entropy) Let p be the density of random variable with variance σ 2, and let φ 2 σ be the density of a N (0, σ 2 ) random variable. Then H(p) H(φ 2 σ) = log(2πeσ2 ) 2 with equality if and only if p is a normal density. Proof: 0 D(p φ 2 σ) = p(x)(log(p(x) + log(2πσ2 ) 2 = H(p) + log(2πeσ2 ) 2 + x 2 2σ 2 log(e))dx = H(p) + H(φ 2 σ) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
27 Second Law of Thermodynamics Let Ω be microstates corresponding to a particular macrostate. Then the entropy of the macrostate is S = k log e Ω. Ω is composed of microstates r of probability p r. If there are v copies of the system, about vp r are microstates of type r. Then, Ω = v v 1!v 2! v k! v v v v 1 1 v v 2 2 v v k k S = k log e Ω k(v log(v) v r log(v r )) = kv p r log(p r ) Entropy is maximized subject to energy constraint at Gibbs states. ( p r = 1 and p r E r = E) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
28 Fisher Information Definition Let Y be a random variable with density g and variance σ 2. Set ρ = g g and let φ be the normal density with the same mean and variance as Y. Fisher Information is I (Y ) = E[(ρ(Y )) 2 ] Standardized Fisher Information is J(Y ) = σ 2 E[(ρ(Y ) ρ φ (Y )) 2 ] Z with distribution N (0, σ 2 ) minimizes Fisher information (I (Z) = 1 σ 2 ) among all distributions with variance σ 2. (Note: J(Z) = 0) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
29 Fisher Information Properties Lemma (de Bruijn) If Z is a normal R.V. independent of Y with the same mean and variance, then D(Y Z) = 1 0 J( ty + 1 tz) 1 2t dt Proof relies on these facts: The normal density satisfies the heat equation: φτ τ = 1 2 Hence, Y + Z also satisfies the heat equation. We can then calculate the derivative of D(Y Z). 2 φ τ /x 2. We also have that if J(Y ) 0 then D(Y Z) 0. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
30 Fisher Information Properties Lemma If U and V are independent, then for β [0, 1] J(U + V ) β 2 J(U) + (1 β 2 )J(V ) J( βu + 1 βv ) βj(u) + (1 β)j(v ) with equality if and only if U and V are normal. In particular, J(X + Y ) J(X ) H(X + Y ) H(X ) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
31 Fisher Information Properties From Lemma, J( βu + 1 βv ) βj(u) + (1 β)j(v ) In particular, with S n = ( n 1 X i)/ n and β = n n+m, nj(s n ) + mj(s m ) (m + n)j(s n+m ) If J(S n ) < for some n, then J(S n ) converges to 0. (Take n = m and assume i.i.d. variables to see monotone convergence of a subsequence.) Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
32 Sketch of Central Limit Theorem Proof Assume i.i.d. random variables J(S n ) converges to 0 Hence D(S n Z) converges to 0 For densities p Sn and φ 2 σ, (p Sn φ σ 2) 2 2D(S n Z) S n N (0, σ 2 ) Can generalize to non i.i.d. variables. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
33 Conclusions Assume errors are uniformally asymptotically negligible (no one dominates). Assume independent errors with distributions that have second moments. Normalize the sum with mean and variance. Then the limit is a standard normal distribution. Normal distribution is stable ( infinitely divisible ). (X = X 1 + X 2 ) Normal distribution maximizes entropy. Convolution (adding R.V. s) increases entropy. Steven Janke (Seminar) The Central Limit Theorem:More of the Story November / 33
Continuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More information1* (10 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2
Math 736-1 Homework Fall 27 1* (1 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2 and let Y be a standard normal random variable. Assume that X and Y are independent. Find the distribution
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationProduct measure and Fubini s theorem
Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω
More informationThe Lindeberg central limit theorem
The Lindeberg central limit theorem Jordan Bell jordan.bell@gmail.com Department of Mathematics, University of Toronto May 29, 205 Convergence in distribution We denote by P d the collection of Borel probability
More informationProbability and Measure
Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationNotes 1 : Measure-theoretic foundations I
Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationProbability and Measure
Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationMath 341: Probability Seventeenth Lecture (11/10/09)
Math 341: Probability Seventeenth Lecture (11/10/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationProbability Theory I: Syllabus and Exercise
Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationMultiple Random Variables
Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More information18.175: Lecture 15 Characteristic functions and central limit theorem
18.175: Lecture 15 Characteristic functions and central limit theorem Scott Sheffield MIT Outline Characteristic functions Outline Characteristic functions Characteristic functions Let X be a random variable.
More informationd(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N
Problem 1. Let f : A R R have the property that for every x A, there exists ɛ > 0 such that f(t) > ɛ if t (x ɛ, x + ɛ) A. If the set A is compact, prove there exists c > 0 such that f(x) > c for all x
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationMathematical Statistical Physics Solution Sketch of the math part of the Exam
Prof. B. Paredes, Prof. P. Pickl Summer Term 2016 Lukas Nickel, Carlos Velasco July 22, 2016 Mathematical Statistical Physics Solution Sketch of the math part of the Exam Family Name: First name: Student
More informationExpectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda
Expectation DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean,
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationExpectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,
More informationMath 341: Probability Eighteenth Lecture (11/12/09)
Math 341: Probability Eighteenth Lecture (11/12/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationLecture 3 - Expectation, inequalities and laws of large numbers
Lecture 3 - Expectation, inequalities and laws of large numbers Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 3 - Expectation, inequalities and laws of large numbersapril 19, 2009 1 / 67 Part
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationTHE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION
THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION JAINUL VAGHASIA Contents. Introduction. Notations 3. Background in Probability Theory 3.. Expectation and Variance 3.. Convergence
More informationEE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm
EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm 1. Feedback does not increase the capacity. Consider a channel with feedback. We assume that all the recieved outputs are sent back immediately
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationProbability Theory Overview and Analysis of Randomized Algorithms
Probability Theory Overview and Analysis of Randomized Algorithms Analysis of Algorithms Prepared by John Reif, Ph.D. Probability Theory Topics a) Random Variables: Binomial and Geometric b) Useful Probabilistic
More informationECE 4400:693 - Information Theory
ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential
More information18.175: Lecture 8 Weak laws and moment-generating/characteristic functions
18.175: Lecture 8 Weak laws and moment-generating/characteristic functions Scott Sheffield MIT 18.175 Lecture 8 1 Outline Moment generating functions Weak law of large numbers: Markov/Chebyshev approach
More information4 Expectation & the Lebesgue Theorems
STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does
More informationLecture 11: Continuous-valued signals and differential entropy
Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components
More informationConditional Distributions
Conditional Distributions The goal is to provide a general definition of the conditional distribution of Y given X, when (X, Y ) are jointly distributed. Let F be a distribution function on R. Let G(,
More informationCourse: ESO-209 Home Work: 1 Instructor: Debasis Kundu
Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear
More information18.175: Lecture 14 Infinite divisibility and so forth
18.175 Lecture 14 18.175: Lecture 14 Infinite divisibility and so forth Scott Sheffield MIT 18.175 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin
More informationProbability and Measure
Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real
More information18.175: Lecture 17 Poisson random variables
18.175: Lecture 17 Poisson random variables Scott Sheffield MIT 1 Outline More on random walks and local CLT Poisson random variable convergence Extend CLT idea to stable random variables 2 Outline More
More informationNormal Random Variables and Probability
Normal Random Variables and Probability An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2015 Discrete vs. Continuous Random Variables Think about the probability of selecting
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationCourse 212: Academic Year Section 1: Metric Spaces
Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationProblem set 2 The central limit theorem.
Problem set 2 The central limit theorem. Math 22a September 6, 204 Due Sept. 23 The purpose of this problem set is to walk through the proof of the central limit theorem of probability theory. Roughly
More informationExercises with solutions (Set D)
Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where
More information. Find E(V ) and var(v ).
Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number
More informationRandom Variables (Continuous Case)
Chapter 6 Random Variables (Continuous Case) Thus far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions
More information1 Assignment 1: Nonlinear dynamics (due September
Assignment : Nonlinear dynamics (due September 4, 28). Consider the ordinary differential equation du/dt = cos(u). Sketch the equilibria and indicate by arrows the increase or decrease of the solutions.
More informationStochastic Models (Lecture #4)
Stochastic Models (Lecture #4) Thomas Verdebout Université libre de Bruxelles (ULB) Today Today, our goal will be to discuss limits of sequences of rv, and to study famous limiting results. Convergence
More informationHomework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015
10-704 Homework 1 Due: Thursday 2/5/2015 Instructions: Turn in your homework in class on Thursday 2/5/2015 1. Information Theory Basics and Inequalities C&T 2.47, 2.29 (a) A deck of n cards in order 1,
More informationOn the Breiman conjecture
Péter Kevei 1 David Mason 2 1 TU Munich 2 University of Delaware 12th GPSD Bochum Outline Introduction Motivation Earlier results Partial solution Sketch of the proof Subsequential limits Further remarks
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More information4 Sums of Independent Random Variables
4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables
More informationMAT 135B Midterm 1 Solutions
MAT 35B Midterm Solutions Last Name (PRINT): First Name (PRINT): Student ID #: Section: Instructions:. Do not open your test until you are told to begin. 2. Use a pen to print your name in the spaces above.
More informationMeasure-theoretic probability
Measure-theoretic probability Koltay L. VEGTMAM144B November 28, 2012 (VEGTMAM144B) Measure-theoretic probability November 28, 2012 1 / 27 The probability space De nition The (Ω, A, P) measure space is
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationProblem Set 6: Solutions Math 201A: Fall a n x n,
Problem Set 6: Solutions Math 201A: Fall 2016 Problem 1. Is (x n ) n=0 a Schauder basis of C([0, 1])? No. If f(x) = a n x n, n=0 where the series converges uniformly on [0, 1], then f has a power series
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationPart II Probability and Measure
Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationMathematics 324 Riemann Zeta Function August 5, 2005
Mathematics 324 Riemann Zeta Function August 5, 25 In this note we give an introduction to the Riemann zeta function, which connects the ideas of real analysis with the arithmetic of the integers. Define
More informationExpectation of Random Variables
1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationStable Process. 2. Multivariate Stable Distributions. July, 2006
Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.
More informationModule 3. Function of a Random Variable and its distribution
Module 3 Function of a Random Variable and its distribution 1. Function of a Random Variable Let Ω, F, be a probability space and let be random variable defined on Ω, F,. Further let h: R R be a given
More informationECE534, Spring 2018: Solutions for Problem Set #3
ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their
More informationChapter 5. Random Variables (Continuous Case) 5.1 Basic definitions
Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More informationX n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)
14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence
More informationChapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye
Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for
More informationStatistics 581, Problem Set 1 Solutions Wellner; 10/3/ (a) The case r = 1 of Chebychev s Inequality is known as Markov s Inequality
Statistics 581, Problem Set 1 Solutions Wellner; 1/3/218 1. (a) The case r = 1 of Chebychev s Inequality is known as Markov s Inequality and is usually written P ( X ɛ) E( X )/ɛ for an arbitrary random
More informationAn Introduction to Laws of Large Numbers
An to Laws of John CVGMI Group Contents 1 Contents 1 2 Contents 1 2 3 Contents 1 2 3 4 Intuition We re working with random variables. What could we observe? {X n } n=1 Intuition We re working with random
More informationSTA 711: Probability & Measure Theory Robert L. Wolpert
STA 711: Probability & Measure Theory Robert L. Wolpert 6 Independence 6.1 Independent Events A collection of events {A i } F in a probability space (Ω,F,P) is called independent if P[ i I A i ] = P[A
More informationEE514A Information Theory I Fall 2013
EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationChapter 4 : Expectation and Moments
ECE5: Analysis of Random Signals Fall 06 Chapter 4 : Expectation and Moments Dr. Salim El Rouayheb Scribe: Serge Kas Hanna, Lu Liu Expected Value of a Random Variable Definition. The expected or average
More informationProbability on a Riemannian Manifold
Probability on a Riemannian Manifold Jennifer Pajda-De La O December 2, 2015 1 Introduction We discuss how we can construct probability theory on a Riemannian manifold. We make comparisons to this and
More informationNotes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770
Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November
More informationODE Final exam - Solutions
ODE Final exam - Solutions May 3, 018 1 Computational questions (30 For all the following ODE s with given initial condition, find the expression of the solution as a function of the time variable t You
More informationEE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.
EE376A - Information Theory Final, Monday March 14th 216 Solutions Instructions: You have three hours, 3.3PM - 6.3PM The exam has 4 questions, totaling 12 points. Please start answering each question on
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationSection 9.1. Expected Values of Sums
Section 9.1 Expected Values of Sums Theorem 9.1 For any set of random variables X 1,..., X n, the sum W n = X 1 + + X n has expected value E [W n ] = E [X 1 ] + E [X 2 ] + + E [X n ]. Proof: Theorem 9.1
More informationUseful Probability Theorems
Useful Probability Theorems Shiu-Tang Li Finished: March 23, 2013 Last updated: November 2, 2013 1 Convergence in distribution Theorem 1.1. TFAE: (i) µ n µ, µ n, µ are probability measures. (ii) F n (x)
More informationLecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs
Lecture Notes 3 Multiple Random Variables Joint, Marginal, and Conditional pmfs Bayes Rule and Independence for pmfs Joint, Marginal, and Conditional pdfs Bayes Rule and Independence for pdfs Functions
More informationMonotonicity of entropy and Fisher information: a quick proof via maximal correlation
Communications in Information and Systems Volume 16, Number 2, 111 115, 2016 Monotonicity of entropy and Fisher information: a quick proof via maximal correlation Thomas A. Courtade A simple proof is given
More informationELEG 3143 Probability & Stochastic Process Ch. 4 Multiple Random Variables
Department o Electrical Engineering University o Arkansas ELEG 3143 Probability & Stochastic Process Ch. 4 Multiple Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Two discrete random variables
More informationChapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic
Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in
More information