STA 711: Probability & Measure Theory Robert L. Wolpert

Size: px
Start display at page:

Download "STA 711: Probability & Measure Theory Robert L. Wolpert"

Transcription

1 STA 711: Probability & Measure Theory Robert L. Wolpert 6 Independence 6.1 Independent Events A collection of events {A i } F in a probability space (Ω,F,P) is called independent if P[ i I A i ] = P[A i ] i I for each finite set I of indices. This is a stronger requirement than pairwise independence, the requirement merely that P[A i A j ] = P[A i ]P[A j ] for each i j. For a simple counter-example, toss two fair coins and let H n be the event Heads on the nth toss for n = 1,2. Then the three events A 1 := H 1, A 2 := H 2, and A 3 := H 1 H 2 (the event that the coins disagree) each have P[A i ] = 1/2 and each pair has P[A i A j ] = (1/2) 2 = 1/4 for i j, but A i = has probability zero and not (1/2) 3 = 1/ Independent Classes of Events Classes {C i } of events are called independent if P[ i I A i ] = P[A i ] i I for each finite I whenever each A i C i. An important tool for simplifying proofs of independence is Theorem 1 (Basic Criterion) If classes {C i } of events are independent and if each C i is a π-system, then {σ(c i )} are independent too. Proof. Let I be a nonempty finite index set and {C i } i I an independent collection of π-systems. Fix i I, set J = I\{i}, and fix A j C j for each j J. Set: { [ L := B F : P B ] } A j = P[B] P[A j ]. Then j J j J Ω L, by the independence of {C j } j J ; B L B c L, by a quick computation; and B n L and {B n } disjoint B n L, quick computation. 1

2 Thus L is a λ-system containing C i, and so by Dynkin s π-λ theorem it contains σ(c i ). Thus σ(c i ) and {A j } j J are independent for each {A j C j }, so {σ(c i ),{C j } j J } are independent π-systems. Repeating the same argument I 1 times (or, more elegantly, mathematical induction on the cardinality I ) completes the proof. 6.3 Independent Random Variables A collection of random variables {X i } on some probability space (Ω,F,P) are called independent if P ( i I [X i B i ] ) = P[X i B i ] i I for each finite set I of indices and each collection of Borel sets {B i B(R)}. This is just the same as the requirement that the σ-algebras F i := σ(x i ) = X i 1 (B) be independent; by the Basic Criterion it is enough to check that the joint CDFs factor, i.e., that P ( i I [X i x i ] ) = i I F i (x i ) for each finite index set I and each x R I, or just for a dense set of such x (Why?). For finitely-many jointly continuous random variables this is equivalent to requiring that the joint density function factor as the product of marginal density functions, while for finitelymany discrete random variables it s equivalent to the usual factorization criterion for the joint pmf. The present definition goes beyond those two cases, however for example, it includes the case of random variables X Bi(7,.3), Y Ex(2.), Z = ζ for ζ No(,1), and C with the Cantor distribution. It also applies to infinite (even uncountable) collections of random variables, where no joint pdf or pmf can exist. Since σ ( g(x) ) σ(x) for any random variable X and Borel function g( ), if {X i } are independent andifg i ( )arearbitraryborelfunctions, itfollowsthat{g i (X i )}areindependent too and, in particular, that if X Y then X g(y) for all Borel functions g( ). 6.4 Two Zero-One Laws First, a little toy example to motivate. Begin with a leather bag containing one gold coin, and n = 1. (a) At nth turn, first add one additional silver coin to the bag, then draw one coin at random. Let A n be the event A n = {Draw gold coin on nth draw}. Whichever coin you draw, replace it; increment n; and repeat. Page 2

3 (b) As above but at nth turn, add n silver coins. Let γ be the probability that you ever draw the gold coin. In each case, is γ =? γ = 1? or < γ < 1? In latter case, give exact asymptotic expression for γ and numerical estimate to four decimals. Why doesn t < γ < 1 violate Borel s zero-one law (Prop. 1 below)? Can you find γ exactly, perhaps with the help of Mathematica or Maple? The Borel-Cantelli Lemmas and Borel s Zero-One Law Our proof below of the Strong Law of Large Numbers for iid bounded random variables relies on the almost-trivial but very useful: Lemma 2 (Borel-Cantelli) Let {A n } be events on some probability space (Ω,F,P) that satisfy P[A n ] <. Then the event that infinitely-many of the {A n } occur (the limsup) has probability zero. Proof. [ P ] [ A m P A m ] P[A m ] as n. This result does not require independence of the {A n }, but its partial converse does: Lemma 3 (Second Borel-Cantelli) Let {A n } be independent events on a probability space (Ω,F,P) that satisfy P[A n ] =. Then the event that infinitely-many of the {A n } occur (the limsup) has probability one. Proof. First recall that 1+x e x for all real x R, positive or not (draw a graph). For each pair of integers 1 n N <, [ N P A c m ] = N ( 1 P[Am ] ) ( N e P[Am] = exp exp ( ) N P[A m ] ) P[A m ] = e = as N ; thus each A c m is a null set, hence so is their union, so Page 3

4 [ P ] [ A m = 1 P 1 ( P A c m ] A c m ) = 1 = 1. Together these two results comprise the Proposition 1 (Borel s Zero-One Law) For independent events {A n }, the event A := limsupa n has probability P(A) = or P(A) = 1, depending on whether the sum P(A n ) is finite or not Kolmogorov s Zero-One Law For any collection {X n } of random variables on a probability space (Ω,F,P), define two sequences of σ-algebras ( past and future ) by: F n := σ{x i : i n} T n := σ{x i : i n+1} and, from them, construct the π-system P and σ-algebra T by P := F n T := T n. In general P will not be a σ-algebra, because it will not be closed under countable unions or intersections, but it is a field and hence a π-system, and generates the σ-algebra F n := σ(p) F. The class T, called the tail σ-field, includes such events as X n converges or limsupx n 1 or, with S n := n 1 X j, 1 n S n Converges or 1 n S n. Theorem 4 (Kolmogorov s Zero-One Law) For independent random variables X n, the tail σ-field T is almost trivial in the sense that every event Λ T has probability P[Λ] = or P[Λ] = 1. Proof. Let A P = F n, and Λ T. Then for some n N, A F n and Λ T n, so A Λ; thus P and T are independent. Since P is a π-system, it follows from the Basic Criterionthatσ(P)andT arealsoindependent. ButF n P soeachx n isσ(p)-measurable, hence T σ(p) and each Λ T must also be in σ(p) T. It follows that: P[Λ] = P[Λ Λ] = P[Λ]P[Λ] = P[Λ] 2, so = P[Λ] ( 1 P[Λ] ) proving that P[Λ] must be zero or one. Page 4

5 6.5 Product Spaces Do independent random variables exist, with arbitrary (marginal) specified distributions? How can they be constructed? One way is to build product probability spaces; let s see how to do that. Let (Ω j,f j,p j ) be a probability space for j = 1,2 and set: Ω = Ω 1 Ω 2 := {(ω 1,ω 2 ) : ω j Ω j } F = F 1 F 2 := σ{a 1 A 2 : A j F j } P := P 1 P 2, the unique extension to F satisfying: P(A 1 A 2 ) = P 1 (A 1 ) P 2 (A 2 ) for A 1 F 1, A 2 F 2. Any random variables X 1 on (Ω 1,F 1,P 1 ) and X 2 on (Ω 2,F 2,P 2 ) can be extended to the common space (Ω,F,P) by defining X 1 (ω 1,ω 2 ) := X 1 (ω 1 ) and X 2 (ω 1,ω 2 ) := X 2 (ω 2 ); it s easy to show that {X j } are independent and have the same marginal distributions as {X j}. Thus, independent random variables do exist with arbitrary distributions. The same construction extends to countable families. 6.6 Fubini s Theorem We now consider how to evaluate probabilities and integrals on product spaces. For any F-measurable random variable X : Ω 1 Ω 2 S (S would be R, for real-valued RVs, but could also be R n or any complete separable metric space), and for any ω 2 Ω 2, the (second) section of X is the (Ω 1,F 1,P 1 ) random variables X ω2 : Ω 1 S defined by X ω2 (ω 1 ) := X(ω 1,ω 2 ). It is not quite obvious, but true, that X ω2 is F 1 -measurable show this first for indicator random variables X = 1 A1 A 2 of product sets, then extend by a π-λ argument to indicators X = 1 A of sets A F, then to simple RVs by linearity, then to the nonnegative RVs X + and X for an arbitrary F-measurable X by monotone limits. Similarly, the first section X ω1 ( ) := X(ω 1, ) is F 2 -measurable for each ω 1 Ω 1. Finally: Fubini s theorem gives conditions (namely, that either X or E X < ) to guarantee that these three integrals are meaningful and equal: { } { } X ω2 dp 1 dp? 2 = XdP =? X ω1 dp 2 dp 1 (1) Ω 2 Ω 1 Ω Ω 1 Ω 2 Toprovethis, firstnotethatit strueforindicatorsx = 1 A1 A 2 oftheπ-systemofmeasurable rectangles (A 1 A 2 ) with each A j F j ; then verify that the class C of events A F for which it holds for X = 1 A is a λ-system. By Dynkin s π-λ theorem it follows that F C so (1) holds for all indicators X = 1 A of events A F, hence for all nonnegative simple functions in E +, and finally for all F-measurable X by the MCT. For X L 1, apply this result separately to X + and X. Page 5

6 Fubini s theorem applies more generally. Each probability measure P j may be replaced by an arbitrary σ-finite measure; or one of them (say, P 2 ) may be replaced by a measurable kernel 1 K(ω 1, dω 2 ) that is a σ-finite measure K(ω 1, ) on F 2 in its second variable for each fixed ω 1, and an F 1 -measurable function K(, B 2 ) in its first variable for each fixed B 2 F 2. Now Fubini s Theorem asserts (under positivity or L 1 conditions) the equality of integrals of X wrt the measure P(dω 1 dω 2 ) = P 1 (dω 1 )K(ω 1,dω 2 ) to the iterated integrals { } ν X (dω 2 ) = XdP = X ω1 (ω 2 )K(ω 1,dω 2 ) P 1 (dω 1 ) Ω 2 Ω Ω 2 for the measure on F 2 given by ν X (dω 2 ) := Ω 1 X ω2 (ω 1 )K(ω 1,dω 2 )P 1 (dω 1 ). Ω 1 As an easy consequence (take Ω 1 = N with counting measure P 1 (B 1 ) := #{B 1 }, and let K(n,B 2 ) = P(X n B 2 ) be the distribution of X n on Ω 2 = R), for any sequence of random variables we may exchange summation and expectation and conclude { } E X n? = {EX n } whenever each X n or when E X n <, but otherwise equality may fail. For an example where interchanging integration order fails, integrate by parts to verify: 1 1 { 1 } y 2 x 2 (x 2 +y 2 ) dx dy = { 2 1 } y 2 x 2 (x 2 +y 2 ) dy dx = { } 1 1+y { 2 } +1 1+x 2 dy = π 4 dx = +π 4. As expected in light of Fubini s Theorem, the integrand isn t nonnegative nor is it in L 1 : 1,1 y 2 x 2 π/2 1, (x 2 +y 2 ) 2 dxdy r 2 sin 2 θ cos 2 θ rdrdθ r 4 ( ) π/2 ( 1 ) = sin 2 (2θ)dθ r 1 dr = (π/4)( ). 1 Measurable kernels come up all the time when studying conditional distributions (as you ll see in week 9 of this course) and, in particular, Markov chains and processes. Page 6

7 6.6.1 A Simple but Useful Consequence of Fubini For any p > and any random variable X, [ ] X E X p = E px p 1 dx It follows that = [ ] = E 1 { X >x} px p 1 dx [ ] E1{ X >x} px p 1 dx = px p 1 P [ X > x ] dx. X L p E X p < n p 1 P[ X > n] <. The case p = 1 is easiest and most important: if S := n= P[ X > n] <, then X L 1 with E X S E X +1. If X takes on only nonnegative integer values then EX = S. For any ǫ >, ǫ P[ X > nǫ] < E X ǫ P[ X > nǫ] 6.7 Hoeffding s Inequality If {X j } are independent and (individually) bounded, so ( j N) ( {a j,b j }) for which P[a j X j b j ] = 1, then ( c > ), S n := n j=1 X j satisfies n= P [ (S n ES n ) c ] / n exp ( 2c 2 b j a j ). 2 If X j are iid and bounded by X j 1, e.g., then P[( X n µ) ǫ] e nǫ2 /2. Wassily Hoeffding proved this improvement on Chebychev s inequality for L random variables in 1963 at UNC. It follows from Hoeffding s Lemma: E[e λ(x j µ j ) ] exp ( λ 2 (b j a j ) 2 /8 ), proved in turn from Jensen s ineq and Taylor s theorem (with remainder). The importance is that the bound decreases exponentially in n as n, while the Chebychev bound only decreases like a power of n. The price is that {X j } must be bounded in L, not merely in L 2. See also related and earlier Bernstein s inequality (1937), Chernoff bounds (1952), and Azuma s inequality (1967). 1 Page 7

8 Here s a proof for the important special case of X j = ±1 with probability 1/2 each (and hence µ = ): P[ X n ǫ] = P[S n nǫ] = P [ e λsn e nλǫ] for any λ > E e λsn e nλǫ = { 1 2 eλ + 1e λ} n 2 e nλǫ { } n e λ2 /2 e nλǫ = exp ( nλ 2 /2 nλǫ ) The exponent is minimized at λ = ǫ, so P[ X n ǫ] exp ( nǫ 2 /2 nǫǫ ) = e nǫ2 /2. by Markov s inequality by independence see footnote 2 The general case isn t much harder, but proving that Ee λx e λ2 /2 is a bit more delicate. By Borel/Cantelli it follows from Hoeffding s inequality that ( X n µ) > ǫ only finitely-many times for each ǫ >, if {X n } L are iid, leading to our first Strong Law of Large Numbers: P[ X n µ] = 1 (why does this follow?). Note that Chebychev s inequality only guarantees the algebraic bound P[ X n ǫ] 1/nǫ 2, instead of Hoeffding s exponential bound. Since 1/nǫ 2 isn t summable in n, Chebychev s bound isn t strong enough to prove a strong LLN, but Hoeffding s is. Hoeffding s inequality is now used commonly in computer science and machine learning, applied to indicators of Bernoulli events (or, equivalently, to binomial random variables), where it gives the bound P [ (p ǫ)n S n (p+ǫ)n ] = P [ X n p ǫ ] 1 2e 2ǫ2 n for S n = j n X j Bi(n,p) for iid Bernoulli X j iid Bi(1, p) variables, showing exponential concentration of probability around the mean. This is far stronger than Chebychev s bound of P [ X n p ǫ ] 1 p(1 p)/ǫ 2 n 2 cosh(λ) = { 1 2 eλ e λ} = λ 2k (2k)! (λ 2 ) k 2 k (k!) = eλ2 /2 Page 8 Last edited: September 28, 215

4 Expectation & the Lebesgue Theorems

4 Expectation & the Lebesgue Theorems STA 205: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on a probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω, does

More information

4 Expectation & the Lebesgue Theorems

4 Expectation & the Lebesgue Theorems STA 7: Probability & Measure Theory Robert L. Wolpert 4 Expectation & the Lebesgue Theorems Let X and {X n : n N} be random variables on the same probability space (Ω,F,P). If X n (ω) X(ω) for each ω Ω,

More information

Independent random variables

Independent random variables CHAPTER 2 Independent random variables 2.1. Product measures Definition 2.1. Let µ i be measures on (Ω i,f i ), 1 i n. Let F F 1... F n be the sigma algebra of subsets of Ω : Ω 1... Ω n generated by all

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

Probability and Measure

Probability and Measure Probability and Measure Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Convergence of Random Variables 1. Convergence Concepts 1.1. Convergence of Real

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Notes 1 : Measure-theoretic foundations I

Notes 1 : Measure-theoretic foundations I Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,

More information

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures 36-752 Spring 2014 Advanced Probability Overview Lecture Notes Set 1: Course Overview, σ-fields, and Measures Instructor: Jing Lei Associated reading: Sec 1.1-1.4 of Ash and Doléans-Dade; Sec 1.1 and A.1

More information

Sample Spaces, Random Variables

Sample Spaces, Random Variables Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted

More information

Probability and Measure

Probability and Measure Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability

More information

7 Convergence in R d and in Metric Spaces

7 Convergence in R d and in Metric Spaces STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a

More information

I. ANALYSIS; PROBABILITY

I. ANALYSIS; PROBABILITY ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so

More information

1 Sequences of events and their limits

1 Sequences of events and their limits O.H. Probability II (MATH 2647 M15 1 Sequences of events and their limits 1.1 Monotone sequences of events Sequences of events arise naturally when a probabilistic experiment is repeated many times. For

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction Chapter 1 Probability Theory: Introduction Basic Probability General In a probability space (Ω, Σ, P), the set Ω is the set of all possible outcomes of a probability experiment. Mathematically, Ω is just

More information

Lecture 11: Random Variables

Lecture 11: Random Variables EE5110: Probability Foundations for Electrical Engineers July-November 2015 Lecture 11: Random Variables Lecturer: Dr. Krishna Jagannathan Scribe: Sudharsan, Gopal, Arjun B, Debayani The study of random

More information

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017. Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 017 Nadia S. Larsen 17 November 017. 1. Construction of the product measure The purpose of these notes is to prove the main

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Midterm Examination. STA 205: Probability and Measure Theory. Wednesday, 2009 Mar 18, 2:50-4:05 pm

Midterm Examination. STA 205: Probability and Measure Theory. Wednesday, 2009 Mar 18, 2:50-4:05 pm Midterm Examination STA 205: Probability and Measure Theory Wednesday, 2009 Mar 18, 2:50-4:05 pm This is a closed-book examination. You may use a single sheet of prepared notes, if you wish, but you may

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

2 Construction & Extension of Measures

2 Construction & Extension of Measures STA 711: Probability & Measure Theory Robert L. Wolpert 2 Construction & Extension of Measures For any finite set Ω = {ω 1,...,ω n }, the power set PΩ is the collection of all subsets of Ω, including the

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Measure and integration

Measure and integration Chapter 5 Measure and integration In calculus you have learned how to calculate the size of different kinds of sets: the length of a curve, the area of a region or a surface, the volume or mass of a solid.

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

1. Probability Measure and Integration Theory in a Nutshell

1. Probability Measure and Integration Theory in a Nutshell 1. Probability Measure and Integration Theory in a Nutshell 1.1. Measurable Space and Measurable Functions Definition 1.1. A measurable space is a tuple (Ω, F) where Ω is a set and F a σ-algebra on Ω,

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5

More information

More on Distribution Function

More on Distribution Function More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution

More information

Probability: Handout

Probability: Handout Probability: Handout Klaus Pötzelberger Vienna University of Economics and Business Institute for Statistics and Mathematics E-mail: Klaus.Poetzelberger@wu.ac.at Contents 1 Axioms of Probability 3 1.1

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define HILBERT SPACES AND THE RADON-NIKODYM THEOREM STEVEN P. LALLEY 1. DEFINITIONS Definition 1. A real inner product space is a real vector space V together with a symmetric, bilinear, positive-definite mapping,

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

REAL ANALYSIS I Spring 2016 Product Measures

REAL ANALYSIS I Spring 2016 Product Measures REAL ANALSIS I Spring 216 Product Measures We assume that (, M, µ), (, N, ν) are σ- finite measure spaces. We want to provide the Cartesian product with a measure space structure in which all sets of the

More information

Lecture 4: Conditional expectation and independence

Lecture 4: Conditional expectation and independence Lecture 4: Conditional expectation and independence In elementry probability, conditional probability P(B A) is defined as P(B A) P(A B)/P(A) for events A and B with P(A) > 0. For two random variables,

More information

Lecture 6 Basic Probability

Lecture 6 Basic Probability Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES Contents 1. Continuous random variables 2. Examples 3. Expected values 4. Joint distributions

More information

Probability Theory. Richard F. Bass

Probability Theory. Richard F. Bass Probability Theory Richard F. Bass ii c Copyright 2014 Richard F. Bass Contents 1 Basic notions 1 1.1 A few definitions from measure theory............. 1 1.2 Definitions............................. 2

More information

Introduction and Preliminaries

Introduction and Preliminaries Chapter 1 Introduction and Preliminaries This chapter serves two purposes. The first purpose is to prepare the readers for the more systematic development in later chapters of methods of real analysis

More information

8 Laws of large numbers

8 Laws of large numbers 8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable

More information

1 Variance of a Random Variable

1 Variance of a Random Variable Indian Institute of Technology Bombay Department of Electrical Engineering Handout 14 EE 325 Probability and Random Processes Lecture Notes 9 August 28, 2014 1 Variance of a Random Variable The expectation

More information

The Essential Equivalence of Pairwise and Mutual Conditional Independence

The Essential Equivalence of Pairwise and Mutual Conditional Independence The Essential Equivalence of Pairwise and Mutual Conditional Independence Peter J. Hammond and Yeneng Sun Probability Theory and Related Fields, forthcoming Abstract For a large collection of random variables,

More information

MATH/STAT 235A Probability Theory Lecture Notes, Fall 2013

MATH/STAT 235A Probability Theory Lecture Notes, Fall 2013 MATH/STAT 235A Probability Theory Lecture Notes, Fall 2013 Dan Romik Department of Mathematics, UC Davis December 30, 2013 Contents Chapter 1: Introduction 6 1.1 What is probability theory?...........................

More information

Section 2: Classes of Sets

Section 2: Classes of Sets Section 2: Classes of Sets Notation: If A, B are subsets of X, then A \ B denotes the set difference, A \ B = {x A : x B}. A B denotes the symmetric difference. A B = (A \ B) (B \ A) = (A B) \ (A B). Remarks

More information

Learning Theory. Sridhar Mahadevan. University of Massachusetts. p. 1/38

Learning Theory. Sridhar Mahadevan. University of Massachusetts. p. 1/38 Learning Theory Sridhar Mahadevan mahadeva@cs.umass.edu University of Massachusetts p. 1/38 Topics Probability theory meet machine learning Concentration inequalities: Chebyshev, Chernoff, Hoeffding, and

More information

Estimates for probabilities of independent events and infinite series

Estimates for probabilities of independent events and infinite series Estimates for probabilities of independent events and infinite series Jürgen Grahl and Shahar evo September 9, 06 arxiv:609.0894v [math.pr] 8 Sep 06 Abstract This paper deals with finite or infinite sequences

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

18.175: Lecture 2 Extension theorems, random variables, distributions

18.175: Lecture 2 Extension theorems, random variables, distributions 18.175: Lecture 2 Extension theorems, random variables, distributions Scott Sheffield MIT Outline Extension theorems Characterizing measures on R d Random variables Outline Extension theorems Characterizing

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

STAT 7032 Probability. Wlodek Bryc

STAT 7032 Probability. Wlodek Bryc STAT 7032 Probability Wlodek Bryc Revised for Spring 2019 Printed: January 14, 2019 File: Grad-Prob-2019.TEX Department of Mathematical Sciences, University of Cincinnati, Cincinnati, OH 45221 E-mail address:

More information

Math 564 Homework 1. Solutions.

Math 564 Homework 1. Solutions. Math 564 Homework 1. Solutions. Problem 1. Prove Proposition 0.2.2. A guide to this problem: start with the open set S = (a, b), for example. First assume that a >, and show that the number a has the properties

More information

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events.

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events. 1 Probability 1.1 Probability spaces We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events. Definition 1.1.

More information

Probability Theory Review

Probability Theory Review Cogsci 118A: Natural Computation I Lecture 2 (01/07/10) Lecturer: Angela Yu Probability Theory Review Scribe: Joseph Schilz Lecture Summary 1. Set theory: terms and operators In this section, we provide

More information

STAT 712 MATHEMATICAL STATISTICS I

STAT 712 MATHEMATICAL STATISTICS I STAT 72 MATHEMATICAL STATISTICS I Fall 207 Lecture Notes Joshua M. Tebbs Department of Statistics University of South Carolina c by Joshua M. Tebbs TABLE OF CONTENTS Contents Probability Theory. Set Theory......................................2

More information

Building Infinite Processes from Finite-Dimensional Distributions

Building Infinite Processes from Finite-Dimensional Distributions Chapter 2 Building Infinite Processes from Finite-Dimensional Distributions Section 2.1 introduces the finite-dimensional distributions of a stochastic process, and shows how they determine its infinite-dimensional

More information

Module 1. Probability

Module 1. Probability Module 1 Probability 1. Introduction In our daily life we come across many processes whose nature cannot be predicted in advance. Such processes are referred to as random processes. The only way to derive

More information

Lecture 2: Review of Basic Probability Theory

Lecture 2: Review of Basic Probability Theory ECE 830 Fall 2010 Statistical Signal Processing instructor: R. Nowak, scribe: R. Nowak Lecture 2: Review of Basic Probability Theory Probabilistic models will be used throughout the course to represent

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS 1.1. The Rutherford-Chadwick-Ellis Experiment. About 90 years ago Ernest Rutherford and his collaborators at the Cavendish Laboratory in Cambridge conducted

More information

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section

More information

x log x, which is strictly convex, and use Jensen s Inequality:

x log x, which is strictly convex, and use Jensen s Inequality: 2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and

More information

Essentials on the Analysis of Randomized Algorithms

Essentials on the Analysis of Randomized Algorithms Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random

More information

Lecture 2: Random Variables and Expectation

Lecture 2: Random Variables and Expectation Econ 514: Probability and Statistics Lecture 2: Random Variables and Expectation Definition of function: Given sets X and Y, a function f with domain X and image Y is a rule that assigns to every x X one

More information

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y. CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Probability Theory for Machine Learning. Chris Cremer September 2015

Probability Theory for Machine Learning. Chris Cremer September 2015 Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares

More information

Lecture 21: Expectation of CRVs, Fatou s Lemma and DCT Integration of Continuous Random Variables

Lecture 21: Expectation of CRVs, Fatou s Lemma and DCT Integration of Continuous Random Variables EE50: Probability Foundations for Electrical Engineers July-November 205 Lecture 2: Expectation of CRVs, Fatou s Lemma and DCT Lecturer: Krishna Jagannathan Scribe: Jainam Doshi In the present lecture,

More information

1 Probability space and random variables

1 Probability space and random variables 1 Probability space and random variables As graduate level, we inevitably need to study probability based on measure theory. It obscures some intuitions in probability, but it also supplements our intuition,

More information

STOR 635 Notes (S13)

STOR 635 Notes (S13) STOR 635 Notes (S13) Jimmy Jin UNC-Chapel Hill Last updated: 1/14/14 Contents 1 Measure theory and probability basics 2 1.1 Algebras and measure.......................... 2 1.2 Integration................................

More information

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME ELIZABETH G. OMBRELLARO Abstract. This paper is expository in nature. It intuitively explains, using a geometrical and measure theory perspective, why

More information

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27 Probability Review Yutian Li Stanford University January 18, 2018 Yutian Li (Stanford University) Probability Review January 18, 2018 1 / 27 Outline 1 Elements of probability 2 Random variables 3 Multiple

More information

Lebesgue Integration on R n

Lebesgue Integration on R n Lebesgue Integration on R n The treatment here is based loosely on that of Jones, Lebesgue Integration on Euclidean Space We give an overview from the perspective of a user of the theory Riemann integration

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor) Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor) Matija Vidmar February 7, 2018 1 Dynkin and π-systems Some basic

More information

2 Probability, random elements, random sets

2 Probability, random elements, random sets Tel Aviv University, 2012 Measurability and continuity 25 2 Probability, random elements, random sets 2a Probability space, measure algebra........ 25 2b Standard models................... 30 2c Random

More information

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias Notes on Measure, Probability and Stochastic Processes João Lopes Dias Departamento de Matemática, ISEG, Universidade de Lisboa, Rua do Quelhas 6, 1200-781 Lisboa, Portugal E-mail address: jldias@iseg.ulisboa.pt

More information

DS-GA 1002 Lecture notes 2 Fall Random variables

DS-GA 1002 Lecture notes 2 Fall Random variables DS-GA 12 Lecture notes 2 Fall 216 1 Introduction Random variables Random variables are a fundamental tool in probabilistic modeling. They allow us to model numerical quantities that are uncertain: the

More information

A D VA N C E D P R O B A B I L - I T Y

A D VA N C E D P R O B A B I L - I T Y A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2

More information

36-752: Lecture 1. We will use measures to say how large sets are. First, we have to decide which sets we will measure.

36-752: Lecture 1. We will use measures to say how large sets are. First, we have to decide which sets we will measure. 0 0 0 -: Lecture How is this course different from your earlier probability courses? There are some problems that simply can t be handled with finite-dimensional sample spaces and random variables that

More information

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past. 1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if

More information

Probability. Lecture Notes. Adolfo J. Rumbos

Probability. Lecture Notes. Adolfo J. Rumbos Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

Joint Distribution of Two or More Random Variables

Joint Distribution of Two or More Random Variables Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few

More information

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019 Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial

More information

Measures. 1 Introduction. These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland.

Measures. 1 Introduction. These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland. Measures These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland. 1 Introduction Our motivation for studying measure theory is to lay a foundation

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik Stochastic Processes Winter Term 2016-2017 Paolo Di Tella Technische Universität Dresden Institut für Stochastik Contents 1 Preliminaries 5 1.1 Uniform integrability.............................. 5 1.2

More information

A PECULIAR COIN-TOSSING MODEL

A PECULIAR COIN-TOSSING MODEL A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or

More information