MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

Similar documents
Lecture 9: Conditional Probability and Independence

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Statistical Inference

Statistics 1 - Lecture Notes Chapter 1

2. AXIOMATIC PROBABILITY

Probability Review I

Lecture 1: Overview of percolation and foundational results from probability theory 30th July, 2nd August and 6th August 2007

the time it takes until a radioactive substance undergoes a decay

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Notes 1 : Measure-theoretic foundations I

Lecture 7. Sums of random variables

1 Sequences of events and their limits

Statistical Theory 1

CMPSCI 240: Reasoning about Uncertainty

Chapter 2 Classical Probability Theories

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Origins of Probability Theory

MA : Introductory Probability

A PECULIAR COIN-TOSSING MODEL

Lecture 11: Random Variables

Lecture 1. ABC of Probability

n if n is even. f (n)=

Compatible probability measures

MATH31011/MATH41011/MATH61011: FOURIER ANALYSIS AND LEBESGUE INTEGRATION. Chapter 2: Countability and Cantor Sets

Probability Theory Review

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Dept. of Linguistics, Indiana University Fall 2015

Deep Learning for Computer Vision

Lecture 7. Bayes formula and independence

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Lecture 16. Lectures 1-15 Review

18.175: Lecture 2 Extension theorems, random variables, distributions

Econ 325: Introduction to Empirical Economics

Topic 5 Basics of Probability

Stat 451: Solutions to Assignment #1

ECE 302: Chapter 02 Probability Model

Module 1. Probability

Basic Measure and Integration Theory. Michael L. Carroll

Measure and integration

STAT:5100 (22S:193) Statistical Inference I

CMPSCI 240: Reasoning about Uncertainty

Do stochastic processes exist?

Axiomatic Foundations of Probability. Definition: Probability Function

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales

Lecture 2. Conditional Probability

Sample Spaces, Random Variables

Single Maths B: Introduction to Probability

PROBABILITY THEORY 1. Basics

Recap of Basic Probability Theory

MATH/STAT 235A Probability Theory Lecture Notes, Fall 2013

Chapter 5 Random vectors, Joint distributions. Lectures 18-23

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Recap of Basic Probability Theory

PROBABILITY VITTORIA SILVESTRI

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

MAKING MONEY FROM FAIR GAMES: EXAMINING THE BOREL-CANTELLI LEMMA

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events.

Problem set 1, Real Analysis I, Spring, 2015.

Recitation 2: Probability

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

EE 178 Lecture Notes 0 Course Introduction. About EE178. About Probability. Course Goals. Course Topics. Lecture Notes EE 178

Math212a1411 Lebesgue measure.

Massachusetts Institute of Technology

Section 21. The Metric Topology (Continued)

ECE353: Probability and Random Processes. Lecture 2 - Set Theory

MAT 271E Probability and Statistics

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Appendix A : Introduction to Probability and stochastic processes

KOLMOGOROV S ZERO-ONE LAW WITH APPLICATIONS

Probability, Random Processes and Inference

Conditional Probability and Bayes

Lecture 6 Feb 5, The Lebesgue integral continued

STAT 7032 Probability Spring Wlodek Bryc

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Introduction to Stochastic Processes

SDS 321: Introduction to Probability and Statistics

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction

FE 5204 Stochastic Differential Equations

Probabilistic models

Measure Theoretic Probability. P.J.C. Spreij

Lecture 2: Convergence of Random Variables

Lecture 1 : The Mathematical Theory of Probability

Vina Nguyen HSSP July 6, 2008

Statistics for Financial Engineering Session 2: Basic Set Theory March 19 th, 2006

Chapter 1: Introduction to Probability Theory

Lecture 1: An introduction to probability theory

Chapter 1 Probability Theory

Section 2: Classes of Sets

5 Measure theory II. (or. lim. Prove the proposition. 5. For fixed F A and φ M define the restriction of φ on F by writing.

Announcements. Topics: To Do:

Conditional Probability

Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya

MTH 202 : Probability and Statistics

Lecture 10. Variance and standard deviation

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

Measures. 1 Introduction. These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland.

Transcription:

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5 and Problem 48 (or problem 43, in the 1st edition), available at http: //athenasc.com/prob-2nd-ch1.pdf. Solutions to the end of chapter problems are available at http://athenasc.com/prob-solved 2ndedition. pdf. These lecture notes provide some additional details and twists. Contents 1. Conditional probability 2. Independence 3. The Borel-Cantelli lemma 1 CONDITIONAL PROBABILITY Definition 1. Consider a probability space (Ω, F, P), and an event B F with P(B) > 0. For every event A F, the conditional probability that A occurs given that B occurs is denoted by P(A B) and is defined by P(A B) P(A B) =. P(B) 1

Theorem 1. Let (Ω, F, P) be a probability space. (a) If B is an event with P(B) > 0, then P(Ω B) = 1, and for any sequence {A i } of disjoint events, we have ( ) P A i B = P(A i B). (b) Suppose that B is an event with P(B) > 0. For every A F, define P B (A) = P(A B). Then, P B is a probability measure on (Ω, F). (c) Let A be an event. If the events B i, i N, form a partition of Ω, and P(B i ) > 0 for every i, then n P(A) = P(A B i )P(B i ). In particular, if B is an event with P(B) > 0 and P(B c ) > 0, then P(A) = P(A B)P(B) + P(A B c )P(B c ). (d) (Bayes rule) Let A be an event with P(A) > 0. If the events B i, i N, form a partition of Ω, and P(B i ) > 0 for every i, then P(B i A) = P(B i) P( A B i ) P(Bi) P( A B i ) = P( A) P( B ) P( A B j ). (e) For any sequence {A i } of events, we have j=1 P( A i ) = P(A 1 ) P(A i A 1 A i 1 ), i=2 as long as all conditional probabilities are well defined. j Proof. (a) We have P(Ω B) = P(Ω B)/P(B) = P(B)/P(B) = 1. Also ( ) P( P ( A i B ) = P B ( A i) = (B A i)). P(B) P(B) 2

Since the sets B A i, i N are disjoint, countable additivity, applied to the right-hand side, yields as claimed. ( ) P(B A i ) P A i B = = P(A i B), P(B) (b) This is immediate from part (a). (c) We have ( ) ( ) P(A) = P(A Ω) = P A ( B i ) = P (A B i ) = P(A B i ) = P(A B i )P(B i ). In the second equality, we used the fact that the sets B i form a partition of Ω. In the next to last equality, we used the fact that the sets B i are disjoint and countable additivity. (d) This follows from the fact P ( i n ) P(A 1 A 2 ) P(A 1 A 2 A 3 ) P(A 1 A n ) =1 A i = P(A1 ) P(A 1 ) P(A 1 A 2 ) P(A1 A n 1 ) n = P(A 1 ) P(A i A 1 A i 1 ). P(B i A) = P(B i A)/P(A) = P(A B i )P(B i )/P(A), and the result from part (c). (e) Note that the sequence of events n A i is decreasing and converges ( to ) A i. By the continuity ( property ) of probability measures, we have P = lim n P n A i. Note that A i i=2 Taking the limit, as n, we obtain the claimed result. 2 INDEPENDENCE Intuitively we call two events A, B independent if the occurrence or nonoccurrence of one does not affect the probability assigned to the other. The following definition formalizes and generalizes the notion of independence. 3

Definition 2. Let (Ω, F.P) be a probability space. (a) Two events, A and B, are said to be independent if P(A B) = P(A)P(B). If P(B) > 0, an equivalent condition is P(A) = P(A B). (b) Let S be an index set (possibly infinite, or even uncountable), and let {A s s S} be a family (set) of events. The events in this family are said to be independent if for every finite subset S 0 of S, we have ( ) P s S0 A s = P(A s ). s S 0 (c) Let F 1 F and F 2 F be two σ-fields. We say that F 1 and F 2 are independent if any two events A 1 F 1 and A 2 F 2 are independent. (d) More generally, let S be an index set, and for every s S, let F s be a σ field contained in F. We say that the σ-fields F s are independent if the following holds. If we pick one event A s from each F s, the events in the resulting family {A s s S} are independent. Example. Consider an infinite sequence of fair coin tosses, under the model constructed in the Lecture 2 notes. The following statements are intuitively obvious (although a formal proof would require a few steps). (a) Let A i be the event that the ith toss resulted in a 1. If i j, the events A i and A j are independent. (b) The events in the (infinite) family {A i i N} are independent. This statement captures the intuitive idea of independent coin tosses. (c) Let F 1 (respectively, F 2 ) be the collection of all events whose occurrence can be decided by looking at the results of the coin tosses at odd (respectively, even) times n. More formally, Let H i be the event that the ith toss resulted in a 1. Let C be the collection of events C = {H i i is odd}, and finally let F 1 = σ(c), so that F 1 is the smallest σ-field that contains all the events H i, for even i. We define F 2 similarly, using even times instead of odd times. Then, the two σ-fields F 1 and F 2 turn out to be independent. This statement captures the intuitive idea that knowing the results of the tosses at odd times provides no information on the results of the tosses at even times. (d) Let F n be the collection of all events whose occurrence can be decided by looking at the results of tosses 2n and 2n + 1. (Note that each F n is a σ-field comprised of finitely many events.) Then, the families F n, n N, are independent. Remark: How can one establish that two σ-fields (e.g., as in the above coin 4

tossing example) are independent? It turns out that one only needs to check independence for smaller collections of sets; see the theorem below (the proof is omitted and can be found in p. 39 of [W]). Theorem: If G 1 and G 2 are two collections of measurable sets, that are closed under intersection (that is, if A, B G i, then A B G i ), if F i = σ(g i ), i = 1, 2, and if P(A B) = P(A)P(B) for every A G 1, B G 2, then F 1 and F 2 are independent. 3 THE BOREL-CANTELLI LEMMA The Borel-Cantelli lemma is a tool that is often used to establish that a certain event has probability zero or one. Given a sequence of events A n, n N, recall that {A n i.o.} (read as A n occurs infinitely often ) is the event consisting of all ω Ω that belong to infinitely many A n, and that {A n i.o.} = lim sup A n = A i. n n=1 Theorem 2. (Borel-Cantelli lemma) Let {A n } be a sequence of events and let A = {A n i.o.}. (a) If n=1 P(A n ) <, then P(A) = 0. (b) If n=1 P(A n ) = and the events A n, n N, are independent, then P(A) = 1. Remark: The result in part (b) is not true without the independence assumption. Indeed, consider an arbitrary event C such that 0 < P(C) < 1 and let A n = C for all n. Then P({A n i.o.}) = P(C) < 1, even though P(A n ) =. The following lemma is useful here and in may other contexts. Lemma 1. Suppose that 0 p i 1 for every i N, and that p i =. Then, (1 p i ) = 0. Proof. Note that log(1 x) is a concave function of its argument, and its derivative at x = 0 is 1. It follows that log(1 x) x, for x [0, 1]. We then n 5

have ( n ) log (1 p i ) = log lim (1 p i ) n k log (1 p i ) k = log(1 p i ) k ( p i ). This is true for every k. By taking the limit as k, we obtain log (1 p i ) =, and (1 p i ) = 0. Proof of Theorem 2. (a) The assumption P(A n ) < implies that lim n=1 n P(A i ) = 0. Note that for every n, we have A A i. Then, the union bound implies that ( ) P(A) P A i P(A n ). We take the limit of both sides as n. Since the right-hand side converges to zero, P(A) must be equal to zero. (b) Let B n = A i, and note that A = n =1 B n. We claim that P(B n c ) = 0. This will imply the desired result because ( ) P(A c ) = P c c n=1 B n P(B n ) = 0. m m ( ) m P( A c P(A c ) = ) = 1 P(A i ). The assumption P(A i ) = implies that P(A i ) =. Using Lemma 1, with the sequence {P(A i ) i n} replacing the sequence {p i }, we obtain m ( ) c P(B n ) = P( A c i ) = lim P( m A c i ) = lim 1 P(A i ) = 0, n=1 Let us fix some n and some m n. We have, using independence, i m i m 6

where the second equality made use of the continuity property of probability measures. 7

MIT OpenCourseWare http://ocw.mit.edu 6.436J / 15.085J Fundamentals of Probability Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.