Lecture 4 : Conditional Probability and Bayes Theorem 0/ 26

Similar documents
LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Econ 325: Introduction to Empirical Economics

Topic 2 Probability. Basic probability Conditional probability and independence Bayes rule Basic reliability

STAT Chapter 3: Probability

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Introduction and basic definitions

CMPSCI 240: Reasoning about Uncertainty

Determining Probabilities. Product Rule for Ordered Pairs/k-Tuples:

Today we ll discuss ways to learn how to think about events that are influenced by chance.

Probability the chance that an uncertain event will occur (always between 0 and 1)

7.1 What is it and why should we care?

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

Statistical Methods for the Social Sciences, Autumn 2012

Independence. P(A) = P(B) = 3 6 = 1 2, and P(C) = 4 6 = 2 3.

the time it takes until a radioactive substance undergoes a decay

Lecture 1 : The Mathematical Theory of Probability

Chapter 7 Wednesday, May 26th

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Properties of Probability

Lecture 2. Conditional Probability

STAT:5100 (22S:193) Statistical Inference I

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Lecture 8: Conditional probability I: definition, independence, the tree method, sampling, chain rule for independent events

STAT 430/510 Probability

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Lecture 4. Selected material from: Ch. 6 Probability

Probability Year 9. Terminology

Notes Week 2 Chapter 3 Probability WEEK 2 page 1

Lecture 3 Probability Basics

Independence. CS 109 Lecture 5 April 6th, 2016

2. AXIOMATIC PROBABILITY

Dept. of Linguistics, Indiana University Fall 2015

Discrete Probability and State Estimation

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Outline. Probability. Math 143. Department of Mathematics and Statistics Calvin College. Spring 2010

Probability Year 10. Terminology

1 Preliminaries Sample Space and Events Interpretation of Probability... 13

Lecture 3 - Axioms of Probability

Probability (Devore Chapter Two)

Lecture 1: Probability Fundamentals

STA Module 4 Probability Concepts. Rev.F08 1

Probability & Random Variables

CMPSCI 240: Reasoning about Uncertainty

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2

Slide 1 Math 1520, Lecture 21

STAT 516: Basic Probability and its Applications

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

Independence Solutions STAT-UB.0103 Statistics for Business Control and Regression Models

P (E) = P (A 1 )P (A 2 )... P (A n ).

3.2 Probability Rules

Origins of Probability Theory

Outline Conditional Probability The Law of Total Probability and Bayes Theorem Independent Events. Week 4 Classical Probability, Part II

ECE353: Probability and Random Processes. Lecture 2 - Set Theory

Announcements. Topics: To Do:

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Probabilistic models

Statistics for Business and Economics

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

Intermediate Math Circles November 8, 2017 Probability II

Fundamentals of Probability CE 311S

Discrete Probability and State Estimation

Statistics 1L03 - Midterm #2 Review

Mean, Median and Mode. Lecture 3 - Axioms of Probability. Where do they come from? Graphically. We start with a set of 21 numbers, Sta102 / BME102

With Question/Answer Animations. Chapter 7

Review of Basic Probability Theory

2.4. Conditional Probability

Lecture 1. ABC of Probability

STAT 285 Fall Assignment 1 Solutions

3 PROBABILITY TOPICS

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Relative Risks (RR) and Odds Ratios (OR) 20

MA : Introductory Probability

Chapter 7: Section 7-1 Probability Theory and Counting Principles

2 Chapter 2: Conditional Probability

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Probability Theory Review

Conditional probability

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

STAT 111 Recitation 1

The set of all outcomes or sample points is called the SAMPLE SPACE of the experiment.

Probability and Statistics Notes

Discrete Probability

EE 178 Lecture Notes 0 Course Introduction. About EE178. About Probability. Course Goals. Course Topics. Lecture Notes EE 178

Probability (Devore Chapter Two)

Probability 1 (MATH 11300) lecture slides

CS626 Data Analysis and Simulation

UNIT 5 ~ Probability: What Are the Chances? 1

Lecture 3: Probability

Math 243 Section 3.1 Introduction to Probability Lab

F71SM STATISTICAL METHODS

6.080 / Great Ideas in Theoretical Computer Science Spring 2008

6.041/6.431 Spring 2009 Quiz 1 Wednesday, March 11, 7:30-9:30 PM. SOLUTIONS

Probability Notes (A) , Fall 2010

Term Definition Example Random Phenomena

Conditional Probability

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Chapter 6: Probability The Study of Randomness

Transcription:

0/ 26

The conditional sample space Motivating examples 1. Roll a fair die once 1 2 3 S = 4 5 6 Let A = 6 appears B = an even number appears So P(A) = 1 6 P(B) = 1 2 1/ 26

Now what about P ( 6 appears given an even number appears ) Philosophical Remark (Ignore this remark unless you intend to be a scientist) At present the above probability does not have a formal mathematical definition but we can still compute it. Soon we will give the formal definition and our computation will be justified. This is the mysterious way mathematics works. Somehow there is a deeper reality underlying the formal theory. Back to Stat 400 The above probability will be written written P(A B) to he read P(A given B). 2/ 26

Now we know an even number occurred so the sample space changes 1 2 3 4 5 6 the conditional sample space So there are only 3 possible outcomes given an even number occurred so P(6 given an even number occurred) = 1 3 The new sample space is called the conditional sample space. 3/ 26

2. Very Important example Suppose you deal two cards (in the usual way without replacement). What is P( ) i.e., P(two hearts in a row). Well, P(first heart) = 13 52. Now what about the second heart? Many of you will come up with 12/51 and P( ) = (13/52)(12/51) 4/ 26

There are TWO theoretical points hidden in the formula. Let s first look at P( } on {{ 2 nd } ) = 12/51 this isn t really correct What we really computed was the conditional probability P( on 2 nd deal on first deal) = 12/51 Why? Given we got a heart on the first deal the conditional sample space is the new deck with 51 cards and 12 hearts so we get P( on 2 nd on 1 st ) = 12/51 5/ 26

The second theoretical point we used was that in the following formula we multiplied the two probablities P( on 1 st ) and P( on 2 nd on 1 st ) together. This is a special case of a the formula for the probability of the intersection of two events that we will state below. P( ) = P( on 1 st )P( on 2 nd on 1 st ) ( )( ) 13 12 = 52 51 The general formula, the multiplicative formula, we will give as a definition shortly is P(A B) = P(A)PB A) Compare this to the additive formula which we already proved P(A B) = P(A) + P(B) P(A B) 6/ 26

Three Basic Questions These three examples will occur repeatedly in today s lecture. The first is the example we just discussed, the second is the reverse of what we just discussed and the third is a tricky variant of finding the probability of a heart on the first with no other information. 1 What is 2 What is P( on 2 nd on 1 } {{ } st ) reverse of pg. 5 P( on 1 st on 2 } {{ nd ) } reverse of pg. 5 3 What is P( on 2 nd with no information on what happened on the 1 st ). 7/ 26

The Formal Mathematical Theory of Conditional Probability (S) = n, (A) = a, (B) = b, (A B) = C Problem Let S be a finite set with the equally - likely probability measure and A and B be events with coordinalities shown in the picture. 8/ 26

Problem Compute P(A B). We are given B occurs so the conditional sample space is B Only part of A is allowed since we know B occurred namely the part A B. So counting elements we get P(A B) = (A B) (B) = c b 9/ 26

We can rewrite this as P(A B) = c b = c /n b/n = P(A B) P(B) so P(A B) = P(A B) P(B) This formula for the equally likely probability measure leads to the following. (*) 10/ 26

Formal Mathematical Definition Let A and B be any two events in a sample space S with P(B) 0. The conditional probability of A given B is written P(A B) and is defined by P(A B) = P(A B) P(B) (*) so, reversing the roles of A and B (so we get the formula that is in the text) if P(A) 0 then P(B A) P(A B) P(B A) = = P(A) P(A) Since A B B A. (**) 11/ 26

We won t prove the next theorem but you could do it and it is useful. Theorem Fix B with P(B) 0. P( B), ()so P(A B) as a function of A), satisfies the axioms (and theorems) of a probability measure - see Lecture 1. For example 1 P(A 1 A 2 B) = P(A 1 B) + P(A 2 B) P(A 1 A 2 B) 2 P(A B) = 1 P(A B) P(A ) (so P(A B) as a function of B) does not satisfy the axioms and theorems. 12/ 26

The Multiplicative Rule for P(A B) Rewrite (**) as P(A B) = P(A)P(B A)( ) ( ) is very important, more important then (**). It complement the formula P(A B) = P(A) + P(B) P(A B) Now we know how P interacts with the basic binary operations and. 13/ 26

More generally P(A B C) = P(A)P(B A)P(C A B) Exercise Write down P(A B C D). Traditional Example An urn contains 5 white chips, 4 black chips and 3 red chips. Four chips are drawn sequentially without replacement. Find P(WRWB). 14/ 26

12 chips P(WRWB) = ( ) ( ) ( ) ( ) 5 3 4 4 12 11 10 9 What did we do formally? The answer is we used the following formula for the intersection of four events P(WRWB) = P(W) P(R W) P(W W R) P(B W R W) 15/ 26

Now we make a computation that reverses the usual order, namely, we compute P( on first on second) By Definition so Now we know from pg. 5. Now we need P(A B) = P(A B) P(B) P( ) P( on first on second) = ( ) on 2 nd with no P other information P( ) = (13/52)(12/51) P( on 2 nd with no other information) = 13/52 16/ 26

By Definition We will prove this later (to some people this is intuitively clear). In fact if you write down any probability statement in this situation, take that statement and everywhere you see first write second and everywhere you see second write first then the resulting event will have the same probability as the event we started with. So back to our problem we have P( on 1 st on 2 nd ) = ( 13/52)( ( 13/52) = P( on 2 nd on 1 } {{ } st ) pg. 5 12/51) = 12 51 This is another instance of the symmetry (in first and second ) stated three lines above. 17/ 26

Bayes Theorem (pg. 72) Bayes Theorem is a truly remarkable theorem. It tells you how to compute P(A B) if you know P(B A) and a few other things. For example - we will get a new way to compute are favorite probability P( as 1 st on 2 nd ) because we know P( on 2 nd on 1 st ). First we will need on preliminary result. 18/ 26

The Law of Total Probability Let A 1, A 2,..., A k be mutually exclusive (A i A j = ) and exhaustive. (A 1 A 2... A k = S = the whole space) Then for any event B P(B) = P(B A 1 )P(A 1 ) + P(B A 2 )P(A 2 ) + + P(B A k )P(A k ) (b) Prove this First prove P(B S) = 1 then use the P(B, C) is satisfies the additivity rule for a probability measure as function of C. Special case k = 2 so we have A and A P(B) = P(B A)P(A) + P(B A )P(A ) (bb) 19/ 26

Now we can prove P( on 2 nd with no other information) = 13/52 Put B = on 2 nd A = heart on 1 st A = a nonheart on 1 st Lets write for nonheart. So, P( on 1 st ) = 39/52 P( on 2 nd / on first) = 13/51 20/ 26

Now P(B) = P(B A)P(A) + P(B A )P(A ) = P( on 2 nd on 1 st )P( on 1 st ) + P( on 2 nd on 1 st )P( on 1 st ) = (12/51)(13/52) + (13/51)(39/52) add fractions = (12)(13) + (13)(39) (51)(52) factor out 13 add to get 51 = (13) (12 + 39) (51)(52) = (13)/(52) Done! = (13)( 51) (51)(52) 21/ 26

Now we can state Bayes Theorem. Bayes Theorem (pg. 73) Let A 1, A 2,..., A k be a collection of n mutually exclusive and exhaustive events with P(A i ) > 0 i = 1, 2,..., k. Then for any event B with P(B) > 0 Again we won t give the proof. P(A j B) = P(B A j)p(a j ) k P(B A i )P(A i ) i=1 22/ 26

Special Case k = 2 Suppose we have two events A and B with P(A) > 0, P(A ) > 0 and P(B > 0). Then P(B A)P(A) )P(A B) = P(B A)P(A) + P(B A )P(A ) Now we will compute (for the last time) Using Bayes Theorem. This is the obvious way to P( on 1 st on 2 nd ) 23/ 26

do it since we know the probability the other way around So let s do it. plugging into ( ) we get P( on 1 st on 2 nd ) = P( on 2 nd on 1 st ) = 12/51 We put A = on 1 st so A = on 1 st and B = on second P( on 2 nd on 1 st )P( on 1 st ) P( on 2 nd on 1 st )P( on 1 st ) + P( on 2 nd on 1 st )P( on 1 st ) 24/ 26

swap numbers use something from high school factor this out 25/ 26

The algebra was hard but the approach was the most natural - a special case of General Principle Compulsory Reading (for your own heath) In case you or someone you love tests positive for a rare (this is the point) disease, read Example 2.31, pg. 81. Misleading (and even bad) statistics is rampant in medicine. 26/ 26