Lecture 2: Probability, conditional probability, and independence

Similar documents
Axiomatic Foundations of Probability. Definition: Probability Function

STAT:5100 (22S:193) Statistical Inference I

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Prof. Thistleton MAT 505 Introduction to Probability Lecture 5

BASICS OF PROBABILITY CHAPTER-1 CS6015-LINEAR ALGEBRA AND RANDOM PROCESSES

Statistical Theory 1

HW MATH425/525 Lecture Notes 1

Conditional Probability and Bayes

STAT 31 Practice Midterm 2 Fall, 2005

Lecture 3: Probability

Discrete Probability. Chemistry & Physics. Medicine

Probability Notes (A) , Fall 2010

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Probability and Statistics Notes

MA : Introductory Probability

Outline Conditional Probability The Law of Total Probability and Bayes Theorem Independent Events. Week 4 Classical Probability, Part II

Basic Statistics and Probability Chapter 3: Probability

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Econ 325: Introduction to Empirical Economics

Lecture 1. ABC of Probability

2. AXIOMATIC PROBABILITY

4/17/2012. NE ( ) # of ways an event can happen NS ( ) # of events in the sample space

Conditional Probability. CS231 Dianna Xu

1 INFO Sep 05

Chapter 3 Conditional Probability and Independence. Wen-Guey Tzeng Computer Science Department National Chiao Tung University

Dynamic Programming Lecture #4

Independence. P(A) = P(B) = 3 6 = 1 2, and P(C) = 4 6 = 2 3.

Announcements. Topics: To Do:

1 Preliminaries Sample Space and Events Interpretation of Probability... 13

Example: Suppose we toss a quarter and observe whether it falls heads or tails, recording the result as 1 for heads and 0 for tails.

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

MAT2377. Ali Karimnezhad. Version September 9, Ali Karimnezhad

Probability, Random Processes and Inference

lecture notes October 22, Probability Theory

Probability Review I

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Probability- describes the pattern of chance outcomes

Probabilistic models

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Introduction and basic definitions

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

STAT 430/510 Probability

Mutually Exclusive Events

Permutation. Permutation. Permutation. Permutation. Permutation

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Executive Assessment. Executive Assessment Math Review. Section 1.0, Arithmetic, includes the following topics:

Conditional Probability

STAT509: Probability

Probability Theory and Applications

Chapter 5 : Probability. Exercise Sheet. SHilal. 1 P a g e

3/15/2010 ENGR 200. Counting

Statistics for Engineers Lecture 2

CIVL 7012/8012. Basic Laws and Axioms of Probability

CMPT 882 Machine Learning

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

Statistical Inference

18.600: Lecture 7 Bayes formula and independence

CS626 Data Analysis and Simulation

1 INFO 2950, 2 4 Feb 10

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

HW2 Solutions, for MATH441, STAT461, STAT561, due September 9th

Chapter 7 Wednesday, May 26th

STAT 201 Chapter 5. Probability

Lecture 7. Bayes formula and independence

Combinatorics and probability

Math 511 Exam #1. Show All Work No Calculators

Introduction to Probability with MATLAB Spring 2014

CMPSCI 240: Reasoning about Uncertainty

Lecture 3 : Probability II. Jonathan Marchini

I - Probability. What is Probability? the chance of an event occuring. 1classical probability. 2empirical probability. 3subjective probability

Lecture 3 - Axioms of Probability

ECE 302: Chapter 02 Probability Model

Chapter 2 Class Notes

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r).

Lecture 16. Lectures 1-15 Review

18.600: Lecture 7 Bayes formula and independence

F71SM STATISTICAL METHODS

Discrete Probability

CLASS NOTES. MATH 334 Introduction to Probability

Dept. of Linguistics, Indiana University Fall 2015

Basics of probability

Chapter 2 Probability

STAT 111 Recitation 1

Lecture Lecture 5

Conditional Probability and Bayes Theorem (2.4) Independence (2.5)

Independence Solutions STAT-UB.0103 Statistics for Business Control and Regression Models

Lecture 4: Probability and Discrete Random Variables

Chapter 2: Probability Part 1

UNIT Explain about the partition of a sampling space theorem?

the time it takes until a radioactive substance undergoes a decay

STAT 516 Answers Homework 2 January 23, 2008 Solutions by Mark Daniel Ward PROBLEMS. = {(a 1, a 2,...) : a i < 6 for all i}

More on conditioning and Mr. Bayes

Determining Probabilities. Product Rule for Ordered Pairs/k-Tuples:

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

Statistics for Business and Economics

Chapter 3 : Conditional Probability and Independence

Lecture 1. Chapter 1. (Part I) Material Covered in This Lecture: Chapter 1, Chapter 2 ( ). 1. What is Statistics?

Discrete Probability Models

Introduction to Probability Theory

Course 003: Basic Econometrics

Transcription:

Lecture 2: Probability, conditional probability, and independence Theorem 1.2.6. Let S = {s 1,s 2,...} and F be all subsets of S. Let p 1,p 2,... be nonnegative numbers that sum to 1. The following defines a probability on F : P(A) = p i, i:s i A A F (sum over the empty set is defined to be 0). Theorem 1.2.8-9. If P is a probability and A and B are events, then a. P(A) 1; b. P(A c ) = 1 P(A); c. P(B A c ) = P(B) P(A B); d. P(A B) = P(A) + P(B) P(A B); e. If A B, then P(A) P(B). UW-Madison (Statistics) Stat 609 Lecture 2 2015 1 / 17

Birth date problem There are n balls in N boxes, n N. What is the probability that there are n boxes having exactly one ball? P = N(N 1) (N n + 1) N! N n = N n (N n)! When there are n persons in a party, what is the probability p n that there are at least two persons having the same birth date? n persons = n balls N = 365 days P(no one has the same birth date with others) = p n = 1 365! 365 n (365 n)! p 20 = 0.4058, p 30 = 0.6963, p 50 = 0.9651, p 60 = 0.9922 Reading assignments: Sections 1.2.3 and 1.2.4 365! 365 n (365 n)! UW-Madison (Statistics) Stat 609 Lecture 2 2015 2 / 17

Union of three events P(A B C) = P(A) + P(B) + P(C) P(A B) P(A C) P(B C) + P(A B C) Example: P(A) = P(B) = P(C) = 1/4, A and B are disjoint, P(B C) = P(A C) = 1/6 P(A B C) =? P(A B C) = 1/4 + 1/4 + 1/4 0 1/6 1/6 + 0 = 7/12 General addition formula For events A 1,...,A n, ( ) n n P A i = P(A i ) P(A i A j ) i<j + P(A i A j A k ) + + ( 1) n 1 P(A 1 A 2 A n ) i<j<k UW-Madison (Statistics) Stat 609 Lecture 2 2015 3 / 17

Theorem 1.2.11. If P is a probability, A, A 1,A 2,..., C 1,C 2,... are events, and C i C j = /0 for any i j and C i = S, then a. P(A) = P(A C i); b. P( A i) P(A i). Property a is useful when P(A C i ) is easy to calculate. Property b is called Boole s inequality. Bonferroni s inequality For n events A 1,..,A n, ( ) ( n n 1 P A i = P Then A c i ) n ( ) n n P A i P(A i ) (n 1) which is called Bonferroni s inequality. P(A c n i ) = n P(A i ) UW-Madison (Statistics) Stat 609 Lecture 2 2015 4 / 17

Conditional probability Sometimes we want to know the probability of event A given that another event B has occurred. Definition 1.3.2. If A and B are events with P(B) > 0, then the conditional probability of A given B is P(A B) P(A B) =. P(B) UW-Madison (Statistics) Stat 609 Lecture 2 2015 5 / 17

The conditional probability can be viewed as: The sample space is reduced to B. Then A occurs iff A B occurs. To have a probability, we need to divide P(A B) by P(B). If A B = /0, then P(A B) = 0 and P(A B) = 0. For convenience, we define P(A B) = 0 when P(B) = 0. P(A B) = P(A B)P(B) = P(B A)P(A). P( B) is a probability: P(/0 B) = 0, P(S B) = P(B B) = 1; P(A c B) = 1 P(A B); P(A C B) = P(A B) + P(C B) P(A C B). Example 1.3.4 (Three prisoners) Three prisoners, A, B, and C, are on death row. The warden knows that one of the three prisoners will be pardoned. Prisoner A tries to get the warden to tell him who had been pardoned, but the warden refuses. Prisoner A then asks which of B or C will be executed. The warden thinks for a while, then tells A that B is to be executed. UW-Madison (Statistics) Stat 609 Lecture 2 2015 6 / 17

Prisoner A s thinking: Let A, B, and C denote the events that A, B, and C is pardoned, respectively. Then P(A) = P(B) = P(C) = 1/3. Given that B will be executed, then either A or C will be pardoned; my chance of being pardoned has risen from 0.33 to 0.5. Prisoner A s calculation: P(A B c ) = 1 2. Warden s reasoning: Each prisoner has a 1/3 chance of being pardoned; either B or C must be executed, so I have given A no information about whether A will be pardoned. Prisoner pardoned Warden tells A Probability A B dies 1/6 A C dies 1/6 B C dies 1/3 C B dies 1/3 P(A warden says B dies) = P(warden says B dies A) P(wanden says B dies) = 1/6 1/6 + 1/3 = 1 3 Reason for the difference: {warden says B dies} B c but the two are not the same; e.g., the warden will not tell A whether A dies or not. UW-Madison (Statistics) Stat 609 Lecture 2 2015 7 / 17

Useful formulas For events A 1, A 2,..., ( ) ( n ) n 1 P A i = P(A 1 )P(A 2 A 1 )P(A 3 A 1 A 2 ) P A n A i For disjoint B 1, B 2,..., with B i = S, P(A) = P(B i )P(A B i ) In particular, P(A) = P(B)P(A B) + P(B c )P(A B c ). Example There are 90 red balls and 10 white balls. We select randomly one ball at a time without replacement. What is the probability that the 1st white ball appears at the 3rd draw? Define A i = {the ith selection is a white ball} B i = {the ith selection is a red ball} P(B 1 B 2 A 3 ) = P(B 1 )P(B 2 B 1 )P(A 3 B 1 B 2 ) = 90 100 89 99 10 98 UW-Madison (Statistics) Stat 609 Lecture 2 2015 8 / 17

Sensitive questions People may not give you a true answer if he/she is asked did you cheat in the last exam"? If we want to know the probability that a person cheats in a given exam, we may use the following procedure. We ask a person to flip a fair coin. If the result is heads", then the question is did you cheat"? If the result is tails", then the question is is your birthday afetr July 1"? No one, except the person being asked, knows which question is asked. The person will give the truthful answer. A = the event that the person answers yes". B = the event that the person is asked the 1st question. P(A) = P(B)P(A B) + P(B c )P(A B c ) = 0.5 P(A B) + 0.5 0.5 P(A B) = 2P(A) 0.5 If we estimate P(A) by asking 10,000 people, then we can estimate P(A B). UW-Madison (Statistics) Stat 609 Lecture 2 2015 9 / 17

Theorem 1.3.5. Bayes formula Let B, A 1,A 2,... be events with A i A j = /0, i j, and A i = S. Then, In particular, Example P(A B) = P(A i B) = P(B A i)p(a i ) P(B A j )P(A j ) j=1 P(B A)P(A) P(B A)P(A) + P(B A c )P(A c ) A = the event that a person has a type of disease, P(A) = 0.0004. B = a medical test indicates that a tested person has the disease, P(B A) = 0.99, P(B A c ) = 0.05. P(A B) =? By Bayes formula, P(A B) = 0.0004 0.99 0.0004 0.99 + 0.9996 0.05 = 0.007808. UW-Madison (Statistics) Stat 609 Lecture 2 2015 10 / 17

Definition 1.3.7. Two events A and B are (statistically) independent iff P(A B) = P(A)P(B). This definition is still good if P(A) = 0 or P(B) = 0. A and B are independent iff P(A B) = P(A) (or P(B A) = P(B)). That is, the probability of the occurrence of A is not affected by the occurrence of B. If A and B are independent, then the following pairs are also independent: A and B c, A c and B, and A c and B c ; e.g., P(A B c ) = P(A) P(A B) = P(A) P(A)P(B) = P(A)P(B c ) Example 1.3.11 (Letters) S consists of the 3! permutations of the letters a, b, and c, along with the three triplets of each letter: aaa bbb ccc S = abc bca cba acb bac cab UW-Madison (Statistics) Stat 609 Lecture 2 2015 11 / 17

Suppose that each element of S has probability 1/9 to occur. Define A i = {ith place in the triple is occupied by a}, i = 1,2,3. By looking at S, we obtain that Thus, P(A 1 ) = P(A 2 ) = P(A 3 ) = 1/3 P(A 1 A 2 ) = P(A 1 A 3 ) = P(A 2 A 3 ) = 1/9 P(A i A j ) = P(A i )P(A j ) i j and A i and A j are independent for any pair of (i,j), i j. It is easy to find events that are not independent, e.g., C = {abc,bca}, P(C) = 2/9, P(A 1 C) = 1/9 P(C)P(A 1 ) = 2/27 If A and B has some relationship, usually they are not independent. If A B, then P(A) = P(A B) = P(A)P(B) iff P(A) = 0 or P(B) = 1. If A B = /0, then P(A B) = P(A)P(B) = 0 iff P(A) = 0 or P(B) = 0. UW-Madison (Statistics) Stat 609 Lecture 2 2015 12 / 17

Definition 1.3.12. A collection of events A 1,...,A n are (mutually) independent iff for any subcollection A i1,...,a ik, P ( A i1 A ik ) = P(Ai1 ) P(A ik ) If we know P(A i ), i = 1,...,n, we may not know P( n A i). But if A 1,...,A n are independent, then ( ) n n P A i = P(A i ) P(A i )P(A j ) i<j + P(A i )P(A j )P(A k ) + + ( 1) n 1 P(A 1 )P(A 2 ) P(A n ) i<j<k If events A, B, C are independent, then so are A B and C, A B and C, and A B and C, e.g., P((A B) C) = P((A C) (B C)) = P(A C) + P(B C) P(A C B) = P(A)P(C) + P(B)P(C) P(A)P(C)P(B) = P(C)[P(A) + P(B) P(A)P(B)] = P(C)P(A B) UW-Madison (Statistics) Stat 609 Lecture 2 2015 13 / 17

Pairwise independence Events A 1,...,A n are pairwise independent iff A i and A j are independent for any pair (i, j). Mutual independence is stronger than pairwise independence. Example 1.3.11 (continued) A 1,A 2,A 3 are pairwise independent. They are not mutually independent, because P(A 1 A 2 A 3 ) = 1 9 1 27 = P(A 1)P(A 2 )P(A 3 ) Example 1.3.10 (Tossing two dice) An experiment consists of tossing two dice. The sample space is S = (1,1) (1,6) (2,1) (2,6) (6,1) (6,6) UW-Madison (Statistics) Stat 609 Lecture 2 2015 14 / 17

Consider Then A = {the same digit appear twice} P(A) = 1/6 B = {the sum is between 7 and 10} P(B) = 1/2 C = {the sum is 2, or 7, or 8} P(C) = 1/3 P(A B C) = P({(4,4)}) = 1 36 = 1 6 1 2 1 3 = P(A)P(B)P(C) However, A, B, and C are not independent, because P(B C) = P({the sum is 7 or 8}) = 11 36 1 6 = P(B)P(C) Independent experiments Two experiments are independent iff any event in experiment 1 and any event in experiment 2 are independent. n experiments are independent iff A 1,...,A n are independent, where A j is any event from experiment j, j = 1,...,n. Independence of experiments is an important concept in statistics. UW-Madison (Statistics) Stat 609 Lecture 2 2015 15 / 17

Conditional independence Events A and B are conditionally independent given event C iff P(A B C) = P(A C)P(B C) Conditional independence is an important concept in statistics. Independence does not imply conditional independence In Example 1.3.11, A i = {ith place in the triple is occupied by a}, i = 1,2,3. A 1 and A 2 are independent, but they are not conditionally independent given A 3, because P(A 1 A 2 A 3 ) = P(A 1 A 2 A 3 ) P(A 3 ) P(A 1 A 3 )P(A 2 A 3 ) = P(A 1 A 3 ) P(A 2 A 3 ) P(A 3 ) P(A 3 ) = 1/9 1/3 = 1 3 = P(A 1 )P(A 2 ) = 1 9 UW-Madison (Statistics) Stat 609 Lecture 2 2015 16 / 17

Conditional independence does not imply independence aaa bbb ccc abd S = abc bca cba acd acb bac cab bcd A i = {ith place in the triple is occupied by a}, i = 1,2,3. B = {d does not appear} A 1 and A 2 are conditionally independent given B, because P(A 1 A 2 B) = 1 9 = 1 3 1 3 = P(A 1 B)P(A 2 B) However, A 1 and A 2 are not independent, because P(A 1 A 2 ) = 1 9 5 12 3 12 = P(A 1)P(A 2 ) Mutual independence implies conditional independence If A, B, C are mutually independent, then A and B are conditionally independent given C, because P(A B C) = P(A B C) P(C) = P(A)P(B) = P(A C)P(B C) UW-Madison (Statistics) Stat 609 Lecture 2 2015 17 / 17