More on conditioning and Mr. Bayes

Similar documents
Lecture 2. Conditional Probability

Conditional Probability. CS231 Dianna Xu

Lecture 3: Probability

Prof. Thistleton MAT 505 Introduction to Probability Lecture 5

BASICS OF PROBABILITY CHAPTER-1 CS6015-LINEAR ALGEBRA AND RANDOM PROCESSES

STAT:5100 (22S:193) Statistical Inference I

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Chapter 3 Conditional Probability and Independence. Wen-Guey Tzeng Computer Science Department National Chiao Tung University

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

1 INFO 2950, 2 4 Feb 10

Lecture 2: Probability, conditional probability, and independence

Conditional Probability

Introduction and basic definitions

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Conditional probability

Discrete Probability. Chemistry & Physics. Medicine

UNIT Explain about the partition of a sampling space theorem?

Basic Statistics and Probability Chapter 3: Probability

Probabilistic models

1 Preliminaries Sample Space and Events Interpretation of Probability... 13

Statistical Theory 1

Topic 4 Probability. Terminology. Sample Space and Event

2. AXIOMATIC PROBABILITY

4-1 BASIC CONCEPTS OF PROBABILITY

Announcements. Topics: To Do:

Dept. of Linguistics, Indiana University Fall 2015

Slide 1 Math 1520, Lecture 21

7.1 What is it and why should we care?

Quantitative Methods for Decision Making

Probability, Statistics, and Bayes Theorem Session 3

Lecture 3 : Probability II. Jonathan Marchini

CIVL 7012/8012. Basic Laws and Axioms of Probability

Lecture 4 Bayes Theorem

M378K In-Class Assignment #1

Conditional Independence

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Lecture 4. Selected material from: Ch. 6 Probability

Combinatorics and probability

IB Math High Level Year 1 Probability Practice 1

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty

2.4 Conditional Probability

Probability. Introduction to Biostatistics

Lecture Slides. Elementary Statistics Tenth Edition. by Mario F. Triola. and the Triola Statistics Series. Slide 1

Single Maths B: Introduction to Probability

CS626 Data Analysis and Simulation

Lecture 16. Lectures 1-15 Review

Outline Conditional Probability The Law of Total Probability and Bayes Theorem Independent Events. Week 4 Classical Probability, Part II

With Question/Answer Animations. Chapter 7

Business Statistics MBA Pokhara University

P(A) = Definitions. Overview. P - denotes a probability. A, B, and C - denote specific events. P (A) - Chapter 3 Probability

Probability and Conditional Probability

Sixth Edition. Chapter 2 Probability. Copyright 2014 John Wiley & Sons, Inc. All rights reserved. Probability

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Denker FALL Probability- Assignment 6

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.

F71SM STATISTICAL METHODS

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Lecture 4 Bayes Theorem

STAT509: Probability

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 4.1-1

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

CSC Discrete Math I, Spring Discrete Probability

Baye s theorem. Baye s Theorem Let E and F be two possible events of an experiment, then P (F ) P (E F ) P (F ) P (E F ) + P (F ) P (E F ).

Basic Probability Theory (I)

Statistics for Engineers Lecture 2

Basic Statistics for SGPE Students Part II: Probability theory 1

2 TWO OR MORE RANDOM VARIABLES

Chapter. Probability

SDS 321: Introduction to Probability and Statistics

Applied Bayesian Statistics

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Lecture 5. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.

Lecture 3 Probability Basics

STATPRO Exercises with Solutions. Problem Set A: Basic Probability

Algebra. Formal fallacies (recap). Example of base rate fallacy: Monty Hall Problem.

A Event has occurred

Basic Probabilistic Reasoning SEG

Unit 2 Introduction to Probability

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Probabilistic models

CINQA Workshop Probability Math 105 Silvia Heubach Department of Mathematics, CSULA Thursday, September 6, 2012

Unit 3 Probability Basics

Chapter 2 Class Notes

Probability, Random Processes and Inference

More Information! Conditional Probability and Inverse Methods Peter Zoogman Harvard Atmospheres Journal Club April 12, 2012

DEPARTMENT OF QUANTITATIVE METHODS & INFORMATION SYSTEMS QM 120. Chapter 4: Experiment, outcomes, and sample space

Topic 2 Probability. Basic probability Conditional probability and independence Bayes rule Basic reliability

Review of Basic Probability

University of Technology, Building and Construction Engineering Department (Undergraduate study) PROBABILITY THEORY

Fundamentals of Machine Learning for Predictive Data Analytics

Probability 5-4 The Multiplication Rules and Conditional Probability

The enumeration of all possible outcomes of an experiment is called the sample space, denoted S. E.g.: S={head, tail}

Introduction to Machine Learning

HW MATH425/525 Lecture Notes 1

Gambler's Ruin. May 22, Gambler's Ruin

Econ 325: Introduction to Empirical Economics

Introduction to Machine Learning

Permutation. Permutation. Permutation. Permutation. Permutation

Introduction to Probability 2017/18 Supplementary Problems

Transcription:

More on conditioning and Mr. Bayes Saad Mneimneh 1 Multiplication rule for conditioning We can generalize the formula P(A,B) P(A B)P(B) to more than two events. For instance, P(A,B,C) P(A)P(B A)P(C A,B). In general, P(A 1,A 2,...A n ) P(A 1 ) P(A 2 A 1 )... P(A n A 1,A 2,...A n 1 ) This is easy to verify by replacing P(A i A 1,...A i 1 ) with its equivalent expression P(A 1,...,A i )/P(A 1,...A i 1 ), as follows: P(A 1 ) P(A 1,A 2 ) P(A 1 ) P(A 1,A 2,A 3 ) P(A 1,A 2 )... P(A 1,...,A n 1 ) P(A 1,...,A n 2 ) P(A 1,...,A n ) P(A 1,...,A n 1 ) In the above expression, the numerators and denominators cancel each other and what remains is P(A 1,...A n ). 2 Conditioning on multiple events Given an event B, consider a partition of the sample space into events A 1,A 2,A 3,... --------------------- A1 A2 A3 ------------- B ------------- --------------------- Now, B (B A 1 ) (B A 2 ) (B A 3 )... These events are exclusive and, therefore,. P(B) P(B A 1 ) + P(B A 2 ) + P(B A 3 )... P(B A 1 )P(A 1 ) + P(B A 2 )P(A 2 ) + P(B A 3 )P(A 3 )...

Therefore, we conclude that: P(B) i P(B A i )P(A i ) 3 Bayes rule From the definition of conditional probability and above, we have: P(A i B) P(B,A i) P(B) P(B A i)p(a i ) i P(B A i)p(a i ) Bayes rule is particularly useful when A i is not observable but B is. For instance, assume we know the probability of A i but we cannot observe the occurrence of such event. However, we can observe an event B, the occurrence of which depends on A i, and assume we can easily compute P(B A i ). Then we can say something about the occurrence of A i. P(A i ) is called the prior, B the observation, and P(A i B) the posterior. Let s revisit the king s sibling paradox. We can imagine the following scenario. -------------- A1: same A2: different non-observable 1/2 1/2 ------- ------- 1/2 1/2 1/2 1/2 MM FF MF FM B B B observable Events A 1 and A 2 correspond to whether both children have the same gender or not, respectively. These events are not observable, but influence event B that at least one of the siblings is a male (observable). Note: physically, the process of having two children is not necessarily carried out this way, i.e. by deciding first whether they will have the same gender or not, but we can model it this way. Now, P(A 1 B) and from the diagram above, P(B A 1 )P(A 1 ) P(B A 1 )P(A 1 ) + P(B A 2 )P(A 2 )

1/2 1/2 P(A 1 B) 1/2 1/2 + 1/2 (1/2 + 1/2) 1/3 Next, we will consider the examples of medical tests, colored balls and bins, evidence in the court room, biology of twins, and the famous Monty Hall. 4 Medical test Consider a medical test that is positive with probability 0.99 if the patient has the disease. Moreover, if the patient does not have the disease, the test is positive with probability 0.05. A patient is tested positive. What is the probability that he has the disease? (note that this is not 0.99) Let A be the event that the patient has the disease, and A c be its complement (not having the disease). Let B be the event that the test is positive. Using Bayes rule: P(A B) P(B A)P(A) P(B A)P(A) + P(B A c )P(A c ) 0.99P(A) 0.99P(A) + 0.05(1 P(A)) 0.99P(A) 0.05 + 0.94P(A) If the disease is rare, say P(A) 0.001, then P(A B) 0.02. Therefore, we have a 98% chance of false positive. For such a rare disease, the test is not good enough. 5 Colored balls and bins Consider a bin with 30 red balls and 10 black balls, and another bin with 20 balls of each color. A ball is picked from a bin with equal probability; therefore, if B i is the event that bin i was chosen, P(B 1 ) P(B 2 ) 1/2. The ball is red, what is the probability that bin 1 was chosen? Let R be the event that the ball is red, then: P(B 1 R) P(R B 1 )P(B 1 ) P(R B 1 )P(B 1 ) + P(R B 2 )P(B 2 ) 30/40 1/2 3/5 0.6 30/40 1/2 + 20/40 1/2 The result is expected since bin 1 contains more red balls. But which is really important, the number, or the fraction of red balls? (think about it).

6 Evidence in the court room Consider using a DNA match as evidence for guilt. If G is the event that the person is guilty, and M is the event of a DNA match, then P(M G) is usually high (close to 1) and P(M G c ) is usually low (close to 0, a match between two random DNAs). This is why a DNA match is considered to be a strong evidence. We are interested in reaching a verdict of guilty or not guilty, given that evidence. Therefore, we need to compute P(G M) and P(G c M). One possibility is to compute the log-odd ratio: log P(G M) P(G c M) and decide based on whether the result is positive (guilty) or non-positive (not guilty). This is not really a new concept, if the logarithm is positive, then P(G M)/P(G c M) > 1 and, hence, P(G M) > P(G c M). So essentially we are deciding based on which is more likely. One advantage of this strategy is that we do not have to compute P(M) (the denominator in Bayes rule) because it is a common factor in P(G M) and P(G c M). P(G M) P(G c M) P(M G)P(G) (1 ǫ)p(g) P(M G c )P(G c ) δ(1 P(G)) 1 P(G) δ 1 P(G) P(G) is of course a controversial measure since one could argue that it is not feasible to determine the probability of crime in a given society (we may however have an estimate of it). Let s say δ 1/10 6, then 1/δ 1000000. P(G M) P(G) P(G c 1000000 M) 1 P(G) Therefore, it takes a really clean society to overcome these odds. Most likely, 1/δ will determine the verdict (guilty). That s the classical approach. However, if the crime rate is really low, there is a chance that the verdict will be not guilty. This is an example where the classical approach and the Bayesian approach lead to different results (later we will see Lindley s paradox). 7 Biology of twins Twins can be either monozygotic (developed from a single egg) or dizygotic. It is always the case that monozygotic twins are of the same sex, whereas dizygotic twins can be of opposite sex. Denote monozygotic by M and dizygotic by D, and let B stand for boy and G stand for girl, then: P(GG M) P(BB M) 1, P(GB M) P(BG M) 0 2 P(GG D) P(BB D) P(GB D) P(BG D) 1 4

Now, assume that we conduct the following experiment: we sample many twins of the same sex, and discover (by looking for different features) that the probability they are dyzygotic is p. What is the proportion of dizygotic twins in the population. P(D {GG,BB}) p P({GG,BB} D)P(D) P({GG,BB}) Therefore, P(D) 2p/(1 + p). 8 Monty Hall 1/2 P(D) 1/2 P(D) + 1 (1 P(D)) There are three boxes: A, B, and C. Only one of them contains money, but all three are equally likely to contain the money. We pick one randomly (with equal probability). Then, one of the remaining two boxes is opened to reveal that it is empty. Finally we are given the choice to stick to the box we have or switch to the remaining one. What is the best strategy? One argument is that all boxes are equally likely to contain the money, so it makes no difference whether we switch or not. Using Bayesian analysis, however, we can show that there is always a benefit in switching. Let (x,y,z) represent the outcome of this game: x: the box containing to money y: the box we pick z: the box that is opened For instance, we are interested in computing: P(x A y A,z B) P(x A,y A,z B)/P(y A,z B) Note that: P(x,y,z) P(x)P(y x)p(z x,y) 1/3 1/3 P(z x,y) 1/9 P(z x,y) P(y,z) x P(y,z x)p(x) x P(y x)p(z x,y)p(x) x 1/3 1/3 P(z x,y)] 1/9 x P(z x,y) Therefore, P(x y,z) P(z x,y) x P(z x,y)

P(x A y A,z B) P(z B x A,y A) P(z B x A,y A) + P(z B x B,y A) + P(z B x C,y A) Similarly, we find that: 1/2 1/2 + 0 + 1 1/3 P(x A y A,z B) 1/3 (from above) P(x B y A,z B) 0 P(x C y A,z B) 2/3 The same result holds if we consider other (valid) permutations of x, y, and z. Therefore, there is always a benefit in switching. Another way to see this is by listing all the outcomes with their probabilities: (A, A, B) 1/18 (B, B, A) 1/18 (C, C, B) 1/18 (A, A, C) 1/18 (B, B, C) 1/18 (C, C, A) 1/18 (A,B,C) 1/9 (B,A,C) 1/9 (C,B,A) 1/9 (A,C,B) 1/9 (B,C,A) 1/9 (C,A,B) 1/9 Clearly, P(x A y A,z B) P(A,A,B)/(P(A,A,B) + P(C,A,B)) 1/3. It is also possible to view the Monty Hall paradox as a tree with three phases, the first of which is non-observable, while the last two are observable (in a similar way to the king s sibling paradox). ------------------------------------------- 1/3 1/3 1/3 A B C ----------- ----------- ----------- 1/3 1/3 1/3 1/3 1/3 1/3 1/3 1/3 1/3 A B C A B C A B C --- --- --- 1/2 1/2 1 1 1 1/2 1/2 1 1 1 1/2 1/2 B C C B C A C A B A A B