Example: Suppose we toss a quarter and observe whether it falls heads or tails, recording the result as 1 for heads and 0 for tails.

Similar documents
Probability- describes the pattern of chance outcomes

Review of Basic Probability

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Probability and Independence Terri Bittner, Ph.D.

Statistical Theory 1

2011 Pearson Education, Inc

Econ 325: Introduction to Empirical Economics

3.2 Probability Rules

The enumeration of all possible outcomes of an experiment is called the sample space, denoted S. E.g.: S={head, tail}

4 Lecture 4 Notes: Introduction to Probability. Probability Rules. Independence and Conditional Probability. Bayes Theorem. Risk and Odds Ratio

Announcements. Topics: To Do:

STA 291 Lecture 8. Probability. Probability Rules. Joint and Marginal Probability. STA Lecture 8 1

Lecture Lecture 5

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

STAT:5100 (22S:193) Statistical Inference I

Probability Theory and Applications

Lecture 4. Selected material from: Ch. 6 Probability

Outline Conditional Probability The Law of Total Probability and Bayes Theorem Independent Events. Week 4 Classical Probability, Part II

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

If S = {O 1, O 2,, O n }, where O i is the i th elementary outcome, and p i is the probability of the i th elementary outcome, then

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

UNIT 5 ~ Probability: What Are the Chances? 1

STA Module 4 Probability Concepts. Rev.F08 1

Lecture 1. ABC of Probability

Review Basic Probability Concept

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Basic Statistics and Probability Chapter 3: Probability

Week 2. Section Texas A& M University. Department of Mathematics Texas A& M University, College Station 22 January-24 January 2019

I - Probability. What is Probability? the chance of an event occuring. 1classical probability. 2empirical probability. 3subjective probability

Chapter 7 Wednesday, May 26th

Lecture notes for probability. Math 124

F71SM STATISTICAL METHODS

Chapter 1 (Basic Probability)

STAT 430/510 Probability

Mean, Median and Mode. Lecture 3 - Axioms of Probability. Where do they come from? Graphically. We start with a set of 21 numbers, Sta102 / BME102

Fundamentals of Probability CE 311S

Homework (due Wed, Oct 27) Chapter 7: #17, 27, 28 Announcements: Midterm exams keys on web. (For a few hours the answer to MC#1 was incorrect on

the time it takes until a radioactive substance undergoes a decay

Single Maths B: Introduction to Probability

Compound Events. The event E = E c (the complement of E) is the event consisting of those outcomes which are not in E.

Lecture 1. Chapter 1. (Part I) Material Covered in This Lecture: Chapter 1, Chapter 2 ( ). 1. What is Statistics?

Independence. P(A) = P(B) = 3 6 = 1 2, and P(C) = 4 6 = 2 3.

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Lecture 2: Probability, conditional probability, and independence

18.600: Lecture 7 Bayes formula and independence

Chapter 4 - Introduction to Probability

Probability and Statistics Notes

Information Science 2

STAT Chapter 3: Probability

HW MATH425/525 Lecture Notes 1

Axioms of Probability

Dept. of Linguistics, Indiana University Fall 2015

Lecture 3 Probability Basics

Lecture 3 - Axioms of Probability

Math Exam 1 Review. NOTE: For reviews of the other sections on Exam 1, refer to the first page of WIR #1 and #2.

Introduction to Probability

8. MORE PROBABILITY; INDEPENDENCE

3 PROBABILITY TOPICS

Probability Theory Review

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Mutually Exclusive Events

18.600: Lecture 7 Bayes formula and independence

MA : Introductory Probability

CIVL Why are we studying probability and statistics? Learning Objectives. Basic Laws and Axioms of Probability

(6, 1), (5, 2), (4, 3), (3, 4), (2, 5), (1, 6)

AMS7: WEEK 2. CLASS 2

CPSC340. Probability. Nando de Freitas September, 2012 University of British Columbia

Lecture 1 : The Mathematical Theory of Probability

Quantitative Methods for Decision Making

Stat 225 Week 1, 8/20/12-8/24/12, Notes: Set Theory

BASICS OF PROBABILITY CHAPTER-1 CS6015-LINEAR ALGEBRA AND RANDOM PROCESSES

Probability Year 9. Terminology

Chapter 6: Probability The Study of Randomness

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

Chapter 5 : Probability. Exercise Sheet. SHilal. 1 P a g e

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Probability COMP 245 STATISTICS. Dr N A Heard. 1 Sample Spaces and Events Sample Spaces Events Combinations of Events...

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

KDF2C QUANTITATIVE TECHNIQUES FOR BUSINESSDECISION. Unit : I - V

1 Probability Theory. 1.1 Introduction

AP Statistics Ch 6 Probability: The Study of Randomness

Lecture 4 Bayes Theorem

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

Section 13.3 Probability

Introduction to Probability 2017/18 Supplementary Problems

2. Probability. Chris Piech and Mehran Sahami. Oct 2017

STA111 - Lecture 1 Welcome to STA111! 1 What is the difference between Probability and Statistics?

Some Basic Concepts of Probability and Information Theory: Pt. 1

Conditional Probability and Bayes

1 INFO 2950, 2 4 Feb 10

Topic 5: Probability. 5.4 Combined Events and Conditional Probability Paper 1

Probability Year 10. Terminology

Executive Assessment. Executive Assessment Math Review. Section 1.0, Arithmetic, includes the following topics:

Basic Probability. Introduction

Ch 14 Randomness and Probability

What is Probability? Probability. Sample Spaces and Events. Simple Event

Probability (Devore Chapter Two)

Lecture 7. Bayes formula and independence

7.1 What is it and why should we care?

Basics of Probability

Transcription:

Example: Suppose we toss a quarter and observe whether it falls heads or tails, recording the result as 1 for heads and 0 for tails. (In Mathematical language, the result of our toss is a random variable, which we will call X, and this variable has two possible values, 0 and 1.) Which of these two values X will actually take is going to be determined by the coin toss. Will we find that X=0 or that X=1? We can t say. Nobody knows. What we can say is that the probability that X will be 0 is ½, and the probability that X will be 1 is also ½. Mathematically, we can write the above statements in symbols: P(coin falls heads) = P(X=1) = 1/2 P(coin falls tails) = P(X=0) = 1/2 Authors: Blume, Greevy Bios 311 Lecture Notes Page 1 of 26

A probability statement is a method of expressing our uncertainty about the outcome of a future event. It does not make sense to talk about the probability of an event that has already occurred (because there is no uncertainty in the outcome, i.e. the outcome has already been determined). Friend 1: Wow, did you see that?! That was incredible! What are the chances of that?! Friend 2: The chance of exactly that is 100%; it just happened. Friend 1: I can t take you anywhere. However if you do not yet know the outcome, then it may make sense to talk about that probability, because it reflects your personal uncertainty about the situation. After a Powerball drawing, I wait a few weeks before checking my ticket. Like Schrödinger's cat, I like to think I haven t lost until I observe the results. Until then, I can maintain my belief that I have a 1 in 175,223,510 chance of winning the jackpot. Authors: Blume, Greevy Bios 311 Lecture Notes Page 2 of 26

Specifically, what does P(X=1)=1/2 mean? The probability of heads represents the tendency of the quarter to fall heads, or the degree of certainty that heads will occur. P(X=1)=0 means no tendency at all to fall heads there is no chance of observing heads. P(X=1)=1 means an overwhelming tendency to fall heads, i.e., perfect certainty as if we were tossing a coin with a head on both sides. The variable X is a special type of variable, one whose value is determined by chance. It is a Random variable. Authors: Blume, Greevy Bios 311 Lecture Notes Page 3 of 26

This experiment (coin toss) can be described by stating that there are two possible values for X, 0 and 1, and that their probabilities are P(X=0)=1/2 and P(X=1)=1/2. Mathematically this description consists of three components: (a) a random variable (b) a set, say S={0,1}, containing all the possible outcomes of the random variable in (a) (c) A listing of the probabilities that the random variable in (a) take each value in S. These three components (a)-(c) are a mathematical representation (a probability model) for the process of tossing a quarter and observing the result. Now lets consider a more complicated experiment. Suppose we toss three coins, a nickel, a dime, and a quarter. We can represent this experiment mathematically with the same three components. Authors: Blume, Greevy Bios 311 Lecture Notes Page 4 of 26

This time let s use Y to represent the result. Again the result is a variable, Y, whose values is determined by chance a random variable. The possible outcomes for this experiment are: (nickel, dime, quarter) (0,0,0) (tails on all three) (0,0,1) (nickel and dime tails, quarter heads) (0,1,0) etc. (0,1,1) (1,0,0) (1,0,1) (1,1,0) (1,1,1) For this experiment, S is the set of these eight possible results: S={(0,0,0), (0,0,1), (0,1,0),, (1,1,1)} So that result of this experiment, Y (the random variable), will have one of these eight values. Either Y=(0,0,0), Y=(0,0,1),, or Y=(1,1,1). The probability of each of the eight possible outcomes is 1/8: P(Y=(0,0,0))=P(Y=(0,0,1))= = P(Y=(1,1,1))=1/8 Authors: Blume, Greevy Bios 311 Lecture Notes Page 5 of 26

Again we have represented the experiment in term of the three elements: (1) a random variable, Y, representing the result (2) a set S, called the sample space, showing the possible values of Y, and (3) a list giving the probability of each value. These three pieces, taken together, form a probability model for this experiment. Any probability model (no matter how complicated) consists of these three components: (1) A random variable (2) A sample space (3) A probability distribution over the sample space. A probability model allows us to analyze the experiment mathematically to answer questions like What is the probability of getting heads on two out of the three tosses? by exact calculations instead of by scratching our heads and guessing. Authors: Blume, Greevy Bios 311 Lecture Notes Page 6 of 26

Using a probability model Now that we have a probability model, let s examine a question we already know the answer to: What is the probability of heads in the first toss? In terms of our probability models, it is the probability that Y takes of the four values, (1,0,0), (1,0,1), (1,1,0), or (1,1,1). Thus the event Heads on the first toss is represented in this probability model by the set of possible outcomes, A={(1,0,0), (1,0,1), (1,1,0), (1,1,1)} The probability of this event (or this set) is simply the sum of the probabilities of its elements: P(A) = P({(1,0,0), (1,0,1), (1,1,0), (1,1,1)}) = P((1,0,0) or (1,0,1) or (1,1,0) or (1,1,1)) = P((1,0,0)) + P((1,0,1)) + P((1,1,0))+P((1,1,1)) = 1/8 + 1/8 + 1/8 + 1/8 = 4/8=1/2 We will see later that the important step that the little arrow points to is justified by one of the basic rules of probability. For now, we ll just accept it. Authors: Blume, Greevy Bios 311 Lecture Notes Page 7 of 26

Now try a question whose answer is not so obvious: What is the probability of heads on one of the last two tosses? This event might be represented by the set B={(0,1,0),(0,0,1),(1,1,0),(1,0,1),(0,1,1),(1,1,1)} But it might also be represented by a different set C={(0,1,0),(0,0,1), (1,1,0),(1,0,1)} Before we can answer the question, we must be more specific: Do we mean heads on at least one of the last two tosses? or heads in exactly one of the last two tosses? These are two different questions, referring to two different events. The first is represented in our probability model by the set B, and the other by set C. Authors: Blume, Greevy Bios 311 Lecture Notes Page 8 of 26

If we mean at least one then we want the probability of event (set) B: P(B) = P({(0,1,0),(0,0,1),,(1,1,1)}) = P((0,1,0) or (0,0,1) or or (1,1,1)) = P((0,1,0)) + P((0,0,1)) + +P((1,1,1)) = 1/8 + 1/8 + 1/8 + 1/8 + 1/8 + 1/8 = 6/8=3/4 If we mean heads on exactly one the we want the probability of event (set) C: P(C) = P({(0,1,0),(0,0,1), (1,0,1),(1,1,0)}) = P((0,1,0) or (0,0,1) or (1,0,1) or (1,1,0)) = P((0,1,0)) + P((0,0,1)) + P((1,0,1)) +P((1,1,0)) = 1/8 + 1/8 + 1/8 + 1/8 = 4/8=1/2 Authors: Blume, Greevy Bios 311 Lecture Notes Page 9 of 26

How about the event Heads on the first toss or heads on at least one of the last two? This event occurs if the result, Y, has any of the values that appear in either set A or set B. A={(1,0,0),(1,0,1),(1,1,0),(1,1,1)} B={(0,1,0),(0,0,1),(1,1,0),(1,0,1),(0,1,1),(1,1,1)} The set A or B consists of A plus the three checked points (0,1,0),(0,0,1), and (0,1,1), that are in B, but not in A. Equivalently, A or B consists of B plus any additional points that are in A but not in B; in this example there is one such point, (1,0,0). You find that the set A or B consists of the same seven points either way: A or B = {(0,1,0),(0,0,1),(1,1,0),(1,0,1),(0,1,1),(1,1,1),(1,0,0)} And the probability of this event is P(A or B) = 7/8. Authors: Blume, Greevy Bios 311 Lecture Notes Page 10 of 26

But wait: How do the probabilities of the events A and B relate to the probability of the event A or B? P(A or B) = P(A) + P(B)? = ½ + ¾ = 1¼ (NO) In order to answer questions like this we need some notation and techniques to help us organize our thinking about sets. Authors: Blume, Greevy Bios 311 Lecture Notes Page 11 of 26

SETS AUB: The union of A and B. The set of points that are in A or in B (or both). AIB: The intersection of A and B. The set of points that are in both A and B. Note: Sometimes when there is no danger of confusion we omit the I and write simply AB to represent this set. A c = Ā: The compliment of A. The set of all points that are not in A. This set is always defined in relation to the set, S, that represents all of the points that are being considered in a particular problem. That is, A c is the set of all point in S that are not in A. Example (tossing three coins continued): A ={(1,0,0),(1,0,1),(1,1,0),(1,1,1)} A c =Ā ={(0,0,0),(0,0,1),(0,1,0),(0,1,1)} Authors: Blume, Greevy Bios 311 Lecture Notes Page 12 of 26

Because every point that is not in A c (not in not A ) is in A, we have (A c ) c = A And because every point in S is either in A or not in A (in A or in Ā = A c ), it is always true that A U (A c ) = S The Empty Set, Ø When two sets have no points in common, their intersection is empty. This is a perfectly good set. It has nothing in it, but it is still a set. It occurs so often that we have a special symbol for it, Ø. Ø: The empty set. A set containing no points. Since a set and its complement have no points in common, it is always true that A I (A c ) = Ø Authors: Blume, Greevy Bios 311 Lecture Notes Page 13 of 26

Subsets When every point in some set A also belongs to some other set, B, then A is a subset of B, written as A B. If A contains a point that is not in B, then A is not a subset of B, written as A B. For the sets in our example, (A is heads in the first toss, B is heads on at least one of the last two tosses, and C is heads on exactly one of the last two tosses ) A={(1,0,0),(1,0,1),(1,1,0),(1,1,1)} B={(0,1,0),(0,0,1),(1,1,0),(1,0,1),(0,1,1),(1,1,1)} C={(0,1,0),(0,0,1),(1,1,0),(1,0,1)} Every point in C is also in B, so C is a subset of B, or C B. But the set A contains a point, (1,0,0) that is not in B, so A B. Note also that every set is a subset of the sample space S. (A S, B S, C S ) Authors: Blume, Greevy Bios 311 Lecture Notes Page 14 of 26

Examples: A I (B c )={(1,0,0)} This set, the intersection of A and B-compliment, or A and not B, consists of the single point, (1,0,0) (because all of the other points are in A are also in B). (A c ) I B={(0,0,1),(0,1,0),(0,1,1)} This is the intersection of A-compliment and B, or not-a and B. It is composed of the points representing both of the events, not-a (not heads on the first toss, which is the same as tails on the first toss ) and B ( heads on at least one of the last two tosses ). Now, if we add to our list the event same result on all three tosses, which corresponds to the set D={(0,0,0),(1,1,1)} We find that C I D = Ø because the set C and D have no points in common. Set like this are called mutually exclusive, or disjoint sets. Two events are mutually exclusive if their intersection is the empty set. Authors: Blume, Greevy Bios 311 Lecture Notes Page 15 of 26

Venn Diagrams Venn diagrams are very handy tools for studying sets. Basically, a Venn diagram is just a picture of the sample space divided into its subsets. They are useful because they illustrate relations that are true of all sets. Here is an example: S A AB c AB A c B B This Venn diagram displays the sets: A, B, AB, AB c, A c B, S. Note: To be perfectly clear we should have written AB c as A I (B c ), so that there could be no confusion with (A I B) c. We will use the rule that unless parentheses tell us otherwise, the little c applies only to the set that it is attached to. Authors: Blume, Greevy Bios 311 Lecture Notes Page 16 of 26

Back to Probability Consider the following generic probability model: (1) A random variable X (2) A sample space S (3) A probability distribution over the sample space. Some fundamental properties of probabilities are: (1) For any event E, 0 P(E) 1 (2) For any two mutually exclusive events, E and F, P(E or F) = P(E U F) = P(E)+P(F) (3) P(S) = 1 (4) P(Ø) = 0 (5) For any event E, P(E c ) = 1 - P(E) (6) For any two events E F, P(E) P(F) Authors: Blume, Greevy Bios 311 Lecture Notes Page 17 of 26

Fundamental Properties continued: S A AB c AB A c B B The Venn Diagram shows us how to relate the probabilities of two events, A and B, to the probability of the event either A or B. (7) For any two events A and B P(A or B) = P(A U B) = P(A) + P(B) P(A I B) = P(A) + P(B) P(A and B) Alternatively rearranging yields: P(A and B) = P(A I B) = P(A) + P(B) P(A U B) = P(A) + P(B) P(A or B) Remember P(AB) = P(A and B) = P(A I B) Authors: Blume, Greevy Bios 311 Lecture Notes Page 18 of 26

To see why this must be true look at the Venn Diagram, which shows that A, B, and A U B can all be written as unions of the three disjoint sets, AB c, AB, and A c B (shaded area). Therefore, A = AB U AB c B = AB U A c B A U B = AB U AB c U A c B (shaded area) P(A U B) = P(AB) + P(AB c ) + P(A c B) = P(AB) + P(AB c ) + P(A c B) +P(AB) P(AB) = P(A) + P(B) P(AB) And we have proven fundamental property number (7), which is one of the most useful properties. Authors: Blume, Greevy Bios 311 Lecture Notes Page 19 of 26

Conditional Probability and Independence A key concept in probability is that of independence: Independent Events: Two events, A and B are independent if P(A I B) = P(A)P(B) Another key concept is that of conditional probability: The Conditional Probability of event A, given event B, is written P(A B) and is defined: Or upon rearrangement P(A B) = P(A I B)/P(B) P(A I B) = P(A B)P(B) S So we have P(A B) = B A Instead of P(A) = A S Authors: Blume, Greevy Bios 311 Lecture Notes Page 20 of 26

Conditional Probability is best illustrated with an example: Smoker Drinker Dieter Total Male 97 45 63 205 Female 80 30 72 182 Other 11 12 13 36 Total 188 87 148 423 So that we have: P(smoker) = 188/423 P(other) = 36/423 P(smoker and other)=11/423 (~ 0.026) P(smoker or other) = 188/423 + 36/423 11/423 = 213/423 = (97+80+11+12+13)/423 P(smoker other) = P(smoker and other)/p(other) = (11/423)/(36/423) = 11/36 Check independence:? P(smoker and other) = P(smoker)P(other) = (188/423)(36/423) ~ 0.038 But 0.038 is not equal to P(smoker and other) = 0.026 (11/423 see above) so the events are not independent. Another option is to check to see if P(smoker other) = P(smoker), which is not the case. Authors: Blume, Greevy Bios 311 Lecture Notes Page 21 of 26

Note: The conditional probability is conditional on the column (or marginal) total. It asks, Out of the others, how many are smokers? Thus we are effectively changing our sample space. Authors: Blume, Greevy Bios 311 Lecture Notes Page 22 of 26

Independence and Mutually Exclusive Mutually exclusive events are highly dependent, because if an event is not in one set, it must be in the other. This is easy to see mathematically: Suppose two events, E and F, are mutually exclusive. This implies that (E and F) = Ø, and that P(E and F) = 0 For E and F to be independent events we need the following equation to hold: P(E and F) = P(E)P(F) But for mutually exclusive events, this would mean that 0=P(E)P(F) so that either P(E) = 0 or P(F) = 0 (neither of which is very interesting) Thus at least one of the events must never occur for two events to be mutually exclusive and independent with each other. Authors: Blume, Greevy Bios 311 Lecture Notes Page 23 of 26

In terms of conditional probabilities, what independence of A and B means is that A s probability of occurrence is unaffected by whether B occurs or not. P(A B) = P(A and B)/P(B) = P(A)P(B)/P(B) by independence = P(A) Likewise, B s probability of occurrence is unaffected by whether or not A occurs: P(B A) = P(B and A)/P(A) = P(B)P(A)/P(A) by independence = P(B) From the way conditional probability is defined, P(A B)=P(A and B)/P(B), we can always write the probability of (A and B) as: P(A and B) = P(A)P(B A) Or P(A and B) = P(B)P(A B) When A and B are independent (and only in that case) P(B A)=P(B) and P(A B)=P(A), the above two equation reduce to P(A and B) = P(A)P(B). Authors: Blume, Greevy Bios 311 Lecture Notes Page 24 of 26

Law of Total Probability Suppose the sample space is divided into any number of disjoint sets, say A 1, A 2,, A n, so that A i I A J = Ø and A 1 U A 2 U U A n = S. Furthermore we can divide any set B into the disjoint subsets: B = BA 1 U BA 2 U U BA n A 2 A 3 A 1 BA 2 BA 3 BA 1 BA 4 A 4 A 5 In this case we can write P(B) = P(BA 1 U BA 2 U U BA n ) = P(BA 1 ) + P(BA 2 ) + + P(BA n ) = P(B A 1 )P(A 1 )+ P(B A 2 )P(A 2 ) + + P(B A n ) P(A n ) Or more generally: P ( B ) = n i = 1 P ( B A ) P ( i A i ) Authors: Blume, Greevy Bios 311 Lecture Notes Page 25 of 26

Final note: We said earlier that probabilities are tendencies. Some people prefer to define what probabilities are in terms of relative frequencies in very long sequences of repetitions of the experiment or process. Still others say that probabilities are nothing more than personal degrees of belief (and some of those people are quite uncomfortable with the idea that a probability might have an objective true value that is independent of any person s beliefs). Thus some think of probabilities as objective (and measurable) characteristics of certain parts of the world, much like temperature or viscosity, while other view probabilities as subjective measures of personal uncertainty, applicable to one s judgment and opinions. In most of this course we take the position that in many important situations in makes sense to think of probabilities as objective quantities, so that it also make sense to speak of the true value of a probability, to attempt to estimate that value, and to speak of the error in an estimate. Authors: Blume, Greevy Bios 311 Lecture Notes Page 26 of 26