Introduction to Probability and Sample Spaces

Similar documents
Lecture 1 : The Mathematical Theory of Probability

Events A and B are said to be independent if the occurrence of A does not affect the probability of B.

Mathematics. ( : Focus on free Education) (Chapter 16) (Probability) (Class XI) Exercise 16.2

Conditional Probability and Bayes Theorem (2.4) Independence (2.5)

STAT 430/510 Probability

Probability Pearson Education, Inc. Slide

Introduction Probability. Math 141. Introduction to Probability and Statistics. Albyn Jones

Axioms of Probability

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

CSC Discrete Math I, Spring Discrete Probability

Statistics Statistical Process Control & Control Charting

324 Stat Lecture Notes (1) Probability

MODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES DISTRIBUTION FUNCTION AND ITS PROPERTIES

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.

Probability: Sets, Sample Spaces, Events

Fundamentals of Probability CE 311S

What is Probability? Probability. Sample Spaces and Events. Simple Event

CS4705. Probability Review and Naïve Bayes. Slides from Dragomir Radev

Sec$on Summary. Assigning Probabilities Probabilities of Complements and Unions of Events Conditional Probability

Steve Smith Tuition: Maths Notes

4/17/2012. NE ( ) # of ways an event can happen NS ( ) # of events in the sample space

Probability. VCE Maths Methods - Unit 2 - Probability

6.2 Introduction to Probability. The Deal. Possible outcomes: STAT1010 Intro to probability. Definitions. Terms: What are the chances of?


Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Deep Learning for Computer Vision

SDS 321: Introduction to Probability and Statistics

Conditional Probability

Probability- describes the pattern of chance outcomes

Econ 325: Introduction to Empirical Economics

Mean, Median and Mode. Lecture 3 - Axioms of Probability. Where do they come from? Graphically. We start with a set of 21 numbers, Sta102 / BME102

Chapter 2 PROBABILITY SAMPLE SPACE

ELEG 3143 Probability & Stochastic Process Ch. 1 Experiments, Models, and Probabilities

CS 441 Discrete Mathematics for CS Lecture 20. Probabilities. CS 441 Discrete mathematics for CS. Probabilities

Lecture 3 - Axioms of Probability

Chapter 7. Inclusion-Exclusion a.k.a. The Sieve Formula

CMPSCI 240: Reasoning about Uncertainty

Probability deals with modeling of random phenomena (phenomena or experiments whose outcomes may vary)

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

27 Binary Arithmetic: An Application to Programming

Probabilistic models

Expected Value. Lecture A Tiefenbruck MWF 9-9:50am Center 212 Lecture B Jones MWF 2-2:50pm Center 214 Lecture C Tiefenbruck MWF 11-11:50am Center 212

EE 178 Lecture Notes 0 Course Introduction. About EE178. About Probability. Course Goals. Course Topics. Lecture Notes EE 178

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 2: Random Experiments. Prof. Vince Calhoun

(i) Given that a student is female, what is the probability of having a GPA of at least 3.0?

Monty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch

Week 2. Section Texas A& M University. Department of Mathematics Texas A& M University, College Station 22 January-24 January 2019

Probability Theory and Applications

Term Definition Example Random Phenomena

PERMUTATIONS, COMBINATIONS AND DISCRETE PROBABILITY

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

Statistical Inference

TA Qinru Shi: Based on poll result, the first Programming Boot Camp will be: this Sunday 5 Feb, 7-8pm Gates 114.

Lecture Lecture 5

LECTURE NOTES by DR. J.S.V.R. KRISHNA PRASAD

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

Marquette University MATH 1700 Class 5 Copyright 2017 by D.B. Rowe

General Info. Grading

AMS7: WEEK 2. CLASS 2

Math 1313 Experiments, Events and Sample Spaces

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

CHAPTER 1 SETS AND EVENTS

Probability 1 (MATH 11300) lecture slides

ECE 450 Lecture 2. Recall: Pr(A B) = Pr(A) + Pr(B) Pr(A B) in general = Pr(A) + Pr(B) if A and B are m.e. Lecture Overview

Properties of Probability

PROBABILITY CHAPTER LEARNING OBJECTIVES UNIT OVERVIEW

Introduction to Probability 2017/18 Supplementary Problems

Chapter 13, Probability from Applied Finite Mathematics by Rupinder Sekhon was developed by OpenStax College, licensed by Rice University, and is

More on Distribution Function

Conditional Probability

First Digit Tally Marks Final Count

MA : Introductory Probability

Probability: Axioms, Properties, Interpretations

CS206 Review Sheet 3 October 24, 2018

Lecture notes for probability. Math 124

Key Concepts. Key Concepts. Event Relations. Event Relations

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Executive Assessment. Executive Assessment Math Review. Section 1.0, Arithmetic, includes the following topics:

9/6/2016. Section 5.1 Probability. Equally Likely Model. The Division Rule: P(A)=#(A)/#(S) Some Popular Randomizers.

Notes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 15 Notes. Class URL:

Origins of Probability Theory

Expected Value 7/7/2006

Independence Solutions STAT-UB.0103 Statistics for Business Control and Regression Models

Topic 5 Basics of Probability

Section 13.3 Probability

ENGI 4421 Introduction to Probability; Sets & Venn Diagrams Page α 2 θ 1 u 3. wear coat. θ 2 = warm u 2 = sweaty! θ 1 = cold u 3 = brrr!

Business Statistics MBA Pokhara University

Recap. The study of randomness and uncertainty Chances, odds, likelihood, expected, probably, on average,... PROBABILITY INFERENTIAL STATISTICS

Standard & Conditional Probability

Lecture 8: Probability

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

ELEG 3143 Probability & Stochastic Process Ch. 1 Probability

4. Probability of an event A for equally likely outcomes:

P [(E and F )] P [F ]

STT When trying to evaluate the likelihood of random events we are using following wording.

STAT 111 Recitation 1

Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

What are the odds? Coin tossing and applications

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Transcription:

2.2 2.3 Introduction to Probability and Sample Spaces Prof. Tesler Math 186 Winter 2019 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 1 / 26

Course overview Probability: Determine likelihood of events Roll a die. The probability of rolling 1 is 1/6. Descriptive statistics: Summarize data Mean, median, standard deviation,... Inferential statistics: Infer a conclusion/prediction from data Test a drug to see if it is safe and effective, and at what dose. Poll to predict the outcome of an election. Repeatedly flip a coin or roll a die to determine if it is fair. Bioinformatics We ll apply these to biological data, such as DN sequences and microarrays. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 2 / 26

Related courses Math 183: Usually uses the same textbook and chapters as Math 186. Focuses on the examples in the book. The mathematical content is the same, but Math 186 has extra material for bioinformatics. Math 180BC plus 181BC: More in-depth: a year of probability and a year of statistics. CSE 103, Econ 120, ECE 109: One quarter intro to probability and statistics, specialized for other areas. Math 283: Graduate version of this course. Review of basic probability and statistics, with a lot more applications in bioinformatics. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 3 / 26

2.2 Sample spaces Flip a coin 3 times. The possible outcomes are HHH HHT HTH HTT THH THT TTH TTT The sample space is the set of all possible outcomes: S = {HHH, HHT, HTH, HTT, THH, THT, TTH, TTT} The size of the sample space is N(S) = 8 Our book s notation S = 8 more common notation in other books We could count this by making a 2 2 2 table: 2 choices for the first flip 2 choices for the second flip 2 choices for the third flip = 2 3 = 8 The number of strings x 1 x 2... x k or sequences (x 1, x 2,..., x k ) of length k with r choices for each entry is r k. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 4 / 26

Rolling two dice Roll two six-sided dice, one red, one green: green 1 2 3 4 5 6 red 1 (1, 1) (1, 2) (1, 3) (1, 4) (1, 5) (1, 6) 2 (2, 1) (2, 2) (2, 3) (2, 4) (2, 5) (2, 6) 3 (3, 1) (3, 2) (3, 3) (3, 4) (3, 5) (3, 6) 4 (4, 1) (4, 2) (4, 3) (4, 4) (4, 5) (4, 6) 5 (5, 1) (5, 2) (5, 3) (5, 4) (5, 5) (5, 6) 6 (6, 1) (6, 2) (6, 3) (6, 4) (6, 5) (6, 6) The sample space is S = {(1, 1), (1, 2),..., (6, 6)} = { (i, j) Z 2 : 1 i 6, 1 j 6 } where Z = integers N(S) = 6 2 = 36 Z 2 = ordered pairs of integers Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 5 / 26

DN sequences codon is a DN sequence of length 3, in the alphabet of nucleotides {, C, G, T }: S = {, C, G, T,..., TTT } How many codons are there? N(S) = 4 3 = 64 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 6 / 26

continuous sample space Consider this disk (filled-in circle): y 3 3 C 2 x 3 Complications 3 S = { (x, y) R 2 : x 2 + y 2 2 2 } The sample space is infinite and continuous. The choices of x and y are dependent. E.g.: at x = 0, we have 2 y 2; at x = 2, we have y = 0. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 7 / 26

Events Flip a coin 3 times. The sample space is S = {HHH, HHT, HTH, HTT, THH, THT, TTH, TTT} n event is a subset of the sample space ( S): = First flip is heads = {HHH, HHT, HTH, HTT} B = Two flips are heads = {HHT, HTH, THH} C = Four flips are heads = (empty set or null set) We can combine these using set operations. For example, The first flip is heads or two flips are heads = { HHH, HHT, HTH, HTT } B = { HHT, HTH, THH } B = { HHH, HHT, HTH, HTT, THH } Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 8 / 26

Using set operations to form new events = First flip is heads = {HHH, HHT, HTH, HTT} B = Two flips are heads THT = {HHT, HTH, THH} TTH Union: ll elements that are in or in B TTT HHH HTT HHT HTH B = {HHH, HHT, HTH, HTT, THH} or B : The first flip is heads or two flips are heads This is inclusive or: one or both conditions are true. Intersection: ll elements that are in both and in B THH B = {HHT, HTH} and B : The first flip is heads and two flips are heads Complement: ll elements of the sample space not in c = {THT, TTH, TTT, THH} Not : The first flip is not heads Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 9 / 26 B

Venn diagram and set sizes = {HHH, HHT, HTH, HTT} B = {HHT, HTH, THH} B = {HHH, HHT, HTH, HTT, THH} B = {HHT, HTH} THT TTH TTT HHH HTT Relation between sizes of union and intersection HHT HTH Notice that N( B) = N() + N(B) N( B) 5 = 4 + 3 2 THH N() + N(B) counts everything in the union, but elements in the intersection are counted twice. Subtract N( B) to compensate. Size of complement N(B c ) = N(S) N(B) 5 = 8 3 B Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 10 / 26

lgebraic rules for set theory Commutative laws B = B B = B ssociative laws ( B) C = (B C) ( B) C = (B C) One may omit parentheses in B C or B C. But don t do that with a mix of and. Distributive laws (B C) = ( B) ( C) (B C) = ( B) ( C) These are like a(b + c) = ab + ac Complements De Morgan s laws c = S c = ( B) c = c B c ( B) c = c B c Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 11 / 26

Distributive laws Visualizing identities using Venn diagrams: (B C) = ( B) ( C) B C (B C) # B# # B# S# C# S# C# B C ( B) ( C) # B# # B# # B# S# C# S# C# S# C# Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 12 / 26

Mutually exclusive sets Two events are mutually exclusive if their intersection is. = First flip is heads = {HHH, HHT, HTH, HTT} B = Two flips are heads = {HHT, HTH, THH} C = One flip is heads = {HTT, THT, TTH} and B are not mutually exclusive, since B = {HHT, HTH}. B and C are mutually exclusive, since B C =. For mutually exclusive events, since N(B C) = 0, we get: N(B C) = N(B) + N(C) Events 1, 2,... are pairwise mutually exclusive when i j = for i j. 1 2 3 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 13 / 26

2.3 Probability functions Historically, there have been several ways of defining probabilities. We ll start with Classical Probability: Classical probability Suppose the sample space has n outcomes (N(S) = n) and all of them are equally likely. Each outcome has a probability 1/n of occurring: P(s) = 1/n for each outcome s S n event S with m outcomes has probability m/n of occurring: P() = m n = N() N(S) Example: Rolling a pair of dice N(S) = n = 36 P(first die is 3) = P ( {(3, 1), (3, 2),..., (3, 6)} ) = 6 36 P(the sum is 8) = P ( {(2, 6), (3, 5), (4, 4), (5, 3), (6, 2)} ) = 5 36 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 14 / 26

Classical probability Drawbacks What if outcomes are not equally likely? What if there are infinitely many outcomes? Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 15 / 26

Empirical probability Use long-term frequencies of different outcomes to estimate their probabilities. Flip a coin a lot of times. Use the fraction of times it comes up heads to estimate the probability of heads. 520 heads out of 1000 flips leads to estimating P(heads) = 0.520. This estimate is only approximate because Due to random variation, the numerator will fluctuate. Precision is limited by the denominator. 1000 flips can only estimate it to three decimals. More on this later in the course in Chapter 5.3. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 16 / 26

Empirical probability E. coli has been sequenced: Position: 1 2 3 4 5 6 7 8 9 10 Base: G C T T T T C T On the forwards strand: # s 1,142,136 P() = 1,142,136 4,639,221 0.2461913326 # C s 1,179,433 P(C) 0.2536578878 # G s 1,176,775 P(G) 0.2542308288 # T s 1,140,877 P(T) 0.2459199508 Total 4,639,221 1 Sample space: set of positions S = {1, 2,..., 4639221} Event is the set of positions with nucleotide (similar for C, G, T). = {1, 9,...} C = {3, 8,...} G = {2,...} T = {4, 5, 6, 7, 10,...} Simplistic model: the sequence is generated from a biased 4-sided die with faces, C, G, T. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 17 / 26

xiomatic probability definition of a probability function P based on events and the following axioms is the most useful. Each event S is assigned a probability that obeys these axioms: xioms for a finite sample space For any event S: P() 0 The total sample space has probability 1: P(S) = 1 For mutually exclusive events and B: dditional axiom for an infinite sample space P( B) = P() + P(B) If 1, 2,... (infinitely many) ( are pairwise mutually exclusive, then ) P i = P( i ) i=1 i=1 i=1 i = 1 2 is like notation, but for unions. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 18 / 26

xiomatic probability additional properties dditional properties of the probability function follow from the axioms. P( c ) = 1 P() Example: P(die roll = 1) = 1 6 P(die roll 1) = 1 1 6 = 5 6 Proof. and c are mutually exclusive, so P( c ) = P() + P( c ). lso, P( c ) = P(S) = 1. Thus, P( c ) = 1 P(). c P( ) = 0 Proof: P( ) = P(S c ) = 1 P(S) = 1 1 = 0 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 19 / 26

xiomatic probability additional properties If B then P() P(B) Proof: Write B = ( c B). and c B are mutually exclusive, so P(B) = P() + P( c B). The first axiom gives P( c B) 0, so P(B) P(). B c U B P() 1 Proof: S so P() P(S) = 1. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 20 / 26

xiomatic probability additional properties 1 2 3 dditive Law If 1, 2,..., n are pairwise ( mutually exclusive, then n ) n P i = P( i ) n i=1 i=1 i=1 i = 1 2 n is like notation, but for unions. Prove for all integers n 1 using induction, based on: P ( ) ( 1 n ) n+1 = P(1 n ) + P( n+1 ) = ( P( 1 ) + + P( n ) ) + P( n+1 ). Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 21 / 26

xiomatic probability additional properties 1 2 3 Induction proves the dditive Law for positive integers n = 1, 2,..., but not for n =, so we have to introduce an additional axiom for that: dditional axiom for an infinite sample space If 1, 2,... (infinitely many) are pairwise mutually exclusive, then ( ) P i = i=1 P( i ) i=1 Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 22 / 26

xiomatic probability additional properties c B U B B U U c U B B De Moivre s Law: P( B) = P() + P(B) P( B) Proof: P() = P( B) + P( B c ) P(B) = P( B) + P( c B) P() + P(B) = P( B) + ( P( B) + P( B c ) + P( c B) ) = P( B) + P( B) c U B c Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 23 / 26

xiomatic probability additional properties dditional properties of the probability function follow from the axioms: P( c ) = 1 P() P( ) = 0 If B then P() P(B) P() 1 dditive Law: If 1, 2,..., n are pairwise mutually exclusive ( i j = for all i j) then ( n ) P i = i=1 n P( i ) i=1 (The first three axioms only lead to a proof of this for finite n, but not n =, so n = has to be handled by an additional axiom.) De Moivre s Law: P( B) = P() + P(B) P( B) Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 24 / 26

Generalizing Sigma ( ) notation Sum over sets instead of over a range of consecutive integers Consider a biased n-sided die, with faces 1,..., n. The probability of face i is q i, with 0 q i 1, and q 1 + + q n = 1: n q i = 1 i=1 For n = 8, the probability of an even number is: P(even number) = q s = q 2 + q 4 + q 6 + q 8 s {2,4,6,8} For a 26-sided die with faces,b,...,z, the total probability is P({, B,..., Z}) = q s = q + q B + + q Z = 1, s {,B,...,Z} and the probability of a vowel is P({, E, I, O, U}) = q s = q + q E + q I + q O + q U. s {,E,I,O,U} Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 25 / 26

Finite discrete sample space Each outcome has a probability between 0 and 1 and the probabilities add up to 1: 0 P(s) 1 for each s S P(s) = 1 For an event S, define P() = s P(s). Examples DN biased n-sided die (previous slide). s S For flipping a coin 3 times, P({HHT, HTH, THH}) = P(HHT) + P(HTH) + P(THH). Generate a random DN sequence by rolling a biased CGT-die. Deviations from what s expected from random rolls of a dice are used to detect structure in the DN sequence, like genes, codons, repeats, etc. Prof. Tesler Ch. 2.3-2.4 Intro to Probability Math 186 / Winter 2019 26 / 26