Discrete Probability

Similar documents
CSC Discrete Math I, Spring Discrete Probability

With Question/Answer Animations. Chapter 7

Chapter Summary. 7.1 Discrete Probability 7.2 Probability Theory 7.3 Bayes Theorem 7.4 Expected value and Variance

Monty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch

Discrete Structures for Computer Science

Sec$on Summary. Assigning Probabilities Probabilities of Complements and Unions of Events Conditional Probability

Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya

Discrete Random Variables

Discrete Random Variables

CS 441 Discrete Mathematics for CS Lecture 19. Probabilities. CS 441 Discrete mathematics for CS. Probabilities

Notes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 16 Notes. Class URL:

Notes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 15 Notes. Class URL:

Probability (10A) Young Won Lim 6/12/17

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Topics in Discrete Mathematics

Conditional Probability. CS231 Dianna Xu

Conditional probability

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

Probabilistic models

Conditional Probability

Probability Distributions. Conditional Probability.

Deviations from the Mean

TOPIC 12 PROBABILITY SCHEMATIC DIAGRAM

Probabilistic models

Quantitative Methods for Decision Making

Denker FALL Probability- Assignment 6

Lecture 4: Probability and Discrete Random Variables

Introduction to Probability 2017/18 Supplementary Problems

What is Probability? Probability. Sample Spaces and Events. Simple Event

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Elementary Discrete Probability

First Digit Tally Marks Final Count

Expectation, inequalities and laws of large numbers

Randomized Algorithms

Discrete Probability

Problems and results for the ninth week Mathematics A3 for Civil Engineering students

ORF 245 Fundamentals of Statistics Chapter 5 Probability

ELEG 3143 Probability & Stochastic Process Ch. 1 Probability

Notes for Math 324, Part 17

STAT:5100 (22S:193) Statistical Inference I

Discrete Distributions

Lower bound for sorting/probability Distributions

Formalizing Probability. Choosing the Sample Space. Probability Measures

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Overview. CSE 21 Day 5. Image/Coimage. Monotonic Lists. Functions Probabilistic analysis

MATH1231 Algebra, 2017 Chapter 9: Probability and Statistics

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Homework 4 Solution, due July 23

Independence, Variance, Bayes Theorem

Notes. Combinatorics. Combinatorics II. Notes. Notes. Slides by Christopher M. Bourke Instructor: Berthe Y. Choueiry. Spring 2006

14 - PROBABILITY Page 1 ( Answers at the end of all questions )

Week 12-13: Discrete Probability

Relationship between probability set function and random variable - 2 -

Midterm 2. Your Exam Room: Name of Person Sitting on Your Left: Name of Person Sitting on Your Right: Name of Person Sitting in Front of You:

Lecture 3. January 7, () Lecture 3 January 7, / 35

Introduction to Probability Theory, Algebra, and Set Theory

CMPSCI 240: Reasoning Under Uncertainty

Probability Distributions. Conditional Probability

Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin,

Probability Notes (A) , Fall 2010

CS 361: Probability & Statistics

Chapter 8 Sequences, Series, and Probability

What is the probability of getting a heads when flipping a coin

Lecture 3 Probability Basics

Section 13.3 Probability

Lecture 04: Conditional Probability. Lisa Yan July 2, 2018

Discrete Random Variable

CS 125 Section #12 (More) Probability and Randomized Algorithms 11/24/14. For random numbers X which only take on nonnegative integer values, E(X) =

Chapter 3 : Conditional Probability and Independence

Senior Math Circles November 19, 2008 Probability II

Expected Value 7/7/2006

STA 2023 EXAM-2 Practice Problems From Chapters 4, 5, & Partly 6. With SOLUTIONS

Statistical Theory 1

STA 2023 EXAM-2 Practice Problems. Ven Mudunuru. From Chapters 4, 5, & Partly 6. With SOLUTIONS

Axiomatic Foundations of Probability. Definition: Probability Function

ACM 116: Lecture 2. Agenda. Independence. Bayes rule. Discrete random variables Bernoulli distribution Binomial distribution

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya

1 True/False. Math 10B with Professor Stankova Worksheet, Discussion #9; Thursday, 2/15/2018 GSI name: Roy Zhao

Great Theoretical Ideas in Computer Science

Question Points Score Total: 137

Some Basic Concepts of Probability and Information Theory: Pt. 1

1. Discrete Distributions

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

6.042/18.062J Mathematics for Computer Science November 28, 2006 Tom Leighton and Ronitt Rubinfeld. Random Variables

Some Special Discrete Distributions

PROBABILITY.

Basic Statistics for SGPE Students Part II: Probability theory 1

Probability and Statistics Notes

CISC 1100/1400 Structures of Comp. Sci./Discrete Structures Chapter 7 Probability. Outline. Terminology and background. Arthur G.

PERMUTATIONS, COMBINATIONS AND DISCRETE PROBABILITY

UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis

Lecture 10. Variance and standard deviation

6.041/6.431 Spring 2009 Quiz 1 Wednesday, March 11, 7:30-9:30 PM. SOLUTIONS

Independence. Conditional Probability. Conditional Probability. Independence. SCONE Lab. Definition: Two events E and F are independent iff

Week 6, 9/24/12-9/28/12, Notes: Bernoulli, Binomial, Hypergeometric, and Poisson Random Variables

Chapter. Probability

Conditional Probability & Independence. Conditional Probabilities

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Solutions 5 Spring 2006

Transcription:

MAT 258 Discrete Mathematics Discrete Probability Kenneth H. Rosen and Kamala Krithivasan Discrete Mathematics 7E Global Edition Chapter 7 Reproduced without explicit consent Fall 2016 Week 11

Probability An experiment is a procedure that yields one of a given set of possible outcomes. The sample space of the expermient is the set of possible outcomes. An event is a subset of the same space. If S is a finite nonempty sample space of equally likely outcomes, and E is an event that is, a subspace of S then the probability of E is p(e) = E. Thus, 0 p(e) 1. S An urn contains four blue balls and five red balls. What is the probability that a ball chosen at random from the urn is blue? 4/9 = 0. 4 What is the probability that when two dice are rolled, the sum of the numbers on the two dice is 7? (1/36+1/36+1/36) 2 = 1/6 = 0.1 6

Lotteries In a lottery, players win a large prize when they pick four digits that match, in the correct order, four digits selected by a random mechanical process. What is the probability that a player wins the large prize? 1/10000 = 0.0001 A smaller prize is won if only three digits are matched. What is the probability that a player wins the small prize instead? 1/1000 9/10 4 = 36/10000 = 0.0036 There are many lotteries that award enormous prizes to people who correctly choose a set of six numbers from out of the first n positive integers, where n is usually between 30 and 60. What is the probability that a person picks the correct six numbers out of 40? one out of C(40,6) = 3838380 0.00000026

Poker Hands Find the probability that a hand of five cards in poker contains four cards of one kind. C(13, 1)C(48, 1) C(52, 5) = 13 48 2598960 0.00024 What is the probability that a poker hand contains a full house that is, three of one kind, and two of another kind? C(13, 1)C(4, 3)C(12, 1)C(4, 2) C(52, 5) = 13 4 12 6 2598960 0.0014

Example 7 What is the probability that the numbers 11, 4, 17, 39, and 23 are drawn in that order from a bin containing 50 balls labeled with the numbers 1, 2,..., 50 if a ball selected is not returned to the bin before the next ball is selected? one out of P(50,5) = 254251200 0.000000003933 What is the probability that the numbers 11, 4, 17, 39, and 23 are drawn in that order from a bin containing 50 balls labeled with the numbers 1, 2,..., 50 if a ball selected is returned to the bin before the next ball is selected? one out of 50 5 = 312500000, for a probability of 0.0000000032

Complementary and Overlapping Events Let E be an event in a sample space S. The probability of the event Ē = S E, the complementary event of E, is given by p(ē) = 1 p(e). A sequence of 10 bits is randomly generated. What is the probability that at least one of these bits is 0? 1 1/(2 10 ) = 0.9990234375 Let E 1 and E 2 be events in the sample space S. Then p(e 1 E 2 ) = p(e 1 )+p(e 2 ) p(e 1 E 2 ). What is the probability that a positive integer selected at random from the set of positive integers not exceeding 100 is divisible by either 2 or 5? 50+20 10 = 0.6 100

Monty Hall Problem Suppose you are a game show contestant. You have a chance to win a large prize. You are asked to select one of three doors to open; the large prize is behind one of the three doors, and the other two doors are losers. Once you select a door, the game show host, who knows what s behind each door, does the following: whether or not you selected the winning door, he opens one of the other two doors that he knows is a losing door (selecting at random if both are losing doors); he asks you whether you would like to switch doors. What strategy should you use: should you change doors or keep your original selection, or does it not matter? There s a 2/3 chance that switching will get the prize, while keeping has a 1/3 chance of winning the prize.

Probability Distribution Let S be the sample space of an experiment with finite or countable number of outcomes. We assign a probability p(s) to each outcome s, such that 0 p(s) 1 and s Sp(s) = 1. The function p from the set of all outcomes of the sample space S is called a probability distribution. What probabilities should be assigned to the outcomes H (heads) and T (tails) when a fair coin is flipped? 1/2 each What probabilities should be assigned to these outcomes when the coin is biased so that heads comes up twice as often as tails? p(h) = 2/3, p(t) = 1/3 Suppose that S is a set with n elements. The uniform distribution assigns the probability 1/n to each element of S.

Probability of an Event The probability of an event E is the sum of the probabilities of the outcomes of E: p(e) = p(s). s E The experiment of selecting an element from a sample space S with a uniform distribution is called selecting an element of S at random. Suppose that a die is biased (or loaded) so that 3 appears twice as often as each other number, but that the other five outcomes are equally likely. What is the probability that an odd number appears when we roll this die? 4/7

Conditional Probability If E 1, E 2,..., is a sequence of pairwise disjoint events in a sample space S, then ( ) p E i = p(e i ). i i Let E and F be events with p(f) > 0. The conditional probability of E given F, denoted by p(e F), is defined as p(e F) p(e F) =. p(f)

Examples 3 4 A bit string of length four is generated at random so that each of the 16 bit strings of length four is equally likely. What is the probability that it contains at least two consecutive 0s, given that its first bit is 0? (Assume 0 and 1 bits are equally likely.) 5/16 = 5/8 = 0.675: 0000, 0001, 0010, 0011, 0100 1/2 What is the conditional probability that a family with two children has two boys, given that they at least have one boy? Assume that each of the possibilities BB, BG, GB, GG are equally likely, where B represents a boy, and G represents a girl, and BG represents a family with an older boy and a younger girl. 1/4 = 1/3 = 0. 3 3/4

Independence The events E and F are independent if and only if p(e F) = p(e)p(f). Suppose E is the event that a randomly generated bit string of length four begins with a 1 and F is the event that this bit string contains an even number of 1s. Are E and F independent, if the 16 bit strings are equally likely? p(e) = p(f) = 1 2, p(e F) = 1 = p(e)p(f), so the events are 4 independent. The events E 1, E 2,..., E n are pairwise independent if and only if p(e i E j ) = p(e i )p(e j ) for i,j N with 1 i < j n. These ( events ) are mutually independent if m p E ij = m p(e ij ) whenever i j N, i 1 < < i m n. j=1 j=1

Children and Gender Suppose that each of the four ways a family can have children (BB, BG, GB, GG) are equally likely. Are the events E, that a family with two children has two boys, and F, that a family with two children has at least one boy, independent? p(e) = 1 4, p(f) = 3 4, p(e F) = 1, so the events are not 4 independent. Are the events E, that a family with three children has children of both sexes, and F, that this family has at most one boy, independent? Assume that the eight ways a family can have children are equally likely. p(e) = 3 4, p(f) = 1 2, p(e F) = 3 8, so the events are independent.

Bernoulli Trials Each performance of an experiment with two possible outcomes (success or failure) is called a Bernoulli trial. If the probability of success is p and the probability of failure is q, then p+q = 1. Bernoulli trials are mutually independent if the conditional probability of success if p, given any information whatsoever about the outcomes of other trials. A coin is biased so that the probability of heads is 2 / 3. What is the probability that exactly four heads come up when the coin is flipped seven times, assuming that the flips are independent? C(7,4)(2/3) 4 (1/3) 3 = 7 6 5 16 6 3 7 = 560 2187 0.256.

Binomial Distribution The probability of exactly k successes in n independent Bernoulli trials, with probability of success p and probability of failure q = 1 p is C(n,k)p k q n k = B(k;p,q). Suppose that the probability that a 0 bit will be generated is 0.9, that the probability that a 1 bit is generated is 0.1, and that bits are generated independently. What is the probability that exactly eight 0 bits are generated when 10 bits are generated? C(10,8)(0.9) 8 (0.1) 2 = 0.1937102445

Random Variables A random variable is a function from the sample space of an experiment to the set of real numbers. That is, a random variable assigns a real number to each possible outcome. Suppose a coin is flipped three times. Let X(t) be the random variable that equals the number of heads that appear when t is the outcome. Then X(t) takes the following values: X(HHH) = 3, X(THH) = X(HTH) = X(HHT) = 2, X(TTH) = X(THT) = X(HTT) = 1, X(TTT) = 0. The distribution of a random variable X on a sample space S is the set of pairs (r,p(x = r)) for all r X(S), where p(x = r) is the probability that X takes the value r. Each of the eight possible outcomes when a fair coin is flipped three times has the probability 1 / 8. So, the distribution of X above is (3,1/8),(2,3/8),(1,3/8),(0,1/8).

Random Variable on Two Standard Dice Let X be the sum of the numbers that appear when a pair of dice are rolled. What are the values of this random variable for the 36 possible outcomes (i,j), where i and j are the numbers that appear on the first die and the second die, respectively, when these two dice are rolled? X 1 2 3 4 5 6 1 2 3 4 5 6 7 2 3 4 5 6 7 8 3 4 5 6 7 8 9 4 5 6 7 8 9 10 5 6 7 8 9 10 11 6 7 8 9 10 11 12

Birthday Problem What is the minimum number of people who need to be in a room so that the probability that at least two of them have the same birthday is greater than 1 / 2? 366 365 364 344 366 23 0.494, so n = 23. Hashing functions map a large universe of keys to a much smaller set of storage locations. A good hashing function yields few collisions. What is the probability that no two (of n) keys are mapped to the same (among m) locations by a hashing function? m m m 1 m n+1 = P(m,n) m m m n Given m locations, how many keys are needed to have about an even probability of a collision occuring? If n 1.177 m, for n possible keys, the possibility of collision is greater than 1 / 2.

Monte Carlo Algorithms A Monte Carlo algorithm for a decision problem uses a sequence of tests. The probability that the algorithm answers the decision problem correctly increases as more tests are carried out. At each step of the algorithm, possible responses are true, which means that the answer to the decision problem is true, and no additional iterations are needed, or unknown, which means that the answer to the decision problem could be either true or false. After running all the iterations in such an algorithm, the final answer produced is true if at least one iteration yields the answer true, and the answer is false if every iteration yields the answer unknown. If the correct answer is false, then the algorithm answers false, because every iteration will yield unknown. However, if the correct answer is true, then the algorithm could answer true or false, because it may be possible that each iteration produced the response unknown.

Quality Testing Suppose that a manufacturer orders processor chips in batches of size n N. The chip maker has tested only some of these batches to make sure that all the chips in the batch are good. In previously untested batches, the probability that a particular chip is bad has been observed to be 0.1 when random testing is done. The PC manufacturer wants to decide whether all the chips in a batch are good. Use a Monte Carlo algorithm to answer Has this batch of chips not been tested by the chip maker? Randomly select a chip from the batch to check. If the chip selected is bad, answer True and stop. Otherwise, answer Unknown and go to the next chip. After k chips, if not True, answer False. The algorithm answers False (this batch has been tested it has no bad chips) incorrectly with probability 0.9 k k = 66 gives probability less than 1 in 1000.

Primality Testing Let n N and n = 1+2 s t, where t is an odd positive integer. n passes Miller s test for base b if b t 1 (mod n) or b 2jt 1 (mod n) for some j, 0 j s 1. It can be shown that a composite integer n passes Miller s test for fewer than n/4 bases between 1 and n. A probabilistic primality test, answering Is n composite? : Randomly select a base b, 1 < b < n. If n fails Miller s test for base b, answer True and stop. Otherwise, answer Unknown and randomly select another base b, 1 < b < n. After k bases, if not True, answer False. The algorithm answers False (n is prime) incorrectly with probability (1/4) k k = 10 gives probability less than 1 in 1000000.

Probabilistic Method If the probability that an element chosen at random from a set S does not have a particular property is less than 1, there exists an element in S with that property. If k is an integer with k 2, then R(k,k) 2 k/2, where R(k,k) is the minimum number of people that are pairwise friends or enemies such that there is a group of k mutual friends or enemies.

Bayes s Theorem Suppose that E and F are events from a sample space S such that p(e) 0 and p(f) 0. Then p(f E) = p(e F)p(F) p(e F)p(F)+p(E F)p( F). We have two boxes. The first contains two green balls and seven red balls; the second contains four green balls and three red balls. Bob selects a ball by first choosing one of the two boxes at random. He then selects one of the balls in this box at random. If Bob has selected a red ball, what is the probability that he selected a ball from the first box? Let E be the event that Bob has chosen a red ball, and F be the event that Bob has selected the a ball from the first box. Then p(f) = 1 2 = p( F), p(e F) = 7 9, p(e F) = 3 7. From Bayes s Theorem, p(f E) = 49/76 0.645.

Example 2 Suppose that one person in 100000 has a particular rare disease for which there is a fairly accurate diagnostic test. This test has 1.0% false negatives and 0.5% false positives. Given this information, can we find the probability that a person who tests positive for the disease has the disease? Let E be the event that the test is positive, and F be the 1 event that the person has the disease. Then p(f) = 100000, p( F) = 99999 99, p(e F) = 100000 100, p(e 1 F) = 200. From Bayes s Theorem, p(f E) 0.002. Find the probability that a person who tests negative for the disease does not have the disease. From above, p(ē F) = 1 100, p(ē 199 F) =. From Bayes s 200 Theorem, p( F Ē) 0.9999999.

Generalized Bayes s Theorem Suppose that E is an event from a sample space S and that F 1,F 2,...,F n are mutually exclusive events such that n F i = S. Assume that p(e) 0 and p(f i ) 0 for i=1 i = 1,2,...,n. Then p(f j E) = p(e F j)p(f j ). n p(e F i )p(f i ) i=1

Bayesian Spam Filter Suppose we have a set of B of messages known to be spam and a set G of messages known not to be spam. Identify the words that occur in B and in G: count the number of messages in the set containing each word w to find the number of messages containing the word w in the sets B and G, n B (w) and n G (w), respectively. Suppose we receive a new e-mail message containing the word w. Let S be the event that the message is spam. Let E be the event that the message contains the word w. If the likelihood of an email is spam is 1/2, then r(w) = p(s E) = p(w) p(w)+q(w) = n B (w)/ B (n B (w)/ B ) +(n G (w)/ G ). If r(w) is greater than a threshold to be set, say 0.9, the message is classified as spam.

Example 3 Suppose that the word Rolex occurs in 250 of 2000 messages known to be spam and in 5 of 1000 messages known not to be spam. Estimate the probability that an incoming message containing the word Rolex is spam, assuming that it is equally likely that an incoming message is spam or not spam. p(w) = 1 1 and q(w) =, so r(w) 0.962. 8 200

Example 4 Suppose a Bayesian spam filter has been trained on a set of 2000 spam messages and 1000 messages that are not spam. The word stock appears in 400 spam messages and 60 messages that are not spam, and the word undervalued appears in 200 spam messages and 25 messages that are not spam. Evaluate the probability that an incoming message containing both the words stock and undervalued is spam, assuming no prior knowledge whether it is spam. p(stock) = 1 5, q(stock) = 3 50, p(undervalued) = 1 10, q(undervalued) = 1 40. So r(stock, undervalued) = p(stock)p(undervalued) p(stock)p(undervalued) + q(stock)q(undervalu 0.93.

Expected Values The expected value of a random variable is the sum over all elements in a sample space of the product of the probability of the element and the value of the random variable at this element it is a weighted average of the values of a random variable. The expected value, also called the expectation or mean, of the random variable X on the sample space S is equal to E(X) = s Sp(s)X(s). The deviation of X at s S is X(s) E(X), the difference between the value of X and the man of X.

Examples 1 2 Let X be the number that comes up when a fair die is rolled. What is the expected value of X? 1+2+3+4+5+6 = 3.5 6 A fair coin is flipped three times. Let S be the sample space of the eight possible outcomes, and let X be the random variable that assigns to an outcome the number of heads in this outcome. What is the expected value of X? 0+1+1+1+2+2+2+3 = 1.5 8

EV of a Random Variable If X is a random variable and p(x = r) is the probability that X(s)=r X = r, so that p(x = r) = p(s), then s S E(X) = r X(S) p(x = r)r. What is the expected value of the sum of the numbers that appear when a pair of fair dice is rolled? (2+12) 1 36 +(3+11) 1 18 +(4+10) 1 12 +(5+9) 1 9 + (6+8) 5 36 +7 1 6 = 7 The expected number of successes when n mutually independent Bernoulli trials are performed, where p is the probability of success on each trial, is np.

Linearity of Expectations If X i, i = 1,2,...,n with n N, are random variables on S, and if a,b R, then E(X 1 +X 2 + +X n ) = E(X 1 )+E(X 2 )+ +E(X n ) E(aX +b) = ae(x)+b Find the expected value of the sum of the numbers that appear when a pair of fair dice. E(X 1 ) = E(X 2 ) = 3.5 so E(X 1 +X 2 ) = 7 Find the expected value of the number of successes when n Bernoulli trials are performed, where p is the probability of success. E(X 1 ) = = E(X n ) = p, so E(X 1 )+ +E(X n ) = np

Example 6 A new employee checks the hats of n people at a restaurant, forgetting to put claim check numbers on the hats. When customers return for their hats, the checker gives them back hats chosen at random from the remaining hats. What is the expected number of hats that are returned correctly? Let X i be a random variable corresponding to whether the ith person has the correct hat (1) or not (0). Then X = n X i corresponds to the number of correctly returned i=1 hats. Note E(X i ) = 1/n, so E(X) = 1.

Example 7 The ordered pair (i,j) is called an inversion in a permutation of the first n positive integers if i < j but j precedes i in the permutation. Let I i,j be the random variable over the set of all the permutations of the first n positive integers corresponding to whether (i,j) is an inversion in the permutation (1) or not (0). Then X = I i,j counts the number of inversions of a 1 i<j n random permutation. Since any particular inversion is as likely as it is not, E(I i,j ) = 1/2, and since there are ( n 2) such pairs, then E(X) = ( ) n 1 2 2 = n(n 1). 4

Average-Case Complexity The average-case complexity of an algorithm is the expected value of the complexity over cases being considered. Linear search on a list with n items has an average-case complexity of O(n) comparisons. Insertion sort on a list with n items has an average-case complexity of O(n 2 ) comparisons.

Geometric Distribution Suppose that the probability that a coin comes up heads is p. The coin is flipped repeatedly until it comes up heads. What is the expected number of flips until this coin comes up heads? p+2(1 p)p+3(1 p) 2 p+ +k(1 p) k 1 p+ = 1+(1 p)+(1 p) 2 + = 1 p A random variable has geometric distribution with parameter p if p(x = k) = (1 p) k 1 p for k = 1,2,3,..., where p is a real number and 0 p 1. If the random variable X has the geometric distribution with parameter p, then E(X) = 1 p.

Independent Random Variables The random variables X and Y on a sample space S are independent if p(x = r 1 and Y = r 2 ) = p(x = r 1 )p(y = r 2 ) for all real numbers r 1 and r 2. Are the random variables X i independent, if X i is the number on the ith die, i = 1,2? For 1 i,j 6, P(X 1 = i and X 2 = j) = 1/36 = p(x 1 = i)p(x 2 = j), so X 1 and X 2 are independent. If X = X 1 +X 2, then p(x 1 = 1) = 1/6, p(x = 12) = 1/36 and p(x 1 = 1 and X = 12) = 0, so X 1 and X are not independent.

Expectation of Independent Random Variables If X and Y are independent random variables on a sample space S, then E(XY) = E(X)E(Y ). Let X and Y be the number of heads and tails, respectively, when a coin is flipped up. Note that E(X) = E(Y) = 1, but E(XY) = 1 2, so E(XY) E(X)E(Y) and X and Y are not independent.

Variance Let X be a random variable on a sample space S. The variance of X is V(X) = s S(X(s) E(X)) 2 p(s) that is, V(X) is the weighted average of the square of the deviation of X. The standard deviation of X is σ(x) = V(X). If X is a random variable on a sample space S, V(X) = E(X 2 ) E(X) 2. If X is a random variable on a sample space S and E(X) = µ, then V(X) = E((X µ) 2 ).

What is the variance of the random variable X((i,j)) = 2i, where i is the number appearing on the first die and j is the number appearing on the second die, when rolling two fair dice? E(X) = 7 and E(X 2 ) = 182/3, so V(X) = E(X 2 ) E(X) 2 = 35/3. Examples 14 16 What is the variance of the random variable X with X(t) = 1 if a Bernoulli trial is a success and X(t) = 0 if it is a failure, where p is the probability of success and q is the probability of failure? Since X(t) 2 = X(t), V(X) = E(X 2 ) E(X) 2 = E(X)(1 E(X)) = pq. What is the variance of the random variable X, where X is the number that comes up when a fair die is rolled? E(X) = 7/2 and E(X 2 ) = 91/6, so V(X) = E(X 2 ) E(X) 2 = 35/12.

Bienaymé s Formula If X i, i = 1,2,...,n, with n N, are pairwise independent random variables on S, then V( n X i ) = n V(X i ). i=1 Find the variance and standard deviation of the random variable X whose value when two fair dice are rolled is X((i,j)) = i+j, where i is the number appearing on the first die and j is the number appearing on the second die. If X 1 ((i,j)) = i and X 2 ((i,j)) = j, these are independent, and V(X) = V(X 1 +X 2 ) = V(X 1 )+V(X 2 ) = 35/6, σ(x) = 35/6. What is the variance of the number of successes when n independent Bernoulli trials are performed, where, on each trial, p is the probability of success and q is the probability of failure? If X i is the random variable on the success (1) or failure (0) of the ith Bernoulli trial, and X = n Xi, then V(X) = npq. i=1

Chebyshev s Inequality Let X be a random variable on a sample space S with probability function p. If r is a positive real number, then p( X(s) E(X) r) V(X) r 2. Suppose that X is the random variable that counts the number of tails when a fair coin is tossed n times. It follows that E(X) = n/2 and V(X) = n/4. Then, p( X(s) n/2 n) 1/4. Let X be the random variable whose value is the number appearing when a fair die is rolled. Then E(X) = 7/2 and V(X) = 35/12. Then, p( X 7/2 r) 35/(12r 2 ). Note, if r = 3, p( X 7/2 3) 35/108 0.324, while p( X 7/2 3) = 0.