Notes 6 Autumn Example (One die: part 1) One fair six-sided die is thrown. X is the number showing.

Similar documents
Steve Smith Tuition: Maths Notes

Probability Theory and Simulation Methods

RVs and their probability distributions

Chap 4 Probability p227 The probability of any outcome in a random phenomenon is the proportion of times the outcome would occur in a long series of

Notes 12 Autumn 2005

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Chapter 1 Probability Theory

Monty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch

What is a random variable

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

Statistical Theory 1

Part (A): Review of Probability [Statistics I revision]

Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

Class 26: review for final exam 18.05, Spring 2014

Recap. The study of randomness and uncertainty Chances, odds, likelihood, expected, probably, on average,... PROBABILITY INFERENTIAL STATISTICS

Sociology 6Z03 Topic 10: Probability (Part I)

Recap of Basic Probability Theory

Recap of Basic Probability Theory

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

OCR Statistics 1 Probability. Section 1: Introducing probability

Discrete Random Variables. David Gerard Many slides borrowed from Linda Collins

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

Discrete Random Variables

Probability, Random Processes and Inference

TOPIC 12: RANDOM VARIABLES AND THEIR DISTRIBUTIONS

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

6.2 Introduction to Probability. The Deal. Possible outcomes: STAT1010 Intro to probability. Definitions. Terms: What are the chances of?

STEP Support Programme. Statistics STEP Questions: Solutions

Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

Homework 4 Solution, due July 23

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

1 Random variables and distributions

HW2 Solutions, for MATH441, STAT461, STAT561, due September 9th

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008

9/6/2016. Section 5.1 Probability. Equally Likely Model. The Division Rule: P(A)=#(A)/#(S) Some Popular Randomizers.

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, < x <. ( ) X s. Real Line

Mathematics. ( : Focus on free Education) (Chapter 16) (Probability) (Class XI) Exercise 16.2

1 Generating functions

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

Solutionbank S1 Edexcel AS and A Level Modular Mathematics

Sec$on Summary. Assigning Probabilities Probabilities of Complements and Unions of Events Conditional Probability

Discrete Random Variables. Discrete Random Variables

Section 13.3 Probability

9. DISCRETE PROBABILITY DISTRIBUTIONS

Dept. of Linguistics, Indiana University Fall 2015

Answers to All Exercises

UNIT 4 MATHEMATICAL METHODS SAMPLE REFERENCE MATERIALS

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r).

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Topic 3: The Expectation of a Random Variable

Outline. Probability. Math 143. Department of Mathematics and Statistics Calvin College. Spring 2010

Edexcel past paper questions

Conditional Probability

Review of Probability. CS1538: Introduction to Simulations

Notice how similar the answers are in i,ii,iii, and iv. Go back and modify your answers so that all the parts look almost identical.

MODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES DISTRIBUTION FUNCTION AND ITS PROPERTIES

Deviations from the Mean

Probability Year 9. Terminology

CS 361: Probability & Statistics

6.041/6.431 Spring 2009 Quiz 1 Wednesday, March 11, 7:30-9:30 PM. SOLUTIONS

Discrete Probability Refresher

Introduction to discrete probability. The rules Sample space (finite except for one example)

The probability of an event is viewed as a numerical measure of the chance that the event will occur.

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Notation: X = random variable; x = particular value; P(X = x) denotes probability that X equals the value x.

Basic Probability. Introduction

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

Chapter 4. Probability-The Study of Randomness

CMPSCI 240: Reasoning Under Uncertainty

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.

Fundamental Tools - Probability Theory II

CHAPTER 3 RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. 3.1 Concept of a Random Variable. 3.2 Discrete Probability Distributions

CSC Discrete Math I, Spring Discrete Probability

Probability Year 10. Terminology

Random variables (discrete)

1 Normal Distribution.

Discrete Probability Distributions

18440: Probability and Random variables Quiz 1 Friday, October 17th, 2014

Random Variables and Probability Distributions

MATH Solutions to Probability Exercises

Discrete Random Variable Practice

Lecture 8: Continuous random variables, expectation and variance

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

CS 441 Discrete Mathematics for CS Lecture 20. Probabilities. CS 441 Discrete mathematics for CS. Probabilities

Introduction to Probability, Fall 2009

SDS 321: Introduction to Probability and Statistics

Notes on Mathematics Groups

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages


18.05 Practice Final Exam

Expected Value. Lecture A Tiefenbruck MWF 9-9:50am Center 212 Lecture B Jones MWF 2-2:50pm Center 214 Lecture C Tiefenbruck MWF 11-11:50am Center 212

P (A) = P (B) = P (C) = P (D) =

Transcription:

MAS 08 Probability I Notes Autumn 005 Random variables A probability space is a sample space S together with a probability function P which satisfies Kolmogorov s aioms. The Holy Roman Empire was, in the words of the historian Voltaire, neither holy, nor Roman, nor an empire. Similarly, a random variable is neither random nor a variable: A random variable is a function from a probability space to the real numbers. Notation We usually use capital letters like X, Y, Z to denote random variables. Eample (Child:part ) One child is born, equally likely to be a boy of a girl. Then S = {B,G}. We can define the random variable X by X(B) = and X(G) = 0. Eample (One die: part ) One fair si-sided die is thrown. X is the number showing. Eample (Three coin tosses: part ) I toss a fair coin three times; I count the number of heads that come up. Here, the sample space is S = {HHH,HHT,HT H,HT T,T HH,T HT,T T H,T T T }; the random variable X is the function which counts the number of Hs in each outcome, so that X(T HH) =, X(T T T ) = 0, etc. If is any real number, the event X = is defined to be the event {s S : X(s) = }, which can also be written as X (). The probability that X =, which is written P(X = ), is just P({s S : X(s) = }). Similarly, other events can be defined in terms of the value(s) taken by X: for eample, P(X y) means P({s S : X(s) y}).

Notation P(X = ) is written p(), or p X () if we need to emphasize the name of the random variable. In handwriting, be careful to distinguish between X and. The list of values (, p()), for those such that there is an outcome in S with X(s) =, is called the distribution of X. The function p() is called the probability mass function of X, often abbreviated to pmf. Often we concentrate on the distribution or on the probability mass function and forget the sample space. Sometimes the pmf is given by a table of values, sometimes by a formula. We can also draw the line graph of a pmf. p(y) p() y values of X Eample (Child: part ) The distribution of X is 0 and the line graph is p() 0 0

Eample (One die: part ) The distribution of X is 3 4 5 and the line graph is p() 0 3 4 5 Eample A biased coin is tossed five times. Each time it has probability p of coming down heads, independently of all other times. Let X be the number of heads. Then P(X = m) = 5 C m p m q 5 m for m = 0,..., 5, where q = p. So the pmf is: m 0 3 4 5 p(m) q 5 5pq 4 0p q 3 0p 3 q 5p 4 q p 5 Different values of p give different distributions. Here are three eamples, together with their line graphs. p = 0.5 m 0 3 4 5 p(m) 0.03 0.5 0.33 0.33 0.5 0.03 0. 0.4 0. 0 3 4 5 3

p = 0. m 0 3 4 5 p(m) 0.00 0.077 0.30 0.34 0.59 0.078 0. 0.4 0. 0 3 4 5 p = 0.9 m 0 3 4 5 p(m) 0.000 0.000 0.008 0.073 0.38 0.590 0. 0.4 0. 0 3 4 5 A random variable X is discrete if either (a) {X(s) : s S} is finite, that is, X takes only finitely many values, or (b) {X(s) : s S} is infinite but the values X can take are separated by gaps: formally, there is some positive number δ such that if and y are two different numbers in {X(s) : s S} then y > δ. 4

For eample, X is discrete if it can take only finitely many values (as in all the eamples above), or if the values of X are integers. Note that is X is discrete then p() =, where the sum is taken over all values that X takes. Eample An eample where the number of values is infinite is the following: you keep taking an eam until you pass it for the first time; X is the number of times you sit the eam. There is no upper limit on the values of X; for eample, you cannot guarantee to pass it even in 00 attempts. So the set of values is {,,3,...}, the set of all positive integers. How do we give the probability mass function if the set of values is infinite? Suppose that your probability of passing is p at each attempt, independently of all previous attempts. The event X = m is made up of just the one outcome FF...FP (with m Fs); this has probability q m p where q = p. So we could just give a formula P(X = m) = q m p for integers m. Alternatively, the following table makes it clear: m 3... n... P(X = m) p qp q p... q n p... When p = /0 the (incomplete) line graph is as follows. 0.... 0 3 4 5 7 8 9 5

Eample (Two dice: part ) I throw two fair si-sided dice. I am interested in the sum of the two numbers. Here the sample space is S = {(i, j) : i, j }, and we can define random variables as follows: X = number on first die Y = number on second die Z = sum of the numbers on the two dice; that is, X(i, j) = i, Y (i, j) = j and Z(i, j) = i + j. Notice that X and Y are different random variables even though they take the same values and their probability mass functions are equal. They are are said to have the same distribution. We write X Y in this case. The target set for Z is the set {,3,...,}. Since each outcome has probability /, the pmf is obtained by counting the number of ways we can achieve each value and dividing by. For eample, 9 = +3 = 5+4 = 4+5 = 3+, so P(Z = 9) = 4. We find: The line graph follows. k 3 4 5 7 8 9 0 P(Z = k) 3 4 5 5 4 3 9 8 0 3 4 5 7 8 9 0 In the first coin-tossing eample, if Y is the number of tails recorded during the eperiment, then X and Y again have the same distribution, even though their actual values are different (indeed, Y = 3 X).

Eample (Sheep: part ) There are 4 sheep in a field. The farmer shears of them. Later, he comes to the field and randomly samples 5 sheep without replacement. Let X be the number of shorn sheep in his sample. Then C m 8 C 5 m P(X = m) = 4 C 5 for m = 0,,, 3, 4, 5. To 4 decimal places the pmf is as follows. m 0 3 4 5 p(m) 0.0 0.430 0.880 0.070 0.004 0.000 0.4 0. 0 0 3 4 5 Epected value Let X be a discrete random variable which takes the values a,..., a n. The epected value (also called the epectation or mean) of X is the number E(X) given by the formula E(X) = n i= a i P(X = a i ). That is, we multiply each value of X by the probability that X takes that value, and sum these terms. Often I will write this sum as p(). I may also write µ X for E(X), and may abbreviate this to µ if X is clear from the contet. The epected value is a kind of generalised average : if each of the values is equally likely, so that each has probability /n, then E(X) = (a + + a n )/n, which is just the average of the values. 7

There is an interpretation of the epected value in terms of mechanics. If we put a mass p i on the ais at position a i for i =,...,n, where p i = P(X = a i ), then the centre of mass of all these masses is at the point E(X). In other words, if we make the line graph out of metal (and do not include the vertical ais) then the graph will balance at the point E(X) on the horizontal ais. If the random variable X takes infinitely many values, say a, a, a 3,..., then we define the epected value of X to be the infinite sum E(X) = i= a i P(X = a i ). Of course, now we have to worry about whether this means anything, that is, whether this infinite series is convergent. This is a question which is discussed at great length in analysis. We won t worry about it too much. Usually, discrete random variables will only have finitely many values; in the few eamples we consider where there are infinitely many values, the series will usually be a geometric series or something similar, which we know how to sum. In the proofs below, we assume that the number of values is finite. Eample (Child: part 3) Eample (One die: part 3) Eample (Sheep: part ) E(X) = 0 + =. E(X) = ( + + 3 + 4 + 5 + ) = = 3.5. E(X) = 0 0.0 + 0.430 + 0.880 + 3 0.070 + 4 0.004 + 5 0.000 =.50 to 4 decimal places. We shall see later that it should be eactly 5/4: what we have here is affected by rounding error. Eample (Two dice: part ) E(Z) = + 3 + 4 3 + + = ( + + + 0 + 30 + 4 + 40 + + 30 + + ) = 7. 8

Variance While E(X) gives the centre of gravity of the distribution, the spread of the disributiion is measured by the average value of (X µ X ). The variance of X is the number Var(X) given by Var(X) = ( µ X ) p(), where µ X = E(X). Sometimes it is written as σ X, or just as σ if X is clear from the contet. Eample (Sheep: part 3) m 0 3 4 5 p(m) 0.0 0.430 0.880 0.070 0.004 0.000 m.5.5 0.5 0.75.75.75 3.75 Var(X) = (.5) 0.0 + + (3.75) 0.000 = 0.7743. (Again, there is rounding error, because the eact value should be 85/( 3), which is 0.7745 to 4 decimal places.) Theorem 3 If E(X) = µ then Var(X) = p() µ. Proof Var(X) = ( µ) p() = = = = ( µ+ µ ) p() [ p() µp() + µ p() ] p() p() µ µp() + µ p() p() + µ p() = p() µ + µ because p() = µ and p() = = p() µ. 9

So now we have two methods of calculating variance. The first is affected by rounding errors in the calculation of µ, and also by rounding errors in the calculation of the values p(). The second method is more badly affected by rounding errors in the p(), and it can cause computer or calculator overflow if X takes very large values; however, the calculations are usually simpler. Two further properties of epected value and variance can be used as a check on your calculations. The epected value of X always lies between the smallest and largest values of X. (Can you prove this?) The variance of X is never negative. (We will prove this in a little while.) 0