RVs and their probability distributions

Similar documents
Recap. The study of randomness and uncertainty Chances, odds, likelihood, expected, probably, on average,... PROBABILITY INFERENTIAL STATISTICS

CS 361: Probability & Statistics

MODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES DISTRIBUTION FUNCTION AND ITS PROPERTIES

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, < x <. ( ) X s. Real Line

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Probability Year 9. Terminology

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

CMPSCI 240: Reasoning Under Uncertainty

Recitation 2: Probability

Chapter 2. Probability. Math 371. University of Hawai i at Mānoa. Summer 2011

Probability Year 10. Terminology

9. DISCRETE PROBABILITY DISTRIBUTIONS

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Dynamic Programming Lecture #4

the time it takes until a radioactive substance undergoes a decay

k P (X = k)

P (A) = P (B) = P (C) = P (D) =

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

P [(E and F )] P [F ]

Elementary Discrete Probability

Review of Probability. CS1538: Introduction to Simulations

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

Probability Theory and Simulation Methods

Recap of Basic Probability Theory

Mathematics. ( : Focus on free Education) (Chapter 16) (Probability) (Class XI) Exercise 16.2

Probability Theory and Applications

Recap of Basic Probability Theory

CS 124 Math Review Section January 29, 2018

Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008

Probability Theory Review

Basic Probability. Introduction

Event A: at least one tail observed A:

Notes 6 Autumn Example (One die: part 1) One fair six-sided die is thrown. X is the number showing.

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2

Introduction to Probability

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation

Chapter 8: An Introduction to Probability and Statistics

Probability theory basics

Probability and Statisitcs

Chapter 2 Random Variables

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

Stochastic Processes

Statistical Inference

MAT Mathematics in Today's World

Probability Theory and Random Variables

Business Statistics. Lecture 3: Random Variables and the Normal Distribution

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Probability and Independence Terri Bittner, Ph.D.

Monty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch

Sample Spaces, Random Variables

1 Presessional Probability

STAT200 Elementary Statistics for applications

Chapter 2: Probability Part 1

Statistics 251: Statistical Methods

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Introduction to Probability

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation

The random variable 1

Lecture 8: Conditional probability I: definition, independence, the tree method, sampling, chain rule for independent events

Discrete Random Variables

4 Lecture 4 Notes: Introduction to Probability. Probability Rules. Independence and Conditional Probability. Bayes Theorem. Risk and Odds Ratio

Discrete Probability. Chemistry & Physics. Medicine

2. Probability. Chris Piech and Mehran Sahami. Oct 2017

Probability: Sets, Sample Spaces, Events

Lecture 1 : The Mathematical Theory of Probability

Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

Randomized Algorithms

1 Random variables and distributions

Discrete Random Variables. Discrete Random Variables

1. Discrete Distributions

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

1 Generating functions

Today we ll discuss ways to learn how to think about events that are influenced by chance.

n N CHAPTER 1 Atoms Thermodynamics Molecules Statistical Thermodynamics (S.T.)

Statistics for Financial Engineering Session 2: Basic Set Theory March 19 th, 2006

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Mutually Exclusive Events

Probabilistic Systems Analysis Spring 2018 Lecture 6. Random Variables: Probability Mass Function and Expectation

Fundamentals of Probability CE 311S

Week 2. Section Texas A& M University. Department of Mathematics Texas A& M University, College Station 22 January-24 January 2019

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

Probability and Probability Distributions. Dr. Mohammed Alahmed

Stats Probability Theory

Chap 4 Probability p227 The probability of any outcome in a random phenomenon is the proportion of times the outcome would occur in a long series of

Review of Basic Probability

1 What are probabilities? 2 Sample Spaces. 3 Events and probability spaces

Probability. Lecture Notes. Adolfo J. Rumbos

Part 3: Parametric Models

Dept. of Linguistics, Indiana University Fall 2015

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

MATH Solutions to Probability Exercises

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya

Random variables (discrete)

Transcription:

RVs and their probability distributions RVs and their probability distributions In these notes, I will use the following notation: The probability distribution (function) on a sample space will be denoted by P (A) - upper case P. Here A represents either an outcome or an event. The probability distribution (function) (pdf) for a discrete (finite or infinite) random variable will be denoted by p(x = x) or p X (x) or just p(x) if there is only one random variable under discussion - note, lower-case p! Here x represents a possible value of the random variable. The probability density function for a continuous random variable will be denoted by f(x). Motivation Why use the sample space S = {HH, HT, T H, T T } for the experiment tossing two balanced coins? What is wrong with using the sample space S = {two heads, one head, no heads}? Another example: why use the sample space consisting of 36 ordered pairs for the experiment rolling two balanced dice when all we care about is the sum of the numbers showing on the two dice? Here the sample space is so large and unwieldy, and if we declared instead that the sample space is S = {, 3, 4, 5, 6, 7, 8, 9, 10, 11, 1} (the possible total number of spots showing), it might be more useful. On the other hand, there is more than one way to throw a total of 7, but only one way to throw a. So I probably would not think that these are equally likely events. If we assume that the two tosses or the two rolls are independent, that means that the probability of getting heads (for instance) on the second toss does not depend on whether or not heads shows on the first toss. How could one coin be influenced by what another coin does? Or equivalently, how can a coin be influenced by whether or not it landed heads up last time I tossed it? If the tossing procedure is anything like reasonable, we can easily believe that the tosses must be independent. And the same goes for the dice. I can use that independence to compute what the probabilities of the four elements of S given in the first sample space, from knowing only that the coin is balanced, by reasoning as follows: the probability of getting heads on the first toss (or the first coin) is 1/. The probability of getting heads on the second toss, given that the first one was heads, is still 1/, because of independence. Hence the probability of getting heads on the first toss and on the second toss is the product 1/ times 1/ or 1/4. Same for all three other outcomes (work it out!) and so, assuming only that the coin is balanced, the sample space which has four events in it is naturally an equally likely outcomes sample space. The other one is not, if we then view its outcomes as events in the first sample space. The same argument works for the experiment rolling two balanced dice. The sample space consisting of 36 ordered pairs is naturally an equally likely outcomes sample space, and the other is not. This is the main reason for preferring such large sample spaces. But then we want to talk about the sum of the numbers on the two dice, or about the number of heads that show! So this is the reason for defining a concept to represent such things. That is a random variable. Random Variables: A random variable is a variable whose value depends on the outcome of a random process (experiment). The values of a random variable in this course will be real numbers. The random variable acts like a function: it assigns a number to each outcome in S. Random variables are traditionally designated by using capital letters, starting with X. RVs and their probability distributions continued on next page... Page 1 of 5

Finite Random Variables A finite random variable is one which has only a finite number of possible values. Probability Distribution Function (pdf) A probability distribution (function) for a finite random variable X is a function p(x = k) which assigns a probability to each possible value k of X. (We will sometimes shorthand this notation by writing just p(k) for p(x = k).) If we already have a probability distribution on the underlying sample space, the pdf for the random variable is defined using that probability distribution. For each possible value k of the RV X, we find all of the outcomes which give that value of the RV and add their probabilities together. Example of a probability distribution function for an RV: Experiment toss two balanced coins. S = {HH, HT, T H, T T } and each outcome has probability 1 4 as argued in the Motivation section. Define a RV on S as follows: X = the number of heads that show Then the possible values of X are 0, 1, and. We define the pdf for X by looking at each value of the random variable as if it described an event and computing the probability of that event as usual, by adding the probabilities of the individual outcomes: p(x = 0) = P (T T ) = 1 4 p(x = 1) = P (HT ) + P (T H) = 1 p(x = ) = P (HH) = 1 4 So, putting this all together (and using shorthand) the pdf for X is given by p(0) = 1 4 p(1) = 1 p() = 1 4 Experiment: Roll two coins which are unbalanced, so that P (HH) = 1, P (HT ) = 1 4, P (T H) = 1 8, P (T T ) = 1 8. Now for X = the number of heads that show, we have: p(x = 0) = P (T T ) = 1 8 p(x = 1) = P (HT ) + P (T H) = 1 4 + 1 8 = 3 8 p(x = ) = P (HH) = 1 RVs and their probability distributions continued on next page... Page of 5

Properties of the pdf of a random variable: The pdf of a finite or discrete RV has similar properties to those of a pdf on a sample space: 0 p(x) 1 for all possible values x p(x) = 1, where the sum is taken over all possible values of X The probability that X = x 1 or X = x is p(x 1 ) + p(x ), etc. (The various values of the RV represent mutually exclusive events and so the probability of their unions are found by adding.) Examples: 1. For example, for our previously defined random variable X= the number of heads that show in tossing two balanced coins, we can compute the probability that X is 1 or : P (X = 1 or X = ) = p(1) + p() = 1 + 1 4 = 3 4. Experiment rolling two balanced dice This has a sample space consisting of 36 ordered pairs describing the two numbers which show. This is the equally likely sample space for this experiment under the classical approach. Each outcome has probability 1 36. It is useful to represent these outcomes as points in the first quadrant of the Cartesian plane. All of the points lie in a 6-by-6 square. We can define the random variable X= the sum of the two numbers that show. You may wish to label each point of the square with its sum. The possible values of X are the whole numbers through 1 inclusive. Note that some sums appear more than once. Then we can compute the pdf for X (details omitted, but it is strongly recommended that you write out the sample space and compute them for yourself): p() = 1 36 p(3) = 36 p(4) = 3 36 (because there are ways to get a sum of 3) (because there are 3 ways to get a sum of 4, etc.) p(5) = 4 36 p(6) = 5 36 p(7) = 6 36 p(8) = 5 36 p(9) = 4 36 p(10) = 3 36 p(11) = 36 p(1) = 1 36 We can use this pdf to find other probabilities. For example, the probability that the sum will be either or 7 is P (X = or X = 7) = p() + p(7) = 1 36 + 6 36 = 7 36. Also, the probability that the sum is at least 10 is P (X 10) = p(10) + p(11) + p(1) = 3 36 + 36 + 1 36 = 6 36 = 1 6 RVs and their probability distributions continued on next page... Page 3 of 5

Infinite Discrete Random Variables: an Example An (infinite) discrete random variable is one which has a countably infinite number of possible values. In other words, the possible values can be out into a one-to-one correspondence with the natural numbers. The definitions and properties given above for the finite case go through without change in case the RV has possible values in an infinite discrete set. We have so far seen exactly one example of an infinite, discrete sample space. Recall that for the experiment toss a balanced coin until heads shows for the first time has an infinite sample space: S = {H, T H, T T H, T T T H, T T T T H,...} This is discrete because it can be put into one-to-one correspondence with the natural numbers. (Informally, you can count the elements, even though the counting never ends.) Following the argument given in the Motivation section, we should arrive at the following probability distribution for this sample space: P (H) = 1 P (T H) = ( 1 )( 1 ) = ( 1 ) (the probability of getting T on the first throw times the probability of getting heads on the second throw) P (T T H) = ( 1 )( 1 )( 1 ) = ( 1 )3 (the probability of getting T on the first throw times the probability of getting T on the second throw times the probability of getting H on the third throw) etc. We are not computing the powers because the object here (since the sample space is infinite) is to get a general formula for the pdf. The general formula can be seen from the process above: P (T...T }{{} H) = ( ) 1 n+1 n tails Verify that this satisfies the requirements to be a pdf. We need to check that all the probabilities are between 0 and 1 (which is obviously true) and that the sum of all the probabilities for all elements of the sample space is 1, which is not so obvious. The sum of all the probabilities is n=0 ( ) n+1 1 which is a geometric series n=0 arn with a = 1 and r = 1. The series converges when r < 1. The sum of this series is then, so the sum of all these probabilities is a 1 r n=0 ( ) n+1 1 1 = 1 1 = 1 RVs and their probability distributions continued on next page... Page 4 of 5

Now we will define a RV on this experiment.: X = the number of tails that show The possible values of X are all the whole numbers 0, 1,,..., so X is a discrete RV. (Note: it is possible to define a finite RV on an infinite sample space. The number of possible values of the RV may be smaller than the number of outcomes in the sample space: it just cannot be larger.) Then from the pdf we just derived, we can find a pdf for X: p(0) = P (H) = 1 p(1) = P (T H) = ( 1 ) p() = P (T T H) = ( 1 )3 p(3) = P (T T T H) = ( 1 )4 etc. As before, we see that p(n) = ( 1 )n+1 and the sum of the probabilities over all possible values of x is 1. This example is important in the historical development of the theory of probability and we will be seeing it again with a different random variable. It is known as the Saint Petersburg Paradox. Exercise: Work out the case of tossing an unbalanced coin until heads shows for the first time. Let s say that the probability of heads on one toss is 3 5 and the probability of tails on one toss is then 5. Find the pdf for the sample space (as a general formula) and then find the pdf for the RV X= the number of tails that show, and show that the sum of all the probabilities is 1. (It is important in doing this that you recreate the thinking of the example we just worked, and not just try to put numbers in the right places.) Page 5 of 5