Expectation of Random Variables

Similar documents
Continuous Random Variables

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Discrete Probability Refresher

Joint Distribution of Two or More Random Variables

Chapter 4. Chapter 4 sections

3 Multiple Discrete Random Variables

Topic 3: The Expectation of a Random Variable

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Convergence of Random Variables

SDS 321: Introduction to Probability and Statistics

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)

Fundamental Tools - Probability Theory II

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

CME 106: Review Probability theory

STAT 430/510: Lecture 16

More on Distribution Function

More than one variable

Convergence of Random Variables

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

6.041/6.431 Fall 2010 Quiz 2 Solutions

SDS 321: Introduction to Probability and Statistics

Analysis of Engineering and Scientific Data. Semester

1 Probability theory. 2 Random variables and probability theory.

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

Chp 4. Expectation and Variance

Random variables (discrete)

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

Probability and Distributions

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Topic 3: The Expectation of a Random Variable

Lecture 2: Repetition of probability theory and statistics

Probability and Statistics

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Continuous Random Variables and Continuous Distributions

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Random Variables. P(x) = P[X(e)] = P(e). (1)

Tom Salisbury

1 Presessional Probability

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Part (A): Review of Probability [Statistics I revision]

3. Probability and Statistics

Chap 2.1 : Random Variables

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SDS 321: Introduction to Probability and Statistics

Random Variables. Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Random Variables - Vocabulary. Random Variables, cont.

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

The mean, variance and covariance. (Chs 3.4.1, 3.4.2)

01 Probability Theory and Statistics Review

Chapter 3: Random Variables 1

1 Random Variable: Topics

Homework 4 Solution, due July 23

1 Review of Probability

Lecture 11. Probability Theory: an Overveiw

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Chapter 4 : Expectation and Moments

Formulas for probability theory and linear models SF2941

Math Spring Practice for the final Exam.

Lecture 22: Variance and Covariance

Probability. Computer Science Tripos, Part IA. R.J. Gibbens. Computer Laboratory University of Cambridge. Easter Term 2008/9

Notes 12 Autumn 2005

1: PROBABILITY REVIEW

Recall that if X 1,...,X n are random variables with finite expectations, then. The X i can be continuous or discrete or of any other type.

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

X = X X n, + X 2

Review of Probability. CS1538: Introduction to Simulations

Practice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality

Mathematical Statistics 1 Math A 6330

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

Limiting Distributions

Chapter 4 continued. Chapter 4 sections

Overview. CSE 21 Day 5. Image/Coimage. Monotonic Lists. Functions Probabilistic analysis

Algorithms for Uncertainty Quantification

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Northwestern University Department of Electrical Engineering and Computer Science

Continuous Random Variables. What continuous random variables are and how to use them. I can give a definition of a continuous random variable.

p. 4-1 Random Variables

Brief Review of Probability

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Chapter 3: Random Variables 1

Refresher on Discrete Probability

Review of Probability Theory

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

ECON Fundamentals of Probability

STAT/MATH 395 PROBABILITY II

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

Lecture 19: Properties of Expectation

Lecture 4: Probability and Discrete Random Variables

1 Review of Probability and Distributions

Probability Distributions

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

Lectures on Elementary Probability. William G. Faris

Bayesian statistics, simulation and software

CMPSCI 240: Reasoning Under Uncertainty

Conditional distributions (discrete case)

Transcription:

1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015

2 / 19 Expectation of Discrete Random Variables Definition The expectation of a discrete random variable X with probability mass function f is defined to be E(X) = xf (x) x:f (x)>0 whenever this sum is absolutely convergent. The expectation is also called the mean value or the expected value of the random variable. Example Bernoulli random variable Ω = {0, 1} where 0 p 1 f (x) = { p if x = 1 1 p if x = 0 E(X) = 1 p + 0 (1 p) = p

3 / 19 More Examples The probability mass function of a binomial random variable X with parameters n and p is P[X = k] = ( n k ) p k (1 p) n k if 0 k n Its expected value is given by n E(X) = kp[x = k] = k=0 ( ) n n k p k (1 p) n k = np k k=0 The probability mass function of a Poisson random variable with parameter λ is given by P[X = k] = λk k! e λ k = 0, 1, 2,... Its expected value is given by E(X) = kp[x = k] = k=0 k=0 k λk k! e λ = λ

Why do we need absolute convergence? A discrete random variable can take a countable number of values The definition of expectation involves a weighted sum of these values The order of the terms in the infinite sum is not specified in the definition The order of the terms can affect the value of the infinite sum Consider the following series 1 1 2 + 1 3 1 4 + 1 5 1 6 + 1 7 1 8 + Its sums to a value less than 5 6 Consider a rearrangement of the above series where two positive terms are followed by one negative term 1 + 1 3 1 2 + 1 5 + 1 7 1 4 + 1 9 + 1 11 1 6 + Since 1 4k 3 + 1 4k 1 1 2k > 0 the rearranged series sums to a value greater than 5 6 4 / 19

Why do we need absolute convergence? A series a i is said to converge absolutely if the series a i converges Theorem: If a i is a series which converges absolutely, then every rearrangement of a i converges, and they all converge to the same sum The previously considered series converges but does not converge absolutely 1 1 2 + 1 3 1 4 + 1 5 1 6 + 1 7 1 8 + Considering only absolutely convergent sums makes the expectation independent of the order of summation 5 / 19

Expectations of Functions of Discrete RVs If X has pmf f and g : R R, then E(g(X)) = x g(x)f (x) whenever this sum is absolutely convergent. Example Suppose X takes values 2, 1, 1, 3 with probabilities 1 4, 1 8, 1 4, 3 8 respectively. Consider Y = X 2. It takes values 1, 4, 9 with probabilities 3 8, 1 4, 3 8 respectively. E(Y ) = y yp(y = y) = 1 3 8 + 4 1 4 + 9 3 8 = 19 4 Alternatively, E(Y ) = E(X 2 ) = x x 2 P(X = x) = 4 1 4 + 1 1 8 + 1 1 4 + 9 3 8 = 19 4 6 / 19

7 / 19 Expectation of Continuous Random Variables Definition The expectation of a continuous random variable with density function f is given by whenever this integral is finite. E(X) = xf (x) dx Example (Uniform Random Variable) f (x) = { 1 b a for a x b 0 otherwise f (x) 1 b a E(X) = a+b 2 a b x

8 / 19 Conditional Expectation Definition For discrete random variables, the conditional expectation of Y given X = x is defined as E(Y X = x) = yf Y X (y x) y For continuous random variables, the conditional expectation of Y given X is given by E(Y X = x) = yf Y X (y x) dy The conditional expectation is a function of the conditioning random variable i.e. ψ(x) = E(Y X) Example For the following joint probability mass function, calculate E(Y ) and E(Y X). Y, X x 1 x 2 x 3 1 y 1 0 0 2 1 1 y 2 0 8 8 y 3 0 1 8 1 8

9 / 19 Law of Iterated Expectation Theorem The conditional expectation E(Y X) satisfies E [E(Y X)] = E(Y ) Example A group of hens lay N eggs where N has a Poisson distribution with parameter λ. Each egg results in a healthy chick with probability p independently of the other eggs. Let K be the number of chicks. Find E(K ).

Some Properties of Expectation If a, b R, then E(aX + by ) = ae(x) + be(y ) If X and Y are independent, E(XY ) = E(X)E(Y ) X and Y are said to be uncorrelated if E(XY ) = E(X)E(Y ) Independent random variables are uncorrelated but uncorrelated random variables need not be independent Example Y and Z are independent random variables such that Z is equally likely to be 1 or 1 and Y is equally likely to be 1 or 2. Let X = YZ. Then X and Y are uncorrelated but not independent. 10 / 19

11 / 19 Expectation via the Distribution Function For a discrete random variable X taking values in {0, 1, 2,...}, the expected value is given by E[X] = P(X i) Proof i=1 P(X i) = i=1 P(X = j) = i=1 j=i j P(X = j) = j=1 i=1 jp(x = j) = E[X] j=1 Example Let X 1,..., X m be m independent discrete random variables taking only non-negative integer values. Let all of them have the same probability mass function P(X = n) = p n for n 0. What is the expected value of the minimum of X 1,..., X m?

12 / 19 Expectation via the Distribution Function For a continuous random variable X taking only non-negative values, the expected value is given by Proof E[X] = 0 P(X x) dx 0 P(X x) dx = = 0 0 x f X (t) dt dx = tf X (t) dt = E[X] t 0 0 f X (t) dx dt

13 / 19 Variance Quantifies the spread of a random variable Let the expectation of X be m 1 = E(X) The variance of X is given by σ 2 = E[(X m 1 ) 2 ] The positive square root of the variance is called the standard deviation Examples Variance of a binomial random variable X with parameters n and p is ( ) n n var(x) = (k np) 2 P[X = k] = k 2 n p k (1 p) n k n 2 p 2 k k=0 = np(1 p) k=0 Variance of a uniform random variable X on [a, b] is var(x) = [ x a + b ] 2 (b a)2 f U (x) dx = 2 12

Properties of Variance var(x) 0 var(x) = E(X 2 ) [E(X)] 2 For a, b R, var(ax + b) = a 2 var(x) var(x + Y ) = var(x) + var(y ) if and only if X and Y are uncorrelated 14 / 19

Probabilistic Inequalities

16 / 19 Markov s Inequality If X is a non-negative random variable and a > 0, then P(X a) E(X) a. Proof We first claim that if X Y, then E(X) E(Y ). Let Y be a random variable such that { a if X a, Y = 0 if X < a. Then X Y and E(X) E(Y ) = ap(x a) = P(X a) E(X) a. Exercise Prove that if E(X 2 ) = 0 then P(X = 0) = 1.

Chebyshev s Inequality Let X be a random variable and a > 0. Then P ( X E(X) a) var(x) a 2. Proof Let Y = (X E(X)) 2. P ( X E(X) a) = P(Y a 2 ) E(Y ) a 2 Setting a = kσ where k > 0 and σ = var(x), we get = var(x) a 2. P ( X E(X) kσ) 1 k 2. Exercises Suppose we have a coin with an unknown probability p of showing heads. We want to estimate p to within an accuracy of ɛ > 0. How can we do it? Prove that P(X = c) = 1 var(x) = 0. 17 / 19

18 / 19 Cauchy-Schwarz Inequality For random variables X and Y, we have E(XY ) E(X 2 ) E(Y 2 ) Equality holds if and only if P(X = cy ) = 1 for some constant c. Proof For any real k, we have E[(kX + Y ) 2 ] 0. This implies k 2 E(X 2 ) + 2kE(XY ) + E(Y 2 ) 0 for all k. The above quadratic must have a non-positive discriminant. [2E(XY )] 2 4E(X 2 )E(Y 2 ) 0.

Questions? 19 / 19