Conditional distributions. Conditional expectation and conditional variance with respect to a variable.

Similar documents
Formulas for probability theory and linear models SF2941

conditional cdf, conditional pdf, total probability theorem?

Quantitative Methods in Economics Conditional Expectations

Chapter 4 Multiple Random Variables

STAT 430/510 Probability

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

Multivariate distributions

Introduction to Probability and Stocastic Processes - Part I

Covariance and Correlation

Chapter 2: Random Variables

EE4601 Communication Systems

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

Chp 4. Expectation and Variance

ACM 116: Lectures 3 4

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

HW4 : Bivariate Distributions (1) Solutions

When Are Two Random Variables Independent?

STAT 430/510: Lecture 15

THE UNIVERSITY OF HONG KONG DEPARTMENT OF STATISTICS AND ACTUARIAL SCIENCE

STA 256: Statistics and Probability I

Continuous r.v practice problems

2 (Statistics) Random variables

Solution to Assignment 3

Continuous Random Variables

Expectation of Random Variables

Multiple Random Variables

18 Bivariate normal distribution I

5 Operations on Multiple Random Variables

Let X and Y denote two random variables. The joint distribution of these random

Bivariate distributions

ENGG2430A-Homework 2

3. Probability and Statistics

Chapter 4 : Expectation and Moments

Bivariate Transformations

STAT 430/510: Lecture 16

Random Variables. P(x) = P[X(e)] = P(e). (1)

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Probability review. September 11, Stoch. Systems Analysis Introduction 1

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

2 Functions of random variables

CHAPTER 1 DISTRIBUTION THEORY 1 CHAPTER 1: DISTRIBUTION THEORY

Problem 1. Problem 2. Problem 3. Problem 4

Conditional distributions (discrete case)

Chapter 5,6 Multiple RandomVariables

Statistics 351 Probability I Fall 2006 (200630) Final Exam Solutions. θ α β Γ(α)Γ(β) (uv)α 1 (v uv) β 1 exp v }

4 Pairs of Random Variables

ECE Lecture #9 Part 2 Overview

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

Random Variables and Their Distributions

Practice Examination # 3

Conditional densities, mass functions, and expectations

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

discrete random variable: probability mass function continuous random variable: probability density function

STAT 516 Midterm Exam 3 Friday, April 18, 2008

4. CONTINUOUS RANDOM VARIABLES

Homework 9 (due November 24, 2009)

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Final Exam # 3. Sta 230: Probability. December 16, 2012

LIST OF FORMULAS FOR STK1100 AND STK1110

Applied Stochastic Models (SS 09)

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

4. Distributions of Functions of Random Variables

1.1 Review of Probability Theory

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

1 Joint and marginal distributions

A review of probability theory

STAT Chapter 5 Continuous Distributions

2 Conditioning. 1 Conditional Distributions

3 Conditional Expectation

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

ECE302 Exam 2 Version A April 21, You must show ALL of your work for full credit. Please leave fractions as fractions, but simplify them, etc.

Econ 508B: Lecture 5

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

1 Review of di erential calculus

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Probability Theory. Patrick Lam

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

SDS 321: Introduction to Probability and Statistics

Lecture 19: Properties of Expectation

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

Linear models and their mathematical foundations: Simple linear regression

Math 510 midterm 3 answers

MATH/STAT 3360, Probability Sample Final Examination Model Solutions

Order Statistics and Distributions

Algorithms for Uncertainty Quantification

Multiple Random Variables

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

Statistics, Data Analysis, and Simulation SS 2015

First Order Differential Equations

Random Signals and Systems. Chapter 3. Jitendra K Tugnait. Department of Electrical & Computer Engineering. Auburn University.

Probability and Statistics

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

Lecture 3. David Aldous. 31 August David Aldous Lecture 3

MAS223 Statistical Inference and Modelling Exercises

1 Random Variable: Topics

Transcription:

Conditional distributions Conditional expectation and conditional variance with respect to a variable Probability Theory and Stochastic Processes, summer semester 07/08 80408 Conditional distributions Conditional distributions of discrete random variables Recall that if both X and Y are discrete random variables and A {x, x, }, B {y, y, } are such that P X A and P Y B then we define the joint probability mass function of X and Y as A B x, y P X x and Y y [0, ] Assuming that P Y y j > 0 we define the conditional distribution of X given Y y j as the distribution with the following probability mass function, called the conditional probability mass function of X given Y y j : P X x i Y y j P X x i and Y y j P Y y j Similarly, assuming that P X x i > 0 we define the conditional distribution of Y given X x i as the distribution with the following probability mass function, called the the conditional probability mass function of Y given X x i : P Y y j X x i P X x i and Y y j P X x i Remark If X and Y are independent then P X x i and Y y j P X x i P Y y j and easy calculation yields P X x i Y y j P X x i and P Y y j X x i P Y y j Example A hen lays N eggs, where N has Poisson distribution with parameter λ Each egg hatches with probability p independently of the other eggs Let K be the number of chicks Find the joint probability mass function of N and K

SOLUTION Since each egg hatches with probability p independently of the other eggs the conditional distribution of K given N n, n, is the binomial distribution with parameters n and p and naturally P K 0 N 0 So for n 0,,, and k 0,,, n we have P N n and K k P K k N k P N n n p k p n k λ λn e k n! n! k!n k! pk p n k e e λ λ n pk k! p n k n k! and P N n and K k 0 otherwise Conditional distributions of continuous random variables λ λn Let X and Y have the joint probability density function f X,Y and let f X and f Y be marginal densities of X and Y respectively Assuming that f Y y > 0 we define the conditional distribution of X given Y y as the distribution with the following density function, called the conditional probability density function of X given Y y: f X Y y x f Y y f X,Y x, y Similarly, assuming that f X x > 0 we define the conditional distribution of Y given X x as the distribution with the following density function, called the conditional probability density function of Y given X x: f Y Xx y f X x f X,Y x, y Remark If X and Y are independent then f X,Y x, y f X xf Y y and easy calculation yields f X Y y x f X x and f Y Xx y f Y y Example Let X, Y has a bivariate normal distribution with the density { x f X,Y x, y π µ exp ρ σ ρ ρ x µ } y µ y µ +, σ σ where, σ > 0, µ, µ R and ρ, Calculate the conditional probability density function of Y given X x SOLUTION To calculate f Y Xx y let us recall that the marginal distributions of X is N µ, σ and f X x exp πσ x µ n!

Now we have f Y Xx y { exp x µ π ρ σ ρ ρ x µ y µ πσ exp { x πσ π µ exp ρ σ ρ We simplify ρ { x µ ρ x µ { y µ ρ σ y µ ρ ρ x µ σ ρ σ and get f Y Xx y y µ σ + x µ ρ x µ y µ σ + ρ y µ ρ σ x µ exp π ρ σ ρ σ σ + y µ ρ x µ y µ σ + } y µ + x µ σ } x µ σ } } y µ + x µ σ y µ ρ σ x µ Thus we recognize that Y X x N µ + ρ σ x µ, ρ σ The knowledge of X reduces the variance of Y by ρ σ Mixed case Sometimes we will deal with the case when X is continuous and Y discrete Let B {y, y, } be such that P Y B We assume that we are given functions f j : R [0, +, j,, such that P X dx and Y y j f j xdx, which means that for any < a < b < + we have P X [a, b] and Y y j xb xa f j xdx The marginal probability density function of X reads as f X x + j f j x b a f j xdx 3

while the marginal probability mass function of Y is given by P Y y j x+ x f j xdx f j xdx The conditional probability density function of X given Y y j reads as f X Y yj x f j x P Y y j f j x f jzdz while the conditional probability mass function of Y given X x for x such that f X x > 0 is given by P Y y j X x f jx f X x f j x + k f kx Example 3 Let Λ be a continuous random variable with Erlangk, α distribution k,,, α > 0, which means that it has density { α k k! f Λ λ λk e α λ if λ 0; 0 if λ < 0 The conditional distribution of variable N given Λ λ > 0 has P oiλ distribution Find the unconditional marginal distribution of N SOLUTION Since N Λ λ P oiλ we have and λ λn P N n Λ λ e n! P N n and Λ dλ P N n Λ λ f Λ λdλ The marginal distribution of N reads as λ λn α k e n! k! λk e α λ dλ Since P N n fλ 0 λ λn e α k n!k! α k n! k! λk e α λ dλ 0 λ n+k e α+λ dλ { α+ n+k n+k! λn+k e α+ λ if λ 0; 0 if λ < 0 4

is the density of Erlangn + k, α + distribution, we get and finally P N n 0 λ n+k e α+λ dλ α k n + k! n!k! α + n+k n + k k n + k! α + n+k k n k α α + α + This is so called negative binomial distribution with parameters k and /α + Conditional expectation Sometimes we are not interested in the conditional distribution of one variable given another variable but in some characteristics of this conditional distribution, for example in its expectation, which may be much easier to estimate Definition The conditional expectation E Y X x is the expectation of the conditional distribution of Y given X x if it is well defined EXAMPLES In Example we have K N n Binn, p thus E K N n n p In Example we have Y X x N µ + ρ σ x µ, ρ σ thus E Y X x µ + ρ σ x µ In Example 3 we have N Λ λ P oiλ thus E N Λ λ λ The conditional expectation E Y X x is always a function of x Let us define E : R R, { E Y X x if E Y X x is well defined; Ex 0 otherwise Now we introduce the conditional expectation of Y given X Definition For two real variables X and Y let E : R R be defined as in The conditional expectation E Y X is the random variable defined as EXAMPLES E Y X : EX In Example we have E K N n n p thus E K N p N In Example we have E Y X x µ + ρ σ x µ thus E Y X µ + ρ σ X µ In Example 3 we have E N Λ λ λ thus E N Λ Λ 5

Properties of EY X The conditional expectation E Y X has the following properties E Y X is a random variable, which may be expressed as a function of variable X, E Y X EX If EY is well defined then also EEY X is well defined and EEY X EY 3 If Y is some function of X, Y F X, then EY X F X 4 If Y is independent from X and EY is well defined then EY X EY 5 If Z is some function of X then EZY X ZEY X 6 Linearity If β and β are some real numbers then: Eβ Y + β Y X β EY X + β EY X We will prove in the case when X and Y have joint probability mass function Let I 0 be the interval or more complicated set of those x for which f X x 0 and I be the interval of those x for which f X x > 0 We write ns EY y f Y ydy y f X,Y x, ydx dy I 0 0 + I yf X,Y x, ydy dx yf X,Y x, ydy EY X xf X xdx I EEX EEY X Decomposition of the variance dx + I y f X,Y x, y dy f X xdx f X x Exf X xdx yf X,Y x, ydy dx Let F : R R and X, Y be two real random variables We have the following important identity: E Y F X E {Y E Y X} + E {E Y X F X} 3 To prove 3 we write E X F X E {Y E Y X + E Y X F X} 4 E Y E Y X + E {Y E Y X E Y X F X} + E E Y X F X 6

We will prove that E {Y E Y X E Y X F X} 0 Since E Y X F X is some function of X we may write E Y X F X GX and from properties -3 and 5-6 of the conditional expectation we get E {Y E Y X E Y X F X} EE Y E Y X GX X E {GXE Y E Y X X} E {GX [E Y X E E Y X X]} E {GX [E Y X E Y X]} E {GX0} 0 This calculation and 4 gives 3 From 3 it follows that E Y F X E {Y E Y X} +E {E Y X F X} E Y E Y X and the equality is attained iff E E Y X F X 0 which holds iff P F X E Y X This may be interpreted that the conditional expectation EY X is the best predictor of Y which is a function of X when the error is measured as E Y F X mean square error Example 4 If X and Y have the bivariate normal distribution as in Example then the best predictor of Y given X, in the sense of the mean square error, is Ŷ E Y X µ + ρ σ X µ Setting in 3 F X EY we get E Y EY E {Y E Y X} + E {E Y X EY } 5 Denoting D Y X E {Y E Y X} X and D E Y X E {E Y X EE Y X} E {E Y X EY } we get the folowing decomposition of the variance of Y σ Y E Y EY ED Y X + D E Y X The quantity D Y X is called the conditional variance of Y given X 7