When Are Two Random Variables Independent?

Similar documents
1 Joint and marginal distributions

2 Functions of random variables

Jointly Distributed Random Variables

Conditional distributions. Conditional expectation and conditional variance with respect to a variable.

Review of Probability Theory

Bivariate distributions

Software Reliability & Testing

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

HW4 : Bivariate Distributions (1) Solutions

More on Distribution Function

3 Multiple Discrete Random Variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

ECE 4400:693 - Information Theory

Lecture 13: p-values and union intersection tests

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Let f(x) = x, but the domain of f is the interval 0 x 1. Note

Mathematics 426 Robert Gross Homework 9 Answers

p. 6-1 Continuous Random Variables p. 6-2

RANDOM VECTOR. Tutorial 6, STAT1301 Fall 2010, 02NOV2010, By Joseph Dong

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

Joint Probability Distributions and Random Samples (Devore Chapter Five)

1 Random Variable: Topics

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Covariance and Correlation

Chapter 4 Multiple Random Variables

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

A review of probability theory

Section 8.1. Vector Notation

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

8. Limit Laws. lim(f g)(x) = lim f(x) lim g(x), (x) = lim x a f(x) g lim x a g(x)

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

ORF 245 Fundamentals of Statistics Joint Distributions

Lecture 8: Channel Capacity, Continuous Random Variables

Basics on Probability. Jingrui He 09/11/2007

Class 8 Review Problems solutions, 18.05, Spring 2014

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)

Quantitative Methods in Economics Conditional Expectations

2 (Statistics) Random variables

3.5 Solving Equations Involving Integers II

EE4601 Communication Systems

Introduction to Probability and Stocastic Processes - Part I

CS145: Probability & Computing

Probability Review. Gonzalo Mateos

Foundations of Mathematics MATH 220 FALL 2017 Lecture Notes

Continuous Random Variables

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

There are two things that are particularly nice about the first basis

Math 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Math 3338: Probability (Fall 2006)

Lecture 14: Multivariate mgf s and chf s

Check Your Understanding of the Lecture Material Finger Exercises with Solutions

STA 256: Statistics and Probability I

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Chapter 5 Random vectors, Joint distributions. Lectures 18-23

Lecture 5 Channel Coding over Continuous Channels

Joint Distributions: Part Two 1

Multivariate distributions

Solution to Assignment 3

Probability. Lecture Notes. Adolfo J. Rumbos

Multivariate probability distributions and linear regression

LESSON 23: EXTREMA OF FUNCTIONS OF 2 VARIABLES OCTOBER 25, 2017

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

Solving Polynomial and Rational Inequalities Algebraically. Approximating Solutions to Inequalities Graphically

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

Week 2: Review of probability and statistics

Joint Probability Distributions, Correlations

Basic counting techniques. Periklis A. Papakonstantinou Rutgers Business School

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

Cauchy Sequences. x n = 1 ( ) 2 1 1, . As you well know, k! n 1. 1 k! = e, = k! k=0. k = k=1

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

SDS 321: Introduction to Probability and Statistics

z = f (x; y) f (x ; y ) f (x; y) f (x; y )

Discrete Distributions

Probability and Distributions

STAT/MATH 395 PROBABILITY II

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

Introduction to Probability Theory

EXAMPLES OF PROOFS BY INDUCTION

conditional cdf, conditional pdf, total probability theorem?

First and Second Order ODEs

Discrete Probability Refresher

MATH 426, TOPOLOGY. p 1.

We introduce methods that are useful in:

Week 10 Worksheet. Math 4653, Section 001 Elementary Probability Fall Ice Breaker Question: Do you prefer waffles or pancakes?

Notes on Probability

STAT 801: Mathematical Statistics. Distribution Theory

Solving Linear and Rational Inequalities Algebraically. Definition 22.1 Two inequalities are equivalent if they have the same solution set.

ECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1

Riemann Sum Comparison

Distributions of Functions of Random Variables

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Power series solutions for 2nd order linear ODE s (not necessarily with constant coefficients) a n z n. n=0

Lecture 4: Probability and Discrete Random Variables

Transcription:

When Are Two Random Variables Independent? 1 Introduction. Almost all of the mathematics of inferential statistics and sampling theory is based on the behavior of mutually independent random variables, for these reasons. 1. In most instances, for all practical purposes, sampling is performed with replacement. 1 2. As a consequence, in an experiment that consists of sampling n members of a population and performing the same measurement on each of them: if Y j is the result of measuring the j th member of the sample, then the random variables {Y 1, Y 2... Y n } are independent and identically distributed. Fortunately, joint distributions of independent random variables are the easiest ones to analyze. This handout focuses on how to characterize independence in the case of two random variables; characterizing independence for three or more random variables is more complicated. The intuitive, unmathematicized idea to be rendered precise is this: Random variables X and Y are independent if they have nothing to do with each other, in the sense that the probabilities associated with X remain the same regardless what the value of Y is and vice-versa). I will need the following useful result about joint CDF s true regardless whether or not X and Y are independent). Recall that F X x) P X x), F Y y) P Y y), and F XY x, y) P X x Y y. Theorem 1 If X and Y are random variables defined on the same sample space, then for any a < b and c < d, P a < X b) c < Y d) F XY b, d) F XY b, c) F XY a, d) + F XY a, c). 1) Exercise 1 Prove equation 1 ). 1 because the population is generally much larger than the sample size the population is often theoretically infinite. 1

2 The Discrete Case. As discussed in class and in handout #5: Definition 1 discrete case) Let X and Y be two discrete random variables defined on the same sample space. X and Y are independent iff for every value x i in the distribution of X and every value y j in the distribution of Y, P X x i Y yj P X x i )P Y y j ); or, equivalently, px i, y j ) p X x i )p Y y j ). 2) Equation 2) wors perfectly for discrete random variables, but it does not wor for random variables in general. Our first tas is to establish a condition which is equivalent to 2) for discrete random variables but which will also apply to other types of random variables. Theorem 2 Let X and Y be discrete random variables defined on the same sample space. Then X and Y are independent F XY x, y) F X x)f Y y) for all points x, y) in the plane. 3) Proof ): Assuming X and Y to be independent, we can calculate as follows. The proof is based on the fact that set intersection distributes over set union in the same way that multiplication distributes over addition see exercise 2) on p.3). F X x)f Y y) P X x)p Y y) P X x i ) P Y y j ) distribute )) P X x i )P Y y j ) xi x partially regroup ) P X x i )P Y y j ) by equation 2) ) because the events are nonoverlapping ) P by exercise 2 on p.3) ) P P X x i ) Y y j )) X x i ) Y y j X x i ) because the events are nonoverlapping ) P by exercise 2 on p.3) ) P 2 P Y y j [ P X x i ) ] Y y)) X x i ) Y y)) X x i ) Y y) X x) Y y) F XY x, y).

The proof in the other direction will be simpler if I allow all integers positive, negative or zero) as indices for the values of X and Y and assume that larger indices go with larger values: 2 x 3 < x 2 < x 1 < x 0 < x 1 < x 2 < x 3 < and y 3 < y 2 < y 1 < y 0 < y 1 < y 2 < y 3 <. Arranging things in this way gives formula 4), which streamlines the proof. Proof ): P X x i ) F X x i ) F X x i 1 ) and P Y y j ) F Y y j ) F Y y j 1 ). 4) P X x i )P Y y j ) F X x i ) F X x i 1 )) F Y y j ) F Y y j 1 )) multiply out ) F X x i )F Y y j ) F X x i 1 )F Y y j ) F X x i )F Y y j 1 ) + F X x i 1 )F Y y j 1 ) by assumption ) F XY x i, y j ) F XY x i 1, y j ) F XY x i, y j 1 ) + F XY x i 1, y j 1 ) by exercise 1 on p.1 ) P x i 1 < X x i yj 1 < Y y j P X x i Y yj. Exercise 2 Show, for any events {B, A 1, A 2, A 3,...}, that ) A B A B. 3 The General Case. 1 Theorem 2 for discrete random variables indicates how to go about characterizing independence in in general. Equation 5) gives a characterization of of independence that is valid in all cases including the discrete case, by Theorem 2). Definition 2 general case) Let X and Y be any random variables defined on the same sample space. Then X and Y are independent F XY x, y) F X x)f Y y) for all points x, y) in the plane. 5) 4 The Continuous Case. Now suppose that X and Y have continuous joint density 3 f XY x, y), so that X and Y separately have marginal densities f X x) 1 f XY x, y) dy and f Y y) f XY x, y) dx. I will need the following fact for the proof of the ) direction of Theorem 4. Theorem 3 x y F XY x, y)) f XY x, y). 6) 2 Full disclosure department: even with this modification, this proof does not wor for all discrete random variables, because it is not always possible to arrange their values consecutively. 3 In general, of course, f XY x, y) need not be continuous, but by assuming continuity I can simplify the proofs. Theorems 4 and 5 are true even if f XY x, y) is not continuous.) 3

Proof. Step 1. For fixed x, put 4 g x ŷ) : Then f XY ˆx, ŷ) dˆx. y F XY x, y) y y f XY ˆx, ŷ) dˆx }{{} g xŷ) y ) g x ŷ) dŷ dy FTC ) g x y) f XY ˆx, y) dˆx. Step 2. Then, x y F XY x, y)) by Step 1 ) ) f XY ˆx, y) dˆx x FTC ) f XY x, y). There are two ways to determine from the joint density function whether or not X and Y are independent. The first of these Theorem 4) parallels equation 2) on p.2; the second characterization Theorem 5) generalizes the first one. Theorem 4 For continuous random variables X and Y with continuous joint density f XY x, y): X and Y are independent f XY x, y) f X x)f Y y) for all x, y). 7) Proof ): F XY x, y) y y F X x)f Y y). f XY ˆx, ŷ) dˆx dŷ f X ˆx)f Y ŷ) dˆx dŷ ) y ) f X ˆx) dˆx f Y ŷ) dŷ 4 The variables ˆx and ŷ are variables of integration. They are logically distinct from the variables x and y, which appear on integral signs. 4

Proof ): f XY x, y) by equation 6) on p.3 ) x y F XY x, y)) y F Xx)F Y y)) by independence ) x x )) d F X x) dy F Y y) x F Xx)f Y y)) f Y y) d dx F Xx)) f Y y)f X x). As mentioned above, Theorem 5 generalizes Theorem 4. The difference between the theorems is that in equation 8), the functions gx) and hy) do not need to be the marginal density functions f X x) and f Y y). Theorem 5 For continuous random variables X and Y with continuous joint density f XY x, y): X and Y are independent there exist continuous nonnegative functions gx) and hy) such that f XY x, y) gx)hy). 8) Proof ): This is immediate from Theorem 4: tae gx) f X x) and hy) f Y y). Proof ): Let f XY x, y) gx)hy). I will show it is also true that f XY x, y) f X x)f Y y), so that X and Y are independent by Theorem 4. Step 1. We have f X x) f XY x, y) dy gx)hy) dy gx) hy) dy gx), 9) }{{ } so that ) ) hy) hy) f XY x, y) gx)hy) gx) f X x). In Step 2, I will show that hy) f Y y) thus completing the proof). Step 2. A calculation parallel to that in equation 9) shows that I will complete the proof by showing that 1 f Y y) hy) gx) dx; gx) dx 1 : f XY x, y) dx dy gx)hy) dx dy ) ) hy) dy gx) dx ) gx) dx. 5