Joint Probability Distributions, Correlations

Similar documents
Joint Probability Distributions, Correlations

Multivariate Distributions CIVL 7012/8012

Continuous Probability Distributions. Uniform Distribution

Bivariate distributions

Ch. 5 Joint Probability Distributions and Random Samples

Continuous Probability Distributions. Uniform Distribution

Lecture 2: Repetition of probability theory and statistics

Lecture 16 : Independence, Covariance and Correlation of Discrete Random Variables

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Algorithms for Uncertainty Quantification

Math Review Sheet, Fall 2008

Joint Distribution of Two or More Random Variables

Learning Objectives for Stat 225

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Random Signals and Systems. Chapter 3. Jitendra K Tugnait. Department of Electrical & Computer Engineering. Auburn University.

Final Exam # 3. Sta 230: Probability. December 16, 2012

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Notes 12 Autumn 2005

Chapter 2. Probability

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Econ 325: Introduction to Empirical Economics

Chapter 5 continued. Chapter 5 sections

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

Recitation 2: Probability

Introduction to Machine Learning

Practice Examination # 3

Random Variables and Their Distributions

Multivariate Random Variable

Lecture 11. Probability Theory: an Overveiw

ENGG2430A-Homework 2

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

18.440: Lecture 28 Lectures Review

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be

Chapter 5 Joint Probability Distributions

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.

Introduction to Statistical Inference Self-study

Stochastic Models of Manufacturing Systems

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Introduction to Probability and Stocastic Processes - Part I

CDA5530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

Continuous Random Variables

Random Variables and Expectations

STAT 430/510: Lecture 16

Basics on Probability. Jingrui He 09/11/2007

Chapter 4 Multiple Random Variables

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

STA 256: Statistics and Probability I

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Review of Statistics

5. Conditional Distributions

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

MAS108 Probability I

4. Distributions of Functions of Random Variables

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

STT 441 Final Exam Fall 2013

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

Probability. Table of contents

MAS223 Statistical Inference and Modelling Exercises

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

6.041/6.431 Fall 2010 Quiz 2 Solutions

18.440: Lecture 28 Lectures Review

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Multivariate distributions

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Statistical Learning Theory

3. Probability and Statistics

CS145: Probability & Computing

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

STAT 430/510: Lecture 15

Class 8 Review Problems solutions, 18.05, Spring 2014

Computational Genomics

EE4601 Communication Systems

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

2 (Statistics) Random variables

3 Multiple Discrete Random Variables

Review of probability

Introduction to Stochastic Processes

Chapter 4 Multiple Random Variables

Chapter 2: Random Variables

We introduce methods that are useful in:

Communication Theory II

Lecture 1. ABC of Probability

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Chapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Problem 1. Problem 2. Problem 3. Problem 4

This does not cover everything on the final. Look at the posted practice problems for other topics.

Single Maths B: Introduction to Probability

1 Presessional Probability

Transcription:

Joint Probability Distributions, Correlations

What we learned so far Events: Working with events as sets: union, intersection, etc. Some events are simple: Head vs Tails, Cancer vs Healthy Some are more complex: 10<Gene expression<100 Some are evcen more complex: Series of dice rolls: 1,3,5,3,2 Conditional probability: P(A B)=P(A B)/P(B) Independent events: P(A B)=P(A) or P(A B)= P(A)*P(B) Bayes theorem: relates P(A B) to P(B A) Random variables: Mean, Variance, Standard deviation. How to work with E(g(X)) Discrete (Uniform, Bernoulli, Binomial, Poisson, Geometric, Negative binomial, Hypergeometric, Power law); PMF: f(x)=prob(x=x); CDF: F(x)=Prob(X x); Continuous (Uniform, Exponential, Erlang, Gamma, Normal, Lognormal); PDF: f(x) such that Prob(X inside A)= A f(x)dx; CDF: F(x)=Prob(X x) Next step: work with multiple random variables

Concept of Joint Probabilities Biological systems are usually described not by a single random variable but by many random variables Example: The expression state of a human cell: 20,000 random variables X i for each of its genes A joint probability distribution describes the behavior of several random variables We will start with just two random variables X and Y and generalize when necessary Chapter 5 Introduction 3

Joint Probability Mass Function Defined The joint probability mass function of the discrete random variables X and Y, denoted as f x, y, satifies: XY (1) f x, y 0 All probabilities are non-negative XY (2) f x, y 1 The sum of all x y XY probabilities is 1 (3) f x, y P X x, Y y (5-1) XY Sec 5 1.1 Joint Probability Distributions 4

Example 5 1: # Repeats vs. Signal Bars You use your cell phone to check your airline reservation. It asks you to speak the name of your departure city to the voice recognition system. Let Y denote the number of times you have to state your departure city. Let X denote the number of bars of signal strength on you cell phone. y = number of times city x = number of bars of signal strength name is stated 1 2 3 1 0.01 0.02 0.25 2 0.02 0.03 0.20 3 0.02 0.10 0.05 4 0.15 0.10 0.05 Figure 5 1 Joint probability distribution of X and Y. The table cells are the probabilities. Observe that more bars relate to less repeating. Probability 0.25 0.20 0.15 0.10 0.05 0.00 Bar Chart of Number of Repeats vs. Cell Phone Bars 1 2 Cell Phone Bars 3 4 Times 3 Times Twice Once Sec 5 1.1 Joint Probability Distributions 5

Marginal Probability Distributions (discrete) For a discrete joint PDF, there are marginal distributions for each random variable, formed by summing the joint PMF over the other variable. f x f x, y X y XY f y f x, y Y x XY Called marginal because they are written in the margins y = number of times city name x = number of bars of signal strength is stated 1 2 3 f Y ( y ) = 1 0.01 0.02 0.25 0.28 2 0.02 0.03 0.20 0.25 3 0.02 0.10 0.05 0.17 4 0.15 0.10 0.05 0.30 f X (x ) = 0.20 0.25 0.55 1.00 Figure 5 6 From the prior example, the joint PMF is shown in green while the two marginal PMFs are shown in purple. Sec 5 1.2 Marginal Probability Distributions 6

Mean & Variance of X and Y are calculated using marginal distributions y = number of times city x = number of bars of signal strength name is stated 1 2 3 f (y ) = y *f (y ) = y 2 *f (y ) = 1 0.01 0.02 0.25 0.28 0.28 0.28 2 0.02 0.03 0.20 0.25 0.50 1.00 3 0.02 0.10 0.05 0.17 0.51 1.53 4 0.15 0.10 0.05 0.30 1.20 4.80 f (x ) = 0.20 0.25 0.55 1.00 2.49 7.61 x *f (x ) = 0.20 0.50 1.65 2.35 x 2 *f (x ) = 0.20 1.00 4.95 6.15 μ X =E(X) = 2.35; σ X2 = V(X) = 6.15 2.35 2 = 6.15 5.52 = 0.6275 μ Y = E(Y) = 2.49; σ Y2 =V(Y) = 7.61 2.49 2 = 7.61 16.20 = 1.4099 Sec 5 1.2 Marginal Probability Distributions 7

Conditional Probability Distributions Recall that P B A P A B P A P(Y=y X=x)=P(X=x,Y=y)/P(X=x)= =f(x,y)/f X (x) From Example 5 1 P(Y=1 X=3) = 0.25/0.55 = 0.455 P(Y=2 X=3) = 0.20/0.55 = 0.364 P(Y=3 X=3) = 0.05/0.55 = 0.091 P(Y=4 X=3) = 0.05/0.55 = 0.091 Sum = 1.00 y = number of times city name is stated x = number of bars of signal strength 1 2 3 f Y (y ) = 1 0.01 0.02 0.25 0.28 2 0.02 0.03 0.20 0.25 3 0.02 0.10 0.05 0.17 4 0.15 0.10 0.05 0.30 f X (x ) = 0.20 0.25 0.55 1.00 Note that there are 12 probabilities conditional on X, and 12 more probabilities conditional upon Y. Sec 5 1.3 Conditional Probability Distributions 8

Joint Random Variable Independence Random variable independence means that knowledge of the values of X does not change any of the probabilities associated with the values of Y. Opposite: Dependence implies that the values of X are influenced by the values of Y Sec 5 1.4 Independence 9

Independence for Discrete Random Variables Remember independence of events (slide 21 lecture 3) : P(A B)=P(A B)/P(B)=P(A) or P(B A)= P(A B)/P(A)=P(B) or P(A B)=P(A) P(B) Random variables independent if any events A that Y=y and B that X=x are independent P(Y=y X=x)=P(Y=y) for any x or P(X=x Y=y)=P(X=x) for any y or P(X=x, Y=y)=P(X=x) P(Y=y) for any x and y

X and Y are Bernoulli variables Y=0 Y=1 X=0 2/6 1/6 X=1 2/6 1/6 What is the marginal P Y (Y=0)? A. 1/6 B. 2/6 C. 3/6 D. 4/6 E. I don t know Get your i clickers 12

X and Y are Bernoulli variables Y=0 Y=1 X=0 2/6 1/6 X=1 2/6 1/6 What is the conditional P(X=0 Y=0)? A. 2/6 B. 1/2 C. 1/6 D. 4/6 E. I don t know Get your i clickers 13

X and Y are Bernoulli variables Y=0 Y=1 X=0 2/6 1/6 X=1 2/6 1/6 Are they independent? A. yes B. no C. I don t know Get your i clickers 14

X and Y are Bernoulli variables Y=0 Y=1 X=0 1/2 0 X=1 0 1/2 Are they independent? A. yes B. no C. I don t know Get your i clickers 15

Credit: XKCD comics

Joint Probability Density Function Defined The joint probability density function for the continuous random variables X and Y, denotes as f XY (x,y), satisfies the following properties: (1) f x, y 0 for all x, y XY (2) f x, y dxdy 1 XY XY (3) P X, Y R f x, y dxdy (5-2) R Figure 5 2 Joint probability density function for the random variables X and Y. Probability that (X, Y) is in the region R is determined by the volume of f XY (x,y) over the region R. Sec 5 1.1 Joint Probability Distributions 17

Joint Probability Mass Function Graph Figure 5 3 Joint probability density function for the continuous random variables X and Y of expression levels of two different genes. Note the asymmetric, narrow ridge shape of the PDF indicating that small values in the X dimension are more likely to occur when small values in the Y dimension occur. Sec 5 1.1 Joint Probability Distributions 18

Marginal Probability Distributions (continuous) Rather than summing a discrete joint PMF, we integrate a continuous joint PDF. The marginal PDFs are used to make probability statements about one variable. If the joint probability density function of random variables X and Y is f XY (x,y), the marginal probability density functions of X and Y are: f x f x, y dy X y XY f y f x, y dx (5-3) Y x XY X Y, f x f x y, Sec 5 1.2 Marginal Probability Distributions 19 y x XY f y f x y XY

Conditional Probability Density Function Defined Given continuous random variables X and Y with joint probability density function f x, y, the conditional probability densiy function of Y given fxy x, y fxy x, y f y = if fx x0 Yx f x f x, y dy y X y XY Compare to discrete: P(Y=y X=x)=f XY (x,y)/f X (x) Yx which satifies the following properties: (1) f 0 Yx (2) f y dy 1 Yx XY Sec 5 1.3 Conditional Probability Distributions X=x is (5-4) (3) PY BX x f ydyfor any set B in the range of Y B 20

Conditional Probability Distributions Conditional probability distributions can be developed for multiple random variables by extension of the ideas used for two random variables. Suppose p = 5 and we wish to find the distribution of X 1, X 2 and X 3 conditional on X 4 =x 4 and X 5 =x 5. f x, x, x XX X 1 2 3 4 5 4 5 1 2 3 for f x, x 0. X X xx 4 5 f x, x, x, x, x XX XX X 1 2 3 4 5 4 5 1 2 3 4 5 f x, x X X 4 5 Sec 5 1.5 More Than Two Random Variables 21

Independence for Continuous Random Variables For random variables X and Y, if any one of the following properties is true, the others are also true. Then X and Y are independent. Y X X Y X A Y B PX APY B (1) f x, y f x f y XY X Y (2) f y f y for all x and y with f x 0 Yx (3) f y f x for all x and y with f y 0 Xy P(Y=y X=x)=P(Y=y) for any x or P(X=x Y=y)=P(X=x) for any y or P(X=x, Y=y)=P(X=x) P(Y=y) for any x and y (4) P, for any sets A and B in the range of X and Y, respectively. (5-7) Sec 5 1.4 Independence 22

X and Y are uniformly distributed in the disc x 2 +y 2 1 Are they independent? A. yes B. no C. I could not figure it out Get your i clickers 27

Credit: XKCD comics

Covariation, Correlations

Covariance Defined Covariance is a number qunatifying between two random variables. average dependence The covariance between the random variables X and Y, denoted as cov XY, or is E X Y E XY XY 5-14) XY X Y X Y The units of are units of X times units of Y. XY Unlike the range of variance, -. XY ( Sec 5 2 Covariance & Correlation 34

Covariance and PMF tables y = number of times city name is stated x = number of bars of signal strength 1 2 3 1 0.01 0.02 0.25 2 0.02 0.03 0.20 3 0.02 0.10 0.05 4 0.15 0.10 0.05 The probability distribution of Example 5 1 is shown. By inspection, note that the larger probabilities occur as X and Y move in opposite directions. This indicates a negative covariance. Sec 5 2 Covariance & Correlation 35

Covariance and Scatter Patterns Figure 5 13 Joint probability distributions and the sign of cov(x, Y). Note that covariance is a measure of linear relationship. Variables with non zero covariance are correlated. Sec 5 2 Covariance & Correlation 36

Independence Implies σ=ρ = 0 but not vice versa If X and Y are independent random variables, σ XY = ρ XY = 0 (5 17) ρ XY = 0 is necessary, but not a sufficient condition for independence. Independent covariance=0 NOT independent covariance=0 Sec 5 2 Covariance & Correlation 37

Correlation is normalized covariance Also called: Pearson correlation coefficient ρ XY =σ XY /σ X σ Y is the covariance normalized to be 1 ρ XY 1 Karl Pearson (1852 1936) English mathematician and biostatistician

Credit: XKCD comics