Correlation analysis 2: Measures of correlation

Size: px
Start display at page:

Download "Correlation analysis 2: Measures of correlation"

Transcription

1 Correlation analsis 2: Measures of correlation Ran Tibshirani Data Mining: / Februar

2 Review: correlation Pearson s correlation is a measure of linear association In the population: for random variables X, Y R, Cor(X, Y ) = Cov(X, Y ) Var(X) Var(Y ) In the sample: for vectors, R n, cor(, ) = If, are have been centered, then cov(, ) = ( 1)T ( ȳ1) var() var() 1 2 ȳ1 2 cor(, ) = T 2 2 2

3 Review: properties of population correlation Recall: Cor(X, Y ) = Cov(X, Y )/ ( Var(X) Var(Y ) ) Properties of Cor: 1. Cor(X, X) = 1 2. Cor(X, Y ) = Cor(Y, X) 3. Cor(aX + b, Y ) = sign(a)cor(x, Y ) for an a, b R 4. 1 Cor(X, Y ) 1, with Cor(X, Y ) > 0 indicating a positive (linear) relationship, < 0 indicating a negative one 5. Cor(X, Y ) = 1 if and onl if Y = ax + b for some a, b R, with a 0 6. If X, Y are independent then Cor(X, Y ) = 0 7. If Cor(X, Y ) = 0 then X, Y need not be independent 8. If (X, Y ) is bivariate normal and Cor(X, Y ) = 0, then X, Y are independent 3

4 Bivariate normal distribution Assume that the random vector Z = (X, Y ) R 2 has a bivariate normal distribution, Z N(µ, Σ), where µ R 2 and Σ R 2 2, ( ) ( ) µx σ 2 µ =, Σ = X ρσ X σ Y µ Y ρσ X σ Y σy 2 Note that we have E[X] = µ X, E[Y ] = µ Y, Var(X) = σ 2 X, Var(Y ) = σ 2 X, Cov(X, Y ) = ρσ Xσ Y, and Cor(X, Y ) = ρ The densit of Z = (X, Y ) is f X,Y (z) = 1 ( 2π det(σ) ep 1 ) 2 (z µ)t Σ 1 (z µ) Fact: ρ = 0 implies that X, Y are independent (Homework 3) 4

5 Review: properties of sample correlation Recall: cor(, ) = ( 1) T ( ȳ1)/ ( 1 2 ȳ1 2 ) Properties of cor: 1. cor(, ) = 1 2. cor(, ) = cor(, ) 3. cor(a + b, ) = sign(a)cor(, ) for an a, b R 4. 1 cor(, ) 1, with cor(, ) > 0 indicating a positive (linear) relationship, < 0 indicating a negative one 5. cor(, ) = 1 if and onl if = a + b for some a, b R, with a 0 6. cor(, ) = 0 if and onl if, are orthogonal 7. If, are centered then cor(, ) = cos θ, where θ is the angle between the vectors, R n cos θ = T 2 2 θ 5

6 Eamples: sample correlation Perfect linear Nois linear cor =

7 Independent Ball cor =

8 Perfect cubic Outliers cor =

9 Rank correlation Spearman s rank correlation is onl defined in the sample. It goes beond linearit, and measures a monotone association between, R n Given vectors, R n, we first define the rank vector r R n that ranks the components of, e.g., if = (0.7, 0.1, 1) then r = (2, 1, 3). We also define the ranks r based on. Rank correlation is now the usual (sample) correlation of r and r : rcor(, ) = cor(r, r ) Ke propert: rcor(, ) = 1 if and onl if there is a monotone function f : R R such that i = f( i ) for each i = 1,... n 9

10 Eamples: rank correlation Perfect linear Nois linear rcor = cor =

11 Independent Ball rcor = cor =

12 Perfect cubic Outliers rcor = cor =

13 13 Perfect quadratic Perfect circle rcor = cor =

14 Maimal correlation Maimal correlation 1 is a notion of population correlation. It has no preference for linearit or monotonicit, and it characterizes independence completel Given two random variables X, Y R, the maimal correlation between X, Y is defined as mcor(x, Y ) = ma f,g Cor(f(X), g(y )) where the maimum is taken over all functions 2 f, g : R R. Note that 0 mcor(x, Y ) 1 Ke propert: mcor(x, Y ) = 0 if and onl if X, Y are independent 1 Gebelein (1947), Das Statitistiche Problem Der Korrelation... ; Reni (1959), On Measures of Dependence 2 Actuall, f, g have to be such that Var(f(X)) > 0, Var(g(Y )) > 0 14

15 15 Review: independence Two random variables X, Y R are called independent if for an sets A, B R P(X A, Y B) = P(X A)P(Y B) If X, Y have densities f X, f Y, and (X, Y ) has a densit f X,Y, then this is equivalent to f X,Y (s, t) = f X (s)f Y (t) for an s, t R. In other words, the joint densit is the product of the marginal densities Important fact: if X, Y are independent, then for an functions f, g, we have E[f(X)g(Y )] = E[f(X)]E[g(Y )] Hence X, Y independent implies that Cor(f(X), g(y )) = 0 for an functions f, g, which means mcor(x, Y ) = 0

16 16 Review: characteristic functions The characteristic function of a random variable X R is the function h X : R C defined as h X (t) = E[ep(itX)] The characteristic function of a random vector (X, Y ) R 2 is h X,Y : R 2 C, h X,Y (s, t) = E [ ep ( i(sx + ty ) )] Characteristic functions completel characterize the distribution of a random variable (hence the name). I.e., if h X (t) = h X (t) for ever t R, then X and X must have the same distribution Important fact: X, Y are independent if and onl if h X,Y (s, t) = h X (s)h Y (t) for an s, t R. In other words, the joint characteristic function is the product of the marginal ones

17 Zero maimal correlation implies independence Suppose that mcor(x, Y ) = 0. Then taking 3 f(x) = ep(isx) and g(y ) = ep(ity ) we have This means that i.e., Cor(ep(isX), ep(ity )) = 0 Cov(ep(isX), ep(ity )) = 0 E[ep(isX) ep(ity )] = E[ep(isX)]E[ep(itY )] This holds for each s, t R, so X, Y are independent 3 Strictl speaking, f, g are supposed to take values in R (not C), but to get around this we can write ep(iθ) = cos θ + i sin θ and then break things up into real and imaginar parts 17

18 18 Maimal correlation in the sample Given two vectors, R n, and functions f, g : R R, write f() R n for the vector with ith component f( i ), i.e., f() = (f( 1 ),... f( n )), and similarl for g() We d like to define maimal correlation in the sample, analogous to its population definition (so that we can use it in practice!). Consider mcor(, ) = ma cor(f(), g()) f,g There s a big problem here as defined, mcor(, ) = 1 for an, R n. (Wh?) We ll derive an algorithm to compute mcor in the population. This inspires an algorithm in the sample, and the sample version mcor is then defined as the output of this algorithm

19 19 Fied points of maimal correlation We define a norm on random variables Z R b Z = E[Z 2 ] Given an function f, we can alwas make f(z) have the following properties: E[f(Z)] = 0, b letting f(z) f(z) E[f(Z)] (centering) f(z) = 1, b letting f(z) f(z)/ f(z) (scaling) Notice that for such an f, we have Var(f(Z)) = f(z) = 1 Therefore, when computing maimal correlation, we can restrict our attention to functions f, g such that E[f(X)] = E[g(Y )] = 0 and f(x) = g(y ) = 1. (Otherwise, we simpl center and scale as needed, and this doesn t change the correlation.) Hence mcor(x, Y ) = ma E[f(X)]=E[g(Y )]=0 f(x) = g(y ) =1 E[f(X)g(Y )]

20 Now notice that for such f, g, E[(f(X) g(y )) 2 ] = E[f 2 (X)] + E[g 2 (Y )] 2E[f(X)g(Y )] = 2 2E[f(X)g(Y )] and hence f, g are optimal functions for mcor if and onl if the are also optimal for the minimization problem min E[f(X)]=E[g(Y )]=0 f(x) = g(y ) =1 E[(f(X) g(y )) 2 ] Ke point: consider just minimizing over the function g (assuming that f is fied and satisties E[f(X)] = 0 and f(x) = 1). To do so, we can minimize the conditional epectation E[(f(X) g()) 2 Y = ] for each fied. And to minimize this over g(), we simpl take g() = E[f(X) Y = ] As a function over Y, this is written as g(y ) = E[f(X) Y ] 20

21 21 Remember though, that we are restricting g to have epectation zero and norm one. Therefore we let g(y ) = E[f(X) Y ]/ E[f(X) Y ] (Check: this satisfies both properties) The eact same arguments, but in reverse, show that minimizing over f for fied g ields f(x) = E[g(Y ) X]/ E[g(Y ) X] These are called the fied point equations of the maimal correlation problem. That is, if there eists f, g that achieved the maimum mcor(x, Y ) = E[f(X)g(Y )], then the must satisf the above equations

22 Alternating conditional epectations algorithm The alternating conditional epectations (ACE) algorithm 4 is motivated directl from these fied point equations. The idea is just to start with a guess at one of the functions, and then iterate the two equations until convergence ACE algorithm: Set f 0 (X) = (X E[X])/ X E[X] For k = 1, 2, 3, Let g k (Y ) = E[f k 1 (X) Y ]/ E[f k 1 (X) Y ] 2. Let f k (X) = E[g k (Y ) X]/ E[g k (Y ) X] 3. Stop if E[f k (X)g k (Y )] = E[f k 1 (X)g k 1 (Y )] Upon convergence, mcor(x, Y ) = E[f k (X)g k (Y )] This has been proven to converge under ver general assumptions 4 Breiman and Friedman (1985), Estimating Optimal Transformations for Multiple Regression and Correlation 22

23 Back to maimal correlation in the sample Given, R n. As in the population case, if we are considering cor(f(), g()) over all functions f, g, we can restrict out attention to functions f, g with 1 T f() = 1 T g() = 0 and f() 2 = g() 2 = 1 where now 2 is the usual Euclidean norm. For such functions f, g, we have cor(f(), g()) = f() T g(). Further, f() g() 2 2 = f() g() 2 2 2f() T g() = 2 2f() T g() for such functions f, g, so maimizing cor(f(), g()) is the same as minimizing f() g() 2 2 We re going to derive a sample version of the ACE algorithm to approimatel minimize f() g() 2 2, and then we define its output as the sample maimal correlation 23

24 24 The ACE algorithm in the sample We define the ACE algorithm analogousl to its definition in the population case. That is, we repeat iterations of the form g() = condep[f() ], center and scale g(), f() = condep[g() ], center and scale f(), where condep[ ] denotes a sample version of the conditional epectation on, and similar for condep[ ] A good question is: how do we compute these sample conditional epectations? (There s not a simple convenient sample analog like there is for unconditional epectation, variance, or covariance) Answer: use a smoother S. We use the notation S( ) to mean that the result is an estimate for = ( 1,... n ) as a function of = ( 1,... n )

25 25 A smoother S( ) could be, e.g., something as simple as a histogram or linear regression. It could also be something more fanc like kernel regression or local linear regression. (We ll see eamples of smoothers later in the course) smoother 1 smoother

26 Assume that we ve picked a smoother S to use in the ACE algorithm. (We d like a smoother that can fit ver general shapes of functions, i.e., not just a histogram or linear regression) ACE algorithm: Set f 0 () = ( 1)/ 1 2 For k = 1, 2, 3, Let G() = S(f k 1 () ), and center and scale, g k () = (G() G()1)/ G() G() Let F () = S(g k () ), and center and scale, f k () = (F () F ()1)/ F () F () Stop if f k () T g k () f k 1 () T g k 1 () 2 is small Upon convergence, define mcor(, ) = f k () T g k () Unforunatel, the ACE algorithm in the sample is onl guaranteed to converge under some restrictive conditions. But in practice, it still tends to perform quite well 26

27 Measures of correlation in R The cor function available in the base R distribution can be used to compute (Pearson s) correlation, and also (Spearman s) rank correlation. E.g., cor(, ) # default is method="pearson" cor(,, method="spearman") # rank correlation The function ace in the package acepack implements the alternating conditional epectations algorithm to compute maimum correlation. E.g., a = ace(, ) cor(a$t, a$t) # maimal correlation Note: this ace implementation doesn t scale the vectors in the wa that discussed in class, so the maimal correlation is not simpl sum(a$t * a$t) 27

28 Recap: measures of correlation In this lecture we reviewed the basic facts about correlation, both in the population (random variables) and in the sample (vectors). If random variables X, Y R are joint normal, then independence and uncorrelatedness are the same thing Rank correlation is simpl the usual sample correlation, ecept replacing vectors, R n with the ranks of their components r X, r Y R n. This aims to capture monotone associations that are not necessaril linear Maimal correlation is defined in the population as the maimum correlation over functions of our two random variables. The maimal correlation equals zero if and onl if the random variables are independent The alternating conditional epectations (ACE) algorithm is an elegant wa to compute maimal correlation in the population, and furthermore, it allows to define maimal correlation in the sample 28

29 Net time: more measures of correlation More about maimal correlation; distance correlation Data Transformed data g() f() Transformation of Transformation of f() g()

Covariance and Correlation Class 7, Jeremy Orloff and Jonathan Bloom

Covariance and Correlation Class 7, Jeremy Orloff and Jonathan Bloom 1 Learning Goals Covariance and Correlation Class 7, 18.05 Jerem Orloff and Jonathan Bloom 1. Understand the meaning of covariance and correlation. 2. Be able to compute the covariance and correlation

More information

11. Regression and Least Squares

11. Regression and Least Squares 11. Regression and Least Squares Prof. Tesler Math 186 Winter 2016 Prof. Tesler Ch. 11: Linear Regression Math 186 / Winter 2016 1 / 23 Regression Given n points ( 1, 1 ), ( 2, 2 ),..., we want to determine

More information

Copyright, 2008, R.E. Kass, E.N. Brown, and U. Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS

Copyright, 2008, R.E. Kass, E.N. Brown, and U. Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS Copright, 8, RE Kass, EN Brown, and U Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS Chapter 6 Random Vectors and Multivariate Distributions 6 Random Vectors In Section?? we etended

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

The data can be downloaded as an Excel file under Econ2130 at

The data can be downloaded as an Excel file under Econ2130 at 1 HG Revised Sept. 018 Supplement to lecture 9 (Tuesda 18 Sept) On the bivariate normal model Eample: daughter s height (Y) vs. mother s height (). Data collected on Econ 130 lectures 010-01. The data

More information

Inference about the Slope and Intercept

Inference about the Slope and Intercept Inference about the Slope and Intercept Recall, we have established that the least square estimates and 0 are linear combinations of the Y i s. Further, we have showed that the are unbiased and have the

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

Summary of Random Variable Concepts March 17, 2000

Summary of Random Variable Concepts March 17, 2000 Summar of Random Variable Concepts March 17, 2000 This is a list of important concepts we have covered, rather than a review that devives or eplains them. Tpes of random variables discrete A random variable

More information

Biostatistics in Research Practice - Regression I

Biostatistics in Research Practice - Regression I Biostatistics in Research Practice - Regression I Simon Crouch 30th Januar 2007 In scientific studies, we often wish to model the relationships between observed variables over a sample of different subjects.

More information

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18 Variance reduction p. 1/18 Variance reduction Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Variance reduction p. 2/18 Example Use simulation to compute I = 1 0 e x dx We

More information

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /

More information

ENGG2430A-Homework 2

ENGG2430A-Homework 2 ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,

More information

18 Bivariate normal distribution I

18 Bivariate normal distribution I 8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer

More information

Triple Integrals. y x

Triple Integrals. y x Triple Integrals. (a) If is an solid (in space), what does the triple integral dv represent? Wh? (b) Suppose the shape of a solid object is described b the solid, and f(,, ) gives the densit of the object

More information

18.440: Lecture 26 Conditional expectation

18.440: Lecture 26 Conditional expectation 18.440: Lecture 26 Conditional expectation Scott Sheffield MIT 1 Outline Conditional probability distributions Conditional expectation Interpretation and examples 2 Outline Conditional probability distributions

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) Last modified: March 7, 2009 Reference: PRP, Sections 3.6 and 3.7. 1. Tail-Sum Theorem

More information

Joint ] X 5) P[ 6) P[ (, ) = y 2. x 1. , y. , ( x, y ) 2, (

Joint ] X 5) P[ 6) P[ (, ) = y 2. x 1. , y. , ( x, y ) 2, ( Two-dimensional Random Vectors Joint Cumulative Distrib bution Functio n F, (, ) P[ and ] Properties: ) F, (, ) = ) F, 3) F, F 4), (, ) = F 5) P[ < 6) P[ < (, ) is a non-decreasing unction (, ) = F ( ),,,

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Probability Densities in Data Mining

Probability Densities in Data Mining Probabilit Densities in Data Mining Note to other teachers and users of these slides. Andrew would be delighted if ou found this source material useful in giving our own lectures. Feel free to use these

More information

Math 53 Homework 4 Solutions

Math 53 Homework 4 Solutions Math 5 Homework 4 Solutions Problem 1. (a) z = is a paraboloid with its highest point at (0,0,) and intersecting the -plane at the circle + = of radius. (or: rotate the parabola z = in the z-plane about

More information

15. Eigenvalues, Eigenvectors

15. Eigenvalues, Eigenvectors 5 Eigenvalues, Eigenvectors Matri of a Linear Transformation Consider a linear ( transformation ) L : a b R 2 R 2 Suppose we know that L and L Then c d because of linearit, we can determine what L does

More information

Covariance and Correlation

Covariance and Correlation Covariance and Correlation ST 370 The probability distribution of a random variable gives complete information about its behavior, but its mean and variance are useful summaries. Similarly, the joint probability

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

Scatter Plot Quadrants. Setting. Data pairs of two attributes X & Y, measured at N sampling units:

Scatter Plot Quadrants. Setting. Data pairs of two attributes X & Y, measured at N sampling units: Geog 20C: Phaedon C Kriakidis Setting Data pairs of two attributes X & Y, measured at sampling units: ṇ and ṇ there are pairs of attribute values {( n, n ),,,} Scatter plot: graph of - versus -values in

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems

More information

Random Vectors. 1 Joint distribution of a random vector. 1 Joint distribution of a random vector

Random Vectors. 1 Joint distribution of a random vector. 1 Joint distribution of a random vector Random Vectors Joint distribution of a random vector Joint distributionof of a random vector Marginal and conditional distributions Previousl, we studied probabilit distributions of a random variable.

More information

NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part I

NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part I NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Comple Analsis II Lecture Notes Part I Chapter 1 Preliminar results/review of Comple Analsis I These are more detailed notes for the results

More information

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane 3.5 Plane Stress This section is concerned with a special two-dimensional state of stress called plane stress. It is important for two reasons: () it arises in real components (particularl in thin components

More information

Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review

Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review You can t see this text! Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Matrix Algebra Review 1 / 54 Outline 1

More information

Chapter 4 continued. Chapter 4 sections

Chapter 4 continued. Chapter 4 sections Chapter 4 sections Chapter 4 continued 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP:

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Correlation analysis. Contents

Correlation analysis. Contents Correlation analysis Contents 1 Correlation analysis 2 1.1 Distribution function and independence of random variables.......... 2 1.2 Measures of statistical links between two random variables...........

More information

Recall Discrete Distribution. 5.2 Continuous Random Variable. A probability histogram. Density Function 3/27/2012

Recall Discrete Distribution. 5.2 Continuous Random Variable. A probability histogram. Density Function 3/27/2012 3/7/ Recall Discrete Distribution 5. Continuous Random Variable For a discrete distribution, for eample Binomial distribution with n=5, and p=.4, the probabilit distribution is f().7776.59.3456.34.768.4.3

More information

Vector Fields. Field (II) Field (V)

Vector Fields. Field (II) Field (V) Math 1a Vector Fields 1. Match the following vector fields to the pictures, below. Eplain our reasoning. (Notice that in some of the pictures all of the vectors have been uniforml scaled so that the picture

More information

Stat 206: Sampling theory, sample moments, mahalanobis

Stat 206: Sampling theory, sample moments, mahalanobis Stat 206: Sampling theory, sample moments, mahalanobis topology James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Notation My notation is different from the book s. This is partly because

More information

6. Vector Random Variables

6. Vector Random Variables 6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following

More information

Lecture 16 - Correlation and Regression

Lecture 16 - Correlation and Regression Lecture 16 - Correlation and Regression Statistics 102 Colin Rundel April 1, 2013 Modeling numerical variables Modeling numerical variables So far we have worked with single numerical and categorical variables,

More information

4.3 Mean-Value Theorem and Monotonicity

4.3 Mean-Value Theorem and Monotonicity .3 Mean-Value Theorem and Monotonicit 1. Mean Value Theorem Theorem: Suppose that f is continuous on the interval a, b and differentiable on the interval a, b. Then there eists a number c in a, b such

More information

H.Algebra 2 Summer Review Packet

H.Algebra 2 Summer Review Packet H.Algebra Summer Review Packet 1 Correlation of Algebra Summer Packet with Algebra 1 Objectives A. Simplifing Polnomial Epressions Objectives: The student will be able to: Use the commutative, associative,

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Let X denote a random variable, and z = h(x) a function of x. Consider the

Let X denote a random variable, and z = h(x) a function of x. Consider the EE385 Class Notes 11/13/01 John Stensb Chapter 5 Moments and Conditional Statistics Let denote a random variable, and z = h(x) a function of x. Consider the transformation Z = h(). We saw that we could

More information

Bivariate Paired Numerical Data

Bivariate Paired Numerical Data Bivariate Paired Numerical Data Pearson s correlation, Spearman s ρ and Kendall s τ, tests of independence University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

matrix-free Elements of Probability Theory 1 Random Variables and Distributions Contents Elements of Probability Theory 2

matrix-free Elements of Probability Theory 1 Random Variables and Distributions Contents Elements of Probability Theory 2 Short Guides to Microeconometrics Fall 2018 Kurt Schmidheiny Unversität Basel Elements of Probability Theory 2 1 Random Variables and Distributions Contents Elements of Probability Theory matrix-free 1

More information

ACM 116: Lectures 3 4

ACM 116: Lectures 3 4 1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance

More information

STOR Lecture 16. Properties of Expectation - I

STOR Lecture 16. Properties of Expectation - I STOR 435.001 Lecture 16 Properties of Expectation - I Jan Hannig UNC Chapel Hill 1 / 22 Motivation Recall we found joint distributions to be pretty complicated objects. Need various tools from combinatorics

More information

Research Design - - Topic 15a Introduction to Multivariate Analyses 2009 R.C. Gardner, Ph.D.

Research Design - - Topic 15a Introduction to Multivariate Analyses 2009 R.C. Gardner, Ph.D. Research Design - - Topic 15a Introduction to Multivariate Analses 009 R.C. Gardner, Ph.D. Major Characteristics of Multivariate Procedures Overview of Multivariate Techniques Bivariate Regression and

More information

f(x) = 2x 2 + 2x - 4

f(x) = 2x 2 + 2x - 4 4-1 Graphing Quadratic Functions What You ll Learn Scan the tet under the Now heading. List two things ou will learn about in the lesson. 1. Active Vocabular 2. New Vocabular Label each bo with the terms

More information

4 Inverse function theorem

4 Inverse function theorem Tel Aviv Universit, 2013/14 Analsis-III,IV 53 4 Inverse function theorem 4a What is the problem................ 53 4b Simple observations before the theorem..... 54 4c The theorem.....................

More information

Elements of Probability Theory

Elements of Probability Theory Short Guides to Microeconometrics Fall 2016 Kurt Schmidheiny Unversität Basel Elements of Probability Theory Contents 1 Random Variables and Distributions 2 1.1 Univariate Random Variables and Distributions......

More information

University of California, Los Angeles Department of Statistics. Joint probability distributions

University of California, Los Angeles Department of Statistics. Joint probability distributions Universit of California, Los Angeles Department of Statistics Statistics 100A Instructor: Nicolas Christou Joint probabilit distributions So far we have considered onl distributions with one random variable.

More information

Two-dimensional Random Vectors

Two-dimensional Random Vectors 1 Two-dimensional Random Vectors Joint Cumulative Distribution Function (joint cd) [ ] F, ( x, ) P xand Properties: 1) F, (, ) = 1 ),, F (, ) = F ( x, ) = 0 3) F, ( x, ) is a non-decreasing unction 4)

More information

CHAPTER 5. Jointly Probability Mass Function for Two Discrete Distributed Random Variables:

CHAPTER 5. Jointly Probability Mass Function for Two Discrete Distributed Random Variables: CHAPTER 5 Jointl Distributed Random Variable There are some situations that experiment contains more than one variable and researcher interested in to stud joint behavior of several variables at the same

More information

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45

Two hours. Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER. 14 January :45 11:45 Two hours Statistical Tables to be provided THE UNIVERSITY OF MANCHESTER PROBABILITY 2 14 January 2015 09:45 11:45 Answer ALL four questions in Section A (40 marks in total) and TWO of the THREE questions

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential CHAPTER : Partial Derivatives.1 Definition of a Partial Derivative. Increments and Differential.3 Chain Rules.4 Local Etrema.5 Absolute Etrema 1 Chapter : Partial Derivatives.1 Definition of a Partial

More information

CONTINUOUS SPATIAL DATA ANALYSIS

CONTINUOUS SPATIAL DATA ANALYSIS CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s

More information

Statistics 3657 : Moment Approximations

Statistics 3657 : Moment Approximations Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Chapter 6: Large Random Samples Sections

Chapter 6: Large Random Samples Sections Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form: 3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by

More information

An Information Theory For Preferences

An Information Theory For Preferences An Information Theor For Preferences Ali E. Abbas Department of Management Science and Engineering, Stanford Universit, Stanford, Ca, 94305 Abstract. Recent literature in the last Maimum Entrop workshop

More information

Chapter 6. Self-Adjusting Data Structures

Chapter 6. Self-Adjusting Data Structures Chapter 6 Self-Adjusting Data Structures Chapter 5 describes a data structure that is able to achieve an epected quer time that is proportional to the entrop of the quer distribution. The most interesting

More information

Regression Analysis. Ordinary Least Squares. The Linear Model

Regression Analysis. Ordinary Least Squares. The Linear Model Regression Analysis Linear regression is one of the most widely used tools in statistics. Suppose we were jobless college students interested in finding out how big (or small) our salaries would be 20

More information

P1 Chapter 4 :: Graphs & Transformations

P1 Chapter 4 :: Graphs & Transformations P1 Chapter 4 :: Graphs & Transformations jfrost@tiffin.kingston.sch.uk www.drfrostmaths.com @DrFrostMaths Last modified: 14 th September 2017 Use of DrFrostMaths for practice Register for free at: www.drfrostmaths.com/homework

More information

Homework Assignments Math /02 Spring 2015

Homework Assignments Math /02 Spring 2015 Homework Assignments Math 1-01/0 Spring 015 Assignment 1 Due date : Frida, Januar Section 5.1, Page 159: #1-, 10, 11, 1; Section 5., Page 16: Find the slope and -intercept, and then plot the line in problems:

More information

Unit 2 Notes Packet on Quadratic Functions and Factoring

Unit 2 Notes Packet on Quadratic Functions and Factoring Name: Period: Unit Notes Packet on Quadratic Functions and Factoring Notes #: Graphing quadratic equations in standard form, verte form, and intercept form. A. Intro to Graphs of Quadratic Equations: a

More information

Problem Set 1. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 20

Problem Set 1. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 20 Problem Set MAS 6J/.6J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 0 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain a

More information

Machine learning - HT Maximum Likelihood

Machine learning - HT Maximum Likelihood Machine learning - HT 2016 3. Maximum Likelihood Varun Kanade University of Oxford January 27, 2016 Outline Probabilistic Framework Formulate linear regression in the language of probability Introduce

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

10.2 The Unit Circle: Cosine and Sine

10.2 The Unit Circle: Cosine and Sine 0. The Unit Circle: Cosine and Sine 77 0. The Unit Circle: Cosine and Sine In Section 0.., we introduced circular motion and derived a formula which describes the linear velocit of an object moving on

More information

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix: Joint Distributions Joint Distributions A bivariate normal distribution generalizes the concept of normal distribution to bivariate random variables It requires a matrix formulation of quadratic forms,

More information

Introduction to bivariate analysis

Introduction to bivariate analysis Introduction to bivariate analysis When one measurement is made on each observation, univariate analysis is applied. If more than one measurement is made on each observation, multivariate analysis is applied.

More information

Business Statistics 41000: Homework # 2 Solutions

Business Statistics 41000: Homework # 2 Solutions Business Statistics 4000: Homework # 2 Solutions Drew Creal February 9, 204 Question #. Discrete Random Variables and Their Distributions (a) The probabilities have to sum to, which means that 0. + 0.3

More information

( ) 2 3x=0 3x(x 3 1)=0 x=0 x=1

( ) 2 3x=0 3x(x 3 1)=0 x=0 x=1 Stewart Calculus ET 5e 05497;4. Partial Derivatives; 4.7 Maimum and Minimum Values. (a) First we compute D(,)= f (,) f (,) [ f (,)] =(4)() () =7. Since D(,)>0 and f (,)>0, f has a local minimum at (,)

More information

What if the characteristic equation has a double root?

What if the characteristic equation has a double root? MA 360 Lecture 17 - Summary of Recurrence Relations Friday, November 30, 018. Objectives: Prove basic facts about basic recurrence relations. Last time, we looked at the relational formula for a sequence

More information

1 Exercises for lecture 1

1 Exercises for lecture 1 1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )

More information

Introduction to bivariate analysis

Introduction to bivariate analysis Introduction to bivariate analysis When one measurement is made on each observation, univariate analysis is applied. If more than one measurement is made on each observation, multivariate analysis is applied.

More information

9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown

9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown 9.07 Introduction to Probabilit and Statistics for Brain and Cognitive Sciences Emer N. Brown I. Objectives Lecture 4: Transformations of Random Variables, Joint Distributions of Random Variables A. Understand

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality. 88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal

More information

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline. MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y

More information

Optimization and Simulation

Optimization and Simulation Optimization and Simulation Variance reduction Michel Bierlaire Transport and Mobility Laboratory School of Architecture, Civil and Environmental Engineering Ecole Polytechnique Fédérale de Lausanne M.

More information

Next is material on matrix rank. Please see the handout

Next is material on matrix rank. Please see the handout B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0

More information

Jointly Distributed Random Variables

Jointly Distributed Random Variables Jointly Distributed Random Variables CE 311S What if there is more than one random variable we are interested in? How should you invest the extra money from your summer internship? To simplify matters,

More information

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours MATH2750 This question paper consists of 8 printed pages, each of which is identified by the reference MATH275. All calculators must carry an approval sticker issued by the School of Mathematics. c UNIVERSITY

More information

Math 20 Spring 2005 Final Exam Practice Problems (Set 2)

Math 20 Spring 2005 Final Exam Practice Problems (Set 2) Math 2 Spring 2 Final Eam Practice Problems (Set 2) 1. Find the etreme values of f(, ) = 2 2 + 3 2 4 on the region {(, ) 2 + 2 16}. 2. Allocation of Funds: A new editor has been allotted $6, to spend on

More information

Regression and Covariance

Regression and Covariance Regression and Covariance James K. Peterson Department of Biological ciences and Department of Mathematical ciences Clemson University April 16, 2014 Outline A Review of Regression Regression and Covariance

More information

Problem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51

Problem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51 Yi Lu Correlation and Covariance Yi Lu ECE 313 2/51 Definition Let X and Y be random variables with finite second moments. the correlation: E[XY ] Yi Lu ECE 313 3/51 Definition Let X and Y be random variables

More information

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y

More information

Chapter 3. Expectations

Chapter 3. Expectations pectations - Chapter. pectations.. Introduction In this chapter we define five mathematical epectations the mean variance standard deviation covariance and correlation coefficient. We appl these general

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Dependence and scatter-plots. MVE-495: Lecture 4 Correlation and Regression

Dependence and scatter-plots. MVE-495: Lecture 4 Correlation and Regression Dependence and scatter-plots MVE-495: Lecture 4 Correlation and Regression It is common for two or more quantitative variables to be measured on the same individuals. Then it is useful to consider what

More information

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't

More information