Section 2 Notes. Elizabeth Stone and Charles Wang. January 15, Expectation and Conditional Expectation of a Random Variable.

Similar documents
X X X E[ ] E X E X. is the ()m n where the ( i,)th. j element is the mean of the ( i,)th., then

ρ < 1 be five real numbers. The

( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model

Chapter 4 Multiple Random Variables

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Chapter 4 Multiple Random Variables

Econometric Methods. Review of Estimation

ENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Lecture 3 Probability review (cont d)

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1

Chapter 5 Properties of a Random Sample

Special Instructions / Useful Data

The expected value of a sum of random variables,, is the sum of the expected values:

Parameter, Statistic and Random Samples

Midterm Exam 1, section 2 (Solution) Thursday, February hour, 15 minutes

ECON 5360 Class Notes GMM

Continuous Distributions

Class 13,14 June 17, 19, 2015

Lecture Note to Rice Chapter 8

Generative classification models

Multivariate Transformation of Variables and Maximum Likelihood Estimation

Dr. Shalabh. Indian Institute of Technology Kanpur

X ε ) = 0, or equivalently, lim

22 Nonparametric Methods.

Summary of the lecture in Biostatistics

CHAPTER 3 POSTERIOR DISTRIBUTIONS

4. Standard Regression Model and Spatial Dependence Tests

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions.

Linear Regression with One Regressor

Law of Large Numbers

Recall MLR 5 Homskedasticity error u has the same variance given any values of the explanatory variables Var(u x1,...,xk) = 2 or E(UU ) = 2 I

TESTS BASED ON MAXIMUM LIKELIHOOD

Qualifying Exam Statistical Theory Problem Solutions August 2005

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model

Maximum Likelihood Estimation

Lecture Notes 2. The ability to manipulate matrices is critical in economics.

Econometrics. 3) Statistical properties of the OLS estimator

Lecture Notes Types of economic variables

ε. Therefore, the estimate

Simulation Output Analysis

STA302/1001-Fall 2008 Midterm Test October 21, 2008

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA

Simple Linear Regression and Correlation.

Introduction to Matrices and Matrix Approach to Simple Linear Regression

Fundamentals of Regression Analysis

Handout #8. X\Y f(x) 0 1/16 1/ / /16 3/ / /16 3/16 0 3/ /16 1/16 1/8 g(y) 1/16 1/4 3/8 1/4 1/16 1

Study of Correlation using Bayes Approach under bivariate Distributions

Multiple Linear Regression Analysis

Example: Multiple linear regression. Least squares regression. Repetition: Simple linear regression. Tron Anders Moger

Midterm Exam 1, section 1 (Solution) Thursday, February hour, 15 minutes

Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests. Soccer Goals in European Premier Leagues

2SLS Estimates ECON In this case, begin with the assumption that E[ i

1 Solution to Problem 6.40

THE ROYAL STATISTICAL SOCIETY 2016 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5

Module 7. Lecture 7: Statistical parameter estimation

PART ONE. Solutions to Exercises

ESS Line Fitting

Overview. Basic concepts of Bayesian learning. Most probable model given data Coin tosses Linear regression Logistic regression

Comparison of Dual to Ratio-Cum-Product Estimators of Population Mean

Point Estimation: definition of estimators

STK3100 and STK4100 Autumn 2018

STA 105-M BASIC STATISTICS (This is a multiple choice paper.)

Part I: Background on the Binomial Distribution

STK3100 and STK4100 Autumn 2017

9 U-STATISTICS. Eh =(m!) 1 Eh(X (1),..., X (m ) ) i.i.d

Machine Learning. Introduction to Regression. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Lecture 2: The Simple Regression Model

ECONOMETRIC THEORY. MODULE VIII Lecture - 26 Heteroskedasticity

A be a probability space. A random vector

6.867 Machine Learning

Useful Statistical Identities, Inequalities and Manipulations

Mean is only appropriate for interval or ratio scales, not ordinal or nominal.

Math Tricks. Basic Probability. x k. (Combination - number of ways to group r of n objects, order not important) (a is constant, 0 < r < 1)

( ) ( ) ( ) f ( ) ( )

Mathematics HL and Further mathematics HL Formula booklet

Extreme Value Theory: An Introduction

Third handout: On the Gini Index

Probability and. Lecture 13: and Correlation

CHAPTER VI Statistical Analysis of Experimental Data

Chapter 2 Supplemental Text Material

7. Joint Distributions

{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution:

Multiple Choice Test. Chapter Adequacy of Models for Regression

= 2. Statistic - function that doesn't depend on any of the known parameters; examples:

Functions of Random Variables

i 2 σ ) i = 1,2,...,n , and = 3.01 = 4.01

Statistics. Correlational. Dr. Ayman Eldeib. Simple Linear Regression and Correlation. SBE 304: Linear Regression & Correlation 1/3/2018

A NEW LOG-NORMAL DISTRIBUTION

Linear Regression Linear Regression with Shrinkage. Some slides are due to Tommi Jaakkola, MIT AI Lab

LINEAR REGRESSION ANALYSIS

Chapter 5 Properties of a Random Sample

Random Variables and Probability Distributions

Fridayʼs lecture" Problem solutions" Joint densities" 1."E(X) xf (x) dx (x,y) dy X,Y Marginal distributions" The distribution of a ratio" Problems"

CS 2750 Machine Learning Lecture 5. Density estimation. Density estimation

Discrete Mathematics and Probability Theory Fall 2016 Seshia and Walrand DIS 10b

DISTURBANCE TERMS. is a scalar and x i

Lecture Notes to Rice Chapter 5

Chapter 14 Logistic Regression Models

Lecture 2: Linear Least Squares Regression

Transcription:

Secto Notes Elzabeth Stoe ad Charles Wag Jauar 5, 9 Jot, Margal, ad Codtoal Probablt Useful Rules/Propertes. P ( x) P P ( x; ) or R f (x; ) d. P ( xj ) P (x; ) P ( ) 3. P ( x; ) P ( xj ) P ( ) 4. Baes Rule: P ( xj ) P ( P jx)p (x) P ( ;x) Expectato ad Codtoal Expectato of a Radom Varable. Expectato e to E () P xp(x) R xf (x)dx f dscrete RV f cotuous RV. E (a + b) ae () + b for a,b costat. E ( + ) E () + E ( ) 3. E (h() + g( )) E (h()) + E (g( )) for arbtrar fuctos g ad h. Codtoal Expectato e to E (j ) P Useful Rules/Propertes. fucto of R xp(xj ) f dscrete RV xfj (x)dx f cotuous RV. Take out what s beg codtoed: E ( j) E ( j) 3. Law of terated expectatos: E () E [E (j )]

Proof. E () xf x dx x f ; d dx () x x f j f d dx (3) x xf j f ddx (4) x xf j f dxd (5) x B xf j dxc @ x f d (6) {z } E(j ) E (E (j )) (7) 4. Usefuless of the Law of Iterated Expectatos B Proposto 3 If E ("j) the Cov ("; f()) for a arbtrar fucto g Proof. WTS: E ("f()) E ("f ()) E [E ("f()j)] E 4f()E ("j) 5 {z } Note: So, f we kow E ("j), we ca easl show that ad " are ucorrelated. (Use the dett fucto) 3 3 Covarace, Varace, ad Correlato 3. Covarace e to 4 (Uvarate) Cov (; ) E [( E ()) ( E ( ))] E ( ) E () E ( ) deoted b x h (Vectors): Cov (; ) E ( E ()) ( E ()) T E T E () E () T deoted b. Cov (a + ; ) Cov (; ) for a costat. Cov (a; b ) abcov (; ) 3. Cov ( + ; ) Cov (; ) + Cov (; ) 4. Cov (aw + b; c + d) accov (W; ) + adcov (W; ) + bccov(; ) + bdcov (; ) 5. V ar () Cov(; ) 6. If? ) Cov (; )

3. Varace h e to 5 (Uvarate) V ar () E ( E ()) E E () deoted b x h (Vectors): Cov (; ) E ( E ()) ( E ()) T E T E () E () T deoted b. V ar ( + ) Cov ( + ; + ) V ar () + V ar( ) + Cov(; ). V ar 3. Multvarate Example: Let W :The, V ar () E h( E ()) ( E ()) T h E (W EW ) E [(W EW ) ( E)] 4 E [(W EW ) ( E)] E h( E) W W W 3 5 3.3 Correlato e to 6 Corr (; ) Cov(; ) p V ar()v ar( ) deoted b x. x [ ; ] Proof. WTS: same as showg Cov(; ) p ad V ar()v ar( ) Cov(; ) p V ar()v ar( ) Cov(; ) p, or Cov (; ) p V ar () V ar ( ) V ar()v ar( ) Proof follows from Cauch-Schwartz equalt.. x or a + b (.e. ad are lear trasformatos of each other) 4 (Weak) Law of Large Numbers ad Cetral Lmt Theorem Proposto 7 (LLN, Uvarate) Let f ; :::; g be a sequece of depedetl ad detcall dstrbuted (d) radom varables wth E ( ) ad V ar ( ) : The, P z! P (LLN, Vector) Let f ; :::; g be a sequece of depedetl ad detcall dstrbuted (d) radom vectors wth E ( ) ad V ar ( ) E T E E T The, P! P Proof. Oe wa to show that a radom varable coverges probablt to a costat s to show that Var of the radom varable coverges to ad the bas goes to Use Chebchev Idequalt: P h E ( ) V ar()+(e[ ]) " " " for a arbtrar " E, ad V ar!! so that P "!! Proposto 8 (CLT,Uvarate) Let f ; :::; g be d wth E ( ) ad V ar ( ) :The, p P z! N ; (or p! N ; ) (CLT,Vector) Let f ; :::; g be d wth E ( ) ad V ar ( ) E P T p! N (; ) (or p! N (; ) ) E E T :The, 3

5 Normal ad Bvarate Normal Radom Varables e to 9 bvarate ormal radom varable s a x vector of radom varables, each beg a ormal radom varable. e.g. If N ; ; N ; ;the N (! ; ) where! ad. Lear trasformatos of ormals are ormall dstrbuted: f ~N ;, the for a; b costats, a + b~n a + b; a. How to d the mea / varace of lear trasformatos of bvarate ormals: If N (! ; ), the for x (matrx of costats) ad b x (vector of costats), + b N ( + b; ) 3. Codtoal dstrbuto: If N ;, the j x N B + (x ) ; @ {z C } 4. ddg/subtractg ormal radom varables results a ormal radom varable: Let N ; ; N ; the + N + ; + + 6 Ch-Squared strbuto e to Let ; ; :::; p be d N (; ), the + + ::: + p p Proposto Suppose W k N k ; V. The T (W ) T V (W ) (k). Clearl,. Pcture 3. Wh s t useful regressos? llows us to test hpothess o multple parameters at the same tme (e.g. Suppose ou have a regresso model Icome + Race+ Geder+" for whch ou have estmates ^ ; ^ ; ^ for the populato parameters. Suppose ou wat to test the hpothess that race N geder have o e ect o come,.e. H : ad ) > USE WL STTISTIC Wald Statstc: We saw class that uder the ull hpothess that true b for some hpotheszed, the Wald Statstc p T ^ b V p ^ b whch s dstrbuted (p). Note: p ^ b s ormall dstrbuted wth mea ad varace V uder the ull. 4

Example to show tuto behd wh Wald statstc s ch-sq dstrbuted: Let ~N (; ) ad N (; ) ad ; depedet. Let s wrte ths matrx otato: The, ~N B @ {z} b ; {z } V I C T ( b) T V ( b) T + () sce, stadard ormal ad depedet of each other. other example: Suppose ow ~N ; ad ~N ; ad ; depedet " # Note that here V so that V The T ( b) T V ( b) " + B C @ {z} N(;) + B C @ {z} N(;) # () 5

sce ; stadard ormal ad depedet of each other. What we do the regresso settg s essetall a more geeralzed verso of ths. (o t eed to kow): T ( b) T V ( b) ( b) T C T C ( b) b Cholesk decomposto of V [C ( b)] T [C ( b)] The matrx C essetall re-weghts the coordates such a wa that elemets of the ew vector C ( b) are stadard ormals ad are depedet of each other, ad therefore the er dot product s just a sum of depedet stadard ormal varables, whch s ch-square dstrbuted. 7 Extra Stu (o t Need to kow) 7. Trlog of Theorems (To help us gure out the lmtg dstrbuto for more complcated sequeces of RVs). Slutsk s Theorem: If! ad! P a; B! P b for a, b o-radom costats, the + B! a + b. Cotuous Mappg Theorem: If! P ad g s a cotuous fucto, the g ( )! P g ( ) 3. elta Method (To be covered later) example usg the above: Showg cosstec of the OLS coe cet: Suppose data s geerated accordg to a + bx + " where E ("j) ( ) Ex " ) ) We have a d sample of datapots, geerated accordg to ths process. Questo: What s the probablt lmt of the OLS coe cet ad the lmtg dstrbuto of the OLS coe cet? ^ ( ) ( ) ( + E) + ( ) E + x x! B LLN, P x x! P E (x x P ) B Cotuous mappg, x x!p [E (x x )] B LLN, P x "! P E (x " ) So, b Slutsk s, ^ P + x x P x "!P! x " 6