When is the OLSE the BLUE? Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 40

Similar documents
The Aitken Model. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 41

11. Linear Mixed-Effects Models. Copyright c 2018 Dan Nettleton (Iowa State University) 11. Statistics / 49

Linear Mixed-Effects Models. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 34

Estimable Functions and Their Least Squares Estimators. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 51

Estimation of the Response Mean. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 27

Estimating Estimable Functions of β. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 17

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X =

2. A Review of Some Key Linear Models Results. Copyright c 2018 Dan Nettleton (Iowa State University) 2. Statistics / 28

The Gauss-Markov Model. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 61

Constraints on Solutions to the Normal Equations. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 41

General Linear Test of a General Linear Hypothesis. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 35

REPEATED MEASURES. Copyright c 2012 (Iowa State University) Statistics / 29

Distributions of Quadratic Forms. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 31

ML and REML Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 58

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

The Multivariate Normal Distribution. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 36

Scheffé s Method. opyright c 2012 Dan Nettleton (Iowa State University) Statistics / 37

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

The Multivariate Normal Distribution. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 36

Likelihood Ratio Test of a General Linear Hypothesis. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 42

STAT 100C: Linear models

THE ANOVA APPROACH TO THE ANALYSIS OF LINEAR MIXED EFFECTS MODELS

Weighted Least Squares

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Regression Analysis Chapter 2 Simple Linear Regression

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Miscellaneous Results, Solving Equations, and Generalized Inverses. opyright c 2012 Dan Nettleton (Iowa State University) Statistics / 51

Gauss Markov & Predictive Distributions

Applied Econometrics (QEM)

Determinants. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 25

Lecture 6: Linear models and Gauss-Markov theorem

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

3 Multiple Linear Regression

Weighted Least Squares

Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = The errors are uncorrelated with common variance:

3. The F Test for Comparing Reduced vs. Full Models. opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics / 43

20. REML Estimation of Variance Components. Copyright c 2018 (Iowa State University) 20. Statistics / 36

Statistics 910, #5 1. Regression Methods

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Lecture 14 Simple Linear Regression

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

Advanced Quantitative Methods: ordinary least squares

Regression Models - Introduction

Preliminary Linear Algebra 1. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 100

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares

variability of the model, represented by σ 2 and not accounted for by Xβ

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Statistical View of Least Squares

MLES & Multivariate Normal Theory

Chapter 5 Matrix Approach to Simple Linear Regression

2.7 Estimation with linear Restriction

Maximum Likelihood Estimation

MS&E 226: Small Data

Time Series Analysis

STAT 100C: Linear models

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key

Chapter 1 Simple Linear Regression (part 6: matrix version)

Simple Linear Regression

Two-Color Microarray Experimental Design Notation. Simple Examples of Analysis for a Single Gene. Microarray Experimental Design Notation

Econ 583 Final Exam Fall 2008

Chapter 3: Multiple Regression. August 14, 2018

MS&E 226: Small Data

2. Regression Review

Chapter 1 Linear Regression with One Predictor

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Generalized Linear Mixed-Effects Models. Copyright c 2015 Dan Nettleton (Iowa State University) Statistics / 58

Simple Linear Regression (Part 3)

Multivariate Regression

STAT5044: Regression and Anova. Inyoung Kim

Well-developed and understood properties

ECON The Simple Regression Model

Estimating the slope and standard error of a long-term. linear trend fitted to adjusted annual temperatures

CLUSTER EFFECTS AND SIMULTANEITY IN MULTILEVEL MODELS

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Apart from this page, you are not permitted to read the contents of this question paper until instructed to do so by an invigilator.

Properties of Summation Operator

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

STAT 540: Data Analysis and Regression

Ordinary Least Squares Regression

The Algebra of the Kronecker Product. Consider the matrix equation Y = AXB where

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Chapter 5 Prediction of Random Variables

Introduction to Estimation Methods for Time Series models. Lecture 1

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

13. The Cochran-Satterthwaite Approximation for Linear Combinations of Mean Squares

Lecture 2: Linear and Mixed Models

STAT5044: Regression and Anova. Inyoung Kim

Measuring the fit of the model - SSR

Linear Algebra Review

Non-Spherical Errors

Regression Shrinkage and Selection via the Lasso

Regression and Statistical Inference

4 Multiple Linear Regression

1/15. Over or under dispersion Problem

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear Model Under General Variance

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Transcription:

When is the OLSE the BLUE? Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 40

When is the Ordinary Least Squares Estimator (OLSE) the Best Linear Unbiased Estimator (BLUE)? Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 2 / 40

We already know the OLSE is the BLUE under the GMM, but are there other situations where the OLSE is the BLUE? Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 3 / 40

Consider an experiment involving 4 plants. Two leaves are randomly selected from each plant. One leaf from each plant is randomly selected for treatment with treatment 1. The other leaf from each plant receives treatment 2. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 4 / 40

Let y ij = the response for the treatment i leaf from plant j (i = 1, 2; j = 1, 2, 3, 4). Suppose y ij = µ i + p j + e ij, where p 1,..., p 4, e 11,..., e 24 are uncorrelated, E(p j ) = E(e ij ) = 0, Var(p j ) = σ 2 p, Var(e ij ) = σ 2 i = 1, 2; j = 1,..., 4. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 5 / 40

Suppose σ 2 p/σ 2 is known to be equal to 2 and y = (y 11, y 21, y 12, y 22, y 13, y 23, y 14, y 24 ). Show that the AM holds. Find OLSE pf µ 1 µ 2 and the BLUE of µ 1 µ 2. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 6 / 40

E(y) = Xβ, where 1 0 0 1 1 0 0 1 X = 1 0 0 1 1 0 0 1 and β = [ µ1 µ 2 ]. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 7 / 40

Var(y ij ) = Var(µ i + p j + e ij ) = Var(p j + e ij ) = Var(p j ) + Var(e ij ) = σp 2 + σ 2 = σ 2 (σp/σ 2 2 + 1) = 3σ 2. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 8 / 40

Cov(y 1j, y 2j ) = Cov(µ 1 + p j + e 1j, µ 2 + p j + e 2j ) = Cov(p j + e 1j, p j + e 2j ) = Cov(p j, p j ) + Cov(p j, e 2j ) + Cov(p j, e 1j ) + Cov(e 1j, e 2j ) = Cov(p j, p j ) = Var(p j ) = σp 2 = σ 2 (σp/σ 2 2 ) = 2σ 2. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 9 / 40

Cov(y ij, y i j ) = 0 if j j because = Cov(p j + e ij, p j + e i j ) = Cov(p j, p j ) + Cov(p j, e i j ) + Cov(p j, e ij ) + Cov(e ij, e i j ) = 0. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 10 / 40

Thus, Var(y) = σ 2 V, where 3 2 0 0 0 0 0 0 2 3 0 0 0 0 0 0 0 0 3 2 0 0 0 0 0 0 2 3 0 0 0 0 V =. 0 0 0 0 3 2 0 0 0 0 0 0 2 3 0 0 0 0 0 0 0 0 3 2 0 0 0 0 0 0 2 3 Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 11 / 40

We can write the model as y = Xβ + ε, where p 1 + e 11 p 1 + e 21 p 2 + e 12 p 2 + e 22 ε =, E(ε) = 0 and Var(ε) = σ 2 V. p 3 + e 13 p 3 + e 23 p 4 + e 14 p 4 + e 24 opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 12 / 40

Note that I 2 2 I 2 2 X = I 2 2 I 2 2 = 1 I. 4 1 2 2 Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 13 / 40

Thus, I X I X = [I, I, I, I] I I = 4I 2 2 (X X) 1 = 1/4I. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 14 / 40

X y = [I, I, I, I]y = [ y1 y 2 ]. Thus, ˆβ OLS = 1 4 I [ y1 y 2 ] = [ȳ1 ȳ 2 ]. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 15 / 40

Thus, the OLSE of µ 1 µ 2 is [1, 1]ˆβ OLS = ȳ 1 ȳ 2. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 16 / 40

To find [ the] GLSE, which we know is the BLUE for this AM, let 3 2 A =. Then 2 3 A 0 0 0 0 A 0 0 V = 0 0 A 0 0 0 0 A = I A. 4 4 Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 17 / 40

V 1 = I A 1 = 4 4 A 1 0 0 0 0 A 1 0 0 0 0 A 1 0 0 0 0 A 1 Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 18 / 40

It follows that X V 1 X = [1 I][I A 1 ][1 I] A 1 0 0 0 I 0 A 1 0 0 I = [I, I, I, I] 0 0 A 1 0 I 0 0 0 A 1 I = 4A 1. Thus, (X V 1 X) 1 = 1 4 A. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 19 / 40

A 1 0 0 0 X V 1 0 A 1 0 0 y = [I, I, I, I] 0 0 A 1 y 0 0 0 0 A 1 = [A 1, A 1, A 1, A 1 ]y = A 1 [I, I, I, I]y. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 20 / 40

Thus, ˆβ GLS = 1 4 AA 1 [I, I, I, I]y = 1 [I, I, I, I]y 4 ] = [ȳ1 ȳ 2 = ˆβ OLS. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 21 / 40

Thus, the GLSE of µ 1 µ 2 and the BLUE of µ 1 µ 2 is [1, 1]ˆβ GLS = ȳ 1 ȳ 2. Thus, OLSE = GLSE = BLUE in this case. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 22 / 40

Although we assumed that σ 2 p/σ 2 = 2 in our example, that assumption was not needed to find the GLSE. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 23 / 40

We have looked at one specific example where the OLSE = GLSE = BLUE. What general conditions must be satisfied for this to hold? Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 24 / 40

Result 4.3: Suppose the AM holds. The estimator t y is the BLUE for E(t y) t y is uncorrelated with all linear unbiased estimators of zero. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 25 / 40

Proof: ( =) Let h y be an arbitrary linear unbiased estimator of E(t y). Then E((h t) y) = E(h y t y) = E(h y) E(t y) = 0. Thus, (h t) y is linear unbiased for zero. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 26 / 40

It follows that Cov(t y, (h t) y) = 0. Var(h y) = Var(h y t y + t y) Var(h y) Var(t y) with equality iff = Var((h t) y) + Var(t y). Var((h t) y) = 0 h = t. t y is the BLUE of E(t y). opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 27 / 40

(= ) Suppose h y is a linear unbiased estimator of zero. If h = 0, then Cov(t y, h y) = Cov(t y, 0) = 0. Now suppose h 0. Let c = Cov(t y, h y) and d = Var(h y) > 0. We need to show c = 0. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 28 / 40

Now consider a y = t y (c/d)h y. E(a y) = E(t y) (c/d)e(h y) = E(t y). Thus, a y is a linear unbiased estimator of E(t y). opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 29 / 40

Var(a y) = Var(t y (c/d)h y) = Var(t y) + c2 d 2 Var(h y) 2Cov(t y, (c/d)h y) = Var(t y) + c2 d 2 d 2(c/d)c = Var(t y) c2 d. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 30 / 40

Now Var(a y) = Var(t y) c2 d c = 0 t y has lowest variance among all linear unbiased estimator of E(t y). opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 31 / 40

Corollary 4.1: Under the AM, the estimator t y is the BLUE of E(t y) Vt C(X). opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 32 / 40

Proof of Corollary 4.1: First note that h y a linear unbiased estimator of zero is equivalent to E(h y) = 0 β R p h Xβ = 0 β R p h X = 0 X h = 0 h N (X ). Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 33 / 40

Thus, by Result 4.3, t y BLUE for E(t y) Cov(h y, t y) = 0 h N (X ) σ 2 h Vt = 0 h N (X ) h Vt = 0 h N (X ) Vt N (X ) = C(X). opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 34 / 40

Result 4.4: Under the AM, the OLSE of c β is the BLUE of c β estimable c β a matrix Q VX = XQ. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 35 / 40

Proof of Result 4.4: ( =) Suppose c β is estimable. Let t = c (X X) X. Then VX = XQ VX[(X X) ] c = XQ[(X X) ] c Vt C(X), which by Cor. 4.1, t y is BLUE of E(t y) c ˆβOLS is BLUE of c β. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 36 / 40

(= ) By Corollary 4.1, c ˆβOLS is BLUE for any estimable c β. VX[(X X) ] c C(X) c C(X ) VX[(X X) ] X a C(X) a R n VP X a C(X) a R n q i VP X x i = Xq i i = 1,..., p, where x i denotes the i th column of X. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 37 / 40

VP X [x 1,..., x p ] = X[q 1,..., q p ] VP X X = XQ, where Q = [q 1,..., q p ] VX = XQ for Q = [q 1,..., q p ]. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 38 / 40

Show that Q VX = XQ in our previous example. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 39 / 40

VX = A 0 0 0 0 A 0 0 0 0 A 0 0 0 0 A I I I I = A A A A = I I I I A = XA = XQ, where Q = A. opyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 40 / 40