WEIGHTED LEAST SQUARES. Model Assumptions for Weighted Least Squares: Recall: We can fit least squares estimates just assuming a linear mean function.

Similar documents
Weighted Least Squares

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Lecture 3: Multiple Regression

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Linear Regression with one Regressor

2.3 Solving Equations Containing Fractions and Decimals

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

STAT Chapter 11: Regression

MS&E 226: Small Data

MS&E 226: Small Data

Advanced Regression Topics: Violation of Assumptions

Multiple Regression Analysis

Chapter 2: simple regression model

Lecture 2: Linear and Mixed Models

Chapter 3. Introduction to Linear Correlation and Regression Part 3

Chapter 27 Summary Inferences for Regression

1. Variance stabilizing transformations; Box-Cox Transformations - Section. 2. Transformations to linearize the model - Section 5.

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1

Part 8: GLMs and Hierarchical LMs and GLMs

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

What to do if Assumptions are Violated?

Ch. 5 Transformations and Weighting

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Lecture 5: Clustering, Linear Regression

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Regression Models - Introduction

10. Alternative case influence statistics

Lecture 5: Clustering, Linear Regression

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Model Estimation Example

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters

Multiple Regression Analysis: Heteroskedasticity

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Simple linear regression

ECON The Simple Regression Model

Simple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com

CAMPBELL COLLABORATION

Chapter 8 Heteroskedasticity

Regression Analysis Tutorial 34 LECTURE / DISCUSSION. Statistical Properties of OLS

Föreläsning /31

BIOSTATISTICS NURS 3324

Weighted Least Squares

The Simple Linear Regression Model

Lecture 2. The Simple Linear Regression Model: Matrix Approach

10 Model Checking and Regression Diagnostics

appstats27.notebook April 06, 2017

Two-Variable Regression Model: The Problem of Estimation

BANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1

Lecture 14 Simple Linear Regression

Correlation Analysis

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

Lecture 5: Clustering, Linear Regression

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

Topic 14: Inference in Multiple Regression

STA Module 10 Comparing Two Proportions

Ordinary Least Squares Regression

ECON 3150/4150 spring term 2014: Exercise set for the first seminar and DIY exercises for the first few weeks of the course

Lectures 5 & 6: Hypothesis Testing

Lecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1

Transformations The bias-variance tradeoff Model selection criteria Remarks. Model selection I. Patrick Breheny. February 17

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Chapter 8 Handout: Interval Estimates and Hypothesis Testing

Time Series Analysis. Smoothing Time Series. 2) assessment of/accounting for seasonality. 3) assessment of/exploiting "serial correlation"

36-463/663: Multilevel & Hierarchical Models

Simple Linear Regression for the Climate Data

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 12th Class 6/23/10

Section 20: Arrow Diagrams on the Integers

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

General Linear Model (Chapter 4)

Econometrics Multiple Regression Analysis: Heteroskedasticity

22 Approximations - the method of least squares (1)

Conditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom

INFERENCE FOR REGRESSION

Lecture 10 Multiple Linear Regression

y response variable x 1, x 2,, x k -- a set of explanatory variables

The Multivariate Gaussian Distribution [DRAFT]

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

, (1) e i = ˆσ 1 h ii. c 2016, Jeffrey S. Simonoff 1

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

Regression Models - Introduction

Graphical Diagnosis. Paul E. Johnson 1 2. (Mostly QQ and Leverage Plots) 1 / Department of Political Science

Multiple Regression Analysis

STA121: Applied Regression Analysis

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Linear Regression with 1 Regressor. Introduction to Econometrics Spring 2012 Ken Simons

Introduction Recursive residuals Structural change CUSUM and CUSUMSQ. Recursive residuals by Tabar and Andrea

Unit 10: Simple Linear Regression and Correlation

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model

22s:152 Applied Linear Regression. 1-way ANOVA visual:

ECON 497: Lecture Notes 10 Page 1 of 1

STAT 705 Chapter 19: Two-way ANOVA

Distributions of linear combinations

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

LECTURE 5 HYPOTHESIS TESTING

LECTURE 15: SIMPLE LINEAR REGRESSION I

Lecture 4: Heteroskedasticity

Transcription:

1 2 WEIGHTED LEAST SQUARES Recall: We can fit least squares estimates just assuming a linear mean function. Without the constant variance assumption, we can still conclude that the coefficient estimators are unbiased, but we can t say anything about their variances; consequently, the inference procedures are not applicable. Moreover: If we fit least squares with non-constant variance, the values with larger variance typically have more influence on the result; values with lower variance typically are fit poorly. Example: Geese Model Assumptions for Weighted Least Squares: 1. E(Y x i ) =! T u i (linear mean function same as for ordinary least squares) 2. Var(Y x i ) = " 2 /w i, where the w i s are known, positive constants (called weights) (Different from OLS!) Observe: w i is inversely proportional to Var(Y x i ). This is sometimes helpful in getting suitable w i s. the w i s aren t unique we could multiply all of them by a constant c, and divide " by c to get an equivalent model. Recall: Sometimes we can find transformations to achieve constant variance. But sometimes we can t do this without messing up the linear mean or normality assumptions. An alternative that works sometimes is weighted least squares.

3 4 For WLS, the error is defined as e i = w i [ Y x i -! T u i ] (Different from OLS!) Exercise: E(e i ) = 0 Var(e i ) = " 2 Reformulating (1) in terms of errors: 1 : Y x i =! T u i + e i / w i Note: WLS is not a universal remedy for nonconstant variance, since weights are needed. But it is useful in many types of situations. Examples: A. If Y x i is the sum of m i independent observations v 1, v 2,, v mi, each with variance " 2, then Var(Y x i ) = =, so we could take w i =. B. If Y x i is the average of m i independent observations v 1, v 2,, v mi, each with variance " 2, then Var(Y x i ) = =, so we could take w i =.

C. Sometimes visual or other evidence suggests a pattern of how Var(Y x i ) depends on x i. e.g., if it looks like Var(Y x i ) is a linear function of x i [Sketch a picture of this!], then we can fit a line to the data points (x i, s i ), where s i = sample standard deviation of observations with x value x i. If we get s ˆ i = ˆ " 0 + ˆ " 1 x i, try w i = Caution: This involves looking at the data to decide on the supposedly known weights, which is iffy. A slightly better approach is to use w i s as above, then iterate by using the standard errors calculate from the first WLS regression. D. Sometimes theoretical considerations may suggest a choice of weights. For example, if theoretical considerations suggest that the conditional distributions are Poisson, then the conditional variances are equal to the conditional means. This suggests taking w i =. 5 E. Weighted least squares is also useful for other purposes. e.g., in calculating the lowess estimate, lines are fit so that points at the ends of the range count less than points at the middle of the range. Fitting WLS: A WLS model may be fit by least squares: Find ˆ " to minimize the weighted residual sum of squares RSS(h) = #w i (y i h T u i ) 2 ˆ " is called the WLS estimate of the coefficients. Comments: a. If all w i = 1, we get. b. The larger w i is, the more the i th observation counts (and the er the variance at x i think of the geese example.) 6

7 8 c. RSS(h) = # [ w i y i h T ( w i u i )] 2, so we could get ˆ " (for WLS) by using OLS to regress the w i y i s on the w i u i s, but, we would need to fit without an intercept, since the first component of w i u i is not 1. In practice, most statistics software has a specific routine to fit a WLS; weights need to be stored. Example: Coin data. Residuals in WLS: Recall errors in WLS: e i = w i [ Y x i -! T u i ]. Analogously, the residuals are defined as Caution: ˆ e i = w i ( y i - ˆ y i ) RSS and! 2 estimates: RSS = $w i ( y i - y ˆ i ) 2 = $ e ˆ 2 i ˆ " 2 = RSS/(n-k) (estimate of! 2, which is not the variance) Example: With the coins data, does ˆ " 2 seem reasonable? Inference for WLS: Proceeds similarly to inference for ordinary least squares. Model assumptions for inference are (1) and (2) above, plus 3) Independence of observations, and 4) Normal conditional distributions. Some software provides only the unweighted residuals y i - y ˆ i ; you need to multiply by the factors w i in order to make residual plots (to be discussed shortly)

9 Cautions in WLS inference: Estimating variances to get weights (as in coins example) introduces more uncertainty. The interpretation of R 2 is questionable some software doesn t even give it. Inference for means and prediction requires a weight (see pp. 209 210 for details) Is prediction appropriate for the coins example? Diagnostics with WLS: As for OLS, except use the WLS (weighted) residuals ˆ e i = w i ( y i - ˆ y i ).