Lecture 4 Hypothesis Testing

Similar documents
Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

x i1 =1 for all i (the constant ).

Professor Chris Murray. Midterm Exam

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Economics 130. Lecture 4 Simple Linear Regression Continued

Chapter 5: Hypothesis Tests, Confidence Intervals & Gauss-Markov Result

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

Statistics for Economics & Business

Chapter 11: Simple Linear Regression and Correlation

a. (All your answers should be in the letter!

x = , so that calculated

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

Answers Problem Set 2 Chem 314A Williamsen Spring 2000

ECON 351* -- Note 23: Tests for Coefficient Differences: Examples Introduction. Sample data: A random sample of 534 paid employees.

Biostatistics 360 F&t Tests and Intervals in Regression 1

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

Basic Business Statistics, 10/e

STAT 3008 Applied Regression Analysis

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.

β0 + β1xi and want to estimate the unknown

Statistics for Business and Economics

Lecture 6 More on Complete Randomized Block Design (RBD)

Lecture 2: Prelude to the big shrink

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Statistics II Final Exam 26/6/18

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Problem of Estimation. Ordinary Least Squares (OLS) Ordinary Least Squares Method. Basic Econometrics in Transportation. Bivariate Regression Analysis

# c i. INFERENCE FOR CONTRASTS (Chapter 4) It's unbiased: Recall: A contrast is a linear combination of effects with coefficients summing to zero:

Chapter 14 Simple Linear Regression

CHAPTER 8. Exercise Solutions

Tests of Single Linear Coefficient Restrictions: t-tests and F-tests. 1. Basic Rules. 2. Testing Single Linear Coefficient Restrictions

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

ECONOMETRICS - FINAL EXAM, 3rd YEAR (GECO & GADE)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

28. SIMPLE LINEAR REGRESSION III

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

Comparison of Regression Lines

Midterm Examination. Regression and Forecasting Models

STATISTICS QUESTIONS. Step by Step Solutions.

F statistic = s2 1 s 2 ( F for Fisher )

Module Contact: Dr Susan Long, ECO Copyright of the University of East Anglia Version 1

STAT 511 FINAL EXAM NAME Spring 2001

18. SIMPLE LINEAR REGRESSION III

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

Statistics Chapter 4

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

Systems of Equations (SUR, GMM, and 3SLS)

NUMERICAL DIFFERENTIATION

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Question 1 carries a weight of 25%; question 2 carries 20%; question 3 carries 25%; and question 4 carries 30%.

Learning Objectives for Chapter 11

Joint Statistical Meetings - Biopharmaceutical Section

STAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression

Modeling and Simulation NETW 707

F8: Heteroscedasticity

Reduced slides. Introduction to Analysis of Variance (ANOVA) Part 1. Single factor

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

e i is a random error

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

January Examinations 2015

17 - LINEAR REGRESSION II

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Chapter 11: I = 2 samples independent samples paired samples Chapter 12: I 3 samples of equal size J one-way layout two-way layout

Introduction to Analysis of Variance (ANOVA) Part 1

Topic 23 - Randomized Complete Block Designs (RCBD)

Properties of Least Squares

Negative Binomial Regression

Basically, if you have a dummy dependent variable you will be estimating a probability.

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

1-FACTOR ANOVA (MOTIVATION) [DEVORE 10.1]

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Econometrics of Panel Data

/ n ) are compared. The logic is: if the two

First Year Examination Department of Statistics, University of Florida

Correlation and Regression

Systematic Error Illustration of Bias. Sources of Systematic Errors. Effects of Systematic Errors 9/23/2009. Instrument Errors Method Errors Personal

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

Introduction to Dummy Variable Regressors. 1. An Example of Dummy Variable Regressors

Convergence of random processes

A Comparative Study for Estimation Parameters in Panel Data Model

Outline. 9. Heteroskedasticity Cross Sectional Analysis. Homoskedastic Case

Topic- 11 The Analysis of Variance

Introduction to Regression

Statistical tables are provided Two Hours UNIVERSITY OF MANCHESTER. Date: Wednesday 4 th June 2008 Time: 1400 to 1600

Lecture 6: Introduction to Linear Regression

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

7.1. Single classification analysis of variance (ANOVA) Why not use multiple 2-sample 2. When to use ANOVA

Goodness of fit and Wilks theorem

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

Using Multivariate Rank Sum Tests to Evaluate Effectiveness of Computer Applications in Teaching Business Statistics

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting.

Transcription:

Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to test whether an elastcty s equal to one. We may wsh to test the hypothess that has no mpact on the dependent varable Y. We may wsh to construct a confdence nterval for our coeffcents.

A hypothess takes the form of a statement of the true value for a coeffcent or for an expresson nvolvng the coeffcent. The hypothess to be tested s called the null hypothess. The hypothess whch t s tested agan s called the alternatve hypothess. Rejectng the null hypothess does not mply acceptng the alternatve We wll now consder testng the smple hypothess that the slope coeffcent s equal to some fxed value.

Settng up the hypothess Consder the smple regresson model: We wsh to test the hypothess that bd where d s some known value (for example zero aganst the hypothess that b s not equal to zero. We wrte ths as follows We wrte Y a + b + H H 0 a : b : b d d u

To test the hypothess we need to know the way that our estmator s dstrbuted. We start wth the smple case where we assume that the error term n the regresson model s a normal random varable wth mean zero and varance σ. Ths s wrtten as u ~ (0, σ ow recall that the OLS estmator can be wrtten as b ˆ b + Thus the OLS estmator s equal to a constant (b plus a weghted sum of normal random varables Weghted sums of normal random varables are also normal w u

The dstrbuton of the OLS slope coeffcent It follows from the above that the OLS coeffcent s a ormal random varable. What s the mean and what s the varance of ths random varable? Snce OLS s unbased the mean s b We have derved the varance and shown t to be Var ( bˆ σ Var ( Snce the OLS estmator s ormally dstrbuted ths means that z bˆ b Var ( bˆ ~ (0,

The dffculty wth usng ths result s that we do not know the varance of the OLS estmator because we do not know σ Ths needs to be estmated An unbased estmator of the varance of the resduals s the resdual sum of squares dvded by the number of observatons mnus the number of estmated parameters. Ths quantty (- n our case s called the degrees of freedom. Thus σˆ uˆ

Return now to hypothess testng. Under the null hypothess bd. Hence t must be the case that z bˆ d Var ( bˆ (0, We now replace the varance by ts estmated value to obtan a test statstc: z * bˆ d σˆ ( Ths test statstc s no longer ormally dstrbuted, but follows the t-dstrbuton wth - degrees of freedom. ~

Testng the Hypothess Thus we have that under the null hypothess z * ~ t bˆ ( σˆ d The next step s to choose the sze of the test (sgnfcance level. Ths s the probablty that we reject a correct hypothess. The conventonal sze s 5%. We say that the sze t, We now fnd the crtcal values α / and α t α /, 0.05

We accept the null hypothess f the test statstc s between the crtcal values correspondng to our chosen sze. Otherwse we reject. The logc of hypothess testng s that f the null hypothess s true then the estmate wll le wthn the crtcal values 00 ( a% of the tme. The ablty of a test to reject a hypothess s called the power of the test.

Confdence Interval We have argued that z * ~ t bˆ d ( σˆ Ths mples that we can construct an nterval such that the chance that the true b les wthn that nterval s some fxed value chosen by us. Call ths value α For a 95% confdence nterval say ths would be 0.95.

From statstcal tables we can fnd crtcal values such that any random varable whch follows a t-dstrbuton falls between these two values wth a probablty of α. Denote these crtcal values by and For a t random varable wth 0 degrees of freedom and a 95% confdence these values are (.8,-.8. Thus t α /, t α /, * pr( t < z < α α / t α / Wth some manpulaton we then get that pr bˆ se( bˆ t < b < bˆ + se( bˆ α ( α / tα / The term n the brackets s the confdence nterval.

Example Consder the regresson of log quantty of butter on the log prce agan regr lbp lpbr umber of obs 5 ------------------------------------------------------------------------------ lbp Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- log prce -.84586.95669-7.04 0.000 -.08437 -.608798 _cons 4.506.600375 8.6 0.000 4.00453 4.843668 ------------------------------------------------------------------------------ The statstc for the hypothess that the elastcty s equal to one s z 0.84 ( 0. 0.6 0..33

Crtcal values for the t dstrbuton wth 5-49 degrees of freedom (5 observatons, coeffcents estmated and sgnfcance level of 0.05 s approxmately (,-(from stat tables Snce -.33 les wthn ths range we accept the null hypothess The 95% confdence nterval s 0.84± 0. (.08, 0.6 Thus the true elastcty les wthn ths range wth 95% probablty. Everythng we have done s of course applcable to the constant as well. The varance formula s dfferent however.

Do we need the assumpton of normalty of the error term to carry out nference (hypothess testng? Under normalty our test s exact. Ths means that the test statstc has exactly a t dstrbuton. We can carry out tests based on asymptotc approxmatons when we have large enough samples. To do ths we wll use Central lmt theorem results that state that n large samples weghted averages are dstrbuted as normal varables.

A Central lmt theorem Suppose we have a set of ndependent random numbers all wth constant varance s and mean µ. Then v,,..., α ( v µ ~ (0, s α Where the symbol ~ reads dstrbuted asymptotcally,.e. as the sample sze tends to nfnty.

Ths extends to weghted sums. Let 0. So we also have that where s the probablty lmt of the sum of squares of the weghts. It s a lmt for sums of random varables. Ths lmt can be estmated n practce by the sum tself: We requre the lmt to be fnte: lm 0, ~ ( w p s wv α lm w p w < lm w p

Applyng the CLT to the slope coeffcent for OLS Recall that the OLS estmator can be wrtten as Ths s a weghted sum of random varables as n the prevous case. u w u u b b ( ( ( ˆ

The Central lmt theorem appled to the OLS estmator We can apply the central lmt theorem to the OLS estmator. Thus accordng to the central lmt theorem we have that Comparng wth the prevous slde the weghts are p u b b ( lm 0, ~ ( ( ˆ ( σ α w ( (

The mplcaton s that the statstc we had before has a normal dstrbuton n large samples rrespectve of how the error term s dstrbuted f t has a constant varance Assumpton - homoskedastcty. z * bˆ ( σˆ (0, ote how the s cancel from the top and bottom. In fact the test statstc s dentcal to the one we used under normalty. The only dfference s that now we wll use the crtcal values of the ormal dstrbuton. For a sze of 5% these are +.96 and.96. d α ~

The expresson on the denomnator s nothng but the standard error of the estmator. The test statstc for the specal case when we are testng that the coeffcent s n fact zero (no mpact on Y s often called the t- statstc. For a large sample test we can accept the hypothess that a coeffcent s zero wth a 5% level of sgnfcance f the t-statstc s between (-.96,.96

Example ------------------------------------------------------------------------------ lmap Coef. Std. Err. t P> t [95% Conf. Interval] -------------+---------------------------------------------------------------- lpsmr -.6856449.84636 -.4 0.00 -.56693 -.45967 _cons 4.83766.534038 7.83 0.000 3.0577 5.56956 ------------------------------------------------------------------------------ Regresson of log margarne purchases on the log prce. Test that the prce effect s zero. Assume large enough sample and use the crtcal values from the ormal dstrbuton. T-statstc -0.69/0.8-.4 95% ormal crtcal values are.96,.96 The hypothess s rejected The 95% confdence nterval s (.6,-0.5 Qute wde whch mples that the coeffcent s not very precsely estmated.

Summary When the error term s normally dstrbuted we can carry out exact tests by comparng the test statstc to crtcal values from the t-dstrbuton If the assumpton of normalty s not beleved to hold we can stll carry out nference when our sample s large enough. In ths case we smply use the normal dstrbuton.