Problem set 1: answers. April 6, 2018

Similar documents
Multiple Regression Analysis

2. Linear regression with multiple regressors

The Simple Regression Model. Part II. The Simple Regression Model

Applied Statistics and Econometrics

Multiple Regression. Midterm results: AVG = 26.5 (88%) A = 27+ B = C =

Answers to Problem Set #4

5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1)

Exercise sheet 6 Models with endogenous explanatory variables

ECON3150/4150 Spring 2015

The Simple Linear Regression Model

6. Assessing studies based on multiple regression

CHAPTER 6: SPECIFICATION VARIABLES

3. Linear Regression With a Single Regressor

Exercises (in progress) Applied Econometrics Part 1

10. Time series regression and forecasting

ECON Introductory Econometrics. Lecture 4: Linear Regression with One Regressor

Applied Economics. Regression with a Binary Dependent Variable. Department of Economics Universidad Carlos III de Madrid

Time Series. Chapter Time Series Data

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Model Specification and Data Problems. Part VIII

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors

Econometrics Review questions for exam

Introduction and Single Predictor Regression. Correlation

Intermediate Econometrics

The Multiple Regression Model Estimation

4. Nonlinear regression functions

Applied Econometrics (QEM)

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

Review of probability and statistics 1 / 31

ECO220Y Simple Regression: Testing the Slope

Simple Linear Regression for the MPG Data

ECON The Simple Regression Model

11. Simultaneous-Equation Models

Simple Linear Regression

Stat 101 Exam 1 Important Formulas and Concepts 1

Linear Regression. Junhui Qian. October 27, 2014

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Monday, November 26: Explanatory Variable Explanatory Premise, Bias, and Large Sample Properties

Solution to Exercise E6.

ECON3150/4150 Spring 2016

1 Correlation and Inference from Regression

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Applied Econometrics (QEM)

WISE International Masters

Lecture 8. Using the CLR Model

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

2) For a normal distribution, the skewness and kurtosis measures are as follows: A) 1.96 and 4 B) 1 and 2 C) 0 and 3 D) 0 and 0

Review of Econometrics

Heteroskedasticity. Part VII. Heteroskedasticity

MATH 1150 Chapter 2 Notation and Terminology

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Economics 471: Econometrics Department of Economics, Finance and Legal Studies University of Alabama

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Chapter 6: Linear Regression With Multiple Regressors

Lecture 8. Using the CLR Model. Relation between patent applications and R&D spending. Variables

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

unadjusted model for baseline cholesterol 22:31 Monday, April 19,

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

ECON3150/4150 Spring 2016

STAT5044: Regression and Anova. Inyoung Kim

Lecture 1: OLS derivations and inference

ECON Introductory Econometrics. Lecture 5: OLS with One Regressor: Hypothesis Tests

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Variables, distributions, and samples (cont.) Phil 12: Logic and Decision Making Fall 2010 UC San Diego 10/18/2010

Preliminary Statistics course. Lecture 1: Descriptive Statistics

Statistical View of Least Squares

Topic 4: Model Specifications

Lectures 5 & 6: Hypothesis Testing

Ch Inference for Linear Regression

Statistics and Econometrics I

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Lab 07 Introduction to Econometrics

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

Graduate Econometrics I: Unbiased Estimation

ECON Introductory Econometrics. Lecture 16: Instrumental variables

Outline. 2. Logarithmic Functional Form and Units of Measurement. Functional Form. I. Functional Form: log II. Units of Measurement

Multivariate Regression: Part I

MS&E 226: Small Data

Lecture 4: Multivariate Regression, Part 2

Econometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012

EMERGING MARKETS - Lecture 2: Methodology refresher

ECON Introductory Econometrics. Lecture 17: Experiments

CORRELATION, ASSOCIATION, CAUSATION, AND GRANGER CAUSATION IN ACCOUNTING RESEARCH

ECON 3150/4150, Spring term Lecture 7

MA Advanced Econometrics: Applying Least Squares to Time Series

Applied Statistics and Econometrics

Research Center for Science Technology and Society of Fuzhou University, International Studies and Trade, Changle Fuzhou , China

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Heteroscedasticity 1

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18

Introduction to Econometrics. Review of Probability & Statistics

Practice Questions for the Final Exam. Theoretical Part

K. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =

Statistics, inference and ordinary least squares. Frank Venmans

Probit Estimation in gretl

ECNS 561 Multiple Regression Analysis

Statistics 910, #5 1. Regression Methods

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold.

Transcription:

Problem set 1: answers April 6, 2018 1

1 Introduction to answers This document provides the answers to problem set 1. If any further clarification is required I may produce some videos where I go through the answers. 2 Crime and Unemployment - practical 1. The result of the histogram of the violence variable is shown below. On graphing this data it is evident that the majority of the states have violent crime rates from 200-600 cases per 100,000 people. The outlier here is The District of Columbia on the far right. 2. The unemployment data histogram is shown below. It is evident that there is quite a considerable spread in the rates of unemployment, with the modal bin being 6.8-7.9%. Page 2

3. The District of Columbia. What makes this state quite so violent?! 4. The summary statistics for the unemployment dataset are shown below. Summary Statistics, using the observations 1 51 for the variable Unemployment (51 valid observations) Mean Median Minimum Maximum 6.7647 6.9000 2.9000 9.6000 Std. Dev. C.V. Skewness Ex. kurtosis 1.5393 0.22755 0.45205 0.39024 5% perc. 95% perc. IQ Range Missing obs. 3.8000 9.1200 2.4000 0 5. The scatterplot of rates of violence against unemployment rates is shown below for the 51 states. To add data labels to each of the points, all that is required is to right click on the plot, and select All data labels. From this chart it is evident that the District of Columbia is a clear outlier, and we might want to treat this point differently from an analysis (regression) perspective. Page 3

6. A correlation of 0.42 does not mean much per se. It is suggestive that there is some positive relationship between the two variables which is relatively strong. 7. The results of the ordinary least squares regression are shown below. Model 1: OLS, using observations 1 51 Dependent variable: Violence Coefficient Std. Error t-ratio p-value const 24.3979 113.900 0.2142 0.8313 Unemployment 53.3478 16.4256 3.248 0.0021 Mean dependent var 385.2804 S.D. dependent var 195.1137 Sum squared resid 1566285 S.E. of regression 178.7876 R 2 0.177141 Adjusted R 2 0.160348 F(1, 49) 10.54851 P-value(F) 0.002102 Log-likelihood 335.8419 Akaike criterion 675.6837 Schwarz criterion 679.5474 Hannan Quinn 677.1601 8. The coefficient on unemployment is approximately 53. This suggests that for a 1% increase in unemployment, on average there tends to be 53 more cases of violence per 100,000 people. Page 4

9. The standard deviation of unemployment can be seen from its summary statistics, and is 1.5. An increase in unemployment by 1.5 will result in an increase in the rate of violence by 1.5 53 82 cases on average per 100,000 people. The standard deviation in Violence is 195. So in terms of standard deviations, the 82.0 cases equates to approximately 0.42 s.d. 10. The results of this regression are shown below. An increase in violence by 195.32 is associated with an average increase in the unemployment rate by 0.66%. Model 2: OLS, using observations 1 51 Dependent variable: Unemployment Coefficient Std. Error t-ratio p-value const 5.48538 0.440645 12.45 0.0000 Violence 0.00332050 0.00102237 3.248 0.0021 Mean dependent var 6.764706 S.D. dependent var 1.539328 Sum squared resid 97.48938 S.E. of regression 1.410524 R 2 0.177141 Adjusted R 2 0.160348 F(1, 49) 10.54851 P-value(F) 0.002102 Log-likelihood 88.88777 Akaike criterion 181.7755 Schwarz criterion 185.6392 Hannan Quinn 183.2520 11. Making Violence the subject of the above regression model yields: Or more neatly. Violence = (1/0.0034) Unemployment (1/0.0034) 5.589 Violence = 294.11 Unemployment 1643.82 ote that this is completely different to the results which we obtained from our OLS regression of Violence on Unemployment. This is because of the fact that the regression of y on x is not the same as the regression of x on y. The former minimises square distances of y from the line, whereas the latter minimises square distances in x. Rearranging the latter regression equation will hence not yield the former. For example this model suggests that a 1.58% increase in the unemployment rate will increase violence rates by 467.6! Very different to the previous estimate. 12. Based on these data alone it is impossible to understand fully the causal mechanism. In my view it is unlikely that violence causes unemployment from a theoretical standpoint, but it is impossible to say whether unemployment causes violence, or whether they are simply both correlated/caused by a third unknown factor. Page 5

13. There are a number of factors which are correlated with violence, which could also be correlated with unemployment. Examples of this might be ethnic fractionalisation, or some measure of geography. These would need to be explictly controlled for in a regression before any conclusions are made, since omitted variable bias in the estimation of the effect of unemployment on violence is likely rife. 3 Theory This section aims to build up your theoretical knowledge of econometrics, and should cover the first 30 videos or so of material from the undergraduate econometrics course. This section will complement the practical part of the problem set, but is not a required part of the course. 1. For a pupil, i, selected at random from a school, the number of years of education of their parents, X i, is given by: X i = µ + ε i ε i iid(0, σ 2 ). Here µ is the mean number of years of education completed by parents. For a sample of students selected independently from the population: (a) The sample mean is given by X = 1 X i. Taking the expectations of both sides yields its expected value of µ. E( X) = E( 1 X i ) = 1 E(X i ) = 1 µ = 1 µ = µ (b) To calculate the variance, just apply it to both sides of the equation, and not that due to the independence of Xs, there is no need to worry about any covariance terms: Var( X) = Var( 1 X i ) = 1 Var(X 2 i ) = 1 2 σ 2 = 1 2 σ 2 = σ2 The reason the 2 term appears is since for any random variable Z, we have that the variance of a constant a times Z is: Var(aZ) = a 2 Var(Z) (c) To prove consistency it is sufficient (actually as it turns out more than sufficient) that the variance of the estimator tends to zero as tends to infinity, and that the expected value of the estimator tends to the true value at the same time. We already know that the expected value of the sample mean is the population mean, µ, for any sample size. The variance of the sample mean is σ 2 / meaning that as tends to infinity, the variance tends to zero. Hence, since these two conditions are satisfied, the sample mean is a consistent estimator of the population mean. (d) Another way to think about the sample mean is as a least squares estimator. The estimator aims to place ˆµ at a value to minimise the following sum: Page 6

S = (X i ˆµ) 2 Where the term in the parenthesis is the error of prediction. The first order conditions for this minimisation are: Which when rearranged yields: S ˆµ = 2 ˆµ (X i ˆµ) = 0 (1) ˆµ = X = 1 X i (e) Yes, it is BLUE. The first part of the proof is proving the conditions for a linear estimator to be unbiased. The second part then proves that if the estimator is unbiased, the variance of a linear estimator cannot better it. First of all we define a linear estimator X = 1 w i X i, and we assume that the weights are made up: w i = 1 + δ i. The 1 here is what the sample mean weights are, so we are saying that our new estimator weights differ from that of the sample mean by the amount δ i. Let s now derive the conditions for the estimator to be unbiased. E( X) = E( 1 (1 + δ i )X i ) = µ + µ δ i Hence we must have that δ i = 0 for our new linear estimator to be unbiased. ow we are going to derive the variance of our new estimator. Var( X) = 1 (1 + δ 2 i ) 2 σ 2 = σ2 (2δ i + δ 2) = Var( X) + σ2 i + σ2 2 δ 2 2 i We got the last expression by using the unbiasedness condition which we derived above. Finally, we note that for non-zero weights the expression δ 2 = η > 0, hence we have that the variance of the new estimator is greater i than that of the sample mean: Var( X) = Var( X) + η Hence this has proved that any other linear estimator apart from the sample mean has a greater sampling variance. Hence we have proved that the sample mean is BLUE. 2. For each of the following state whether or not the estimator is biased, consistent, both or neither, when used to estimate the population mean: (a) X = 1 X 1 i Consistent. The estimator is biased due to the dividing through by -1. But the expected value of this estimator converges towards µ as tends to infinity. (b) ˆX = 2 /2 X i This is both unbiased and consistent. It is not BLUE however since it only uses half the sample s data, the sample mean betters it. Page 7

(c) Assuming is even. X = 2 /2 (X i + µ) + 2 i=/2+1 (X i µ) This is unbiased and consistent - it is also BLUE. If you work through the maths it is actually the sample mean! (d) Y (µ, σ 2 ) This is an odd estimator, in that it does not depend on the sample properties. It is unbiased however, but is not consistent! A very rare type of estimator. Albeit from a rather silly example. (e) Z = w i X i Where: w i = 1 Unbiased and consistent. 3. Examine the following economic model Y i = α + βx i + ε i (a) See the videos starting with this one which walk through the solution to this question. http://tinyurl.com/pdhn9dk (b) The regression of X on Y yields the following estimator for the slope parameter: ˆδ = (X i X)(Y i Ȳ) (Y i Ȳ) 2 (c) o. See the question in the practical part above. The regression of Y on X minimises the sum of square distances of points on the line from each Y. The regression of X on Y, minimises the sum of square distances of points on the line from each X. These are not equivalent. You cannot get to the other slope parameter by inverting the original. The expression ˆδ ˆβ can be found however: ˆδ ˆβ = ( [ (X i X)(Y i Ȳ)] 2 (Y i Ȳ) 2 )( (X i X) 2 ) (d) This expression of the RHS actually looks similar to something we have discussed in the videos - it is the square of the correlation coefficient! Hence we have that the correlation coefficient is simply the geometric mean of the two slope regression coefficients: ˆρ = ( ˆδ ˆβ) 1 2 4. There are two populations of individuals called samies varies respectively. The height of individuals in the samies is given by: X i µ + ε i And the height of individuals in the varies is given by: Where ε i iid(0, σ 2 ). Y i µ + 2ε i (a) Yes. See the logic in the second question. All you need to do to prove this is take the expectation of both sides. Page 8

(b) Yes on both accounts. Again see the logic above. (c) The variance of the sample means are Var( X) = σ2 4σ2 and Var(Ȳ) =. Hence the samies sample mean is the most efficient of the two estimators. (d) You have a sample of individuals from each population. Yes. Since they are both, on their own unbiased, then mean of the two estimators will also be unbiased. However, it is not the best estimator possible. See the next part. (e) In order to construct a BLUE estimator we first of all need to ensure that it is unbiased, then minimise its variance over choice of parameters. Let us construct an estimator, Z which is a linear combination of the two sample means. Z = α X + βȳ In order for this estimator to be unbiased we must have that its expectation is equal to the population mean µ. This means that: E( Z) = αe( X) + βe(ȳ) = αµ + βµ = µ Which implies that α + β = 1. Hence we can rewrite our estimator as: Z = α X + (1 α)ȳ ow finding the variance of our estimator (noting that the covariance between the two estimators is zero): Var( Z) = α 2 Var( X) + (1 α) 2 Var(Ȳ) = σ2 [α2 + 4(1 α) 2 ] = σ2 [5α2 8α + 4] In order to find the best unbiased estimator, we need to minimise this expression over choice of α. Differentiating, and setting this equal to zero: (Var( Z))= α = σ2 [10α 8] = 0 If you then solve this, one finds that α = 4. Hence the linear combination of 5 the two sample means that has the lowest variance, and is still unbiased is given by: Z = 4 X + 1 5 5Ȳ The intuition behind this is that this estimator gives most weight to the sample mean which has the lowest variance, minimising predictive error. In principle one could give zero weight to the second sample mean, but this would unnecessarily dispense with observations, reducing efficiency. Page 9