STA 431s17 Assignment Eight 1

Similar documents
Instrumental Variables Again 1

Structural Equa+on Models: The General Case. STA431: Spring 2013

STA 302f16 Assignment Five 1

STA 2101/442 Assignment 3 1

STA442/2101: Assignment 5

Please note, you CANNOT petition to re-write an examination once the exam has begun. Qn. # Value Score

Structural Equa+on Models: The General Case. STA431: Spring 2015

STA 2101/442 Assignment 2 1

STA 2101/442 Assignment Four 1

STA441: Spring Multiple Regression. This slide show is a free open source document. See the last slide for copyright information.

The Bootstrap 1. STA442/2101 Fall Sampling distributions Bootstrap Distribution-free regression example

The Multinomial Model

STA441: Spring Multiple Regression. More than one explanatory variable at the same time

UNIVERSITY OF TORONTO Faculty of Arts and Science

Contingency Tables Part One 1

Introduction 1. STA442/2101 Fall See last slide for copyright information. 1 / 33

Generalized Linear Models 1

STA 2201/442 Assignment 2

An Introduction to Path Analysis

Factorial ANOVA. STA305 Spring More than one categorical explanatory variable

Interactions and Factorial ANOVA

An Introduction to Mplus and Path Analysis

Interactions and Factorial ANOVA

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

Multivariate Regression (Chapter 10)

The Multivariate Normal Distribution 1

Categorical Data Analysis 1

1 Motivation for Instrumental Variable (IV) Regression

UNIVERSITY OF TORONTO MISSISSAUGA April 2009 Examinations STA431H5S Professor Jerry Brunner Duration: 3 hours

Power and Sample Size for Log-linear Models 1

Deciphering Math Notation. Billy Skorupski Associate Professor, School of Education

Factorial ANOVA. More than one categorical explanatory variable. See last slide for copyright information 1

Instrumental Variables

The Multivariate Normal Distribution 1

REVIEW 8/2/2017 陈芳华东师大英语系

UNIVERSITY OF TORONTO. Faculty of Arts and Science APRIL 2010 EXAMINATIONS STA 303 H1S / STA 1002 HS. Duration - 3 hours. Aids Allowed: Calculator

Wooldridge, Introductory Econometrics, 3d ed. Chapter 9: More on specification and data problems

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Lecture 4: Testing Stuff

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Introduction to Confirmatory Factor Analysis

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

ECO375 Tutorial 8 Instrumental Variables

1 Correlation and Inference from Regression

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

BTRY 4830/6830: Quantitative Genomics and Genetics Fall 2014

Next is material on matrix rank. Please see the handout

Section 3: Simple Linear Regression

Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS

Do not copy, post, or distribute

Interactions. Interactions. Lectures 1 & 2. Linear Relationships. y = a + bx. Slope. Intercept

Chapter 4: Factor Analysis

Final Exam. Economics 835: Econometrics. Fall 2010

Joint Distributions: Part Two 1

Resemblance among relatives

Introduction to Structural Equation Modeling

SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora

1 The Classic Bivariate Least Squares Model

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

EDF 7405 Advanced Quantitative Methods in Educational Research. Data are available on IQ of the child and seven potential predictors.

CHAPTER 6: SPECIFICATION VARIABLES

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

STAT 135 Lab 11 Tests for Categorical Data (Fisher s Exact test, χ 2 tests for Homogeneity and Independence) and Linear Regression

Statistical Distribution Assumptions of General Linear Models

Ron Heck, Fall Week 3: Notes Building a Two-Level Model

AGEC 621 Lecture 16 David Bessler

STANDARDS OF LEARNING CONTENT REVIEW NOTES. ALGEBRA I Part II 1 st Nine Weeks,

Chapter 6. Logistic Regression. 6.1 A linear model for the log odds

Statistics Introductory Correlation

Hypothesis Testing for Var-Cov Components

Chapter 1 Statistical Inference

Regression Part II. One- factor ANOVA Another dummy variable coding scheme Contrasts Mul?ple comparisons Interac?ons

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1

Introduction to Bayesian Statistics 1

Evolution of quantitative traits

Multiple Linear Regression

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

MS&E 226: Small Data

ECO220Y Simple Regression: Testing the Slope

AMS 7 Correlation and Regression Lecture 8

Regression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur

Assignment 1. SEM 2: Structural Equation Modeling

Equation Number 1 Dependent Variable.. Y W's Childbearing expectations

Hypothesis Testing hypothesis testing approach

Final Exam. Question 1 (20 points) 2 (25 points) 3 (30 points) 4 (25 points) 5 (10 points) 6 (40 points) Total (150 points) Bonus question (10)

Biostatistics for physicists fall Correlation Linear regression Analysis of variance

14.74 Lecture 10: The returns to human capital: education

Statistics 5100 Spring 2018 Exam 1

Confidence intervals

Chapter 12 - Lecture 2 Inferences about regression coefficient

Path Diagrams. James H. Steiger. Department of Psychology and Human Development Vanderbilt University

Regression and correlation. Correlation & Regression, I. Regression & correlation. Regression vs. correlation. Involve bivariate, paired data, X & Y

Inference using structural equations with latent variables

Exam Applied Statistical Regression. Good Luck!

For more information about how to cite these materials visit

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

ECON2228 Notes 8. Christopher F Baum. Boston College Economics. cfb (BC Econ) ECON2228 Notes / 35

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Transcription:

STA 43s7 Assignment Eight The first three questions of this assignment are about how instrumental variables can help with measurement error and omitted variables at the same time; see Lecture slide set 3. When instrumental variables are not available, sometimes extra response variables allow us to purchase identifiability. Questions 4 through 7 explore this possibility. The non-computer questions on this assignment are for practice, and will not be handed in. For the SAS part of this assignment (Question 7) please bring hard copy of your log file and your results file to the quiz. There may be one or more questions about them, and you may be asked to hand the printouts in with the quiz.. In the model pictured below, the explanatory variable X is measured with error as well as being correlated with omitted variables. Z is an instrumental variable. e W Y ε β Z X κ Φ 2 (a) Give the model equations in centered form. The centering can be invisible. (b) Guided by the symbols on the path diagram, provide notation for the variances and covariances of the error terms and exogenous variables. (c) Let θ denote the vector of parameters you have witten down so far. These are the parameters that will appear in the covariance matrix of the observable data. What is θ? (d) Does this model pass the test of the Parameter Count Rule? Answer Yes or No and give the numbers. (Notice that we are only trying to identify the parameters in θ, which is a function of the full parameter vector. The full parameter vector has intercepts and unknown probability distributions.) (e) Calculate the covariance matrix Σ of D i, a single observable data vector. (f) Is the parameter β identifiable provided φ 2 0? Answer Yes or No. If the answer is Yes, prove it. If the answer is No, give a simple numerical example of two parameter vectors with different β values, yielding the same covariance matrix Σ. This assignment was prepared by Jerry Brunner, Department of Statistical Sciences, University of Toronto. It is licensed under a Creative Commons Attribution - ShareAlike 3.0 Unported License. Use any part of it as you like and share the result freely. The L A TEX source code is available from the course website: http://www.utstat.toronto.edu/ brunner/oldclass/43s7

(g) Why is is reasonable to assume φ 2 0? (h) Now let s make the model more realistic and scary. The response variable is measured with error, so V = Y + e 2. Furthermore, because of omitted variables, all the error terms might be correlated with one another and with X. i. Do your best to make a path diagram of the new model. You need not write symbols on the curved double-headed arrows you have added. ii. Show that β is still identifiable. 2. Here is a model with two explanatory variables and two instrumental variables. The path diagram looks busy, but it has features that make sense once you think about them. The instrumental and explanatory variables have covariance matrix Φ = [φ ij ], so that for example V ar(x ) = φ 33. No doubt there are omitted explanatory variables that are correlated with X and X 2, and affect Y. That is the source of c = Cov(X, ɛ) and c 2 = Cov(X 2, ɛ). The variables in the latent regression model are all measured (once) with error. Because of omitted variables in the measurement process, the measurement errors are correlated, with 3 3 covariance matrix Ω = [ω ij ]. Ω e 2 e e 3 W V W 2 Y β β 2 Z Z 2 X ε X 2 c c 2 Φ 2 ω 2

(a) Give the model equations in centered form. The centering can be invisible. (b) How many parameters appear in the covariance matrix of the observable data? Scanning from the bottom, I get 0+2+2+6=20. (c) Does this model pass the test of the Parameter Count Rule? Answer Yes or No and give the numbers. (d) The next step would be to calculate the covariance matrix Σ of the observable data vector, but that s a big job. To save work and also to reveal the essential features of the problem, please just calculate cov ( (Z, Z 2 ), (W, W 2, V ) ). (e) Are the parameters β and β 2 identifiable? Answer Yes or No. If the answer is Yes, prove it. You don t have to finish solving for β and β 2. You can stop once you have two linear equations in two unknowns, where the coefficients are σ ij quantities. Presumably it s possible to solve two linear equations in two unknowns. To prove identifiability, you don t have to actually recover the parameters from the covariance matrix. All you have to do is show it can be done. In Σ, please maintain the order Z, Z 2, W, W 2, V so we will have the same answer. 3. Here is a matrix version of instrumental variables. Independently for i =,..., n, the centered model equations are Y i = βx i + ɛ i W i = X i + e i, V i = Y i + e i,2. The random vectors X i and Y i are latent, while W i and V i are observable. In addition, there is a vector of observable instrumental variables Z i. The random vectors X i and Z i are p, while Y i is q. This determines the sizes of all the matrices. The variances and covariances are as follows: cov(x( i ) = ) Φ x, cov(z i ) = Φ z, cov(z i, X i ) = Φ zx, ei, cov(ɛ i ) = Ψ, cov(x i, ɛ i ) = C, and cov = Ω. All variance-covariance matrices e i,2 are positive definite (why not), and in addition, the p p matrix of covariances Φ zx has an inverse. Covariances that are not specified are zero; in particular, the instrumental variables have zero covariance with the error terms. Collecting Z i, W i, V i into a single long data vector D i, we write its variance-covariance matrix as a partitioned matrix: Σ Σ 2 Σ 3 Σ = Σ 22 Σ 23 Σ 33, where cov(z i, W i ) = Σ 2, and so on. (a) Give the dimensions (number of rows and number of columns) of the following matrices: β, Ψ, Ω, Σ 23. 3

(b) This problem fails the test of the Parameter Count Rule, though you are not required to show it. Fortunately, all we care about is β. Doing as little work as possible, prove that β is identifiable by showing how it can be recovered from the Σ ij matrices. (c) Give the formula for an estimator of β and show that it is consistent. 4. When instrumental variables are not available, sometimes identifiability can be obtained by adding more response variables to the model. Independently for i =,..., n, let W i = X i + e i Y i, = β X i + ɛ i, Y i,2 = β 2 X i + ɛ i,2 where X i, e i, ɛ i, and ɛ i,2 are all independent, V ar(x i ) = φ, V ar(e i ) = ω, V ar(ɛ i, ) = ψ, V ar(ɛ i,2 ) = ψ 2, and all the expected values are zero. The explanatory variable X i is latent, while W i, Y i, and Y i,2 are observable (a) Make a path diagram for this model (b) What are the unknown parameters in this model? (c) Let θ denote the vector of Greek-letter unknowns that apear in the covariance matrix of the observable data. What is θ? (d) Does this model pass the test of the Parameter Count Rule? Answer Yes or No and give the numbers. (e) Calculate the variance-covariance matrix of the observable variables. Show your work. (f) The parameter of primary interest is β. Is β identifiable at points in the parameter space where β = 0? Why or why not? (g) Is ω identifiable where β = 0? (h) Give a simple numerical example to show that β is not identifiable at points in the parameter space where β 0 and β 2 = 0. (i) Is β identifiable at points in the parameter space where β 2 0? Answer Yes or No and prove your answer. (j) Show that the entire parameter vector is identifiable at points in the parameter space where β 0 and β 2 0. (k) Propose a Method of Moments estimator of β. (l) How do you know that your estimator cannot be consistent in a technical sense? (m) For what points in the parameter space will your estimator converge in probability to β? 4

(n) How do you know that your Method of Moments estimator is also the Maximum Likelihood estimator (assuming normality)? (o) Explain why the likelihood rato test of H 0 : β = 0 will fail. Hint: What will happen when you try to locate θ 0? (p) Since the parameter of primary interest is β, it s important to be able to test H 0 : β = 0. So at points in the parameter space where β 2 0, what two equality constraints on the elements of Σ are implied by H 0 : β = 0? Why is this unexpected? (q) Assuming β 0 and β 2 0, you can use the model to deduce more than one testable inequality constraints on the variances and covariances. Give at least one example. 5. In the model of Question 4, suppose that X and the extra response variable Y 2 are influenced by common omitted variables, so that there is non-zero covariance between X and ɛ 2. Here is a path diagram. e ε W Y Y 2 ε 2 β β 2 X κ (a) How do you know that the full set of parameters (that is, the ones that appear in Σ) cannot possibly be identifiable in most of the parameter space? (b) Calculate the variance-covariance matrix of the observable variables. How does your answer compare to the one in Question 4? (c) Primary interest is still in β. Propose a Method of Moments estimator of β. Is it the same as the one in Question 4, or different? (d) For what set of points in the current parameter space is β identifiable? (e) If you had data in hand, what null hypothesis could you test about the σ ij quantities to verify the identifiability of β? (f) Suppose you want to test H 0 : β = 0, which is likely the main objective. 5

i. If you rejected the null hypothesis in Question 5e, what null hypothesis would you test about the σ ij quantities to test H 0 : β = 0? ii. If you failed to reject the null hypothesis in Question 5e, could you still test H 0 : β = 0? What is the test on σ ij quantities in this case? iii. If you rejected H 0 : β = 0, naturally you would want to state whether β is positive or negative. Is this possible? 6. Question 7 will use the Longitudinal IQ Data. IQ is short for Intelligence Quotient, and IQ tests are attempts to measure intelligence. A score of 00 is considered average. Most IQ tests have many sub-parts, including vocabulary tests, math tests, logical puzzles, tests of spatial reasoning, and so on. What the better tests probably succeed in doing is to measure one kind of intelligence potential for doing well in school. Of course, they measure it with error. In the Longitudinal IQ Data, the IQs of adopted children were measured at ages 2, 4, 8 and 3. The birth mother s IQ was assessed at the time of adoption, and the adoptive mother s education (in years) was also recorded. The variables are Adoptive mother s education Birth mother s IQ IQ at age 2 IQ at age 4 IQ at age 8 IQ at age 3 Here is a path diagram for just the IQ part of the data. e ε ε 2 ε 3 ε 4 W Y Y 2 Y 3 Y 4 β β 2 β 3 β 4 X 6 e Y 2 ε 2

X is birth mother s true academic ability, W is birth mother s measured IQ, Y is child s measured IQ at age 2, and so on. Of course child s ability is a latent variable too, but we ve re-parameterized, making the response variables appear to be observable. This model is far from perfect, but we ll go with it for now. (a) Write the model equations in scalar form. (b) Let θ denote the vector of Greek-letter unknowns that apear in the covariance matrix of the observable data. What is θ? (c) Does this model pass the test of the Parameter Count Rule? Answer Yes or No and give the numbers. (d) Calculate the variance-covariance matrix of the observable variables. Show your work. (e) The parameter vector is identifiable except on a set of volume zero in the parameter space. Show how β 2 can be recovered from σ ij quantities. (f) If β = 0, is it identifiable? (g) How many βs need to be non-zero for all the parameters to be identifiable? (h) How many degrees of freedom will there be in the likelihood ratio test for model fit? The answer is a number. (i) Suppose you want to test whether all four regression coefficients are equal, using a likelihood ratio test. What are the degrees of freedom for this test? 7. The longitudinal IQ data described in Question 6 are given in the file origiq.data.txt. These data are taken from The Statistical Sleuth by F. Ramsey and D. Schafer, and are used without permission. (a) Start by reading the data. There are n = 62 cases, so please verify that you are reading the correct number of cases. Now run proc corr to produce a correlation matrix of all the variables (with tests), including adoptive mother s education. (b) Do the test on the correlations support omitting adoptive mother s education from the model? (c) How are the proc corr results helpful in justifying your identifiability condition from the Question 6? (d) Please fit the model of Question 6. We ll call this the full model. (e) Sticking strictly to the α = 0.05 significance level, does the full model fit the data adequately? Answer Yes or No 2, and give a value of G 2, the degrees of freedom and the p-value. These numbers are all directly on your printout. Do the degrees of freedom agree with your answer to Question 6h? 2 My p-value of 0.0707 is too close to 0.05 for comfort. This model is hard to defend. It says the only reason that kids scores at different ages are correlated is the mother s academic potential as an adult. Just for starters, these children do technically have fathers. Also, the model fails to distinguish between genotype and phenotype, and it has other weaknesses. 7

(f) Now fit the reduced model in which all the regression coefficients are equal. Using a calculator (or proc IML if you want to), calculate the likelihood ratio test comparing the full and reduced models. Obtain G 2, a number. (g) What are the degrees of freedom for this test? Compare your answer to Question 6i. (h) Using this table of critical values, do you reject H 0 at α = 0.05? Answer Yes or No. Does birth mother s IQ seem to affect her child s IQ to the same degree at different ages? > df = :8 > CriticalValue = qchisq(0.95,df) > round(rbind(df,criticalvalue),3) [,] [,2] [,3] [,4] [,5] [,6] [,7] [,8] df.000 2.000 3.000 4.000 5.00 6.000 7.000 8.000 CriticalValue 3.84 5.99 7.85 9.488.07 2.592 4.067 5.507 (i) Just one more thing. Please go back to the proc calis run where you fit the full model, and use the simtests statement to do a Wald test of H 0 : β = β 2 = β 3 = β 4. Do you reach the same conclusion? SAS makes it easy to do Wald tests and hard to do likelihood ratio tests, but likelihood ratio tests perform better for relatively small samples like this, so we ll believe the likelihood ratio test. Bring your log file and your results file to the quiz. You may be asked for numbers from your printouts, and you may be asked to hand them in. There are lots of There must be no error messages, and no notes or warnings about invalid data on your log file. If you see a reference to the Moore-Penrose inverse, you ve made a coding error. 8