ISQS 6348 Final Open notes, no books. Points out of 100 in parentheses. Y 1 ε 2

Similar documents
UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

Lecture 6 More on Complete Randomized Block Design (RBD)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

STAT 511 FINAL EXAM NAME Spring 2001

First Year Examination Department of Statistics, University of Florida

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Statistics MINITAB - Lab 2

Comparison of Regression Lines

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

a. (All your answers should be in the letter!

Negative Binomial Regression

Professor Chris Murray. Midterm Exam

/ n ) are compared. The logic is: if the two

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 12 Analysis of Covariance

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Scatter Plot x

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 11: Simple Linear Regression and Correlation

x i1 =1 for all i (the constant ).

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

e i is a random error

Answers Problem Set 2 Chem 314A Williamsen Spring 2000

Exam. Econometrics - Exam 1

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

Economics 130. Lecture 4 Simple Linear Regression Continued

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Biostatistics. Chapter 11 Simple Linear Correlation and Regression. Jing Li

Goodness of fit and Wilks theorem

β0 + β1xi. You are interested in estimating the unknown parameters β

Outline. Multivariate Parametric Methods. Multivariate Data. Basic Multivariate Statistics. Steven J Zeil

Chapter 8 Indicator Variables

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Linear Regression Analysis: Terminology and Notation

β0 + β1xi. You are interested in estimating the unknown parameters β

Module Contact: Dr Susan Long, ECO Copyright of the University of East Anglia Version 1

Modeling and Simulation NETW 707

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

CHAPTER 8. Exercise Solutions

STAT 3008 Applied Regression Analysis

STATISTICS QUESTIONS. Step by Step Solutions.

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Joint Statistical Meetings - Biopharmaceutical Section

Linear Approximation with Regularization and Moving Least Squares

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics )

Statistics II Final Exam 26/6/18

Polynomial Regression Models

x = , so that calculated

Q1: Calculate the mean, median, sample variance, and standard deviation of 25, 40, 05, 70, 05, 40, 70.

Two-factor model. Statistical Models. Least Squares estimation in LM two-factor model. Rats

Lecture 4 Hypothesis Testing

Statistics for Economics & Business

Tests of Single Linear Coefficient Restrictions: t-tests and F-tests. 1. Basic Rules. 2. Testing Single Linear Coefficient Restrictions

Online Appendix to: Axiomatization and measurement of Quasi-hyperbolic Discounting

Statistics Chapter 4

Lecture 3: Probability Distributions

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

ANOVA. The Observations y ij

# c i. INFERENCE FOR CONTRASTS (Chapter 4) It's unbiased: Recall: A contrast is a linear combination of effects with coefficients summing to zero:

Question 1 carries a weight of 25%; question 2 carries 20%; question 3 carries 25%; and question 4 carries 30%.

SIMPLE LINEAR REGRESSION

Composite Hypotheses testing

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Homework Assignment 3 Due in class, Thursday October 15

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Measuring the Strength of Association

NUMERICAL DIFFERENTIATION

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Effects of Ignoring Correlations When Computing Sample Chi-Square. John W. Fowler February 26, 2012

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

2.3 Nilpotent endomorphisms

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Errors for Linear Systems

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 13: Multiple Regression

Accepted Manuscript. Guillaume Marrelec, Habib Benali. S (08)00059-X DOI: /j.spl Reference: STAPRO 4923.

β0 + β1xi and want to estimate the unknown

January Examinations 2015

Midterm Examination. Regression and Forecasting Models

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

Chapter 14 Simple Linear Regression

Important Instructions to the Examiners:

Lecture Notes on Linear Regression

Generalized Linear Methods

The topics in this section concern with the second course objective. Correlation is a linear relation between two random variables.

LECTURE 9 CANONICAL CORRELATION ANALYSIS

Chap 10: Diagnostics, p384

Transcription:

ISQS 6348 Fnal Open notes, no books. Ponts out of 100 n parentheses. 1. The followng path dagram s gven: ε 1 Y 1 ε F Y 1.A. (10) Wrte down the usual model and assumptons that are mpled by ths dagram. Soluton: Y 1 = βf + ε 1 Y = ε where Var(F) = 1, F s uncorrelated wth the ε s, and Cov(ε 1,ε )=φ. 1.B. (10) Fnd the covarance matrx of that s mpled by the model and assumptons. Y F 1 0 0 β 0 β 1 0 β 1 0 Soluton: = ε1 Y 0 0 1, so Cov = 0 ψ1 φ 1 ε Y 0 0 1 0 0 φ ψ 0 1

β + ψ1 φ =. φ ψ 4. 1. 1.C. (10) Suppose the covarance matrx s Cov =. Usng 1.B., dentfy the Y 1..3 parameters of your model n 1.A. that are dentfable, and state whch parameters are not dentfable. β + ψ1 φ 4. 1. Soluton: Equatng to, we can dentfy φ = 1. and φ ψ 1..3 ψ =.3. However, there are nfntely many choces for β and ψ 1 that gve β + ψ 1 =4., so β and ψ 1 are not dentfable. d.... (0) Suppose Y N0(µ,Σ), =1,,n. How does ths statement relate to the data matrx, and to numercal and graphcal summares of the data n the data matrx? By data matrx, I mean my usual depcton of the data set, where I show OBS n the frst column. State the varous connectons. Soluton: The data matrx looks lke ths: OBS Y 1 Y Y 0 1 y 11 y 1 y 1,0 y 1 y y,0 3 y 31 y 3 y 3,0 n y n1 y n y n,0 d... The statement Y N0(µ,Σ) states that the twenty-dmensonal row vectors of the data matrx are randomly sampled from a multvarate normal dstrbuton wth some 0-dmensonal mean µ and 0x0 covarance matrx Σ. For example, (y 11 y 1 y 1,0 ) s the result of one random selecton from N 0 (µ,σ), (y 1 y y,0 ) s the result of another, ndependent random selecton from N 0 (µ,σ), etc. If you look at the 0 hstograms for each of the varables, they wll show bell shapes. If you look at the 0*19/ scatterplots of pars of the varables, such as (Y 1, Y ), (Y 1, Y 3 ) etc., they wll show ellptcal shapes. If you construct the χ q-q plot of the n Mahalanobs dstances aganst ther expected Ch-Square quantles (df=0), t wll show the appearance of a 45 degree lne.

If you calculate the 0-dmensonal sample mean vector Y you calculate the 0x0 sample covarance matrx to Σ, both by the large of large numbers. n = 1 n = Y / n t wll be close to µ, and f = 1 S = ( Y Y)( Y Y)'/( n 1), t wll be close 3. Consder the followng SAS code: data measurement; do = 1 to 10000; F = rannor(0); e1 = rannor(0); e = rannor(0); e3 = rannor(0); x1 =.5 * F + sqrt(1-.5**) * e1 ; x =.5 * F + sqrt(1-.5**) * e ; x3 =.5 * F + sqrt(1-.5**) * e3; AverageX = (X1 + X + X3)/3; output; end; run; 3.A. (10) Fnd the true relablty of AverageX usng theory. Soluton: The relablty s the squared correlaton between AverageX and what t s measurng, namely, F. Frst we fnd the covarance matrx of AverageX and F, Note that AverageX =.5F + (1/3) sqrt(1-.5**)e1 + (1/3) sqrt(1-.5**)e + (1/3) sqrt(1-.5**)e3, so F AverageX.5.75 / 3.75 / 3.75 / 3 ε1 = F 1 0 0 0 ε ε3 and hence 1 0 0 0.5 1 AverageX.5.75 / 3.75 / 3.75 / 3 0 1 0 0.75 / 3 0 Cov = F 1 0 0 0 0 0 1 0.75/3 0 0 0 0 1.75 / 3 0

.5 1.5.75 / 3.75 / 3.75 / 3.75 / 3 0.5.5 = =. 1 0 0 0.75/3 0.5 1.75 / 3 0 So.5 Corr(AverageX, F) = =.5, and hence the relablty s.5 1 { Corr(AverageX, F) } = (.5) =.5 3.B. (5) How could you use the data n the fle WORK.MEASUREMENT to fnd the usual estmated lower bound on the true relablty of AverageX? Soluton: proc corr data=work.measurement alpha; var x1 x x3; run; In other words, calculate Cronbach s Alpha usng the X data. Note that ths gves a lower bound on the relablty of the summate, but the relablty of the summate s dentcal to the relablty of the average, so t also gves a lower bound on the relablty of the average. 3.C. (5) How could you use the data n the fle WORK.MEASUREMENT to fnd an estmate of the true relablty of AverageX? Soluton: True relablty s the squared correlaton between the measure and the object beng measured. You can estmate ths by calculatng the sample correlaton coeffcent f you have both quanttes. In ths case you do, snce you smulated the data yourself. So fnd the correlaton of AverageX wth F; these varables already exst n the fle work.measurement: proc corr data=work.measurement; var AverageX F; run; The output would gve you the estmated correlaton between the AverageX and F; square t to get the estmated relablty. 4. The followng output was obtaned from PROC CALIS. I show ths just to refresh you r memory. Do not use the output drectly n your answers.

Answer the followng questons genercally, rather than wth specfc reference to the numerc values shown n the output. Gve formulas and explanatons. 4.A. (5) What s the null hypothess tested by the ch-square statstc? Soluton: The null hypothess s H 0 : Σ=Σ(θ), e that the true covarance matrx obeys the restrctons mpled by the model. 4.B. (5) How are the degrees of freedom for the ch-square test obtaned? Soluton: Typcally df = p(p+1)/ {number of parameters n θ}, e, the number of free covarance parameters n the unrestrcted model mnus the number of model parameters used to determne the covarance matrx n the restrcted model. 4.C. (10) How are RMSEA and the ch-square statstc smlar? How are they dfferent? Soluton: Smlar: Both are related to the Ch Square statstc. RMSEA s a normalzaton of the ch-square statstc. Dfferent: The RMSEA s used to remove the dependency of the Chsquare statstc on the sample sze. We know that E(χ ) = df + N, where N s the sample sze and denotes dscrepancy between Σ and Σ(θ) (whch does not depend on N). RMSEA provdes an estmate of /df. 4.D. (10) How are GFI and RMR smlar? How are they dfferent? Soluton: Smlar: They both measure dscrepancy between the observed (unrestrcted) covarance matrx ( ˆΣ ) and the ftted (restrcted) covarance matrx ( Σ ( ˆ θ )). Dfferent: Smaller values of RMR ndcate less dscrepancy, whle larger GFI (closer to 1.0) ndcate less dscrepancy. Also, RMR s unweghted by the sze of the covarances (hence s an absolute measures) whle GFI s weghted by the sze of the covarances (hence s a relatve measure).