Concordia University (5+5)Q 1.

Similar documents
PART I. (a) Describe all the assumptions for a normal error regression model with one predictor variable,

[4+3+3] Q 1. (a) Describe the normal regression model through origin. Show that the least square estimator of the regression parameter is given by

Formal Statement of Simple Linear Regression Model

Ch 2: Simple Linear Regression

STAT 360-Linear Models

(ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box.

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, October 2013

Linear regression. We have that the estimated mean in linear regression is. ˆµ Y X=x = ˆβ 0 + ˆβ 1 x. The standard error of ˆµ Y X=x is.

STAT420 Midterm Exam. University of Illinois Urbana-Champaign October 19 (Friday), :00 4:15p. SOLUTIONS (Yellow)

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Correlation Analysis

MATH 644: Regression Analysis Methods

Ch 13 & 14 - Regression Analysis

Section 3: Simple Linear Regression

Basic Business Statistics, 10/e

LI EAR REGRESSIO A D CORRELATIO

Chapter 1. Linear Regression with One Predictor Variable

Multiple Regression Examples

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

Chapter 3. Diagnostics and Remedial Measures

Chapter 4. Regression Models. Learning Objectives

6. Multiple Linear Regression

Confidence Interval for the mean response

Inference in Normal Regression Model. Dr. Frank Wood

TMA4255 Applied Statistics V2016 (5)

Analysis of Bivariate Data

Diagnostics and Remedial Measures

Simple and Multiple Linear Regression

Linear models and their mathematical foundations: Simple linear regression

Model Building Chap 5 p251

Inference for Regression Inference about the Regression Model and Using the Regression Line

Lecture 6 Multiple Linear Regression, cont.

3. Diagnostics and Remedial Measures

Math 3330: Solution to midterm Exam

Inference for Regression

ST430 Exam 2 Solutions

Chapter 2. Continued. Proofs For ANOVA Proof of ANOVA Identity. the product term in the above equation can be simplified as n

The simple linear regression model discussed in Chapter 13 was written as

The Multiple Regression Model

SMAM 314 Practice Final Examination Winter 2003

Basic Business Statistics 6 th Edition

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Remedial Measures, Brown-Forsythe test, F test

BNAD 276 Lecture 10 Simple Linear Regression Model

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

Matrix Approach to Simple Linear Regression: An Overview

Chapter 4: Regression Models

Business Statistics. Chapter 14 Introduction to Linear Regression and Correlation Analysis QMIS 220. Dr. Mohammad Zainal

Sociology 6Z03 Review II

F-tests and Nested Models

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

W&M CSCI 688: Design of Experiments Homework 2. Megan Rose Bryant

Lecture 15. Hypothesis testing in the linear model

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

Regression Analysis II

28. SIMPLE LINEAR REGRESSION III

Section 4: Multiple Linear Regression

Chapter 14 Student Lecture Notes 14-1

Measuring the fit of the model - SSR

Multiple Linear Regression

2.1: Inferences about β 1

Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest.

Ch 3: Multiple Linear Regression

Diagnostics and Remedial Measures: An Overview

Simple Linear Regression

Chapter 14. Linear least squares

INFERENCE FOR REGRESSION

Applied Regression Analysis. Section 2: Multiple Linear Regression

Inferences for Regression

Lecture 5: Linear Regression

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

STA 4210 Practise set 2b

Lecture 3: Inference in SLR

SIMPLE REGRESSION ANALYSIS. Business Statistics

Apart from this page, you are not permitted to read the contents of this question paper until instructed to do so by an invigilator.

Simple Linear Regression

Simple Linear Regression

Inference for the Regression Coefficient

1-Way ANOVA MATH 143. Spring Department of Mathematics and Statistics Calvin College

CS 5014: Research Methods in Computer Science

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #6

Chapter 7 Student Lecture Notes 7-1

Lecture 1 Linear Regression with One Predictor Variable.p2

Statistics for Managers using Microsoft Excel 6 th Edition

Class time (Please Circle): 11:10am-12:25pm. or 12:45pm-2:00pm

Regression Models. Chapter 4. Introduction. Introduction. Introduction

Inference for Regression Simple Linear Regression

STAT 212 Business Statistics II 1

Mathematics for Economics MA course

STA121: Applied Regression Analysis

Regression Analysis. Regression: Methodology for studying the relationship among two or more variables

STAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Multiple Regression Methods

We like to capture and represent the relationship between a set of possible causes and their response, by using a statistical predictive model.

Applied Regression Analysis

SSR = The sum of squared errors measures how much Y varies around the regression line n. It happily turns out that SSR + SSE = SSTO.

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Chapter Learning Objectives. Regression Analysis. Correlation. Simple Linear Regression. Chapter 12. Simple Linear Regression

Transcription:

(5+5)Q 1. Concordia University Department of Mathematics and Statistics Course Number Section Statistics 360/1 40 Examination Date Time Pages Mid Term Test May 26, 2004 Two Hours 3 Instructor Course Examiner Marks Y.P. Chaubey Y.P. Chaubey 80 Special Instructions: Closed Book Exam 1. Calculators are permitted. 2. Full Credit will be given only for answering questions systematically. 3. Answer ANY THREE questions. 4. Tables showing the required percentage points appear on the last page (a) Consider a simple linear model with one dependent variable (Y ) and one predictor variable (X). Given the independent observations (X i, Y i ), i = 1, 2,..., n, show that the least squares estimators of the regression parameters β 0, β 1 are given by b 1 = S xy /, b 0 = Ȳ b 1 X where, n S xy = (X i X)(Y i Ȳ ). i=1 (b.) Shown below are the number of galleys for a manuscript (X) and the dollar cost of correcting typographical errors (Y ) in a random sample of recent orders handled by a firm specializing in technical manuscripts. Y X 128.0 7.0 213.0 12.0 75.0 4.0 250.0 14.0 446.0 25.0 540.0 30.0 Show that the best fitted regression line is given by y = 1.60 + 17.9x. 1

(5+5)Q 2. (a.) Express Ŷh = b 0 + b 1 X h as a linear function of Y s and hence deduce for a normal linear regression model that Ŷ h N(µ h, σ 2 {Ŷh}) where ( 1 µ h = β 0 + β 1 X h, σ 2 {Ŷh} = σ 2 n + (X h X) 2 ). (5+5)Q 3. (b.) The following output is obtained in studying the relation between the asking price(x) and selling price (Y ) in a survey of 51 houses. Two new houses are available on the market with asking prices X = $40, 000. Find a 90% confidence interval for the expected selling price for such a house. Note: the value of X required in your computations may be obtained using the formulae; s 2 {b 1 } = MSE. and s 2 {b 0 } = MSE { 1 n + X 2 } = MSE n The regression equation is y = 5076 + 0.836 x + X 2 s 2 {b 1 }. Predictor Coef StDev T P Constant 5076 2193 2.31 0.025 x 0.83571 0.03266 25.59 0.000 S = 5517 R-Sq = 93.0% R-Sq(adj) = 92.9% (a) Use the result of Q. 2(a) to deduce that b 0 N(β 0, σ 2 ( 1 n + X 2 ). Outline the steps and facts in establishing that (b 0 β 0 ) MSE ( 1 + X 2 n ) t (n 2) 2

(5+5)Q 4. (b.) For the data in Q. 1 (b), an analyst fitted the simpler model Y i = βx i + ɛ i. Test an appropriate hypothesis to justify the use of the alternative model (use α = 0.05.) (a) Define the sum of squares denoted by SST O, SSR and SSE and establish the identity (6+4)Q 5. SST O = SSR + SSE Show that SSR can be computed from the formula SSR = b 1 S xy. (b) Compute the above sum of squares for the data in Q. 1 (b). Present the ANOVA table for testing the null hypothesis H 0 : β 1 = 0 at 5% level of significance. Please, specify the alternative, test statistic, decision rule and conclusion clearly. (a.) Explain the General Linear Test and derive the F statistic of ANOVA from the general Linear Test for testing the null hypothesis, H 0 : β 1 = 0. (b.) Consider testing the null hypothesis, H 0 : β 0 = β 1. Show that the sum of squares due to error under the reduced model is given by where SSE(r) = (Y i Ŷi(r)) 2, Ŷ i (r) = b(x i + 1), b = i Y i (X i + 1) i(x i + 1) 2. Give the expression for the F statistic along with the corresponding degrees of freedom. Simplify, the expression as much as you can. (6+4)Q 6. (a). For a simple linear model, prove that in the sample the residuals e i = Y i Ŷi, i = 1,..., n are uncorrelated with Ŷi, i = 1,..., n. Explain, how this information is used to detect departures from the model. 3

(5+5)Q 7. (b). Let r denote the sample correlation coefficient between X and Y. Show that the t statistic for testing H 0 : β 1 = 0 can be written as t = r (n 2). 1 r 2 (a.) Consider the data setup for testing Lack of Fit of the linear regression model as Y ij = β 0 + β 1 X j + ɛ ij, j = 1,..., c; i = 1,..., n j. Define SSP E and SSLF and prove the identity SSE = SSP E + SSLF. Use the General Linear Test approach to justify the test statistic F = MSLF MSP E (b.) Consider the following data on (X, Y ). Use the lack of fit test to determine if the simple linear model is appropriate here (Use α = 0.05.) X Y 75 28, 125 160,150 150 152 200 124,104 You may use the following regression output in aid to your computations. The regression equation is ynew = 65.4 + 0.367 xnew Predictor Coef SE Coef T P Constant 65.42 72.48 0.90 0.418 x 0.3674 0.4756 0.77 0.483 S = 51.60 R-Sq = 13.0% R-Sq(adj) = 0.0% Analysis of Variance Source DF SS MS F P Regression 1 1589 1589 0.60 0.483 Residual Error 4 10649 2662 Total 5 12238 4

(5+5)Q 8. (a). Let Ŷi(f) = b 0f + b 1f X i and Ŷi(r) = b 0r + b 1r X i denote the fitted values of Y i under the full model and reduced model respectively, prove that SSE(r) SSE(f) = i (Ŷi(r) Ŷi(f)) 2. (5+5)Q 9. (ii) Show that the estimators of regression parameters in a simple linear model obtained through the Maximum Likelihood method are the same as those obtained by the Least Square method if Y 1, Y 2,..., Y n are normally distributed? (a.) Prove that MSE = SSE n 2 is unbiased for σ2. (b.) Outline the steps and facts in establishing that Ŷ h Y h(new) t(n 2). s 2 {Ŷh} + MSE Use the above result in providing a 95% prediction interval for predicting the cost when the number of galleys is 10 for the data in Q.2 (b). 5