Question Possible Points Score Total 100

Similar documents
AMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression

SMAM 319 Exam 1 Name. 1.Pick the best choice for the multiple choice questions below (10 points 2 each)

STAT 512 MidTerm I (2/21/2013) Spring 2013 INSTRUCTIONS

AMS 7 Correlation and Regression Lecture 8

Inference for Regression Simple Linear Regression

STATISTICS 110/201 PRACTICE FINAL EXAM

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #6

Analysis of Bivariate Data

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, October 2013

Elementary Probability and Statistics Midterm I April 22, 2010

1. An article on peanut butter in Consumer reports reported the following scores for various brands

Apart from this page, you are not permitted to read the contents of this question paper until instructed to do so by an invigilator.

1 Introduction to Minitab

Test 3 Practice Test A. NOTE: Ignore Q10 (not covered)

Inference for the Regression Coefficient

FRANKLIN UNIVERSITY PROFICIENCY EXAM (FUPE) STUDY GUIDE

Sociology 593 Exam 1 February 17, 1995

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)

Inference for Regression Inference about the Regression Model and Using the Regression Line, with Details. Section 10.1, 2, 3

Lecture 18: Simple Linear Regression

STAT 3900/4950 MIDTERM TWO Name: Spring, 2015 (print: first last ) Covered topics: Two-way ANOVA, ANCOVA, SLR, MLR and correlation analysis

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

Ch 13 & 14 - Regression Analysis

Analysis of Variance. Source DF Squares Square F Value Pr > F. Model <.0001 Error Corrected Total

This document contains 3 sets of practice problems.

Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014

Predict y from (possibly) many predictors x. Model Criticism Study the importance of columns

Model Building Chap 5 p251

MATH 644: Regression Analysis Methods

Multiple Regression Examples

Regression Models - Introduction

Chapter 14 Student Lecture Notes 14-1

Confidence Intervals, Testing and ANOVA Summary

9 Correlation and Regression

Statistical Techniques II EXST7015 Simple Linear Regression

Correlation & Simple Regression

STAT 350: Summer Semester Midterm 1: Solutions

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) HIGHER CERTIFICATE IN STATISTICS, 2000

Math 3330: Solution to midterm Exam

with the usual assumptions about the error term. The two values of X 1 X 2 0 1

Unit 6 - Introduction to linear regression

What If There Are More Than. Two Factor Levels?

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

AP Statistics Unit 6 Note Packet Linear Regression. Scatterplots and Correlation

Chapter 15 Multiple Regression

Inferences for Regression

Algebra 2 CP Semester 1 PRACTICE Exam

Box-Cox Transformations

Department of Mathematics & Statistics STAT 2593 Final Examination 17 April, 2000

Basic Business Statistics 6 th Edition

Multiple Linear Regression

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters

Can you tell the relationship between students SAT scores and their college grades?

Chapter 14. Multiple Regression Models. Multiple Regression Models. Multiple Regression Models

INFERENCE FOR REGRESSION

Mrs. Poyner/Mr. Page Chapter 3 page 1

Data Set 8: Laysan Finch Beak Widths

SMAM 314 Computer Assignment 5 due Nov 8,2012 Data Set 1. For each of the following data sets use Minitab to 1. Make a scatterplot.

Models with qualitative explanatory variables p216

PART I. (a) Describe all the assumptions for a normal error regression model with one predictor variable,

2. Outliers and inference for regression

SMAM 319 Exam1 Name. a B.The equation of a line is 3x + y =6. The slope is a. -3 b.3 c.6 d.1/3 e.-1/3

Stat 101: Lecture 6. Summer 2006

(ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box.

SCHOOL OF MATHEMATICS AND STATISTICS Autumn Semester

ST430 Exam 1 with Answers

(4) 1. Create dummy variables for Town. Name these dummy variables A and B. These 0,1 variables now indicate the location of the house.

Inference for Regression Inference about the Regression Model and Using the Regression Line

ANCOVA. ANCOVA allows the inclusion of a 3rd source of variation into the F-formula (called the covariate) and changes the F-formula

TAMS38 Experimental Design and Biostatistics, 4 p / 6 hp Examination on 19 April 2017, 8 12

Regression analysis is a tool for building mathematical and statistical models that characterize relationships between variables Finds a linear

Simple Linear Regression for the Climate Data

Ecn Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman. Midterm 2. Name: ID Number: Section:

Sociology Research Statistics I Final Exam Answer Key December 15, 1993

Swarthmore Honors Exam 2012: Statistics

Analysis of Variance

O2. The following printout concerns a best subsets regression. Questions follow.

Model Selection Procedures

Unit 6 - Simple linear regression

Regression. Marc H. Mehlman University of New Haven

Any of 27 linear and nonlinear models may be fit. The output parallels that of the Simple Regression procedure.

Analysis of Covariance. The following example illustrates a case where the covariate is affected by the treatments.

MASM22/FMSN30: Linear and Logistic Regression, 7.5 hp FMSN40:... with Data Gathering, 9 hp

ST430 Exam 2 Solutions

No books, notes, any calculator, or electronic devices are allowed on this exam. Show all of your steps in each answer to receive a full credit.

11 Correlation and Regression

Correlation and Linear Regression

Lecture 3: Inference in SLR

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

STAT 350. Assignment 4

Chapter 12 - Part I: Correlation Analysis

23. Inference for regression

Simple Linear Regression

Chapter 4: Regression Models

Inference with Simple Regression

Non-parametric (Distribution-free) approaches p188 CN

Business Statistics. Lecture 10: Correlation and Linear Regression

This is a multiple choice and short answer practice exam. It does not count towards your grade. You may use the tables in your book.

STAT 212: BUSINESS STATISTICS II Third Exam Tuesday Dec 12, 6:00 PM

Transcription:

Midterm I NAME: Instructions: 1. For hypothesis testing, the significant level is set at α = 0.05. 2. This exam is open book. You may use textbooks, notebooks, and a calculator. 3. Do all your work in the spaces provided. If you need additional space, use the back of the preceding page, indicating clearly that you have done so. 4. To get full credit, you must show your work. Partial credit will be awarded. 5. Do not dwell too long on any one question. Answer as many questions as you can. 6. Note that some questions have multiple parts. For some questions, these parts are independent, and so, for example, you can work on parts (b) or (c) separately from part (a). For grader s use: Question Possible Points Score 1 40 2 36 3 24 Total 100 1

1. Agricultural researchers are interested in the growth of Golden Acre cabbage. The following data are recorded on the height (y) in cm of one cabbage after the number of weeks (x) since the first observation. In total, there are n = 8 observations. x: 0 1 2 3 4 5 6 7 y: 4.2 5.6 6.3 8.2 10.5 10.8 15.3 17.6 (a) The simple linear regression (SLR) model y = b 0 + b 1 x + e is fitted to the complete data set. The following is the corresponding studentized residual plot. Perform model diagnosis (i.e., check the four model assumptions and any potential outliers). For violation against any of the model assumptions, suggest possible remedies. (You do not need to carry out the remedies you are to suggest.) MTB > plot residual yhat Plot A A 1.0+ A residual A 0.0+ A A A 1.0+ A ++++++yhat 2.5 5.0 7.5 10.0 12.5 15.0 2

(b) The largest residual corresponds to the observation x = 5, y = 10.8. Suppose this observation is deleted and the SLR is fitted to the reduced data set. Given that ŷ est = 13.03, se(ŷ est ) = 0.39, and SSError = 3.50, all of which are based on the reduced data set, test formally whether this observation is an outlier. (c) Consider fitting the following two models to the complete data set: I SLR: y = b 0 + b 1 x + e II Polynomial regression: y = b 0 + b 1 x + b 2 x 2 + e Given that SSTotal = 156.39, R 2 = 95.1% for Model I, and R 2 = 98.3% for Model II, test formally H o : Model I (i.e., y = b 0 + b 1 x + e) versus H a : Model II (i.e., y = b 0 + b 1 x + b 2 x 2 + e). 3

2. In another experiment, the growth of three different cabbages (A, B, and C) at biweekly intervals are recorded. The data are tabulated below with y = the height in cm and x = the number of weeks since the first observation. A B C x y x y x y 0 3.2 0 6.3 0 3.9 2 8.7 2 10.4 2 10.4 4 11.1 4 14.1 4 18.9 6 16.1 6 16.4 6 26.0 The following (edited) Minitab printouts may be of use: MTB > print w1w5 y Data Display Row w1 w2 w3 w4 w5 y 1 0 1 0 0 0 3.2 2 2 1 0 2 0 8.7 3 4 1 0 4 0 11.1 4 6 1 0 6 0 16.1 5 0 0 1 0 0 6.3 6 2 0 1 0 2 10.4 7 4 0 1 0 4 14.1 8 6 0 1 0 6 16.4 9 0 0 0 0 0 3.9 10 2 0 0 0 0 10.4 11 4 0 0 0 0 18.9 12 6 0 0 0 0 26.0 MTB > regress y 5 w1 w2 w3 w4 w5 Regression Analysis: y versus w1, w2, w3, w4, w5 y = 3.58 + 3.74 w1 + 0.030 w2 + 3.12 w3 1.68 w4 2.04 w5 Constant 3.5800 0.6124 5.85 0.001 w1 3.7400 0.1637 22.85 0.000 w2 0.0300 0.8661 0.03 0.973 w3 3.1200 0.8661 3.60 0.011 w4 1.6850 0.2315 7.28 0.000 w5 2.0400 0.2315 8.81 0.000 S = 0.7320 RSq = 99.3% RSq(adj) = 98.8% Analysis of Variance Source DF SS MS F P Regression 5 473.147 94.629 176.60 0.000 Residual Error 6 3.215 0.536 Total 11 476.362 Source DF Seq SS w1 1 374.500 w2 1 33.135 w3 1 18.000 w4 1 5.896 w5 1 41.616 4

MTB > regress y 3 w1 w4 w5 Regression Analysis: y versus w1, w4, w5 y = 4.63 + 3.51 w1 1.68 w4 1.37 w5 Constant 4.6300 0.6013 7.70 0.000 w1 3.5150 0.2104 16.70 0.000 w4 1.6786 0.2353 7.14 0.000 w5 1.3714 0.2353 5.83 0.000 S = 1.245 RSq = 97.4% RSq(adj) = 96.4% Analysis of Variance Source DF SS MS F P Regression 3 463.96 154.65 99.80 0.000 Residual Error 8 12.40 1.55 Total 11 476.36 Source DF Seq SS w1 1 374.50 w4 1 36.80 w5 1 52.66 Consider the model y = b 0 + b 1 w 1 + b 2 w 2 + b 3 w 3 + b 4 w 4 + b 5 w 5 + e. (a) Briefly interpret b 3 and test H 0 : (b 3 = 0 b 0, b 1, b 2, b 4, b 5 ). 5

(b) Test whether the intercepts of the regression lines relating height to time for the three cabbages are equal (without assuming that the slopes are equal). (c) Test whether the slopes of the regression lines relating height to time for the three cabbages are equal (without assuming that the intercepts are equal). 6

3. It is of interest to study the possible relationships between a dependent variable y and three independent variables x 1, x 2, and x 3, based on n = 30 observations. The following (edited) Minitab printouts may be of use: MTB > regress y 1 x1 y = 0.94 + 0.898 x1 Constant 0.938 1.440 0.65 0.520 x1 0.89840 0.08122 11.06 0.000 S = 3.801 RSq = 81.4% RSq(adj) = 80.7% MTB > regress y 1 x2 y = 2.92 + 0.762 x2 Constant 2.922 1.627 1.80 0.083 x2 0.76207 0.08863 8.60 0.000 S = 4.617 RSq = 72.5% RSq(adj) = 71.5% MTB > regress y 1 x3 y = 10.1 + 0.274 x3 Constant 10.136 1.936 5.23 0.000 x3 0.27449 0.08015 3.42 0.002 S = 7.394 RSq = 29.5% RSq(adj) = 27.0% MTB > regress y 2 x1 x2 y = 0.95 + 0.751 x1 + 0.145 x2 Constant 0.953 1.449 0.66 0.516 x1 0.7506 0.2022 3.71 0.001 x2 0.1452 0.1817 0.80 0.431 S = 3.826 RSq = 81.8% RSq(adj) = 80.5% MTB > regress y 2 x1 x3 y = 0.94 + 0.895 x1 + 0.0030 x3 Constant 0.940 1.467 0.64 0.527 x1 0.8949 0.1032 8.67 0.000 x3 0.00298 0.05235 0.06 0.955 S = 3.870 RSq = 81.4% RSq(adj) = 80.0% MTB > regress y 2 x2 x3 7

y = 2.87 + 0.802 x2 0.0334 x3 Constant 2.869 1.654 1.73 0.094 x2 0.8023 0.1225 6.55 0.000 x3 0.03342 0.06918 0.48 0.633 S = 4.681 RSq = 72.8% RSq(adj) = 70.7% MTB > regress y 3 x1 x2 x3 y = 0.94 + 0.746 x1 + 0.169 x2 0.0168 x3 Constant 0.939 1.475 0.64 0.530 x1 0.7458 0.2064 3.61 0.001 x2 0.1693 0.2027 0.84 0.411 x3 0.01677 0.05771 0.29 0.774 S = 3.892 RSq = 81.9% RSq(adj) = 79.8% (a) Perform a backward elimination to choose the best model. Give all the steps in the backward elimination. (b) The following italicized statement is either True or False. Indicate whether the statement is true or false. Give a clear justification of your answer. For this data set, the model chosen by a forward selection is the same as the model chosen by a stepwise procedure. 8