Analysis of Variance. Source DF Squares Square F Value Pr > F. Model <.0001 Error Corrected Total

Similar documents
Table 1: Fish Biomass data set on 26 streams

Simple Linear Regression (Part 3)

STAT 3A03 Applied Regression With SAS Fall 2017

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

(ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box.

Math 3330: Solution to midterm Exam

Lecture 1 Linear Regression with One Predictor Variable.p2

Multiple Linear Regression

Ch 2: Simple Linear Regression

T-test: means of Spock's judge versus all other judges 1 12:10 Wednesday, January 5, judge1 N Mean Std Dev Std Err Minimum Maximum

Chapter 1 Linear Regression with One Predictor

Confidence Intervals, Testing and ANOVA Summary

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

In Class Review Exercises Vartanian: SW 540

Simple Linear Regression

Simple linear regression

Chapter 2 Inferences in Simple Linear Regression

Lecture 11: Simple Linear Regression

Lecture 6 Multiple Linear Regression, cont.

1 Independent Practice: Hypothesis tests for one parameter:

a. The least squares estimators of intercept and slope are (from JMP output): b 0 = 6.25 b 1 =

unadjusted model for baseline cholesterol 22:31 Monday, April 19,

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

EE290H F05. Spanos. Lecture 5: Comparison of Treatments and ANOVA

AMS 7 Correlation and Regression Lecture 8

Statistics 512: Solution to Homework#11. Problems 1-3 refer to the soybean sausage dataset of Problem 20.8 (ch21pr08.dat).

Chapter 12: Multiple Regression

STATISTICS 479 Exam II (100 points)

Review 6. n 1 = 85 n 2 = 75 x 1 = x 2 = s 1 = 38.7 s 2 = 39.2

Simple Linear Regression

Multiple Linear Regression

General Linear Model (Chapter 4)

ST Correlation and Regression

ANALYSES OF NCGS DATA FOR ALCOHOL STATUS CATEGORIES 1 22:46 Sunday, March 2, 2003

Simple Linear Regression

Inferences for Regression

Chapter 8 Quantitative and Qualitative Predictors

Lecture 3: Inference in SLR

Objectives Simple linear regression. Statistical model for linear regression. Estimating the regression parameters

3 Variables: Cyberloafing Conscientiousness Age

R 2 and F -Tests and ANOVA

STOR 455 STATISTICAL METHODS I

Overview Scatter Plot Example

Answer Keys to Homework#10

Booklet of Code and Output for STAC32 Final Exam

Assignment 9 Answer Keys

UNIVERSITY EXAMINATIONS NJORO CAMPUS SECOND SEMESTER 2011/2012

ST 512-Practice Exam I - Osborne Directions: Answer questions as directed. For true/false questions, circle either true or false.

STATISTICS 110/201 PRACTICE FINAL EXAM

Correlation Analysis

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

Chapter 8 (More on Assumptions for the Simple Linear Regression)

SAS Procedures Inference about the Line ffl model statement in proc reg has many options ffl To construct confidence intervals use alpha=, clm, cli, c

Paper: ST-161. Techniques for Evidence-Based Decision Making Using SAS Ian Stockwell, The Hilltop UMBC, Baltimore, MD

SIMPLE LINEAR REGRESSION

Lecture 18 Miscellaneous Topics in Multiple Regression

df=degrees of freedom = n - 1

The program for the following sections follows.

Lecture 18: Simple Linear Regression

Question Possible Points Score Total 100

Inference for Regression

Inference for Regression Inference about the Regression Model and Using the Regression Line

Ch 3: Multiple Linear Regression

Stat 302 Statistical Software and Its Applications SAS: Simple Linear Regression

Chapter 14 Simple Linear Regression (A)

Statistical Techniques II EXST7015 Simple Linear Regression

Lecture 12 Inference in MLR

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #6

Correlation and the Analysis of Variance Approach to Simple Linear Regression

Inference for Regression Simple Linear Regression

Inference for Regression Inference about the Regression Model and Using the Regression Line, with Details. Section 10.1, 2, 3

Statistical Modelling in Stata 5: Linear Models

Econometrics Homework 1

Lecture 11 Multiple Linear Regression

Lecture 13 Extra Sums of Squares

IES 612/STA 4-573/STA Winter 2008 Week 1--IES 612-STA STA doc

Lectures 5 & 6: Hypothesis Testing

STAT 350: Summer Semester Midterm 1: Solutions

MS&E 226: Small Data

Chapter 10. Correlation and Regression. McGraw-Hill, Bluman, 7th ed., Chapter 10 1

Stat 500 Midterm 2 12 November 2009 page 0 of 11

Residuals from regression on original data 1

Basic Business Statistics, 10/e

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

Ordinary Least Squares Regression Explained: Vartanian

Use of Dummy (Indicator) Variables in Applied Econometrics

Empirical Application of Simple Regression (Chapter 2)

Simple Linear Regression

STAT 3A03 Applied Regression Analysis With SAS Fall 2017

Figure 1: The fitted line using the shipment route-number of ampules data. STAT5044: Regression and ANOVA The Solution of Homework #2 Inyoung Kim

Inference with Simple Regression

Six Sigma Black Belt Study Guides

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

STA 101 Final Review

Biostatistics 380 Multiple Regression 1. Multiple Regression

CHAPTER 2: Assumptions and Properties of Ordinary Least Squares, and Inference in the Linear Regression Model

Models for Clustered Data

CHAPTER EIGHT Linear Regression

Models for Clustered Data

Transcription:

Math 221: Linear Regression and Prediction Intervals S. K. Hyde Chapter 23 (Moore, 5th Ed.) (Neter, Kutner, Nachsheim, and Wasserman) The Toluca Company manufactures refrigeration equipment as well as many replacement parts. In the past, one of the replacement parts has been produced periodically in lots of varying sizes. When a cost improvement program was undertaken, company officials wished to determine the optimum lot size for producing this part. The production of this part involves setting up the production process (which must be done no matter what is the lot size) and machining and assembly operations. One key input for the model to ascertain the optimum lot size was the relationship between lot size and labor hours requires to produce the lot. To determine this relationship, data on lot size and work hours for 25 recent production runs were utilized. The production conditions were stable during the six-months period in which the 25 runs were made and were expected to continue to be the same during the next three years, the planning period for which the cost improvement program was being conducted. The output from SAS is below. Lot Hours Lot Hours Lot Hours Lot Hours 80 399 80 352 110 435 90 468 30 121 100 353 100 420 40 244 50 221 50 157 30 212 80 342 90 376 40 160 50 268 70 323 70 361 70 252 90 377 60 224 90 389 110 421 120 546 20 113 30 273 Table 1: Data on Lot Size and Work Hours Toluca Company The REG Procedure Model: MODEL1 Dependent Variable: y Work Hrs Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model 1 252378 252378 105.88 <.0001 Error 23 54825 2383.71562 Corrected Total 24 307203 Root MSE 48.82331 R-Square 0.8215 Dependent Mean 312.28000 Adj R-Sq 0.8138 Coeff Var 15.63447 Parameter Estimates Parameter Standard Variable Label DF Estimate Error t Value Pr > t Intercept Intercept 1 62.36586 26.17743 2.38 0.0259 x Lot Size 1 3.57020 0.34697 10.29 <.0001 Obs yhat stdi lower upper 9 419.386 50.8666 332.207 506.565

Linear Regression and Prediction Intervals, page 2 1. Find a 99% confidence interval for the true regression slope. 2. Test the hypothesis of no linear relationship. H 0 : H a : α = Test Statistic: P -value: Decision: (Circle one:) Reject H 0 or Fail to Reject H 0 Conclusion: 3. Predict the number of work hours needed for a lot size of 100.

Linear Regression and Prediction Intervals, page 3 4. Find a 90% prediction interval for the work hours needed for a lot size of 100. 5. Find a 99% prediction interval for the work hours needed for a lot size of 100.

Linear Regression and Prediction Intervals, page 4 Solution: 1. Find a 99% confidence interval for the true regression slope. The degrees of freedom for the critical value are found by either looking for the degrees of freedom for the ERROR on the printout (df = 23) or from the formula df = n 2 = 23. The critical value is then found in the t-table is 2.807. The estimate and standard error of the estimate for the true regression slope can be found on the printout. They are 3.57020 and 0.34697 respectively. Hence, a 99% confidence interval for the true regression slope is b ± m 2. Test the hypothesis of no linear relationship. b ± t SE b 3.57020 ± 2.807(.34697) (2.5963, 4.5441) This test can be performed easily as it is the same as the test that the true regression slope is zero. Hence, H 0 : β = 0 H a : β 0 α = 0.05 Test Statistic: t = 10.29 P -value: P -value < 0.0001 Decision: Since P -value < 0.05, then Reject H 0 Conclusion: There is a significant linear relationship between the lot size and the number of work hours. 3. Predict the number of work hours needed for a lot size of 100. The predicted number of work hours is W = 62.36586 + 3.57020(100) = 419.38586 Note that the printout also gives the prediction and is labeled as yhat. The ŷ given is for observation 9 or the ordered pair (100, 353). 4. Find a 90% prediction interval for the work hours needed for a lot size of 100. The easiest way to do a prediction interval is to have the computer do it. I had the computer compute a 90% prediction interval. The prediction interval for y when x is 100 is given as 332.207 < y x=100 < 506.565 To compute from the data, first compute the standard error of ŷ. In order to do so, x, Σx, and Σx 2 is needed. They are x = 70, Σx = 1750, and Σx 2 = 142300. The standard error for the

Linear Regression and Prediction Intervals, page 5 prediction of an individual y is SEŷ = s = 48.82331 1 + 1 n + n(x x) 2 n(σx 2 ) (Σx) 2 1 + 1 25 + 25(100 70) 2 25(142300) (1750) 2 = 50.8666387 Hence a 90% prediction interval for y when x = 100 is ŷ ± t SEŷ = 419.38586 ± 1.714(50.8666387) = (332.20044, 506.57128) Note that the computer prediction interval is more accurate since they use a better critical value. Notice that the number 50.8666387 is also given on the printout (under the label stdi). 5. Find a 99% prediction interval for the work hours needed for a lot size of 100. This can easily be modified by just changing the critical value. For a 99% confidence interval with 23 degrees of freedom, t = 2.807. Hence, ŷ ± t SEŷ = 419.38586 ± 2.807(50.8666387) = (276.60, 562.17)