Homework Solution (#5)

Similar documents
Objectives of Multiple Regression

Chapter 13 Student Lecture Notes 13-1

b. There appears to be a positive relationship between X and Y; that is, as X increases, so does Y.

Multiple Choice Test. Chapter Adequacy of Models for Regression

Chapter Business Statistics: A First Course Fifth Edition. Learning Objectives. Correlation vs. Regression. In this chapter, you learn:

Multiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades

Correlation and Simple Linear Regression

ENGI 3423 Simple Linear Regression Page 12-01

Correlation and Regression Analysis

Chapter Two. An Introduction to Regression ( )

STA302/1001-Fall 2008 Midterm Test October 21, 2008

12.2 Estimating Model parameters Assumptions: ox and y are related according to the simple linear regression model

Statistics. Correlational. Dr. Ayman Eldeib. Simple Linear Regression and Correlation. SBE 304: Linear Regression & Correlation 1/3/2018

Statistics MINITAB - Lab 5

Probability and. Lecture 13: and Correlation

Simple Linear Regression


Summary of the lecture in Biostatistics

Midterm Exam 1, section 2 (Solution) Thursday, February hour, 15 minutes

ε. Therefore, the estimate

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model

Simple Linear Regression

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE

: At least two means differ SST

Lecture 1: Introduction to Regression

Econometric Methods. Review of Estimation

Lecture Notes Types of economic variables

Lecture 1: Introduction to Regression

Midterm Exam 1, section 1 (Solution) Thursday, February hour, 15 minutes

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #1

Statistics: Unlocking the Power of Data Lock 5

Mean is only appropriate for interval or ratio scales, not ordinal or nominal.

ECON 482 / WH Hong The Simple Regression Model 1. Definition of the Simple Regression Model

residual. (Note that usually in descriptions of regression analysis, upper-case

Chapter 13, Part A Analysis of Variance and Experimental Design. Introduction to Analysis of Variance. Introduction to Analysis of Variance

Applied Statistics and Probability for Engineers, 5 th edition February 23, b) y ˆ = (85) =

Multiple Linear Regression Analysis

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions.

ESS Line Fitting

Previous lecture. Lecture 8. Learning outcomes of this lecture. Today. Statistical test and Scales of measurement. Correlation

Simple Linear Regression and Correlation.

Simple Linear Regression - Scalar Form

Linear Regression with One Regressor

Lecture 8: Linear Regression

Example: Multiple linear regression. Least squares regression. Repetition: Simple linear regression. Tron Anders Moger

Lecture 2: Linear Least Squares Regression

i 2 σ ) i = 1,2,...,n , and = 3.01 = 4.01

Simple Linear Regression and Correlation. Applied Statistics and Probability for Engineers. Chapter 11 Simple Linear Regression and Correlation

ENGI 4421 Propagation of Error Page 8-01

( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model

Maximum Likelihood Estimation

CLASS NOTES. for. PBAF 528: Quantitative Methods II SPRING Instructor: Jean Swanson. Daniel J. Evans School of Public Affairs

Comparison of Dual to Ratio-Cum-Product Estimators of Population Mean

Lecture Notes Forecasting the process of estimating or predicting unknown situations

Functions of Random Variables

Section l h l Stem=Tens. 8l Leaf=Ones. 8h l 03. 9h 58

Third handout: On the Gini Index

Johns Hopkins University Department of Biostatistics Math Review for Introductory Courses

CHAPTER 6. d. With success = observation greater than 10, x = # of successes = 4, and

Johns Hopkins University Department of Biostatistics Math Review for Introductory Courses

Analyzing Two-Dimensional Data. Analyzing Two-Dimensional Data

4. Standard Regression Model and Spatial Dependence Tests

Lecture 3. Sampling, sampling distributions, and parameter estimation

Chapter 14 Logistic Regression Models

Lecture 2: The Simple Regression Model

Chapter 11 The Analysis of Variance

LINEAR REGRESSION ANALYSIS

Sampling Theory MODULE V LECTURE - 14 RATIO AND PRODUCT METHODS OF ESTIMATION

Chapter 2 Supplemental Text Material

Multivariate Transformation of Variables and Maximum Likelihood Estimation

Chapter 5 Transformation and Weighting to Correct Model Inadequacies

MEASURES OF DISPERSION

Chapter 11 Systematic Sampling

9.1 Introduction to the probit and logit models

2SLS Estimates ECON In this case, begin with the assumption that E[ i

{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution:

The equation is sometimes presented in form Y = a + b x. This is reasonable, but it s not the notation we use.

Chapter 8. Inferences about More Than Two Population Central Values

Regression. Linear Regression. A Simple Data Display. A Batch of Data. The Mean is 220. A Value of 474. STAT Handout Module 15 1 st of June 2009

The number of observed cases The number of parameters. ith case of the dichotomous dependent variable. the ith case of the jth parameter

Chapter Statistics Background of Regression Analysis

STA 105-M BASIC STATISTICS (This is a multiple choice paper.)

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA

Suggested Answers, Problem Set 4 ECON The R 2 for the unrestricted model is by definition u u u u

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1

REVIEW OF SIMPLE LINEAR REGRESSION SIMPLE LINEAR REGRESSION

Module 7. Lecture 7: Statistical parameter estimation

ECONOMETRIC THEORY. MODULE VIII Lecture - 26 Heteroskedasticity

Model Fitting, RANSAC. Jana Kosecka

C. Statistics. X = n geometric the n th root of the product of numerical data ln X GM = or ln GM = X 2. X n X 1

best estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best

St John s College. Preliminary Examinations July 2014 Mathematics Paper 1. Examiner: G Evans Time: 3 hrs Moderator: D Grigoratos Marks: 150

CS 2750 Machine Learning. Lecture 8. Linear regression. CS 2750 Machine Learning. Linear regression. is a linear combination of input components x

CHAPTER 2. = y ˆ β x (.1022) So we can write

Can we take the Mysticism Out of the Pearson Coefficient of Linear Correlation?

University of Belgrade. Faculty of Mathematics. Master thesis Regression and Correlation

Analysis of Variance with Weibull Data

Recall MLR 5 Homskedasticity error u has the same variance given any values of the explanatory variables Var(u x1,...,xk) = 2 or E(UU ) = 2 I

Wu-Hausman Test: But if X and ε are independent, βˆ. ECON 324 Page 1

CHAPTER 8 REGRESSION AND CORRELATION

Transcription:

Homework Soluto (# Chapter : #6,, 8(b, 3, 4, 44, 49, 3, 9 ad 7 Chapter. Smple Lear Regresso ad Correlato.6 (6 th edto 7, old edto Page 9 Rafall volume ( vs Ruoff volume ( : 9 8 7 6 4 3 : a. Yes, the scatter-plot shows a strog lear relatoshp betwee rafall volume ad ruoff volume, thus t supports the use of the smple lear regresso model. ( 798 b. 3., 4. 867, S 634,86.4, ( 643 ( 798( 643 S 4,999 4,43.7, ad S,3 7,4.4. S 7,4.4 ˆ.8697 S,86.4 ad ˆ 4.867 (.8697 3.. 78. c..78 +.8697( 4. 7 µ. d. S ˆ 4,43.7 (.8697( 7,34.4 S 37. 7 SSE. SSE 37.7 s ˆ σ.4. 3 SSE 37.7 e. r. 973. So 97.3% of the observed varato SST 4,43.7 ruoff volume ca be attrbuted to the smple lear regresso relatoshp betwee ruoff ad rafall.. (6 th edto 9, old edto Page a. Verfcato 49.4 b. ˆ. 663486, ˆ 84748 744.6. c. Predcted: ˆ + ˆ ( 7. 7

ˆ ˆ µ Y d. + ( 7. 7.8 (6 th edto, old edto Page ˆ b. We wsh b ad b to mmze f(b, b Σ [ ( b + b ( ] f b to f b to elds b + b Σ( Σ, b Σ ( + b Σ( Σ( Σ(. Sce Σ(, b Σ ( Σ( ( [ Because ( Σ( Thus ˆ * Y ad ˆ * ˆ.. Equatg, ad sce Σ ], b ˆ..3 (old edto Page 9 a. ˆ t s :.49 ±.36.3.6,.7. The terval cotas ± ( ( α /, ˆ the value, dcatg s a possble value for, ad showg lttle usefuless of the model. b. The p-value assocated wth the model s., whch eceeds stadard levels of α. Ths dcates least square s ot a good wa to predct age from trasparet dete cotet..33 ( 6 th edto 8 a ˆ t s :.748 ±.6(.8.8,.34. The terval cotas ± ( α /, ˆ the value, dcatg s a possble value for, ad showg lttle usefuless of the model. b H., H :. : a > P-value ˆ.748. P( >. H P( T > P( T >.84. 8 >., so we.8 caot reject the ull hpothess at sgfcace level...4 (6 th edto 9, old edto Page 9 E. ( ( Σ ˆ Σ E ˆ We use the fact that ˆ s ubased for E( Σ ( ˆ E ( ΣY Σ( + E +.44 (6 th edto 36, old edto Page 6 a. The mea of the data Eercse. s 4.. Sce 4 s closer to 4. tha s 6, the quatt ( 4 must be smaller tha ( 6. Therefore, sce these quattes are the ol oes that are dfferet the two s ˆ. values, the value for 4 must ecessarl be smaller tha the for 6. Sad brefl, the closer s to, the smaller the value of s ˆ. s ˆ s ˆ

b. From the prtout Eercse., the error degrees of freedom s df. t.,.6, so the terval estmate whe 4 s : 7.9 ± (.6(79. 7.9 ±.369 ( 7.3,7. 96. We estmate, wth a hgh degree of cofdece, that the true average stregth for all beams whose MoE s 4 GPa s betwee 7.3 MPa ad 7.96 MPa. c. From the prtout Eercse., s.867, so the 9% predcto terval s ˆ ± t ˆ ( ( (., s + s 7.9 ±.6.867 +. 79 7.9 ±.8 (.77,9. 43. Note that the predcto terval s almost tmes as wde as the cofdece terval. d. For two 9% tervals, the smultaeous cofdece level s at least ( (. 9%.49 (6 th edto 37, old edto Page 7 9% CI: (46., 97.7; mdpot 9.9; t. 36 ; ( 7.,8 9.9 + (.36 ˆ ˆ ˆ ( 97. ˆ + ˆ ˆ + 99% CI: 9.9 ± ( 3.3( 9.4 ( 43.3,68. s, so s ( 9. 4..3 (6 th edto 38, old edto Page 7 Choce a wll be the smallest, wth d beg largest. a s less tha b ad c (obvousl, ad b ad c are both smaller tha d. Nothg ca be sad about the relatoshp betwee b ad c..9 (6 th edto 47, old edto Page 37 ( 9 ( 47.9 a. S,97 4,7, S 3.674 3.337, ad 8 8 ( 9( 47.9 S 3.9 339.86667, so 8 339.86667 r.966. There s a ver strog postve correlato 4,7 3.337 betwee the two varables. b. Because the assocato betwee the varables s postve, the specme wth the larger shear force wll ted to have a larger percet dr fber weght. c. Chagg the uts of measuremet o ether (or both varables wll have o effect o the calculated value of r, because a chage uts wll affect both the umerator ad deomator of r b eactl the same multplcatve costat. r.966. d. ( 933 r e. H o : ρ vs H a : ρ >. t ; Reject H o at level. f r.966 6 t t.,6.83. t 4.94. 83, so H o should be rejected. The.966 data dcates a postve lear relatoshp betwee the two varables.

.7 (6 th edto 49, old edto Page 39 a. Sample sze 8 b. ˆ 36.97638 ( 8. 43964. Whe 3., ˆ 8. 64. c. Yes, the model utlt test s statstcall sgfcat at the level.. d. Notce the sg of the slope s egatve, hece r r.934.97 e. Frst check to see f the value 4 falls wth the rage of values used to geerate the least-squares regresso equato. If t does ot, ths equato should ot be used. Furthermore, for ths partcular model a value of 4 elds a g value of 9.8, whch s a mpossble value for. Chapter 3: #9, 9 ad 9 Chapter 3. Nolear ad Multple Regresso 3.9 (6 th edto 63, old edto Page Both a scatter plot ad resdual plot (based o the smple lear regresso model for the frst data set suggest that a smple lear regresso model s reasoable, wth o patter or fluetal data pots whch would dcate that the model should be modfed. However, scatter plots for the other three data sets reveal dffcultes. Scatter Plot for Data Set # Scatter Plot for Data Set # 9 9 8 8 7 7 6 6 4 4 4 9 4 3 4 9 4

Scatter Plot for Data Set #3 Scatter Plot for Data Set #4 3 9 8 7 6 4 9 4 3 9 8 7 6 For data set #, a quadratc fucto would clearl provde a much better ft. For data set #3, the relatoshp s perfectl lear ecept oe outler, whch has obvousl greatl flueced the ft eve though ts value s ot uusuall large or small. The sgs of the resduals here (correspodg to creasg are + + + + - - - - - + -, ad a resdual plot would reflect ths patter ad suggest a careful look at the chose model. For data set #4 t s clear that the slope of the least squares le has bee determed etrel b the outler, so ths pot s etremel fluetal (ad ts value does le far from the remag oes. 3.9 (6 th edto 7, old edto Page 6 a. No, there s defte curvature the plot. b. Y + ( + ε where ad l(lfetme. Plottg vs. temp gves a plot, whch has a proouced lear appearace (ad fact r.94 for the straght le ft. c..873, 3. 64,. 3783, 879. 88, Σ Σ Σ Σ.79, from whch ˆ 373. 448 ad ˆ. 4 (values read from computer output. Wth,.44 so ˆ ˆ.4 + 373.448(.44 6. 7748 ad thus ˆ e 87.. d. For the trasformed data, SSE.3987, ad 6, Σ 3 8.4469, 6. 837,. 389, from whch SSPE.3694, SSLF.. 3..993/.993, f. 33. Comparg ths to F.,, 8. 68, t s clear that.3694 / H o caot be rejected. 3.9 (6 th edto 8, old edto Page 7 a. From computer output: ŷ :.89.66 4.7 94.6 8.69 ŷ : -.89.34 4.9-8.6 3.3

3.37 SSE, s. 69, s 7.9. ( Σ 3.37 b. SST Σ 63, so R. 96. 63 c. H wll be rejected favor of H : f ether t t. 4.33 or Thus (.89 +... + ( 3.3 3. 37 :.84 f t 4.33. Wth t 3. 83, H o caot be rejected; the data does ot.48 argue strogl for the cluso of the quadratc term. d. To obta jot cofdece of at least 9%, we compute a 98% C.I. for each coeffcet usg t., 6.96. For the C.I. s 8.6 ± ( 6.96( 4. ( 9.87,3. 99 ( a etremel wde terval, ad for the C.I. s.84 ± ( 6.96(.48 (.8,.. e. t.9 ad ˆ + 4 ˆ + 6 ˆ 4. 7, so the C.I. s 4.7±.9.., a, ( ( (.8,9.34. 4.7± 4.63 f. If we kew ˆ ˆ ˆ,,, the value of whch mamzes ˆ + ˆ + ˆ would be obtaed b settg the dervatve of ths to ad solvg: ˆ +. The estmate of ths s.9. ˆ Etra Problems: A & B A. There are ma problems wth cocludg that ths data teds to cofrm the clam. I ll epla some here. a. The regresso effect or regresso to the mea. Ths pheomeo tells us that sce we re eamg ol the lowest schools we should epect o average that the wll all move closer to the mea. Ths s a well documeted pheomeo gog all the wa back to Galto s data o the heght of sos/daughters ad predctg these heghts as a fucto of the heghts of parets. b. Ths data does t sa athg about what mprovemets, f a, were made b the other schools. If those schools also creased ther scores the we should adjust the gas made b the lower to reflect the overall tred of scorg hgher o the tests. A better wa (though stll ot perfect to coduct ths stud would be to make t a trul radom test. Take the test scores from the frst ear ad radoml choose some schools to tell ther scores (ad hece ther rak overall ad wthhold the scores from the remag schools. I ths maer ou would be able to eame the effect that kowledge of the eamato score has o the et ear s eamato. Wth the curret aalss there s o evdece to suggest that

kowledge of the score had a effect o crease scores. The stud orgazers dd t attempt to measure a dfferece betwee schools wth kowledge ad schools wthout kowledge of ther scores from the prevous ear. B. a. 3 Respose Salar ($g Whole Model Actual b Predcted Plot Salar ($g Actual 3-3 3 Salar ($g Predcted P<. RSq.79 RMSE4.764 Summar of Ft RSquare.7866 RSquare Adj.777343 Root Mea Square Error 4.76383 Mea of Respose 3.94 Observatos (or Sum Wgts Aalss of Varace Source DF Sum of Squares Mea Square F Rato Model 478.97 478.9 84.789 Error 3 4.7 7.44 Prob > F C. Total 4 88.77 <. Parameter Estmates Term Estmate Std Error t Rato Prob> t Itercept 6.7443.67943.86 <. Pck # -.9897.7468-9. <. Resdual b Predcted Plot Salar ($g Resdual - - 3 3 Salar ($g Predcted b. I $, s: tpcal pck: ($9.97, $3.49 dvdual pck: ($.9, $.

c. Both the Y b predcted ad the resdual plot show a sstematc varato. We ca hopefull accout for ths volato of the regresso requremets b trasformg the data. d. I have omtted the regresso tables for space cocers but here are the approprate cofdece ad predcto tervals ( $, s: e.666967, e.3773 $7.9, $.76 A.: Tpcal pck: ( ( Idvdual pck: ( e.4467378, e.996698 ( $4., $. A.: Tpcal pck: ($6.64, $.6 Idvdual pck: ($.4, $4.3 A.3: Tpcal pck: ($8.43, $.76 Idvdual pck: ($4.3, $.7 e. Because we kow that salares wll, geeral, ol cotue to decrease t seems ulkel that usg a quadratc regresso wll be accurate ear the rght tal of our data set. A rse salares s t what we epect to happe ad so the predctve power of A. ear the later pcks the draft should t be trusted.