Chapter 14 Simple Linear Regression Page 1. Introduction to regression analysis 14-2

Size: px
Start display at page:

Download "Chapter 14 Simple Linear Regression Page 1. Introduction to regression analysis 14-2"

Transcription

1 Chapter 4 Smple Lnear Regresson Page. Introducton to regresson analyss 4- The Regresson Equaton. Lnear Functons Estmaton and nterpretaton of model parameters Inference on the model parameters 4-5. Sums of Squares and the ANOVA table An example 4-7. Estmaton and Predcton Standardzed regresson coeffcents Addtonal concerns and observatons A. Karpnsk

2 The Regresson Equaton. Overvew of regresson analyss Regresson analyss s generally used when both the ndependent and the dependent varables are contnuous. (But modfcatons exst to handle categorcal ndependent varables and dchotomous dependent varables.) Type of Analyss Independent Varable Dependent Varable ANOVA Categorcal Contnuous Regresson Categorcal Analyss (Contngency Table Analyss) Contnuous or Categorcal Contnuous Categorcal Goals of regresson analyss: o To descrbe the relatonshp between two varables o To model responses on a dependent varable o To predct a dependent varable usng one or more ndependent varables o To statstcally control the effects of varables whle examnng the relatonshp between the ndependent and dependent varable Regresson analyss s usually performed on observatonal data. In these cases, we descrbe, model, predct, and control, but we cannot make any causal clams regardng these relatonshps 4- A. Karpnsk

3 Termnology n regresson analyss o As n ANOVA, we wll develop a model to explan the data DATA = MODEL + ERROR o The model assumes greater mportance n regresson. Unlke ANOVA, we are usually nterested n the model parameters o The goal of most regresson models s to use the nformaton contaned n a set of varables to predct a response. As a result, we use slghtly dfferent termnology n regresson, compared to ANOVA. ANOVA Dependent varable Independent varables REGRESSION Dependent varable or Response varable or Outcome varable Independent varables or Predctor varables 4-3 A. Karpnsk

4 Smple Lnear Regresson The Regresson Equaton. Lnear Functons The goal of smple lnear regresson s to descrbe an outcome varable (Y) as a lnear functon of a predctor varable (). The end result wll be a model that defnes the equaton of a straght lne Y = b + a Where b = the y-ntercept a = the slope o Let s consder a smple example: Y = + 3 The y-ntercept s - The lne crosses the y-axs at y = The slope of the lne s /3 The slope s a measure of the steepness of the lne The slope s the change n y assocated wth a unt change n x Two data ponts A straght lne through two ponts y-axs x-axs y-axs x-axs 4-4 A. Karpnsk

5 o Let s revew the method covered n hgh school algebra for determnng the lne that falls through ponts: (, ) & (5, ) Frst, we compute the slope of the lne y y slope = x x ( ) slope = 5 ( ) = = We nterpret the slope as the change n y assocated wth a unt change n x In ths example, for every unt ncrease n x, y wll ncrease by.333 x y We compute the y-ntercept by fndng the value of y when x = We can use: the equaton for the slope of a lne and the coordnates of ether known pont to solve for, y ) ( ( x, y) Let s use (5, ) y.333 = (5) = y y =.667 y =.667 Fnally, we use the slope and the ntercept to wrte the equaton of the lne through the ponts Y = b + a Y = ( ) 4-5 A. Karpnsk

6 3. Estmaton and nterpretaton of model parameters Wth real data, the ponts rarely fall drectly on a straght lne. Regresson s a technque to estmate the slope and the y-ntercept from nosy data Because not every pont wll fall on the regresson lne, there wll be error n our model DATA = MODEL + ERROR o The DATA, or the outcome we want to predct s the Y varable o The MODEL s the equaton of the regresson lne, b + b b = the populaton value of the ntercept b = the populaton value of the slope = the predctor varable o The ERROR s devaton of the observed data from our regresson lne. We refer to the ndvdual error terms as resduals o The full smple lnear regresson model s gven by the followng equaton: DATA = MODEL + ERROR Y b + b + ε = Some key characterstcs of ths model o We can only model lnear relatonshps between the outcome varable and the predctor varable o The model can be expanded to nclude the lnear relatonshps between multple predctor varables and a sngle outcome Y = b + b + b b k k + ε 4-6 A. Karpnsk

7 Predcted values and resduals o Wth real data, we need to estmate the value of the slope and the ntercept. (Detals on the estmaton process wll follow shortly.) Y = bˆ + bˆ + ε ˆb = the estmated value of the ntercept ˆb = the estmated value of the slope o Based on the model, we have a best guess as to the partcpant s response on the outcome varable ˆ = bˆ + bˆ Y In other words, we use the equaton of the lne we have developed to estmate how each partcpant responded on the outcome varable Yˆ s called the predcted value or ftted value for the th partcpant o If the actual response of the partcpant devates from our predcted value, then we have some ERROR n the model. We defne the resdual to be the devaton of the observed value from the predcted value. DATA = MODEL + ERROR Y = ( bˆ + bˆ ) + e Y = Y ˆ + e e = Y Yˆ o If we want to know f our model s a good model, we can examne the resduals. If we have many large resduals, then there are many observatons that are not predcted well by the model. We say that the model has a poor ft. If most of the resduals are small, then our model s very good at explanng responses on the Y varable. Ths model would have a good ft. 4-7 A. Karpnsk

8 o Let s consder a smple example to llustrate these ponts Y We notce that a straght lne can be drawn that goes drectly through three of the 5 observed data ponts. Let s use ths lne as our best guess lne ~ Y = + Now we can calculate predcted values and resduals Y Yˆ e A. Karpnsk

9 In the prevous example, we eyeballed a regresson lne. We would lke to have a better method of estmatng the regresson lne. Let s consder desrable two propertes of a good regresson lne. The sum of the resduals should be zero ( yˆ) = y If we have ths property, then the average resdual would be zero In other words, the average devaton from the predcted lne would be zero. Overall, we would lke the resduals to be as small as possble We already requre the resduals to sum to zero, by property (). So, let s requre the sum of the squared resduals to be as small as possble. Ths approach has the added beneft of penalzng large resduals more than small resduals ( y yˆ ) = mnmum o Estmatng a regresson lne usng these two propertes s called the ordnary least squares (OLS) estmaton procedure o Estmates of the ntercept and slope are called the ordnary least squares (OLS) estmates o To solve for these estmates, we can use the followng procedure We want to mnmze SSE = ( Y ˆ Y ) = ( Y b b ) We take the dervatves of SSE wth respect to b and b, set each equal to zero, and solve for b and b SSE SSE = and = b b We ll skp the detals and jump to the fnal estmates SS ˆ Y bˆ Y b SS = b = Where SS SS Y = = n n ( ( ) )( Y Y ) 4-9 A. Karpnsk

10 Now, let s return to our example and examne the least squares regresson lne 5 4 Eyeball Lne 3 LS Lne Let s compare the least squares regresson lne to our eyeball regresson lne ~ Y = + Y ˆ = Y Data Eyeball Least Squares Y ~ e ~ ~e Yˆ ê ê e e. o For both models, we satsfy the condton that the resduals sum to zero o But the least squares regresson lne produces the model wth the smallest squared resduals Note that other regresson lnes are possble o We could mnmze the absolute value of the resduals o We could mnmze the shortest dstance to the regresson lne 4- A. Karpnsk

11 4. Inference on the model parameters We have learned how to estmate the model parameters, but also want to perform statstcal tests on those parameters SS ˆ Y bˆ Y b SS = ˆ b = Frst, let s estmate the amount of error n the model, σ o Intutvely, the greater the amount of error n a sample, the more dffcult t wll be to estmate the model parameters. o The error n the model s captured n the resduals, e We need to calculate the varance of the resduals Recall a varance s the average squared devaton from the mean When appled to resduals, we obtan But we know ε = ( ε ε ) Var( ε ) = N Varˆ( ( ˆ ε ) ( ˆ ) Y Y ) = ˆ σ = = N N ε ε = SSResdual N Why use N-? A general heurstc s to use N (number of parameters ftted) In ths case, we have estmated two parameter: the slope and the ntercept Recall that for Var(), we dvded by N-. We only estmated one parameter (the grand mean) Ths heurstc also appled for ANOVA. 4- A. Karpnsk

12 And so we are left wth ( ˆ ε ) SSresd Varˆ( ε ) = = N N # of parameters = MSresd And we are justfed usng nvolvng the regresson model MSresd as the error term for tests o Interpretng MSresd: Resduals measure devaton from regresson lne (the predcted values) The varance of the resduals captures the average squared devaton from the regresson lne So we can nterpret MSresd as a measure of average devaton from the regresson lne. SPSS labels MSresd as standard error of the estmate Now that we have an estmate of the error varance, we can proceed wth statstcal tests of the model parameters We can perform a t-test usng our famlar t-test formula t ~ estmate standard error of the estmate o We know how to calculate the estmates of the slope and the ntercept. All we need are standard errors of the estmates 4- A. Karpnsk

13 Inferences about the slope, ˆb o Dervng the samplng dstrbuton of ˆb tedous. We ll skp the detals (see an advanced regresson textbook, f nterested) and the end result s: std. error ( bˆ ) MSresd ( ) = o Thus, we can conduct the followng statstcal test: H : b = H : b t( N ) ~ bˆ standard error ( bˆ ) o We can also easly compute confdence ntervals around ˆb estmate ± tα /, df bˆ * standard error of MSresd ( ) ± tα /, df * estmate o Conclusons If the test s sgnfcant, then we conclude that there s a sgnfcant lnear relatonshp between and Y For every one-unt change n, there s a unt change n Y ˆb If the test s not sgnfcant, then there s no sgnfcant lnear relatonshp between and Y Utlzng the lnear relatonshp between and Y does not sgnfcantly mprove our ablty to predct Y, compared to usng the grand mean. There may stll exst a sgnfcant non-lnear relatonshp between and Y 4-3 A. Karpnsk

14 Inferences about the ntercept, b o b tells us the predcted value of Y when = o The test of b s automatcally computed and dsplayed, but be careful not to msnterpret ts sgnfcance! o Only rarely are we nterested n the value of the ntercept o Agan, we ll skp the detals concernng the dervaton of the samplng dstrbuton of ˆb (see an advanced regresson textbook, f nterested) and the end result s: std. error ( bˆ ) ( = MSresd N ) o Thus, we can conduct the followng statstcal test: H H : b : b = t( N ) ~ bˆ standard error ( bˆ ) o We can also easly compute confdence ntervals around ˆb estmate ± tα /, df bˆ * standard error of ( estmate ) ± tα /, df * MSresd N 4-4 A. Karpnsk

15 5. Sums of Squares n Regresson and the ANOVA table Total Sums of Squares (SST) o In ANOVA, the total sums of squares were the sum of the squared devatons from the grand mean o We wll use ths same defnton n regresson. SST s the sum of the squared devatons from the grand mean of Y 5 SST = n = ( Y Y ) 4 3 Mean(Y) Sums of Squares Regresson o In ANOVA, we had a sum of squares for the model. Ths SS captured the mprovement n our predcton of Y based on all the terms n the model o In regresson, we can also examne how much we mprove our predcton (compared to the grand mean) by usng the regresson lne to predct new observatons If we had not conducted a regresson, then our best guess for a new value of Y would be the mean of Y, Y But we can use the regresson lne to make better predctons of new observatons ˆ = bˆ + bˆ Y 4-5 A. Karpnsk

16 The devaton of the regresson best guess (the predcted value) from the grand mean s the SS Regresson. 5 SSReg = n = ( Yˆ Y ) 4 3 LS Lne Mean(Y) Sums of Squares Error / Resdual o The resduals are the devatons of the predcted values from the observed values e = Y Yˆ o The SS Resdual s the amount of the total SS that we cannot predct from the regresson model n SSResd = ( Y Yˆ ) 5 = 4 3 LS Lne Mean(Y) A. Karpnsk

17 Sums of Squares parttonng o We have three SS components and we can partton them n the followng manner n = SST = SSreg + n SSresd ( Y ) = ( ˆ ) + ( ˆ Y Y Y Y Y ) = o In ANOVA, we had a smlar partton SST = SSmodel + SSerror It turns out that ANOVA s a specal case of regresson. If we set up a regresson wth categorcal predctors, then we wll fnd SSreg = SSmodel SSresd = SSerror Every analyss we conducted n ANOVA, can be conducted n regresson. But regresson provdes a much more general statstcal framework (and thus s frequently called the general lnear model ). Where there are sums of squares, there s an ANOVA table. o Based on the SS decomposton, we can construct an ANOVA table Source SS df MS F Regresson n SSReg = Yˆ (# of parameters) SSreg MSreg ( Y ) - = df MSresd Resdual n SSResd = Y Yˆ N SSresd ( ) (# of parameters) df = Total SST = n = ( Y Y ) N- 4-7 A. Karpnsk

18 o The Regresson test examnes all of the slope parameters n the model smultaneously. Do these parameters sgnfcantly mprove our ablty to predct Y, compared to usng the grand mean to predct Y? H : b = b =... = bk = : b j ' s = H Not all o For smple lnear regresson, we only have one slope parameter. Ths test becomes a test of the slope of b H : b = : b H In other words, for smple lnear regresson, the Regresson F-test wll be dentcal to the t-test of the b parameter Ths relatonshp wll not hold for multple regresson, when more than one predctor s entered nto the model Calculatng a measure of varance n Y accounted for by o SS Total s a measure of the total varablty n Y SST = n = ( Y Y ) SST Var ( Y ) = N o The SS Regresson s the part of the total varablty that we can explan usng our regresson lne o As a result, we can consder the followng rato, R to be a measure of the proporton of the sample varance n Y that s explaned by R = SSReg SSTotal R s analogous to η n ANOVA 4-8 A. Karpnsk

19 o But n ANOVA, we preferred a measure varance accounted for n the populaton ( ω ) rather than n the sample ( η ). o The regresson equvalent of ω s called the Adjusted R. Any varable (even a completely random varable) s unlkely to have SSReg exactly equal to zero. Thus, any varable we use wll explan some of the varance n the sample Adjusted R corrects for ths overestmaton by penalzng R for the number of varables n the regresson equaton What happens f we take the square root of R? R = SSReg SSTotal o R s nterpreted as the overall correlaton between all the predctor varables and the outcome varable o When only one predctor s n the model, R s the correlaton between and Y, r Y 4-9 A. Karpnsk

20 6. An example Predctng the amount of damage caused by a fre from the dstance of the fre from the nearest fre staton Fre Damage Data Fre Damage Dstance from Staton (Thousands (Mles) of Dollars) Dstance from Staton (Mles) Fre Damage (Thousands of Dollars) Always plot the data frst!!! 4. dollars mles A. Karpnsk

21 In SPSS, we use the Regresson command to obtan a regresson analyss REGRESSION /DEPENDENT dollars /METHOD=ENTER mles. Varables Entered/Removed b Model Varables Entered Varables Removed Method MILES a. Enter a. All requested varables entered. b. Dependent Varable: DOLLARS Ths box tells us that MILES was entered as the only predctor Ths box gves us measures of the varance accounted for by the model Model Summary Model R R Square Adjusted R Square Std. Error of the Estmate.96 a a. Predctors: (Constant), MILES MSE ANOVA b Here s our old frend the ANOVA table Model Sum of Squares df Mean Square F Sg. Regresson a Resdual Total a. Predctors: (Constant), MILES b. Dependent Varable: DOLLARS Unstandardzed Coeffcents Coeffcents a Standardzed Coeffcents Model B Std. Error Beta t Sg. (Constant) MILES a. Dependent Varable: DOLLARS These are the tests of the ntercept and the slope 4- A. Karpnsk

22 o From ths table, we read that bˆ =. 78 and that bˆ = Usng ths nformaton we can wrte the regresson equaton Y ˆ = * o To test the slope: H H : b : b = We fnd a sgnfcant lnear relatonshp between the dstance from the fre, and the amount of damage caused by the fre, t ( 3) =.53, p <.. For every mle from the fre staton, the fre caused an addtonal $4,99 n damage o Note that the t-test for ˆβ s dentcal to the Regresson test on the ANOVA table because we only have one predctor n ths case. o In ths case, the test of the ntercept s not meanngful You can also easly obtan 95% confdence ntervals around the parameter estmates REGRESSION /STATISTICS coeff r anova c /DEPENDENT dollars /METHOD=ENTER mles. o COEFF, R and ANOVA are defaults COEFF prnts the estmates of b and b R prnts R and Adjusted R ANOVA prnts the regresson ANOVA table 4- A. Karpnsk

23 o Addng CI to the STATISTICS command wll prnt the confdence ntervals for all model parameters Model Unstandardzed Coeffcents Coeffcents a Standard zed Coeffcen ts 95% Confdence Interval for B B Std. Error Beta t Sg. Lower Bound Upper Bound (Constant) MILES a. Dependent Varable: DOLLARS b ˆ = 4.99 t ( 3) =.53, p <. 95% CI for ˆb : bˆ * ( ˆ ) ± tα /, df std.error b (.393) (4.7, 5.77) 7. Estmaton and predcton One of the goals of regresson analyss s to allow us to estmate or predct new values of Y based on observed values. There are two knds of Y values we may want to predct o Case I: We may want to estmate the mean value of Y, Ŷ, for a specfc value of In ths case, we are attemptng to estmate the mean result of many events at a sngle value of For example, what s the average damage caused by (all) fres that are 5.8 mles from a fre staton? 4-3 A. Karpnsk

24 o Case II: We may also want to predct a partcular value of Y, Yˆ, for a specfc value of In ths case, we are attemptng to predct the outcome of a sngle event at a sngle value of For example, what would be the predcted damage caused by a (sngle) fre that s 5.8 mles from a fre staton? In ether case, we can use our regresson equaton to obtan an estmated mean value or partcular value of Y Y ˆ = * o For a fre 5.8 mles from a staton, we substtute = 5. 8 nto the regresson equaton Yˆ = * 5.8 Yˆ = 38.8 The dfference n these two uses of the regresson model les n the accuracy (varance) of our estmate of the predcton Case I: Varance of the estmate the mean value of Y, Yˆ, at p o When we attempt to estmate a mean value, there s one source of varablty: the varablty due to the regresson lne We know the equaton of the regresson lne: ˆ = bˆ + bˆ Y Var Yˆ) = Var( bˆ + bˆ ( Skppng a few detals, we arrve at the followng equaton ) ˆ σ Yˆ = Var( Yˆ) = MSE N ( p ) + S 4-4 A. Karpnsk

25 o And thus, the equaton for the confdence nterval of the estmate of the mean value of Y, Yˆ s Yˆ ± t σ ˆ α /, N Yˆ Yˆ ± tα /, N MSE N ( p ) + S Case II: Varance of the predcton of a partcular value of Y, Yˆ, at p o When we attempt to predct a sngle value, there are now two sources of varablty: the varablty due to the regresson lne and varablty of Y around ts mean Predcton lmts f the mean value of Y s at the lower bound Predcton lmts f the mean value of Y s at the upper bound Ŷ Confdence nterval for mean value of Y Predcton nterval for a sngle value of Y o The varance for the predcton nterval of a sngle value must nclude these two forms of varablty ˆ σ Y ˆ ˆ σ = Yˆ σ + ε σ Yˆ ( p ) = MSE + + N S 4-5 A. Karpnsk

26 o And thus, the equaton for the predcton nterval of the estmate of a partcular value of Y, Yˆ s Yˆ ± t σ Yˆ ± tα /, N ˆ α /, N Y ˆ ( p ) MSE + + N S Luckly, we can get SPSS to perform most of the ntermedate calculatons for us, but we need to be sneaky o Add a new lne to the data fle wth a mssng value for Y and = p o Ask SPSS to save the predcted value and the standard error of the predcted value when you run the regresson REGRESSION /MISSING LISTWISE /DEPENDENT dollars /METHOD=ENTER mles /SAVE PRED (pred) SEPRED (sepred). 4-6 A. Karpnsk

27 We wll have two new varables n the data fle PRED Yˆ for the value SEPRED for the value σˆ Yˆ DOLLARS MILES PRED SEPRED For = 5. 8 p Y ˆ = 38.8 ˆ ˆ =. 56 σ Y o Use the formulas to compute the confdence and predcton ntervals To calculate a 95% confdence nterval around the mean value, Ŷ ˆ Y ± t ˆ α /, N σ Yˆ 38.8± t.5, 3 (.56) 38.8± (.6)(.56) ( 36.3, 4.3) To calculate a 95% predcton nterval around the sngle value, Yˆ Yˆ ± t ˆ Y ± t α /,N ˆ α /, N σ Y ˆ ˆ σ ε + ˆ σ Y ˆ 38.8± t.5, (.56) 38.8± (.6)(.589) (3.8, 45.54) 4-7 A. Karpnsk

28 The regresson lne can be used for predcton and estmaton, but not for extrapolaton In other words, the regresson lne s only vald for s wthn the range of the observed s SPSS can be used to graph confdence ntervals and predcton ntervals Confdence Bands Predcton Bands dollars 3. dollars mles mles 8. Standardzed regresson coeffcents To nterpret the slope parameter, we must return to the orgnal scale of the data b = 56 suggests that for every one unt change n the varable, Y changes by 56 unts. Ths dependence on unts can make for dffculty n comparng the effects of on Y across dfferent studes o If one researcher measures self-esteem usng a 7 pont scale and another uses a 4 pont scale, they wll obtan dfferent estmates of b o If one researcher measures length n centmeters and another uses nches, they wll obtan dfferent estmates of b One soluton to ths problem s to use standardzed regresson coeffcents β = b σ σ Y 4-8 A. Karpnsk

29 To understand how to nterpret standardzed regresson coeffcents, t s helpful to see how they can be obtaned drectly o Transform both Y and nto z-scores, z Y and z compute zmles = (mles - 3.8)/.576. compute zdollar = (dollars - 6.4)/ o Regress z Y on z REGRESSION /DEPENDENT zdollar /METHOD=ENTER zmles. Model Unstandardzed Coeffcents (Constant) ZMILES a. Dependent Varable: ZDOLLAR Coeffcents a Standard zed Coeffcen ts B Std. Error Beta t Sg. 4.3E b = β =.96 o Compare ths result to the regresson on the raw data REGRESSION /DEPENDENT dollars /METHOD=ENTER mles. Model (Constant) MILES Unstandardzed Coeffcents a. Dependent Varable: DOLLARS Coeffcents a Standard zed Coeffcen ts B Std. Error Beta t Sg β = A. Karpnsk

30 o To nterpret standardzed beta coeffcents, we need to thnk n terms of z-scores A standard devaton change n (mles), s assocated wth a.96 standard devaton change n Y (dollars) For smple lnear regresson (wth only predctor), β = ry Wth more than predctor, standardzed coeffcents should not be nterpreted as correlatons. It s possble to have standardzed coeffcents greater than. 9. Addtonal concerns and observatons Standard assumptons of regresson analyss ε ~ NID (, σ ) o All observatons are ndependent and randomly selected from the populaton (or equvalently, the resdual terms, ε s, are ndependent) o The resduals are normally dstrbuted at each level of o The varance of the resduals s constant across all levels of Addtonally, we assume that the regresson model s a sutable proxy for the correct (but unknown) model: o The relatonshp between and Y must be lnear o No mportant varables have been omtted from the model o No outlers or nfluental observatons These assumptons can be examned by lookng at the resduals 4-3 A. Karpnsk

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

Introduction to Regression

Introduction to Regression Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε Chapter 3 Secton 3.1 Model Assumptons: Multple Regresson Model Predcton Equaton Std. Devaton of Error Correlaton Matrx Smple Lnear Regresson: 1.) Lnearty.) Constant Varance 3.) Independent Errors 4.) Normalty

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Statistics for Business and Economics

Statistics for Business and Economics Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear

More information

Statistics MINITAB - Lab 2

Statistics MINITAB - Lab 2 Statstcs 20080 MINITAB - Lab 2 1. Smple Lnear Regresson In smple lnear regresson we attempt to model a lnear relatonshp between two varables wth a straght lne and make statstcal nferences concernng that

More information

Chapter 15 - Multiple Regression

Chapter 15 - Multiple Regression Chapter - Multple Regresson Chapter - Multple Regresson Multple Regresson Model The equaton that descrbes how the dependent varable y s related to the ndependent varables x, x,... x p and an error term

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an

More information

Learning Objectives for Chapter 11

Learning Objectives for Chapter 11 Chapter : Lnear Regresson and Correlaton Methods Hldebrand, Ott and Gray Basc Statstcal Ideas for Managers Second Edton Learnng Objectves for Chapter Usng the scatterplot n regresson analyss Usng the method

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students. PPOL 59-3 Problem Set Exercses n Smple Regresson Due n class /8/7 In ths problem set, you are asked to compute varous statstcs by hand to gve you a better sense of the mechancs of the Pearson correlaton

More information

Lecture 3 Stat102, Spring 2007

Lecture 3 Stat102, Spring 2007 Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

STATISTICS QUESTIONS. Step by Step Solutions.

STATISTICS QUESTIONS. Step by Step Solutions. STATISTICS QUESTIONS Step by Step Solutons www.mathcracker.com 9//016 Problem 1: A researcher s nterested n the effects of famly sze on delnquency for a group of offenders and examnes famles wth one to

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

Correlation and Regression

Correlation and Regression Correlaton and Regresson otes prepared by Pamela Peterson Drake Index Basc terms and concepts... Smple regresson...5 Multple Regresson...3 Regresson termnology...0 Regresson formulas... Basc terms and

More information

The Ordinary Least Squares (OLS) Estimator

The Ordinary Least Squares (OLS) Estimator The Ordnary Least Squares (OLS) Estmator 1 Regresson Analyss Regresson Analyss: a statstcal technque for nvestgatng and modelng the relatonshp between varables. Applcatons: Engneerng, the physcal and chemcal

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 1 Chapters 14, 15 & 16 Professor Ahmad, Ph.D. Department of Management Revsed August 005 Chapter 14 Formulas Smple Lnear Regresson Model: y =

More information

18. SIMPLE LINEAR REGRESSION III

18. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III US Domestc Beers: Calores vs. % Alcohol Ftted Values and Resduals To each observed x, there corresponds a y-value on the ftted lne, y ˆ ˆ = α + x. The are called ftted values.

More information

17 - LINEAR REGRESSION II

17 - LINEAR REGRESSION II Topc 7 Lnear Regresson II 7- Topc 7 - LINEAR REGRESSION II Testng and Estmaton Inferences about β Recall that we estmate Yˆ ˆ β + ˆ βx. 0 μ Y X x β0 + βx usng To estmate σ σ squared error Y X x ε s ε we

More information

28. SIMPLE LINEAR REGRESSION III

28. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

β0 + β1xi. You are interested in estimating the unknown parameters β

β0 + β1xi. You are interested in estimating the unknown parameters β Ordnary Least Squares (OLS): Smple Lnear Regresson (SLR) Analytcs The SLR Setup Sample Statstcs Ordnary Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals) wth OLS

More information

Chapter 14 Simple Linear Regression

Chapter 14 Simple Linear Regression Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng

More information

Introduction to Analysis of Variance (ANOVA) Part 1

Introduction to Analysis of Variance (ANOVA) Part 1 Introducton to Analss of Varance (ANOVA) Part 1 Sngle factor The logc of Analss of Varance Is the varance explaned b the model >> than the resdual varance In regresson models Varance explaned b regresson

More information

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott

More information

Reduced slides. Introduction to Analysis of Variance (ANOVA) Part 1. Single factor

Reduced slides. Introduction to Analysis of Variance (ANOVA) Part 1. Single factor Reduced sldes Introducton to Analss of Varance (ANOVA) Part 1 Sngle factor 1 The logc of Analss of Varance Is the varance explaned b the model >> than the resdual varance In regresson models Varance explaned

More information

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9 Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Reminder: Nested models. Lecture 9: Interactions, Quadratic terms and Splines. Effect Modification. Model 1

Reminder: Nested models. Lecture 9: Interactions, Quadratic terms and Splines. Effect Modification. Model 1 Lecture 9: Interactons, Quadratc terms and Splnes An Manchakul amancha@jhsph.edu 3 Aprl 7 Remnder: Nested models Parent model contans one set of varables Extended model adds one or more new varables to

More information

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction ECONOMICS 35* -- NOTE 7 ECON 35* -- NOTE 7 Interval Estmaton n the Classcal Normal Lnear Regresson Model Ths note outlnes the basc elements of nterval estmaton n the Classcal Normal Lnear Regresson Model

More information

Statistics II Final Exam 26/6/18

Statistics II Final Exam 26/6/18 Statstcs II Fnal Exam 26/6/18 Academc Year 2017/18 Solutons Exam duraton: 2 h 30 mn 1. (3 ponts) A town hall s conductng a study to determne the amount of leftover food produced by the restaurants n the

More information

Linear Regression Analysis: Terminology and Notation

Linear Regression Analysis: Terminology and Notation ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented

More information

Activity #13: Simple Linear Regression. actgpa.sav; beer.sav;

Activity #13: Simple Linear Regression. actgpa.sav; beer.sav; ctvty #3: Smple Lnear Regresson Resources: actgpa.sav; beer.sav; http://mathworld.wolfram.com/leastfttng.html In the last actvty, we learned how to quantfy the strength of the lnear relatonshp between

More information

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting.

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting. The Practce of Statstcs, nd ed. Chapter 14 Inference for Regresson Introducton In chapter 3 we used a least-squares regresson lne (LSRL) to represent a lnear relatonshp etween two quanttatve explanator

More information

Sociology 301. Bivariate Regression. Clarification. Regression. Liying Luo Last exam (Exam #4) is on May 17, in class.

Sociology 301. Bivariate Regression. Clarification. Regression. Liying Luo Last exam (Exam #4) is on May 17, in class. Socology 30 Bvarate Regresson Lyng Luo 04.28 Clarfcaton Last exam (Exam #4) s on May 7, n class. No exam n the fnal exam week (May 24). Regresson Regresson Analyss: the procedure for esjmajng and tesjng

More information

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors Multple Lnear and Polynomal Regresson wth Statstcal Analyss Gven a set of data of measured (or observed) values of a dependent varable: y versus n ndependent varables x 1, x, x n, multple lnear regresson

More information

Introduction to Dummy Variable Regressors. 1. An Example of Dummy Variable Regressors

Introduction to Dummy Variable Regressors. 1. An Example of Dummy Variable Regressors ECONOMICS 5* -- Introducton to Dummy Varable Regressors ECON 5* -- Introducton to NOTE Introducton to Dummy Varable Regressors. An Example of Dummy Varable Regressors A model of North Amercan car prces

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

Topic 7: Analysis of Variance

Topic 7: Analysis of Variance Topc 7: Analyss of Varance Outlne Parttonng sums of squares Breakdown the degrees of freedom Expected mean squares (EMS) F test ANOVA table General lnear test Pearson Correlaton / R 2 Analyss of Varance

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Biostatistics 360 F&t Tests and Intervals in Regression 1

Biostatistics 360 F&t Tests and Intervals in Regression 1 Bostatstcs 360 F&t Tests and Intervals n Regresson ORIGIN Model: Y = X + Corrected Sums of Squares: X X bar where: s the y ntercept of the regresson lne (translaton) s the slope of the regresson lne (scalng

More information

β0 + β1xi. You are interested in estimating the unknown parameters β

β0 + β1xi. You are interested in estimating the unknown parameters β Revsed: v3 Ordnar Least Squares (OLS): Smple Lnear Regresson (SLR) Analtcs The SLR Setup Sample Statstcs Ordnar Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals)

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 14 Multiple Regression Models

Statistics for Managers Using Microsoft Excel/SPSS Chapter 14 Multiple Regression Models Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 14 Multple Regresson Models 1999 Prentce-Hall, Inc. Chap. 14-1 Chapter Topcs The Multple Regresson Model Contrbuton of Indvdual Independent Varables

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

SIMPLE LINEAR REGRESSION

SIMPLE LINEAR REGRESSION Smple Lnear Regresson and Correlaton Introducton Prevousl, our attenton has been focused on one varable whch we desgnated b x. Frequentl, t s desrable to learn somethng about the relatonshp between two

More information

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis Resource Allocaton and Decson Analss (ECON 800) Sprng 04 Foundatons of Regresson Analss Readng: Regresson Analss (ECON 800 Coursepak, Page 3) Defntons and Concepts: Regresson Analss statstcal technques

More information

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION 014-015 MTH35/MH3510 Regresson Analyss December 014 TIME ALLOWED: HOURS INSTRUCTIONS TO CANDIDATES 1. Ths examnaton paper contans FOUR (4) questons

More information

Properties of Least Squares

Properties of Least Squares Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

January Examinations 2015

January Examinations 2015 24/5 Canddates Only January Examnatons 25 DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR STUDENT CANDIDATE NO.. Department Module Code Module Ttle Exam Duraton (n words)

More information

PubH 7405: REGRESSION ANALYSIS. SLR: INFERENCES, Part II

PubH 7405: REGRESSION ANALYSIS. SLR: INFERENCES, Part II PubH 7405: REGRESSION ANALSIS SLR: INFERENCES, Part II We cover te topc of nference n two sessons; te frst sesson focused on nferences concernng te slope and te ntercept; ts s a contnuaton on estmatng

More information

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

Chapter 3. Two-Variable Regression Model: The Problem of Estimation Chapter 3. Two-Varable Regresson Model: The Problem of Estmaton Ordnary Least Squares Method (OLS) Recall that, PRF: Y = β 1 + β X + u Thus, snce PRF s not drectly observable, t s estmated by SRF; that

More information

β0 + β1xi and want to estimate the unknown

β0 + β1xi and want to estimate the unknown SLR Models Estmaton Those OLS Estmates Estmators (e ante) v. estmates (e post) The Smple Lnear Regresson (SLR) Condtons -4 An Asde: The Populaton Regresson Functon B and B are Lnear Estmators (condtonal

More information

a. (All your answers should be in the letter!

a. (All your answers should be in the letter! Econ 301 Blkent Unversty Taskn Econometrcs Department of Economcs Md Term Exam I November 8, 015 Name For each hypothess testng n the exam complete the followng steps: Indcate the test statstc, ts crtcal

More information

Lecture 2: Prelude to the big shrink

Lecture 2: Prelude to the big shrink Lecture 2: Prelude to the bg shrnk Last tme A slght detour wth vsualzaton tools (hey, t was the frst day... why not start out wth somethng pretty to look at?) Then, we consdered a smple 120a-style regresson

More information

experimenteel en correlationeel onderzoek

experimenteel en correlationeel onderzoek expermenteel en correlatoneel onderzoek lecture 6: one-way analyss of varance Leary. Introducton to Behavoral Research Methods. pages 246 271 (chapters 10 and 11): conceptual statstcs Moore, McCabe, and

More information

Chapter 8 Multivariate Regression Analysis

Chapter 8 Multivariate Regression Analysis Chapter 8 Multvarate Regresson Analyss 8.3 Multple Regresson wth K Independent Varables 8.4 Sgnfcance tests of Parameters Populaton Regresson Model For K ndependent varables, the populaton regresson and

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

Lab 4: Two-level Random Intercept Model

Lab 4: Two-level Random Intercept Model BIO 656 Lab4 009 Lab 4: Two-level Random Intercept Model Data: Peak expratory flow rate (pefr) measured twce, usng two dfferent nstruments, for 17 subjects. (from Chapter 1 of Multlevel and Longtudnal

More information

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X). 11.4.1 Estmaton of Multple Regresson Coeffcents In multple lnear regresson, we essentally solve n equatons for the p unnown parameters. hus n must e equal to or greater than p and n practce n should e

More information

ANOVA. The Observations y ij

ANOVA. The Observations y ij ANOVA Stands for ANalyss Of VArance But t s a test of dfferences n means The dea: The Observatons y j Treatment group = 1 = 2 = k y 11 y 21 y k,1 y 12 y 22 y k,2 y 1, n1 y 2, n2 y k, nk means: m 1 m 2

More information

T E C O L O T E R E S E A R C H, I N C.

T E C O L O T E R E S E A R C H, I N C. T E C O L O T E R E S E A R C H, I N C. B rdg n g En g neern g a nd Econo mcs S nce 1973 THE MINIMUM-UNBIASED-PERCENTAGE ERROR (MUPE) METHOD IN CER DEVELOPMENT Thrd Jont Annual ISPA/SCEA Internatonal Conference

More information

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva Econ 39 - Statstcal Propertes of the OLS estmator Sanjaya DeSlva September, 008 1 Overvew Recall that the true regresson model s Y = β 0 + β 1 X + u (1) Applyng the OLS method to a sample of data, we estmate

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Diagnostics in Poisson Regression. Models - Residual Analysis

Diagnostics in Poisson Regression. Models - Residual Analysis Dagnostcs n Posson Regresson Models - Resdual Analyss 1 Outlne Dagnostcs n Posson Regresson Models - Resdual Analyss Example 3: Recall of Stressful Events contnued 2 Resdual Analyss Resduals represent

More information

This column is a continuation of our previous column

This column is a continuation of our previous column Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard

More information

Laboratory 3: Method of Least Squares

Laboratory 3: Method of Least Squares Laboratory 3: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly they are correlated wth

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3. Outlne 3. Multple Regresson Analyss: Estmaton I. Motvaton II. Mechancs and Interpretaton of OLS Read Wooldrdge (013), Chapter 3. III. Expected Values of the OLS IV. Varances of the OLS V. The Gauss Markov

More information

Interpreting Slope Coefficients in Multiple Linear Regression Models: An Example

Interpreting Slope Coefficients in Multiple Linear Regression Models: An Example CONOMICS 5* -- Introducton to NOT CON 5* -- Introducton to NOT : Multple Lnear Regresson Models Interpretng Slope Coeffcents n Multple Lnear Regresson Models: An xample Consder the followng smple lnear

More information

Topic 23 - Randomized Complete Block Designs (RCBD)

Topic 23 - Randomized Complete Block Designs (RCBD) Topc 3 ANOVA (III) 3-1 Topc 3 - Randomzed Complete Block Desgns (RCBD) Defn: A Randomzed Complete Block Desgn s a varant of the completely randomzed desgn (CRD) that we recently learned. In ths desgn,

More information

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5).

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5). (out of 15 ponts) STAT 3340 Assgnment 1 solutons (10) (10) 1. Fnd the equaton of the lne whch passes through the ponts (1,1) and (4,5). β 1 = (5 1)/(4 1) = 4/3 equaton for the lne s y y 0 = β 1 (x x 0

More information

JAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger

JAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger JAB Chan Long-tal clams development ASTIN - September 2005 B.Verder A. Klnger Outlne Chan Ladder : comments A frst soluton: Munch Chan Ladder JAB Chan Chan Ladder: Comments Black lne: average pad to ncurred

More information

Regression Analysis. Regression Analysis

Regression Analysis. Regression Analysis Regresson Analyss Smple Regresson Multvarate Regresson Stepwse Regresson Replcaton and Predcton Error 1 Regresson Analyss In general, we "ft" a model by mnmzng a metrc that represents the error. n mn (y

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VIII LECTURE - 34 ANALYSIS OF VARIANCE IN RANDOM-EFFECTS MODEL AND MIXED-EFFECTS EFFECTS MODEL Dr Shalabh Department of Mathematcs and Statstcs Indan

More information

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y) Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,

More information

Topic- 11 The Analysis of Variance

Topic- 11 The Analysis of Variance Topc- 11 The Analyss of Varance Expermental Desgn The samplng plan or expermental desgn determnes the way that a sample s selected. In an observatonal study, the expermenter observes data that already

More information

Statistics Chapter 4

Statistics Chapter 4 Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment

More information

Chapter 3 Describing Data Using Numerical Measures

Chapter 3 Describing Data Using Numerical Measures Chapter 3 Student Lecture Notes 3-1 Chapter 3 Descrbng Data Usng Numercal Measures Fall 2006 Fundamentals of Busness Statstcs 1 Chapter Goals To establsh the usefulness of summary measures of data. The

More information

Chapter 15 Student Lecture Notes 15-1

Chapter 15 Student Lecture Notes 15-1 Chapter 15 Student Lecture Notes 15-1 Basc Busness Statstcs (9 th Edton) Chapter 15 Multple Regresson Model Buldng 004 Prentce-Hall, Inc. Chap 15-1 Chapter Topcs The Quadratc Regresson Model Usng Transformatons

More information