WORKSHOP ON THE ANALYSIS OF NEURAL DATA 2001 MARINE BIOLOGICAL LABORATORY WOODS HOLE, MASSACHUSETTS

Size: px
Start display at page:

Download "WORKSHOP ON THE ANALYSIS OF NEURAL DATA 2001 MARINE BIOLOGICAL LABORATORY WOODS HOLE, MASSACHUSETTS"

Transcription

1 Emery. Brown, M.D., Ph.D. WORKSHOP O THE AALYSIS OF EURAL DATA 00 MARIE BIOLOGICAL LABORATORY WOODS HOLE, MASSACHUSETTS A REVIEW OF STATISTICS PART 3: REGRESSIO AALYSIS AD THE GEERALIZED LIEAR MODEL EMERY. BROW EUROSCIECE STATISTICS RESEARCH LABORATORY DEPARTMET OF AESTHESIA AD CRITICAL CARE MASSACHUSETTS GEERAL HOSPITAL DIVISIO OF HEALTH SCIECES AD TECHOLOGY HARVARD MEDICAL SCHOOL / MIT

2 Regresson Analyss A. Smple Regresson B. Model Assumptons C. Model Fttng D. Propertes of Parameter Estmates E. Model Goodness-of-Ft F-test R Analyss of Resduals F. The Geometry of Regresson (Method of Least-Squares) Page

3 A. Smple Regresson Assume we have a data consstng of pars of two varables and we denote them as ( x, y ), ( x, y ),, ( xn, y n). For example, x and y mght be measurements of heght and weght of a set of ndvduals from a well-defned cohort. Let s assume that there s a lnear relaton between x and y. We assume that the lnear relaton may be wrtten as y = α + β x Example. For example let us consder ths example taken from Draper and Smth (98), Appled Regresson Analyss Fgure. Relaton between Monthly Steam Producton and Mean Atmospherc Temperature. The varable y s the amount of steam produced per month n a plant and varable x s the mean atmospherc temperature. There s an obvous negatve relaton. Page 3

4 B. Model Assumptons We assume ) E[ y x] = α + β x ) The x s are fxed non-random covarates ) The y s are ndependent Gaussan random varables wth mean α + β x and varance σ C. Model Fttng Our objectve s to estmate the parameters α, β and σ. Because y s assumed to have a Gaussan dstrbuton condtonal on x, a logcal approach s to use maxmum lkelhood estmaton. For these data the jont probablty densty (lkelhood) s (,,, ) f y α β σ x = ( α β, σ ) f y + x = πσ exp ( y α βx ) σ The log lkelhood s logf ( y x, α, β, σ ) log( πσ ) = ( y α βx) σ Page 4

5 Dfferentatng wth respect to α and β yelds ( y x α β σ ) log f,,, α ( y x α β σ ) log f,,, β ( y α βx ) = ( y α β ) = Settng the dervatve equal to zero yelds the normal equatons x x α α + β x = y x + βx = xy Or n matrx form x y α β = x x xy x y α β = x x xy = Page 5

6 The soluton for β and α are We may wrte any estmate as xy x / x x y y βˆ = = x x / αˆ = y β x ( )( ) ( x x) ( ) yˆ = αˆ + βˆx = y βˆx + βˆx = y+ βˆ x x If we go back and dfferentate the log lkelhood wth respect to σ lkelhood estmate Remarks ( ˆ ) σˆ = y αˆ βx /, we obtan the maxmum. Choosng α and β by the maxmum lkelhood s equvalent to the method of least square n ths case. By the method of least squares, we mnmze the sum of the squared ntal devatons of the data from the regresson lne.. The estmate of ˆα shows that every regresson lne goes through the pont ( x, y. ) 3. The resduals are y yˆ and are the components n the data whch the model does not explan. We note that ˆ ( ) y = y+ β x x ( y y ) ( y y) β ( x x) ˆ = + = 0 Page 6

7 4. The Pythagorean Relaton ( ) ( ˆ ) y yˆ = y y y y ( y yˆ ) = {( y y) ( yˆ y) } ( y y) ( y y) ( y y)( yˆ y) = +.B. ( y y) β( x x) β ( y y)( x x) = ( x x) = β ( yˆ ) y, = and ( y yˆ) = ( y y) ( yˆ y). Or ( ) ( ) y ˆ ( ˆ ) y = y y + y y Total sum of squares (TSS) = Explaned sum of squares (ESS) + Resdual sum of squares (RSS) Page 7

8 5. Correlaton and Regresson ote that the correlaton coeffcent for x and y s r xy = ( x x)( y y) ( x x) ( y y) and recall that β Hence, β = ( x x)( y y) ( x x) ( y y) = ( x x ) r xy The regresson coeffcent s a scale verson of the correlaton coeffcent. D. Propertes of the Parameter Estmates The varances of the parameter estmates Var ( βˆ ) = σ ( x x) Page 8

9 Var ( αˆ ) = = σ x ( x x) If we estmate σ by σˆ then ρ confdence ntervals for the parameters based on the t- dstrbuton are t βˆ ± n, ρ / σˆ ( x x) x αˆ ± tn, ρ / ( x x ) σˆ We also nvert the above statstcs to test hypotheses about the regresson coeffcents by constructng a t-test. To construct a confdence nterval for a predcted y k value at a gven x value, x k, we note that the var ( y ) k σ = + ( x x) k and hence an approxmate ρ confdence nterval s ( x x) σ Example. (contnued) yˆ ( x x) k k ± tn, ρ / + ˆ σ ( x x) Page 9

10 Fgure. Ft and Confdence Intervals for a Smple Lnear Regresson Model E. Model Goodness-of-Ft A crucal (f not the most crucal) step n statstcal analyses of data s measurng goodness of ft. That s, how well does the model agree wth the data. We dscuss three graphcal and two statstcal measures of goodness-of-ft.. F-test Gven the null hypothess H0 : β = 0,.e. that there s no lnear relaton n x and y, we can test ths explctly usng an F-test defned as F /( p ) ( ) ESS RSS / p y yˆ ( y y) ( p ) p, p = = ( ) ( p ) Page 0

11 where ESS s the explaned sum of squares; RSS s the resdual sum. We reject ths null hypothess for large values of the F statstc. Ths suggests that f the amount of the varance n the data that the regresson explans s large relatve to the amount whch s unexplaned then we reject the null hypothess.. R-Squared Another measure of goodness-of-ft s the R. It s defned as R ESS TSS = = ( yˆ y) ( y y) The R measures the fracton of varance n the data explaned by the regresson equaton. The R-Squared and the F-statstc are related as ( ) ( ) ( ) ( ) F p / p R = + F p / p Clearly, as the F-statstc ncreases, so does the R. Example. Data Analyss Summary (Contnued) Parameter Estmate Standard Error t-statstc ˆα = ˆβ = TSS = 63.8 ESS = RSS = 7.3 F-statstc (, 3) = 57.5, R = Factod: The square of the t-statstc (n-p degrees of freedom) s the F-statstc (, n-p). 3. Graphcal Measures of Goodness-of-Ft. Plot the Raw Data Is the relaton lnear?. Plot the Ft vs. the ResdualsIs there a relaton between the explaned and the unexplaned structure n the data? 3. Plot the Resduals Lack of ft of the model Page

12 Fgure 3. Plot of Resduals versus x. Page

13 Fgure 4. Plot of the Resduals aganst the Predcted y values. F. The Geometry of Regresson Analyss (Method of Least-Squares) Regresson analyss has an ntutvely appealng geometrc nterpretaton. z 3 y ε z z ŷ x x Page 3

14 R = ŷ y F statstc ŷ ε I. Dervaton and Propertes of the Least Squares Estmates The procedure used to estmate the parameters n the lnear regresson models n ths study s least squares. In ths secton we frst derve the least squares estmates, and then by makng some dstrbutal assumptons about our model, we next derve the statstcal propertes of the estmates. We begn by assumng a model of the form Y = Xβ + ε () where Y s an n l vector of dependent observatons. X s an n p matrx of explanatory observatons,.e. the matrx contans p n l dmensonal vectors. β s an n l vector of parameters. ε s an n l vector of errors. In order to estmate our vector of parameters β usng least squares, we assume that X s of rank p or, n other words, that the p column vectors n X are lnearly ndependent. Therefore, we want to fnd the estmates of β, ˆβ, whch mnmzes ˆ ε ' ˆ ε = ( Y Xβˆ)'( Y Xβˆ) = YY ' βˆ' XY ' Y' Xβˆ + βˆ' X' Xβˆ (.) Because ˆ ' β X ' Y s a scalar t equals ts transpose Y ' Xβˆ. Consequently (.) smplfes to ˆ ε' ˆ ε = YY ' βˆ' XY ' + βˆ' Xβˆ To fnd the least squares estmates of β we dfferentate (.3) wth respect to ˆβ and set the dervatve equal to zero. (.3) Page 4

15 ( ˆ ε' ˆ ε) = X ' Y + ( X ' X) βˆ βˆ (.4) Solvng for ˆβ gves That βˆ = ( X ' X ) X ' Y ( βˆ' X ' Xβˆ) βˆ = ( X ' X) βˆ (.5) follows from the fact that βˆ' ( X ' X) βˆ s a quadratc form. Furthermore, we know that ( X ' X) has an nverse snce we ntally assumed X had rank p. Hence, ( X ' X) s postve defnte of rank p and has nverse ( X ' X). The second order condtons for the mnmzaton of ε' ε follows drectly snce ( X ' X) s postve defnte. In order to test hypotheses about our data we need to make some dstrbutonal assumptons about our model. The standard assumptons for the lnear regresson model are: I. ε ~ ( 0, Iσ ) common varance σ.. The ε s are ndependent, normally dstrbuted random errors wth mean 0 and II. The pn-dmensonal vectors comprsng X are taken as fxed constants. From these two assumptons we see that Y beng a lnear functon of ε, s also normally dstrbuted wth mean X β and varance Iσ. Usng these dstrbutonal propertes we now derve the dstrbutons of the least squares estmates. Proposton. The least squares estmate of varance σ ( X ' X). β, βˆ, s normally dstrbuted wth means β and Proof: To show that β s normally dstrbuted we have only to notce that β s a lnear functon of Y, snce lnear functons of ndependent normally dstrbuted random varables are also normally dstrbuted. To fnd the mean of the dstrbuton of β consder Page 5

16 { } ( ) ( ) ( ) ( ) ( ) E βˆ = E X ' X X ' Y = X ' X X ' Xβ + X ' X X ' E ε (.6) But from Assumpton I, we know that E( ε ) = 0, and snce ( X ' X) X ' X = I, t follows that E ( βˆ ) = β. To determne the varance of ˆβ, we consder { " } Var( βˆ ) = Var ( X ' X ) X ' Y = ( X ' X) X" Var( YY ') X( X ' X) But Var( YY ') = σ I, so we can rewrte (7) as βˆ = σ Var( ) ( X ' X ) ( X ' X )( X ' X ) I = ( X ' X) σ I Proposton. The least squares estmate of Y, Yˆ, where Yˆ = Xβˆ s ( Xβσ, X( X ' X) X ') Proof: the fact that Ŷ has a normal dstrbuton follows drectly from the fact that t s a lnear combnaton of Y. We fnd the mean of Ŷ by consderng { βˆ } { β ε } E( Yˆ ) = E X = E X( X ' X) X ' X + X( X ' X) X ' E( ) Because ( X ' X) ( X ' X) = I and E( ε) = 0, we get EY ( ˆ ) = Xβˆ. (.7) (.8). (.9). To fnd the varance of Ŷ we wrte Var( Yˆ ) = Var{ X ( X ' X ) X ' Y} = X( X ' X) X ' Var( YY ') X( X ' X) X ' (.0) Snce we know that Var( YY ') = σ I, substtutng ths nto (.0) and rearrangng terms gves Page 6

17 Var( Yˆ ) = X ( X ' X ) ( X ' X )( X ' X ) X ' σ I = X( X ' X) X ' σ I (.) Proposton.3 The least squares estmate of ε, ˆ ε, where ˆ ε = Y Yˆ s (0,( I H) σ ), where H = X( X ' X) X '. To complete the proof of ths Proposton, we need the fact that the matrx I H s symmetrc and dempotent. A matrx A s sad to be dempotent s A = A and A s symmetrc f A' = A. Proof: The fact that ˆε has a normal dstrbuton follows drectly snce t s a lnear functon of Y. In order to fnd the mean of ˆε to consder { ˆ } { } E( ˆ ε ) = E Y Y = E( I X( X ' X) X ' Y) (.) We recall that EY ( ) = Xβ. Therefore, (.) smplfes to E( ˆ ε) = X β X( X ' X" ) X ' Xβ = Xβ Xβ = 0 (.3) We fnd the varance of ˆε by frst rewrtng ˆε as Hence, But I H ˆ ε = Y Yˆ = ( I H) Y Var( ˆ ε ) = ( I H ) Var( YY ')( I H )' s symmetrc, and Var( YY ') = σ I, so from (.5) we get (.4) (.5) Var( ε) = ( I H)( I H ) σ I. (.6) We recall that I H s dempotent and fnally we have Var( ε) = ( I H ) σ We notce that all the estmates just derved are unbased n that the expected values of ther dstrbutons are the parameters they estmate. ow we can fnd an unbased estmate of σ. In the next Proposton we demonstrate that such an estmate s σ, where Page 7

18 Proposton.4 σˆ YY ' βˆ XY ' = ( n p) s an unbased estmate of σ. In order to complete the proof of ths Proposton we need the followng two results from the theory of matrx algebra, and the dstrbuton theory of quadratc forms: For two conformng matrces A and B trace ( AB ) = trace ( BA). If A s a symmetrc matrx and e s a vector of uncorrelated random varables wth common varance σ and mean 0, Proof: We consder frst EeAe ( ' ) = σ trace ( A ). E ( σˆ ) { ' βˆ ' X ' Y ' } E Y Y = ( n p) (.7) Substtutng for β ˆ ' gves E ( σˆ ) { ' ' ( ' ) } E Y Y Y X X X XY = (.8) ( n p) If we now substtute for Y ' and Y, (.8) becomes E( σˆ {( β ' ' + ε ')( β + ε) ( β ' ' + ε ') ( ' ) ( β + ε} E X X X X X X X X 0 = n p (.9) Expandng (.9), takng expectatons and smplfyng yelds { ε' ε ε' ( ' " ) ' ε} E I X X X X E( σ ) = ( n p) (.0) Page 8

19 In equaton (.0) we have the expectaton of two quadratc forms. Applyng the frst of two results stated ntally, we get E( σˆ ) = { σ trace( I) σ trace ( X( X ' X) X ')} (.) ( n p) Snce I s an nxn dentty matrx ts trace s n. Applyng the second of the two results, we stated, we fnd that trace ( X( X ' X) X ') = trace ( X ' X( X ' X) ). (.) Because ( X ' X) and ( X ' X) are pxp, trace ( X ' X( X ' X) ) Substtutng back nto (.) gves = p. σ n σ p ˆ E( σ ) = = σ n p (.3) Havng found the least squares estmates for our lnear regresson model and derved ther statstcal propertes, we now would lke to demonstrate an mportant optmalty property assocated wth our estmate ˆβ. Specfcally, we want to show that ˆβ s the best lnear unbased estmate of β. That s, n the class of lnear unbased estmates of β, βˆ has the smallest varance-covarance matrx. Ths well-known property of the least squares estmates s known as the Gauss-Markov Proposton. It s stated and proved below as Proposton.5. Proposton.5 In the class of lnear unbased estmates of β, the least squares estmate ˆβ the smallest varance. has Proof: Consder an alternatve lnear estmate of β, and call t β *, where β * = AY Let A= ( XX) X + C. We now want to determne C such that β * s unbased wth the smallest varances of all lnear estmates. Therefore, we consder E( β ) = AE( Y) = ( X ' X) ( X ' X) β + CXC' = β CXC ' (.4) Page 9

20 In order for β * to be unbased we requre that CX = 0. { } Var( β*) = E ( β * β)( β * β ') ow we want to fnd the varance of β * (.5) Substtutng for β * n (.5) gves { } ( ) { } Var( β ) = E ( X ' X ) X C ( X β ε ) β X ' X X C X β ε β ' (.6) Expandng terms, takng expectatons and smplfyng yelds Var( β ) = ( X ' X) X ' E( εε ') X( X ' X) + Cεε ' C' = σ ( X ' X) + CC' (.7) We now notce that Var( β*) Var( β) 0 (.8) The equalty n (.8) holds f and only f we choose CC ' = 0. Ths mples that the elements of C must all be zero. We have demonstrated that the least squares estmate of β, has the ˆβ smallest varance-covarance matrx among the class of all lnear unbased estmates. We note however that ˆβ s not the only estmate n ths class wth the smallest varance. Moreover, t s possble to construct based lnear estmates or unbased nonlnear estmates whose varance s smaller than ˆβ. By makng smlar arguments for Ŷ and ˆε, we can also prove that they beng lnear functons of ˆβ, they are the b.l.u.e. of Y and ε, respectvely. We now llustrate one fnal mathematcal property of our least squares estmates, whch s helpful for understandng the theory behnd ths method of estmaton. Proposton.6 ˆ ε '( Yˆ ) = 0 To prove ths Proposton, we need the same result for the matrx X( X ' X) X that we stated n the proof of Proposton.3 for the matrx I X( X ' X) X '. That s X( X ' X) X ' s also symmetrc and dempotent. Proof: Agan let X( X ' X) X ' = H and recall from (.4) that ˆ ( ) ε = I H Y. Hence, ( ˆ) ( ) ( ) ˆ ε ' Y = I H Y ' HY = Y ' I H ' HY (.9) Page 0

21 snce Y ˆ = HY. But both I H and H are symmetrc and dempotent, so that ˆ ε '( Y) = Y '( H H) Y = 0 (.30) What ths result says s that the vector of estmated errors s orthogonal to the vector of predcted values of the dependent varable. In other words, regresson decomposes Y nto two orthogonal components Ŷ and ˆε. The predcted values of Y, Ŷ, s the orthogonal projecton of Y on to the subspace determned by X. We llustrate ths geometrcally n Fgure A. for the case when n=3 and p=. We return to ths geometrc nterpretaton of regresson n ths next secton, where we dscuss the F-test. I. The F-Test Our prmary method for testng hypotheses about our regresson models s the F-test. In ths secton, we outlne ts dervaton and gve ts geometrc nterpretaton. Let s wrte β as β β = β where β s a k l column vector (.) β s a ( p k) l column vector If we wsh to test the hypothess that β = 0 we can do so by usng the followng steps to construct the F-test. ) Compute the Regresson Sum of Squares (SSR 0 ) under H 0 that β = 0. ) Subtract (SSR 0 ) from the Regresson Sum of Squares (SSR A ) obtaned under the alternatve hypothess H A β 0. 3) Dvde ths dfference by k, the number of parameters n β. 4) Dvde ths new statstc by the Mean-Squared Error under H A (MSE A ). If the normalty assumptons regardng ε stated n Secton I hold, under H 0 the rato ( ) SSR SSR k A MSE A 0 / (.) has an F-dstrbuton wth k and n-p degrees of freedom. We reject H 0 at the α level of sgnfcance f our observed F-statstc exceeds the ( α ) quantle of the F-dstrbuton wth k, n- Page

22 p degrees of freedom. Although we do not prove here that ths statstc has an F-dstrbuton, we show geometrcally why ths rato s reasonable for the test of ths hypothess. For the geometrc argument, we agan consder the case where n = 3 p =, shown n Fgure A.. Here the null hypothess s that the coeffcent of X s zero, and therefore K =. From Fgure A. we see that ####$ SSR = OA ny A #####$ SSR = OB ny 0 (.3) MSE A #####$ = AC To test the hypothess that β = 0, we fnd that the approprate F statstc s #####$ #####$ F = AB AC, (.4) ###$ We notce that AB s the dfference between the projectons of Y under H A and Y under H 0. Ths geometrc descrpton of the F-test shows that H 0 s rejected f Y s sgnfcantly closer to the subspace spanned by X and X than the one spanned by X alone. G. Generalzed Lnear Model (GLM) Lnear Model Y = Xβ + ε ε (0, σ I) y = µ + ε where µ s the mean of y n X n p β p Components of the Generalzed Lnear Model. Random Component for y s chosen from the exponental famly. Systematc Component are covarates x, x,, xp produce a lnear predctor η gven by p η = x β. j= j j 3. The lnk between the random and systematc components η = g( µ ) where g s monotonc dfferentable functon. Page

23 Revst the Exponental Famly of Dstrbuton Gven y a data vector from a probablty dstrbuton belongng to the exponental famly. We may express the probablty densty or mass functon of y as { } f ( y, θ, φ) = exp yθ b( θ))/ a( φ) + c( y, φ). y a(), b () and c () functons to be specfed f φ s known then we have an exponental famly model wth a canoncal parameter. Example : Gaussan Model { } fy ( y, θφ, ) = exp ( y µ ) /σ πσ = exp ( yµ µ / ) / σ ( y / σ + log( πσ )), where θ µ, φ σ, b( θ) θ /, c( y, φ) { y / σ log(πσ } = = = = +. Example 3: Bnomal Model n Pr( Y = y) = p ( p) ( n y) y y n y p n = exp ylog + nlog( p) + log p ( n y) y n = log( /( )), =, ( ) = log( ), (, ) = log ( n y ) y where θ p p φ b θ n p c y φ Example 4: Posson Model Pr( Y = y) = λ exp( λ) / y y ( y λ λ y ) = exp log log where θ = log λ, φ =, b( θ) = λ, c( y, φ) = log y Mean and Varance of y (We wll skp ths stuff n the lecture; we can revst t n a tutoral). Let '( θ, φ, y) = log f ( y, θ, φ). Two standard propertes of lkelhood functons y Page 3

24 The expected value of the score functon s zero ' E = 0 θ. () ' ' E + E = 0 θ θ. () The expected Fsher Informaton s the negatve of the varance of the score functon '( θ, y) = { y( θ) b( θ) }/ a( φ) + c( y, φ), (3) whence ' = { y b ( θ )}/ a( φ) θ (4) θ ' = b ( θ )/ a( φ). (5) Eqs. () and (4) mply ' 0 = E = { µ b ( θ )}/ a( φ) θ or Ey [ ] = µ = b ( θ ). Usng Eqs. 4 and 5 n Eq., we obtan b ( θ ) Var( y) 0 = +, a φ ( ) a ( φ) Hence Var( y) = b ( θ ) a( φ). The varance depends on the mean and a functon, whch depends on the dsperson parameter; b ( θ ) s the varance functon. Page 4

25 Model Summary Table Gaussan Posson Bnomal Gamma Inverse Gaussan Canoncal µ log λ p log lnk: g ( µ ) p λ λ MAI POIT. WE MAY WRITE EXPLICITLY THE LIK OF THE SYSTEMATIC AD RADOM µ = β. I THIS WAY, WE EXTED THE SIMPLE REGRESSIO MODEL WITH STRUCTURE AS g( ) X GAUSSIA ERRORS TO REGRESSIO MODELS FOR THE ETIRE EXPOETIAL FAMILY OF PROBABILITY DISTRIBUTIOS. Devance (Goodness of Ft Crteron) Let '( µ ; y) = log f( y; θ) and defne the devance as D*( y; µ ) = '( y; y) '( µ ; y) ote that the second expresson s the log lkelhood of the data. Hence mnmzng D*( y ; µ ) s equvalent to maxmzng the log lkelhood. otce that n the Gaussan case, the devance s D*( y; µ ˆ) = ( y µ ˆ ) / σ whch s just the resdual sum of squares (RSS). Analyss of Devance. Instead of analyss of varance we use analyss of devance, whch s addtve for nested sets of models ft by maxmum lkelhood. D*( y ; µ ) s analyzed as an approxmately ch-squared quantty. Devance Resduals. The devance resduals are defned as D*( y; µ ˆ) d = where we defne r d, = sgn( d) d. Plots of these quanttes can be analyzed n the same way the regresson resduals are analyzed. For generalzed lnear models n whch there s a natural tme dependence n the observatons we wll show that a global measure of model goodness of ft may be constructed usng the tmerescalng Proposton (Berman, 983; Ogata, 988; Brown et al., 00). Page 5

26 Remarks. GLM allows generalzaton (extenson) of lnear models to non-gaussan settng.. A unfed computatonal approach. 3. Can use non-canoncal lnk and varance functons. 4. Wdely dstrbuton Splus and non Matlab. 5. ow there s Bayesan GLM. 6. Assess goodness-of-ft wth an analyss of devance. euroscence Data Analyss Examples Usng GLM. Random Threshold Model (Brllnger (988, 99)) Smplfed verson J k x µ jxt j kyt k pt α β x j= k= pt log = + + xt = 0 spke from neuron otherwse xat t yt = 0 spke from neuron otherwse y at t. pt x probablty of a spke at t from neuron x.. Prospectve and Retrospectve Encodng by Hppocampal Place Cells (Frank, Brown and Wlson, 000) A rat runnng on a W-maze, multple sngle unt actvty s recorded from CA regon of the hppocampus and the entorhnal cortex. t ndexes tral P t probablty of gong rght on an outbound journey from the center arm of the W-maze. pt log = µ + α jλj( t) pt J, where λ j () t s the frng rate of the neuron at locaton j pror to the choce pont on pass t. 3. Cockroach Cercal System Response to Wnd Stmul (Davdowtz and Rmberg, 000). n = number of spkes at tme t (dscrete tme bns of 0msec) t j= Page 6

27 vx () t x drecton wnd velocty vy () t y drecton wnd velocty a t x drecton wnd acceleraton a t y drecton wnd acceleraton x () y () Posson lnk functon log λ ( t) = µ + α v ( t) + α v ( t) + β a ( t) + β a ( t), x y x y λ () t spke rate at tme t. References Draper R, Smth H. (98). Appled Regresson Analyss, nd ed., Wley: ew York. McCullagh, P. elder JA (989). Generalzed Lnear Models, nd ed, Chapman and Hall: London. Page 7

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X). 11.4.1 Estmaton of Multple Regresson Coeffcents In multple lnear regresson, we essentally solve n equatons for the p unnown parameters. hus n must e equal to or greater than p and n practce n should e

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of Chapter 7 Generalzed and Weghted Least Squares Estmaton The usual lnear regresson model assumes that all the random error components are dentcally and ndependently dstrbuted wth constant varance. When

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values Fall 007 Soluton to Mdterm Examnaton STAT 7 Dr. Goel. [0 ponts] For the general lnear model = X + ε, wth uncorrelated errors havng mean zero and varance σ, suppose that the desgn matrx X s not necessarly

More information

Diagnostics in Poisson Regression. Models - Residual Analysis

Diagnostics in Poisson Regression. Models - Residual Analysis Dagnostcs n Posson Regresson Models - Resdual Analyss 1 Outlne Dagnostcs n Posson Regresson Models - Resdual Analyss Example 3: Recall of Stressful Events contnued 2 Resdual Analyss Resduals represent

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y) Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction ECONOMICS 35* -- NOTE 7 ECON 35* -- NOTE 7 Interval Estmaton n the Classcal Normal Lnear Regresson Model Ths note outlnes the basc elements of nterval estmaton n the Classcal Normal Lnear Regresson Model

More information

β0 + β1xi. You are interested in estimating the unknown parameters β

β0 + β1xi. You are interested in estimating the unknown parameters β Ordnary Least Squares (OLS): Smple Lnear Regresson (SLR) Analytcs The SLR Setup Sample Statstcs Ordnary Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals) wth OLS

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Learning Objectives for Chapter 11

Learning Objectives for Chapter 11 Chapter : Lnear Regresson and Correlaton Methods Hldebrand, Ott and Gray Basc Statstcal Ideas for Managers Second Edton Learnng Objectves for Chapter Usng the scatterplot n regresson analyss Usng the method

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

Chapter 3. Two-Variable Regression Model: The Problem of Estimation Chapter 3. Two-Varable Regresson Model: The Problem of Estmaton Ordnary Least Squares Method (OLS) Recall that, PRF: Y = β 1 + β X + u Thus, snce PRF s not drectly observable, t s estmated by SRF; that

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Professor Chris Murray. Midterm Exam

Professor Chris Murray. Midterm Exam Econ 7 Econometrcs Sprng 4 Professor Chrs Murray McElhnney D cjmurray@uh.edu Mdterm Exam Wrte your answers on one sde of the blank whte paper that I have gven you.. Do not wrte your answers on ths exam.

More information

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede Fall 0 Analyss of Expermental easurements B. Esensten/rev. S. Errede We now reformulate the lnear Least Squares ethod n more general terms, sutable for (eventually extendng to the non-lnear case, and also

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva Econ 39 - Statstcal Propertes of the OLS estmator Sanjaya DeSlva September, 008 1 Overvew Recall that the true regresson model s Y = β 0 + β 1 X + u (1) Applyng the OLS method to a sample of data, we estmate

More information

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students. PPOL 59-3 Problem Set Exercses n Smple Regresson Due n class /8/7 In ths problem set, you are asked to compute varous statstcs by hand to gve you a better sense of the mechancs of the Pearson correlaton

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson

More information

a. (All your answers should be in the letter!

a. (All your answers should be in the letter! Econ 301 Blkent Unversty Taskn Econometrcs Department of Economcs Md Term Exam I November 8, 015 Name For each hypothess testng n the exam complete the followng steps: Indcate the test statstc, ts crtcal

More information

STAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression

STAT 405 BIOSTATISTICS (Fall 2016) Handout 15 Introduction to Logistic Regression STAT 45 BIOSTATISTICS (Fall 26) Handout 5 Introducton to Logstc Regresson Ths handout covers materal found n Secton 3.7 of your text. You may also want to revew regresson technques n Chapter. In ths handout,

More information

Properties of Least Squares

Properties of Least Squares Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures

More information

Statistics for Business and Economics

Statistics for Business and Economics Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear

More information

The Ordinary Least Squares (OLS) Estimator

The Ordinary Least Squares (OLS) Estimator The Ordnary Least Squares (OLS) Estmator 1 Regresson Analyss Regresson Analyss: a statstcal technque for nvestgatng and modelng the relatonshp between varables. Applcatons: Engneerng, the physcal and chemcal

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear

More information

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9 Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,

More information

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

Lecture 16 Statistical Analysis in Biomaterials Research (Part II) 3.051J/0.340J 1 Lecture 16 Statstcal Analyss n Bomaterals Research (Part II) C. F Dstrbuton Allows comparson of varablty of behavor between populatons usng test of hypothess: σ x = σ x amed for Brtsh statstcan

More information

Rockefeller College University at Albany

Rockefeller College University at Albany Rockefeller College Unverst at Alban PAD 705 Handout: Maxmum Lkelhood Estmaton Orgnal b Davd A. Wse John F. Kenned School of Government, Harvard Unverst Modfcatons b R. Karl Rethemeer Up to ths pont n

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

β0 + β1xi. You are interested in estimating the unknown parameters β

β0 + β1xi. You are interested in estimating the unknown parameters β Revsed: v3 Ordnar Least Squares (OLS): Smple Lnear Regresson (SLR) Analtcs The SLR Setup Sample Statstcs Ordnar Least Squares (OLS): FOCs and SOCs Back to OLS and Sample Statstcs Predctons (and Resduals)

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Linear Regression Analysis: Terminology and Notation

Linear Regression Analysis: Terminology and Notation ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented

More information

17 - LINEAR REGRESSION II

17 - LINEAR REGRESSION II Topc 7 Lnear Regresson II 7- Topc 7 - LINEAR REGRESSION II Testng and Estmaton Inferences about β Recall that we estmate Yˆ ˆ β + ˆ βx. 0 μ Y X x β0 + βx usng To estmate σ σ squared error Y X x ε s ε we

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

Lecture 2: Prelude to the big shrink

Lecture 2: Prelude to the big shrink Lecture 2: Prelude to the bg shrnk Last tme A slght detour wth vsualzaton tools (hey, t was the frst day... why not start out wth somethng pretty to look at?) Then, we consdered a smple 120a-style regresson

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

Chapter 4: Regression With One Regressor

Chapter 4: Regression With One Regressor Chapter 4: Regresson Wth One Regressor Copyrght 2011 Pearson Addson-Wesley. All rghts reserved. 1-1 Outlne 1. Fttng a lne to data 2. The ordnary least squares (OLS) lne/regresson 3. Measures of ft 4. Populaton

More information

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION 014-015 MTH35/MH3510 Regresson Analyss December 014 TIME ALLOWED: HOURS INSTRUCTIONS TO CANDIDATES 1. Ths examnaton paper contans FOUR (4) questons

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis Resource Allocaton and Decson Analss (ECON 800) Sprng 04 Foundatons of Regresson Analss Readng: Regresson Analss (ECON 800 Coursepak, Page 3) Defntons and Concepts: Regresson Analss statstcal technques

More information

Lecture 3 Stat102, Spring 2007

Lecture 3 Stat102, Spring 2007 Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture

More information

Solutions Homework 4 March 5, 2018

Solutions Homework 4 March 5, 2018 1 Solutons Homework 4 March 5, 018 Soluton to Exercse 5.1.8: Let a IR be a translaton and c > 0 be a re-scalng. ˆb1 (cx + a) cx n + a (cx 1 + a) c x n x 1 cˆb 1 (x), whch shows ˆb 1 s locaton nvarant and

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

Estimation: Part 2. Chapter GREG estimation

Estimation: Part 2. Chapter GREG estimation Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the

More information

Introduction to Generalized Linear Models

Introduction to Generalized Linear Models INTRODUCTION TO STATISTICAL MODELLING TRINITY 00 Introducton to Generalzed Lnear Models I. Motvaton In ths lecture we extend the deas of lnear regresson to the more general dea of a generalzed lnear model

More information

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3. Outlne 3. Multple Regresson Analyss: Estmaton I. Motvaton II. Mechancs and Interpretaton of OLS Read Wooldrdge (013), Chapter 3. III. Expected Values of the OLS IV. Varances of the OLS V. The Gauss Markov

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experments- MODULE LECTURE - 6 EXPERMENTAL DESGN MODELS Dr. Shalabh Department of Mathematcs and Statstcs ndan nsttute of Technology Kanpur Two-way classfcaton wth nteractons

More information

4.3 Poisson Regression

4.3 Poisson Regression of teratvely reweghted least squares regressons (the IRLS algorthm). We do wthout gvng further detals, but nstead focus on the practcal applcaton. > glm(survval~log(weght)+age, famly="bnomal", data=baby)

More information

Statistics Chapter 4

Statistics Chapter 4 Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment

More information

Chapter 14: Logit and Probit Models for Categorical Response Variables

Chapter 14: Logit and Probit Models for Categorical Response Variables Chapter 4: Logt and Probt Models for Categorcal Response Varables Sect 4. Models for Dchotomous Data We wll dscuss only ths secton of Chap 4, whch s manly about Logstc Regresson, a specal case of the famly

More information

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε Chapter 3 Secton 3.1 Model Assumptons: Multple Regresson Model Predcton Equaton Std. Devaton of Error Correlaton Matrx Smple Lnear Regresson: 1.) Lnearty.) Constant Varance 3.) Independent Errors 4.) Normalty

More information

28. SIMPLE LINEAR REGRESSION III

28. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III Ftted Values and Resduals US Domestc Beers: Calores vs. % Alcohol To each observed x, there corresponds a y-value on the ftted lne, y ˆ = βˆ + βˆ x. The are called ftted

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

SIMPLE LINEAR REGRESSION

SIMPLE LINEAR REGRESSION Smple Lnear Regresson and Correlaton Introducton Prevousl, our attenton has been focused on one varable whch we desgnated b x. Frequentl, t s desrable to learn somethng about the relatonshp between two

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

Chapter 12 Analysis of Covariance

Chapter 12 Analysis of Covariance Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty

More information

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes 25/6 Canddates Only January Examnatons 26 Student Number: Desk Number:...... DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR Department Module Code Module Ttle Exam Duraton

More information

Basically, if you have a dummy dependent variable you will be estimating a probability.

Basically, if you have a dummy dependent variable you will be estimating a probability. ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy

More information

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 1 Chapters 14, 15 & 16 Professor Ahmad, Ph.D. Department of Management Revsed August 005 Chapter 14 Formulas Smple Lnear Regresson Model: y =

More information

18. SIMPLE LINEAR REGRESSION III

18. SIMPLE LINEAR REGRESSION III 8. SIMPLE LINEAR REGRESSION III US Domestc Beers: Calores vs. % Alcohol Ftted Values and Resduals To each observed x, there corresponds a y-value on the ftted lne, y ˆ ˆ = α + x. The are called ftted values.

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5).

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5). (out of 15 ponts) STAT 3340 Assgnment 1 solutons (10) (10) 1. Fnd the equaton of the lne whch passes through the ponts (1,1) and (4,5). β 1 = (5 1)/(4 1) = 4/3 equaton for the lne s y y 0 = β 1 (x x 0

More information