Profile Analysis Multivariate Regression

Size: px
Start display at page:

Download "Profile Analysis Multivariate Regression"

Transcription

1 Lecture 8 October 12, 2005 Analysis Lecture #8-10/12/2005 Slide 1 of 68

2 Today s Lecture Profile analysis Today s Lecture Schedule : regression review multiple regression is due Thursday, October 27th, at 5pm in my mailbox Lecture #8-10/12/2005 Slide 2 of 68

3 Schedule 10/19 - Case studies in statistics Today s Lecture Schedule 10/26 - Principal components 10/27 - due 11/2 - Analysis of covariance structures 11/9 - Canonical correlation 11/16 - Discrimination/classification 11/23 - Thanksgiving break 11/30 - Clustering/distance methods/mds (Final handed out - Due 12/15) 12/7 - No class Lecture #8-10/12/2005 Slide 3 of 68

4 Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We begin with a definition of profile analysis (aka growth curves or repeated measures) For each subject we have a vector of observations that are commensurate: We assume that the observations are measured in the same units We assume that the observations have about the same variances Typically, these analyses are done with data that has been collected over time and we are interested in seeing whether there is a general linear trend (an ordering is not necessary though) Lecture #8-10/12/2005 Slide 4 of 68

5 Single Sample Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We begin our discussion of profile analyses with a reminder that we performed such an analysis with a single sample: the repeated measures T 2 (example using the dogs data) In this case we a set of p observations with mean vector µ = {µ 1, µ 2,,µ p } Now we want to test the Null hypothesis H o : µ 1 = µ 2 = = µ p If all you knew was ANOVA, this particular hypothesis would be analyzed using an ANOVA, BUT that would assume that the observations are independent across the p variables We know that is not true So how are we going to do this? Lecture #8-10/12/2005 Slide 5 of 68

6 Single Sample Recall from the dogs data that we can rewrite our hypothesis that all means are equal to be phrased as H 0 : µ 1 µ 2 = µ 2 µ 3 = = µ p 1 µ p = 0 Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? H 0 : µ 1 µ 2 µ 2 µ 3 µ p 1 µ p = Notice that the only way this can occur is when all means are equal Lecture #8-10/12/2005 Slide 6 of 68

7 Single Sample Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We begin by defining z = CX where C is a ((p 1) x p) contrast matrix defined as: Cy = y = y 1 y 2 y 2 y 3 y p 1 y p Then we compute are new mean vector z = C x and covariance matrix S z = CS x C There are now p 1 rows of z S z is size p 1 p 1 Lecture #8-10/12/2005 Slide 7 of 68

8 Single Sample Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? From there we compute Hotelling s T 2 using our old equation and our new variables z Which is the same as T 2 = n( z µ 0 ) (S z ) 1 ( z µ 0 ) T 2 = n(c x µ 0 ) (CS x C ) 1 (C x µ 0 ) T 2 in this case has p 1 variables and n 1 degrees of freedom Lecture #8-10/12/2005 Slide 8 of 68

9 Two Samples Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Now imagine that of having only a single mean vector that we are interested in testing, we have two Previously we only used a test to see if we believed that the means were all equal, well now with two vectors we can test three things 1 Are the two profiles parallel? 1a Is H 01 : µ 1i µ 1i 1 = µ 2i µ 2i 1 (i = 2, 3,,p)? 2 If they are parallel, are they coincident? 2a Is H 02 : µ 1i = µ 2i (i = 1, 2, 3,, p)? 3 If they are coincident, are they level? 3a Is H 03 : µ 1i = µ 12 = = µ 1p = µ 2i = µ 22 = = µ 2p? Lecture #8-10/12/2005 Slide 9 of 68

10 Example Data Set Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? To demonstrate profile analysis, we will use the example data set from Johnson and Wichern (p 320): As part of a larger study of love and marriage, E Hatfield, a sociologist, survey adults with respect to their marriage contributions and outcomes and their levels of passionate and companionate love Recently married males and females were asked to respond to the following questions using an 8-point scale (8 representing extremely positive) How would you describe your contributions to the marriage? How would you describe your outcomes from the marriage? Lecture #8-10/12/2005 Slide 10 of 68

11 Example Data Set Continuing on: Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Subjects were also asked to respond to the following questions, using the 5-point scale shown: What is the level of passionate love that you feel for your partner? What is the level of compassionate love that you feel for your partner? Lecture #8-10/12/2005 Slide 11 of 68

12 Example Data Set Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Lecture #8-10/12/2005 Slide 12 of 68

13 Example Data Set Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Lecture #8-10/12/2005 Slide 13 of 68

14 Part 1: Parallel Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We begin by asking whether the two population mean vectors are parallel If they are parallel, it could mean that: One group performed uniformly better than the other group Both groups are equal To test this we need to think what it would mean that they are parallel It would mean that the difference between any two means in group one is the same as the difference between the same two means in group two Lecture #8-10/12/2005 Slide 14 of 68

15 Parallel Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We have already done something like this in the single profile case In the single case we said that all differences should equal zero (this amounted to a single mean vector test) Now we have mean differences for each group and instead of test whether they equal zero we will test if they are EQUAL the Null hypothesis is: H o : µ 12 µ 11 µ 13 µ 12 µ 1p µ 1(p 1) = µ 22 µ 21 µ 23 µ 22 µ 2p µ 2(p 1) Lecture #8-10/12/2005 Slide 15 of 68

16 Parallel Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? We compute our regular T 2 using the contrast matrix C, with p 1 contrasts: T 2 = ( x 1 x 2 ) C [( 1 n n 2 ) CS p C ] 1 C( x 1 x 2 ) Which is compared against: (n 1 + n 2 2)(p 1) n 1 + n 2 p F p 1,n1 +n 2 p(α) Lecture #8-10/12/2005 Slide 16 of 68

17 Coincident Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Now, especially if we found out that the profiles are parallel, we are going to test whether they are, ON AVERAGE the same height Recall our null hypothesis: Is H 02 : µ 1i = µ 2i (i = 1, 2, 3,, p)? We could also rewrite this as: H 0 : (µ 11 + µ µ 1p ) = (µ 21 + µ µ 2p ) Lecture #8-10/12/2005 Slide 17 of 68

18 Coincident Profiles? To test this hypothesis, we again us T 2 : Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? [( 1 T 2 = 1 ( x 1 x 2 ) + 1 ) 1 1 S p 1] 1 ( x 1 x 2 ) n 1 n 2 This statistic is nearly the same as the univariate t-test comparing the overall average across all variables for each group T 2 is compared with: t 2 n 1 +n 2 +2 ( α ) = F 1,n1 +n 2 2 2(α) Lecture #8-10/12/2005 Slide 18 of 68

19 Level Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? Finally, we can text to see if the two profiles are, on average, flat (ie, this is the same as saying looking for a main effect across variables ON AVERAGE) What is nice about this is that we have already learned how to test for whether a single profile is flat H 0 : µ 1 = µ 2 = = µ p Now we want to phrase this same test in a way so we will test H 0 : 1 2 (µ 11 + µ 21 ) = 1 2 (µ 12 + µ 22 ) = = 1 2 (µ 1p + µ 2p ) We could rewrite this as another hypothesis test H 0 : 1 2 (µ 1 + µ 2 ) = 0 Lecture #8-10/12/2005 Slide 19 of 68

20 Level Profiles? Single Sample Two Samples Parallel Profiles? Coincident Profiles? Level Profiles? So to test this we will use the overall mean and test whether it is flat (this is conceptually the same as pooling our two groups and performing a one vector test) This means we define the overall mean x as x = n 1 x 1 + n 2 x 2 n 1 + n 2 Our estimate of variance will be the pooled variance covariance matrix S p We perform our regular test to see if it is flat T 2 = (n 1 + n 2 ) xc ( CS p C ) 1 C x This is compared against (n 1+n 2 1)(p 1) n 1 +n 2 p+1 F p 1,n1 +n 2 p+1(α) Lecture #8-10/12/2005 Slide 20 of 68

21 Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Recall the multiple regression equation (for the i th observation, prediction of Y i by r variables z ir ): Y j = β 0 + β 1 z j1 + β 2 z j2 + + β r z jr + ǫ j In univariate multiple regression, we assume the following: All n observations are independent E(ǫ j ) = 0 - the error terms have a zero mean Var(ǫ j ) = σ 2 - the error terms have a constant variance Cov(ǫ j, ǫ k ) = 0 - the error terms are uncorrelated between observations Lecture #8-10/12/2005 Slide 21 of 68

22 Analysis with Matrices Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The equation above can be expressed more compactly by a set of matrices: y is of size (n 1) Z is of size (n (r + 1)) β is of size ((r + 1) 1) ǫ is of size (n 1) y = Zβ + ǫ Lecture #8-10/12/2005 Slide 22 of 68

23 with Matrices Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n y = Z β + ǫ (n 1) (n (r + 1)) ((r + 1) 1) (n 1) Lecture #8-10/12/2005 Slide 23 of 68

24 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1

25 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0

26 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0 + β 1 z 11

27 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0 + β 1 z 11 + β 2 z 12

28 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0 + β 1 z 11 + β 2 z 12 +

29 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0 + β 1 z 11 + β 2 z β r z 1r

30 Lecture #8-10/12/2005 Slide 24 of 68 with Matrices Working the matrix multiplication and addition for a single case gives: Y 1 Y 2 Y 3 Y 4 Y n = 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr β 0 β 1 β 2 β r + ǫ 1 ǫ 2 ǫ 3 ǫ 4 ǫ n Y 1 = β 0 + β 1 z 11 + β 2 z β r z 1r +e 1

31 Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph with Matrices The matrix of predictors, Z, has the first column containing all ones This represents the intercept parameter β 0 This is also an introduction to setting columns of the Z matrix to represent design and or group controls (as in ANOVA) Also note that the assumptions from the univariate regression model can be reexpressed: E(ǫ) = 0 - the error terms have a zero mean vector Cov(ǫ) = E(ǫǫ ) = σ 2 I: The error terms have a constant variance (all equal to σ 2 The error terms are uncorrelated between observations Lecture #8-10/12/2005 Slide 25 of 68

32 Distribution of Errors Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph We often place distributional assumptions on our error terms, allowing for the development of tractable hypothesis tests With matrices, the distributional assumptions are no different, except for things are approached in a multivariate fashion: ǫ N n (0, σ 2 I) Having a multivariate normal distribution with uncorrelated variables (from I N ) is identical to saying: for all i observations ǫ j N(0, σ 2 ) Lecture #8-10/12/2005 Slide 26 of 68

33 Estimation with Matrices Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph estimates are typically found via least squares (called L 2 estimates) In least squares regression, the estimates are found by minimizing: n ǫ 2 j = j=1 n (Y j Ŷ j ) 2 = j=1 n (Y j β 0 + β 1 z j1 + + β r z jr ) 2 j=1 As you could guess, we could accomplish all of this via matrices Equivalently: n ǫ 2 j = j=1 n (Y j z j β) 2 = (y Zβ) (y Zβ) = ǫ ǫ j=1 Lecture #8-10/12/2005 Slide 27 of 68

34 Estimation with Matrices Thankfully, there are people who figured out the equation for b that minimizes ǫ ǫ b = (Z Z) 1 Z y b = ( Z Z ) 1 Z y ((r + 1) 1) ((r + 1) n) (n (r + 1)) ((r + 1) n) (n 1) This equation is what I have been talking about for quite some time, the General Linear Model For many types of data, in many differing analyses, this equation will provide estimates: Multiple ANOVA Analysis of Covariance (ANCOVA) Multiple with Curvilinear relationships in Z Lecture #8-10/12/2005 Slide 28 of 68

35 Example Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Imagine that you are trying to determine best type of birthday cake to get for someone who s birthday is fast approaching Given your knowledge of the subject, you have identified two variables that seem to play a primary role in the ratings people assign to other cakes: Moisture content Sweetness You are interested in determining how moisture content and sweetness play a role in cake preference ratings You collect a random sample of cakes from locations here in Lawrence (like Target) You collect a random sample of cake eaters here in Lawrence Lecture #8-10/12/2005 Slide 29 of 68

36 The Cook Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Chef Bob Lecture #8-10/12/2005 Slide 30 of 68

37 The Cake Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Cake Lecture #8-10/12/2005 Slide 31 of 68

38 The Data Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Rating Moisture Sweetness Y z1 z Lecture #8-10/12/2005 Slide 32 of 68

39 The Code proc glm data=cake; model rating=moisture sweetness; Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 33 of 68

40 The Output Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 34 of 68

41 Sums of Squares Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The Sum of Squares for the can be obtained by matrix calculations: SS reg = b Z y ( N i=1 Y )2 N = b Z y 1 N (y 1) 2 Lecture #8-10/12/2005 Slide 35 of 68

42 Residual Sums of Squares Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Hypothesis tests in multiple regression proceed similar to those in ANOVA in that a single, omnibus test is performed prior to inspecting for significance of each of the model parameters The hypothesis test is constructed by comparing the sum of squares due to regression (divided by the number of variables in the model) to the sum of squares due to error (divided by the number of observations minus the number of variables minus one) Lecture #8-10/12/2005 Slide 36 of 68

43 Residual Sums of Squares Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The Residual Sum of Squares is also obtained by matrix calculations: SS res = n e 2 j = e e = y y b Z y j=1 Lecture #8-10/12/2005 Slide 37 of 68

44 Omnibus Hypothesis Test The overall hypothesis test, testing: Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph is given by: F = H 0 : β = 0 SS reg /r SS res /(n r 1) This is compared with F r,n r 1 (α) If this is rejected, inspection of the model parameters ensues, testing each for being equal to zero Lecture #8-10/12/2005 Slide 38 of 68

45 Variance of Estimators Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The covariance matrix of the regression parameters contains useful information regarding the standard errors of the estimates (found along the diagonal) To find the covariance matrix of the estimates, the Residual Sum of Squares is needed (dividing this term by the residual degrees of freedom gives the error variance, or σ 2 e): Cov(b) = σ 2 e(z Z) 1 = SS res n r 1 (Z Z) 1 = 1 n r 1 e e(z Z) 1 Lecture #8-10/12/2005 Slide 39 of 68

46 Squared Correlation Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The Squared Multiple Correlation Coefficient, or R 2, is found from: R 2 = SS reg n j=1 (Y j Ȳ )2 This is also obtainable by matrix calculations: R 2 = b Z y 1 n (y 1) 2 y y 1 n (y 1) 2 Lecture #8-10/12/2005 Slide 40 of 68

47 Prediction Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph To make things easy, let s begin with trying to predict a cake s rating based on a single z variable: moisture content From the results of the analysis, for any size of X (any living space size), we will predict the sale price of the house to be: Ŷ = z But knowing the predicted value is just the beginning Lecture #8-10/12/2005 Slide 41 of 68

48 What Does the Predicted Value Represent? Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph The predicted value given by a regression equation is the expected value of Y conditional on z This means that for every value of z, the mean value [E(Y z)] is given by Ŷ Recall from basic statistics that if a variable x is distributed with a mean µ x and variance σ 2 x then: As sample size, n, goes to infinity, the distribution of x is N(µ x, σ2 x n ) (Central Limit Theorem) Using the CLT, we are able to build a confidence bound around our sample mean so that we can be 100(α/2)% certain that the true µ x lies in the interval The same principal can be applied to of the distribution of Y z Ŷ, which is the mean Lecture #8-10/12/2005 Slide 42 of 68

49 Confidence Interval for Mean Predicted Value Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph A confidence interval for the mean predicted value of a regression is given by: ( α ) z 0b ± t n r 1 z 0 2 (Z Z) 1 z 0 s 2 where z 0 is the set of values of z used to predict Y from As the value of z 0 gets farther from its mean, the prediction interval increases in range You can get mean prediction values form SAS with the following code: *SAS Example #2- mean prediction; proc glm data=cake; model rating=moisture/ p clm; r Lecture #8-10/12/2005 Slide 43 of 68

50 Predicted Value Example Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 44 of 68

51 Graph of the Mean Predicted Value Interval Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 45 of 68

52 CI for Single Predicted Values Often times, the mean value is not of interest in prediction Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Imagine you buy a cake with a moisture rating of 6 You want to know how well that cake will be liked Furthermore, you want to know what range of ratings you should expect from this cake Clearly, in your case, the CI for the mean isn t very helpful The CI for a single predicted value would be z 0b ± t n r 1 ( α 2 ) (1 + z 0 (Z Z) 1 z 0 )s 2 Lecture #8-10/12/2005 Slide 46 of 68

53 SE of a Single Predicted Value This formula is very similar to that for the mean Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Again, because of (z i z) 2, as z gets farther from its mean, the CI gets wider Ŷ = z = (6) = In SAS: *SAS Example #3- individual prediction; proc glm data=cake; model rating=moisture/ p cli; r Lecture #8-10/12/2005 Slide 47 of 68

54 Predicted Value Example Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 48 of 68

55 Graph of the Mean Predicted Value Interval Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 49 of 68

56 Graph of the Both Intervals Matrix Forms Error Distribution Estimation Sums of Squares b Variance Squared Correlation Mean Prediction CI Mean CI Graph Single Value CI Single CI Graph Lecture #8-10/12/2005 Slide 50 of 68

57 Hypothesis Tests Hypothesis Tests Predictions Everything about univariate regression can be generalized to multivariate regression through use of matrices Instead of predicting a single column vector y, we now predict a matrix of observations Y Y is of size (n m) Z is of size (n (r + 1)) β is of size ((r + 1) m) Y = Zβ + ǫ ǫ is of size (n m) Lecture #8-10/12/2005 Slide 51 of 68

58 Lecture #8-10/12/2005 Slide 52 of 68 with Matrices Y 11 Y 12 Y 1m Y 21 Y 22 Y 2m Y 31 Y 32 Y 3m Y 41 Y 42 Y 4m Y n1 Y n2 Y nm 1 z 11 z 12 z 1r 1 z 21 z 22 z 2r 1 z 31 z 32 z 3r 1 z 41 z 42 z 4r 1 z n1 z n2 z nr Y Z (n m) (n (r + 1))

59 with Matrices β 01 β 02 β 0m β 11 β 12 β 1m β 21 β 22 β 2m β r1 β r2 β rm ǫ 11 ǫ 12 ǫ 1m ǫ 21 ǫ 22 ǫ 2m ǫ 31 ǫ 32 ǫ 3m ǫ 41 ǫ 42 ǫ 4m ǫ n1 ǫ n2 ǫ nm β ǫ ((r + 1) m) (n m) Lecture #8-10/12/2005 Slide 53 of 68

60 Estimation with Matrices Just as with univariate regression, the least squares estimates are found by: B = (Z Z) 1 Z Y B = ( Z Z ) 1 Z Y ((r + 1) m) ((r + 1) n) (n (r + 1)) ((r + 1) n) (n m) Every other formula used for linear regression can be extended in a similar fashion Lecture #8-10/12/2005 Slide 54 of 68

61 Hypothesis Tests Hypothesis Tests Hypothesis Tests Predictions Like MANOVA, the multivariate regression model can be tested for an overall indication that some regression coefficient is equal to zero (one out of the many within the β matrix) Again, the hypothesis test information can be found from within the MANOVA output from SAS To introduce this, take SAS example #4from problem 725 of Johnson and Wichern (p 423): Amitriptyline is prescribed by some physicians as an antidepressant However, there are also conjectured side effects that seem to be related to the use of the drug: irregular heartbeat, abnormal blood pressures, and irregular waves on the electrocardiogram, among other things Data are gathered on 17 patients who were admitted to the hospital after an amitriptyline overdose Lecture #8-10/12/2005 Slide 55 of 68

62 Hypothesis Tests Variables: Hypothesis Tests Hypothesis Tests Predictions Y 1 : TOT - Total TCAD Plasma level Y 2 : AMI - Amount of amitriptyline present in TCAD plasma level (AMI) Z 1 : GEN - Gender: 1 if female, 0 if male Z 2 : AMT - Amount of antidepressants taken at time of overdose Z 3 : PR - PR wave measurement Z 4 : DIAP - Diastolic blood pressure Z 5 : QRS - QRS wave measurement Lecture #8-10/12/2005 Slide 56 of 68

63 Hypothesis Tests The SAS code for multivariate regression is: Hypothesis Tests Hypothesis Tests Predictions data battery; infile C:\T7-6dat ; input tot ami gen amt pr diap qrs; run; proc glm data=battery; model tot ami=gen amt pr diap qrs; manova h=gen amt pr diap qrs/printe; run; Lecture #8-10/12/2005 Slide 57 of 68

64 SAS Output Hypothesis Tests Hypothesis Tests Predictions H 0 : β 5 = 0 Lecture #8-10/12/2005 Slide 58 of 68

65 SAS Output Hypothesis Tests Hypothesis Tests Predictions Lecture #8-10/12/2005 Slide 59 of 68

66 Predictions Hypothesis Tests Hypothesis Tests Predictions Just as with univariate regression, two types of predictions can be made for multivariate regression: Predictions for the mean response vector Predictions for a single response Now, these predictions are in the form of a (you guessed it)confidence ellipse These ellipses can then be projected to be simultaneous confidence intervals Lecture #8-10/12/2005 Slide 60 of 68

67 Predictions for the Mean Response Hypothesis Tests Hypothesis Tests Predictions For a vector of observations z 0, the 100 (1 α)% simultaneous confidence intervals for the mean prediction are found by: z 0 B (i) ± ( ) m(n r 1) z 0 n r m (Z Z) 1 n z 0 n r 1 ˆσ ii Here,ˆσ ii is the i th diagonal element of ˆΣ, where: ˆΣ = 1 n (Y ZB) (Y ZB) Lecture #8-10/12/2005 Slide 61 of 68

68 Predictions for the Individual Response Hypothesis Tests Hypothesis Tests Predictions For a vector of observations z 0, the 100 (1 α)% simultaneous confidence intervals for the individual prediction are found by: z 0 B (i) ± ( ) m(n r 1) (1 + z 0 n r m (Z Z) 1 n z 0 ) n r 1 ˆσ ii Here,ˆσ ii is the i th diagonal element of ˆΣ, where: ˆΣ = 1 n (Y ZB) (Y ZB) Lecture #8-10/12/2005 Slide 62 of 68

69 Final Exam Two problems General Info Data Analysis #1 Data Analysis #2 - Asthma 1 Crash test data 2 Asthma data For both please: Double-space Have less than 10 pages of text per problem Lecture #8-10/12/2005 Slide 63 of 68

70 Data Analysis #1 General Info Data Analysis #1 Data Analysis #2 - Asthma Crash test dummy data Lecture #8-10/12/2005 Slide 64 of 68

71 Data Analysis #2 - Asthma General Info Data Analysis #1 Data Analysis #2 - Asthma Lecture #8-10/12/2005 Slide 65 of 68

72 Data Analysis #2 - Air Pollution General Info Data Analysis #1 Data Analysis #2 - Asthma Lecture #8-10/12/2005 Slide 66 of 68

73 Final Thought Profile analysis is helpful in determining the nature of differences between two groups regression is very similar to MANOVA in its application A hint for the test: use of IML is not necessary Final Thought Next Class Lecture #8-10/12/2005 Slide 67 of 68

74 Next Time Statistical Case Studies Reporting multivariate statistical analyses A SAS Macro for testing for multivariate normality I bring in some data sets and we analyze them together If you have data you would like analyzed (featuring topics we have used up to this point), send them to me via Final Thought Next Class Lecture #8-10/12/2005 Slide 68 of 68

Multivariate Linear Regression Models

Multivariate Linear Regression Models Multivariate Linear Regression Models Regression analysis is used to predict the value of one or more responses from a set of predictors. It can also be used to estimate the linear association between

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation Using the least squares estimator for β we can obtain predicted values and compute residuals: Ŷ = Z ˆβ = Z(Z Z) 1 Z Y ˆɛ = Y Ŷ = Y Z(Z Z) 1 Z Y = [I Z(Z Z) 1 Z ]Y. The usual decomposition

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Regression With a Categorical Independent Variable

Regression With a Categorical Independent Variable Regression With a Categorical Independent Variable Lecture 15 March 17, 2005 Applied Regression Analysis Lecture #15-3/17/2005 Slide 1 of 29 Today s Lecture» Today s Lecture» Midterm Note» Example Regression

More information

Regression With a Categorical Independent Variable

Regression With a Categorical Independent Variable Regression With a Independent Variable Lecture 10 November 5, 2008 ERSH 8320 Lecture #10-11/5/2008 Slide 1 of 54 Today s Lecture Today s Lecture Chapter 11: Regression with a single categorical independent

More information

Regression With a Categorical Independent Variable: Mean Comparisons

Regression With a Categorical Independent Variable: Mean Comparisons Regression With a Categorical Independent Variable: Mean Lecture 16 March 29, 2005 Applied Regression Analysis Lecture #16-3/29/2005 Slide 1 of 43 Today s Lecture comparisons among means. Today s Lecture

More information

Mean Vector Inferences

Mean Vector Inferences Mean Vector Inferences Lecture 5 September 21, 2005 Multivariate Analysis Lecture #5-9/21/2005 Slide 1 of 34 Today s Lecture Inferences about a Mean Vector (Chapter 5). Univariate versions of mean vector

More information

Multivariate analysis of variance and covariance

Multivariate analysis of variance and covariance Introduction Multivariate analysis of variance and covariance Univariate ANOVA: have observations from several groups, numerical dependent variable. Ask whether dependent variable has same mean for each

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 1: August 22, 2012

More information

Regression With a Categorical Independent Variable

Regression With a Categorical Independent Variable Regression ith a Independent Variable ERSH 8320 Slide 1 of 34 Today s Lecture Regression with a single categorical independent variable. Today s Lecture Coding procedures for analysis. Dummy coding. Relationship

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model EPSY 905: Multivariate Analysis Lecture 1 20 January 2016 EPSY 905: Lecture 1 -

More information

Review of the General Linear Model

Review of the General Linear Model Review of the General Linear Model EPSY 905: Multivariate Analysis Online Lecture #2 Learning Objectives Types of distributions: Ø Conditional distributions The General Linear Model Ø Regression Ø Analysis

More information

ST505/S697R: Fall Homework 2 Solution.

ST505/S697R: Fall Homework 2 Solution. ST505/S69R: Fall 2012. Homework 2 Solution. 1. 1a; problem 1.22 Below is the summary information (edited) from the regression (using R output); code at end of solution as is code and output for SAS. a)

More information

Multivariate Analysis - Overview

Multivariate Analysis - Overview Multivariate Analysis - Overview In general: - Analysis of Multivariate data, i.e. each observation has two or more variables as predictor variables, Analyses Analysis of Treatment Means (Single (multivariate)

More information

Applied Multivariate Statistical Modeling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur

Applied Multivariate Statistical Modeling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur Applied Multivariate Statistical Modeling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur Lecture - 29 Multivariate Linear Regression- Model

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

A Re-Introduction to General Linear Models (GLM)

A Re-Introduction to General Linear Models (GLM) A Re-Introduction to General Linear Models (GLM) Today s Class: You do know the GLM Estimation (where the numbers in the output come from): From least squares to restricted maximum likelihood (REML) Reviewing

More information

Lecture 3: Inference in SLR

Lecture 3: Inference in SLR Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals

More information

Statistical Foundations:

Statistical Foundations: Statistical Foundations: t distributions, t-tests tests Psychology 790 Lecture #12 10/03/2006 Today sclass The t-distribution t ib ti in its full glory. Why we use it for nearly everything. Confidence

More information

Appendix A: Review of the General Linear Model

Appendix A: Review of the General Linear Model Appendix A: Review of the General Linear Model The generallinear modelis an important toolin many fmri data analyses. As the name general suggests, this model can be used for many different types of analyses,

More information

Interactions among Continuous Predictors

Interactions among Continuous Predictors Interactions among Continuous Predictors Today s Class: Simple main effects within two-way interactions Conquering TEST/ESTIMATE/LINCOM statements Regions of significance Three-way interactions (and beyond

More information

General Linear Model (Chapter 4)

General Linear Model (Chapter 4) General Linear Model (Chapter 4) Outcome variable is considered continuous Simple linear regression Scatterplots OLS is BLUE under basic assumptions MSE estimates residual variance testing regression coefficients

More information

Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models

Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models EPSY 905: Multivariate Analysis Spring 2016 Lecture #12 April 20, 2016 EPSY 905: RM ANOVA, MANOVA, and Mixed Models

More information

Rejection regions for the bivariate case

Rejection regions for the bivariate case Rejection regions for the bivariate case The rejection region for the T 2 test (and similarly for Z 2 when Σ is known) is the region outside of an ellipse, for which there is a (1-α)% chance that the test

More information

Categorical Predictor Variables

Categorical Predictor Variables Categorical Predictor Variables We often wish to use categorical (or qualitative) variables as covariates in a regression model. For binary variables (taking on only 2 values, e.g. sex), it is relatively

More information

STA442/2101: Assignment 5

STA442/2101: Assignment 5 STA442/2101: Assignment 5 Craig Burkett Quiz on: Oct 23 rd, 2015 The questions are practice for the quiz next week, and are not to be handed in. I would like you to bring in all of the code you used to

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

A Introduction to Matrix Algebra and the Multivariate Normal Distribution

A Introduction to Matrix Algebra and the Multivariate Normal Distribution A Introduction to Matrix Algebra and the Multivariate Normal Distribution PRE 905: Multivariate Analysis Spring 2014 Lecture 6 PRE 905: Lecture 7 Matrix Algebra and the MVN Distribution Today s Class An

More information

STAT 501 EXAM I NAME Spring 1999

STAT 501 EXAM I NAME Spring 1999 STAT 501 EXAM I NAME Spring 1999 Instructions: You may use only your calculator and the attached tables and formula sheet. You can detach the tables and formula sheet from the rest of this exam. Show your

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

Lecture 5: Hypothesis tests for more than one sample

Lecture 5: Hypothesis tests for more than one sample 1/23 Lecture 5: Hypothesis tests for more than one sample Måns Thulin Department of Mathematics, Uppsala University thulin@math.uu.se Multivariate Methods 8/4 2011 2/23 Outline Paired comparisons Repeated

More information

Review of Multiple Regression

Review of Multiple Regression Ronald H. Heck 1 Let s begin with a little review of multiple regression this week. Linear models [e.g., correlation, t-tests, analysis of variance (ANOVA), multiple regression, path analysis, multivariate

More information

Topic 28: Unequal Replication in Two-Way ANOVA

Topic 28: Unequal Replication in Two-Way ANOVA Topic 28: Unequal Replication in Two-Way ANOVA Outline Two-way ANOVA with unequal numbers of observations in the cells Data and model Regression approach Parameter estimates Previous analyses with constant

More information

STA 303H1F: Two-way Analysis of Variance Practice Problems

STA 303H1F: Two-way Analysis of Variance Practice Problems STA 303H1F: Two-way Analysis of Variance Practice Problems 1. In the Pygmalion example from lecture, why are the average scores of the platoon used as the response variable, rather than the scores of the

More information

Correlation & Simple Regression

Correlation & Simple Regression Chapter 11 Correlation & Simple Regression The previous chapter dealt with inference for two categorical variables. In this chapter, we would like to examine the relationship between two quantitative variables.

More information

STA 101 Final Review

STA 101 Final Review STA 101 Final Review Statistics 101 Thomas Leininger June 24, 2013 Announcements All work (besides projects) should be returned to you and should be entered on Sakai. Office Hour: 2 3pm today (Old Chem

More information

STA 302f16 Assignment Five 1

STA 302f16 Assignment Five 1 STA 30f16 Assignment Five 1 Except for Problem??, these problems are preparation for the quiz in tutorial on Thursday October 0th, and are not to be handed in As usual, at times you may be asked to prove

More information

Multivariate Regression (Chapter 10)

Multivariate Regression (Chapter 10) Multivariate Regression (Chapter 10) This week we ll cover multivariate regression and maybe a bit of canonical correlation. Today we ll mostly review univariate multivariate regression. With multivariate

More information

Business Statistics. Lecture 9: Simple Regression

Business Statistics. Lecture 9: Simple Regression Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals

More information

Simple, Marginal, and Interaction Effects in General Linear Models

Simple, Marginal, and Interaction Effects in General Linear Models Simple, Marginal, and Interaction Effects in General Linear Models PRE 905: Multivariate Analysis Lecture 3 Today s Class Centering and Coding Predictors Interpreting Parameters in the Model for the Means

More information

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima

Hypothesis testing Goodness of fit Multicollinearity Prediction. Applied Statistics. Lecturer: Serena Arima Applied Statistics Lecturer: Serena Arima Hypothesis testing for the linear model Under the Gauss-Markov assumptions and the normality of the error terms, we saw that β N(β, σ 2 (X X ) 1 ) and hence s

More information

Applied Multivariate and Longitudinal Data Analysis

Applied Multivariate and Longitudinal Data Analysis Applied Multivariate and Longitudinal Data Analysis Chapter 2: Inference about the mean vector(s) Ana-Maria Staicu SAS Hall 5220; 919-515-0644; astaicu@ncsu.edu 1 In this chapter we will discuss inference

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

Simple linear regression

Simple linear regression Simple linear regression Biometry 755 Spring 2008 Simple linear regression p. 1/40 Overview of regression analysis Evaluate relationship between one or more independent variables (X 1,...,X k ) and a single

More information

Chapter 7, continued: MANOVA

Chapter 7, continued: MANOVA Chapter 7, continued: MANOVA The Multivariate Analysis of Variance (MANOVA) technique extends Hotelling T 2 test that compares two mean vectors to the setting in which there are m 2 groups. We wish to

More information

Simple, Marginal, and Interaction Effects in General Linear Models: Part 1

Simple, Marginal, and Interaction Effects in General Linear Models: Part 1 Simple, Marginal, and Interaction Effects in General Linear Models: Part 1 PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 2: August 24, 2012 PSYC 943: Lecture 2 Today s Class Centering and

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Comparisons of Two Means Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c

More information

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

WISE International Masters

WISE International Masters WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are

More information

Regression and Models with Multiple Factors. Ch. 17, 18

Regression and Models with Multiple Factors. Ch. 17, 18 Regression and Models with Multiple Factors Ch. 17, 18 Mass 15 20 25 Scatter Plot 70 75 80 Snout-Vent Length Mass 15 20 25 Linear Regression 70 75 80 Snout-Vent Length Least-squares The method of least

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Lecture 2 January 27, 2005 Lecture #2-1/27/2005 Slide 1 of 46 Today s Lecture Simple linear regression. Partitioning the sum of squares. Tests of significance.. Regression diagnostics

More information

Sections 7.1, 7.2, 7.4, & 7.6

Sections 7.1, 7.2, 7.4, & 7.6 Sections 7.1, 7.2, 7.4, & 7.6 Adapted from Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1 / 25 Chapter 7 example: Body fat n = 20 healthy females 25 34

More information

22s:152 Applied Linear Regression. Take random samples from each of m populations.

22s:152 Applied Linear Regression. Take random samples from each of m populations. 22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each

More information

STA441: Spring Multiple Regression. This slide show is a free open source document. See the last slide for copyright information.

STA441: Spring Multiple Regression. This slide show is a free open source document. See the last slide for copyright information. STA441: Spring 2018 Multiple Regression This slide show is a free open source document. See the last slide for copyright information. 1 Least Squares Plane 2 Statistical MODEL There are p-1 explanatory

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Regression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur

Regression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur Regression Analysis and Forecasting Prof. Shalabh Department of Mathematics and Statistics Indian Institute of Technology-Kanpur Lecture 10 Software Implementation in Simple Linear Regression Model using

More information

Linear Regression and Non-Linear Regression

Linear Regression and Non-Linear Regression Linear Regression and Non-Linear Regression Andrew Long October 7, 2016 Abstract There are two fundamental ideas at work here. The first is finding a best-fit to an over-determined problem (over-determined

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Introduction Edps/Psych/Stat/ 584 Applied Multivariate Statistics Carolyn J Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board of Trustees,

More information

Lectures on Simple Linear Regression Stat 431, Summer 2012

Lectures on Simple Linear Regression Stat 431, Summer 2012 Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population

More information

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model 1 Linear Regression 2 Linear Regression In this lecture we will study a particular type of regression model: the linear regression model We will first consider the case of the model with one predictor

More information

Final Exam Bus 320 Spring 2000 Russell

Final Exam Bus 320 Spring 2000 Russell Name Final Exam Bus 320 Spring 2000 Russell Do not turn over this page until you are told to do so. You will have 3 hours minutes to complete this exam. The exam has a total of 100 points and is divided

More information

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA 22s:152 Applied Linear Regression Chapter 8: ANOVA NOTE: We will meet in the lab on Monday October 10. One-way ANOVA Focuses on testing for differences among group means. Take random samples from each

More information

Lecture 6 Multiple Linear Regression, cont.

Lecture 6 Multiple Linear Regression, cont. Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression

More information

Lecture 12 Inference in MLR

Lecture 12 Inference in MLR Lecture 12 Inference in MLR STAT 512 Spring 2011 Background Reading KNNL: 6.6-6.7 12-1 Topic Overview Review MLR Model Inference about Regression Parameters Estimation of Mean Response Prediction 12-2

More information

MITOCW ocw f99-lec30_300k

MITOCW ocw f99-lec30_300k MITOCW ocw-18.06-f99-lec30_300k OK, this is the lecture on linear transformations. Actually, linear algebra courses used to begin with this lecture, so you could say I'm beginning this course again by

More information

Hypothesis Testing for Var-Cov Components

Hypothesis Testing for Var-Cov Components Hypothesis Testing for Var-Cov Components When the specification of coefficients as fixed, random or non-randomly varying is considered, a null hypothesis of the form is considered, where Additional output

More information

STATISTICS 141 Final Review

STATISTICS 141 Final Review STATISTICS 141 Final Review Bin Zou bzou@ualberta.ca Department of Mathematical & Statistical Sciences University of Alberta Winter 2015 Bin Zou (bzou@ualberta.ca) STAT 141 Final Review Winter 2015 1 /

More information

STA 437: Applied Multivariate Statistics

STA 437: Applied Multivariate Statistics Al Nosedal. University of Toronto. Winter 2015 1 Chapter 5. Tests on One or Two Mean Vectors If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition Chapter 5. Tests

More information

UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Thursday, August 30, 2018

UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Thursday, August 30, 2018 UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Thursday, August 30, 2018 Work all problems. 60 points are needed to pass at the Masters Level and 75

More information

An R # Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM

An R # Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM An R Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM Lloyd J. Edwards, Ph.D. UNC-CH Department of Biostatistics email: Lloyd_Edwards@unc.edu Presented to the Department

More information

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /1/2016 1/46

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /1/2016 1/46 BIO5312 Biostatistics Lecture 10:Regression and Correlation Methods Dr. Junchao Xia Center of Biophysics and Computational Biology Fall 2016 11/1/2016 1/46 Outline In this lecture, we will discuss topics

More information

Note: Please use the actual date you accessed this material in your citation.

Note: Please use the actual date you accessed this material in your citation. MIT OpenCourseWare http://ocw.mit.edu 18.06 Linear Algebra, Spring 2005 Please use the following citation format: Gilbert Strang, 18.06 Linear Algebra, Spring 2005. (Massachusetts Institute of Technology:

More information

Applied Regression Analysis. Section 2: Multiple Linear Regression

Applied Regression Analysis. Section 2: Multiple Linear Regression Applied Regression Analysis Section 2: Multiple Linear Regression 1 The Multiple Regression Model Many problems involve more than one independent variable or factor which affects the dependent or response

More information

Confidence Intervals, Testing and ANOVA Summary

Confidence Intervals, Testing and ANOVA Summary Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0

More information

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math. Regression, part II I. What does it all mean? A) Notice that so far all we ve done is math. 1) One can calculate the Least Squares Regression Line for anything, regardless of any assumptions. 2) But, if

More information

df=degrees of freedom = n - 1

df=degrees of freedom = n - 1 One sample t-test test of the mean Assumptions: Independent, random samples Approximately normal distribution (from intro class: σ is unknown, need to calculate and use s (sample standard deviation)) Hypotheses:

More information

(ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box.

(ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box. FINAL EXAM ** Two different ways to submit your answer sheet (i) Use MS-Word and place it in a drop-box. (ii) Scan your answer sheets INTO ONE FILE only, and submit it in the drop-box. Deadline: December

More information

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700 Class 4 Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science Copyright 013 by D.B. Rowe 1 Agenda: Recap Chapter 9. and 9.3 Lecture Chapter 10.1-10.3 Review Exam 6 Problem Solving

More information

Ordinary Least Squares Regression Explained: Vartanian

Ordinary Least Squares Regression Explained: Vartanian Ordinary Least Squares Regression Explained: Vartanian When to Use Ordinary Least Squares Regression Analysis A. Variable types. When you have an interval/ratio scale dependent variable.. When your independent

More information

General Principles Within-Cases Factors Only Within and Between. Within Cases ANOVA. Part One

General Principles Within-Cases Factors Only Within and Between. Within Cases ANOVA. Part One Within Cases ANOVA Part One 1 / 25 Within Cases A case contributes a DV value for every value of a categorical IV It is natural to expect data from the same case to be correlated - NOT independent For

More information

Inference in Regression Analysis

Inference in Regression Analysis Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value

More information

Course Review. Kin 304W Week 14: April 9, 2013

Course Review. Kin 304W Week 14: April 9, 2013 Course Review Kin 304W Week 14: April 9, 2013 1 Today s Outline Format of Kin 304W Final Exam Course Review Hand back marked Project Part II 2 Kin 304W Final Exam Saturday, Thursday, April 18, 3:30-6:30

More information

Factorial designs. Experiments

Factorial designs. Experiments Chapter 5: Factorial designs Petter Mostad mostad@chalmers.se Experiments Actively making changes and observing the result, to find causal relationships. Many types of experimental plans Measuring response

More information

BIOMETRICS INFORMATION

BIOMETRICS INFORMATION BIOMETRICS INFORMATION Index of Pamphlet Topics (for pamphlets #1 to #60) as of December, 2000 Adjusted R-square ANCOVA: Analysis of Covariance 13: ANCOVA: Analysis of Covariance ANOVA: Analysis of Variance

More information

Correlation and Linear Regression

Correlation and Linear Regression Correlation and Linear Regression Correlation: Relationships between Variables So far, nearly all of our discussion of inferential statistics has focused on testing for differences between group means

More information

A Re-Introduction to General Linear Models

A Re-Introduction to General Linear Models A Re-Introduction to General Linear Models Today s Class: Big picture overview Why we are using restricted maximum likelihood within MIXED instead of least squares within GLM Linear model interpretation

More information

2.1 Linear regression with matrices

2.1 Linear regression with matrices 21 Linear regression with matrices The values of the independent variables are united into the matrix X (design matrix), the values of the outcome and the coefficient are represented by the vectors Y and

More information

Data Analyses in Multivariate Regression Chii-Dean Joey Lin, SDSU, San Diego, CA

Data Analyses in Multivariate Regression Chii-Dean Joey Lin, SDSU, San Diego, CA Data Analyses in Multivariate Regression Chii-Dean Joey Lin, SDSU, San Diego, CA ABSTRACT Regression analysis is one of the most used statistical methodologies. It can be used to describe or predict causal

More information

Overview. Overview. Overview. Specific Examples. General Examples. Bivariate Regression & Correlation

Overview. Overview. Overview. Specific Examples. General Examples. Bivariate Regression & Correlation Bivariate Regression & Correlation Overview The Scatter Diagram Two Examples: Education & Prestige Correlation Coefficient Bivariate Linear Regression Line SPSS Output Interpretation Covariance ou already

More information

Homework 2: Simple Linear Regression

Homework 2: Simple Linear Regression STAT 4385 Applied Regression Analysis Homework : Simple Linear Regression (Simple Linear Regression) Thirty (n = 30) College graduates who have recently entered the job market. For each student, the CGPA

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Example. Multiple Regression. Review of ANOVA & Simple Regression /749 Experimental Design for Behavioral and Social Sciences

Example. Multiple Regression. Review of ANOVA & Simple Regression /749 Experimental Design for Behavioral and Social Sciences 36-309/749 Experimental Design for Behavioral and Social Sciences Sep. 29, 2015 Lecture 5: Multiple Regression Review of ANOVA & Simple Regression Both Quantitative outcome Independent, Gaussian errors

More information

1. (Problem 3.4 in OLRT)

1. (Problem 3.4 in OLRT) STAT:5201 Homework 5 Solutions 1. (Problem 3.4 in OLRT) The relationship of the untransformed data is shown below. There does appear to be a decrease in adenine with increased caffeine intake. This is

More information

Chapter 14. Simple Linear Regression Preliminary Remarks The Simple Linear Regression Model

Chapter 14. Simple Linear Regression Preliminary Remarks The Simple Linear Regression Model Chapter 14 Simple Linear Regression 14.1 Preliminary Remarks We have only three lectures to introduce the ideas of regression. Worse, they are the last three lectures when you have many other things academic

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

LECTURE 5. Introduction to Econometrics. Hypothesis testing

LECTURE 5. Introduction to Econometrics. Hypothesis testing LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will

More information

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay Lecture 5: Multivariate Multiple Linear Regression The model is Y n m = Z n (r+1) β (r+1) m + ɛ

More information