GLM Repeated Measures Notation The GLM (general linear model) procedure provides analysis of variance when the same measurement or measurements are made several times on each subject or case (repeated measures). Algorithms for GLM not discussed in this chapter are in GLM Univariate and Multivariate. Distribution functions are discussed in Appendix 1. The notation listed in the GLM chapter is used in this chapter. Additional notation conventions are defined below: t The number of within-subjects factors. c The number of measures. r k The number of levels of the kth within-subjects factor. r, k k = 1, K, t. M k The contrast matrix of the kth within-subjects factor, k = 1, K, t. It is a square matrix with dimension r k. Each element in the first column is usually equal to 1/r k. For a polynomial contrast each element is 1/ r k, or, for a user-specified contrast, a non-zero constant The other columns have zero column sums. Number of Variables t It is required that c rk = r, the number of dependent variables in the model. k= 1 Covariance Structure As usual in GLM, the data matrix is related to the parameter matrix B as Y = XB+ E. The rows of E are uncorrelated and the ith row has the distribution 1 Nr( 0, wi Σ ). Repeated measures analysis has two additional assumptions: 1
GLM Repeated Measures Σ = ΣC Σ1 L Σt where Σ C is the covariance matrix of the measures and is the Kronecker product operator. The Huynh and Feldt (1970) condition: Suppose σ05 k rs is the 0rs, 5-th element of Σ k 1k = 1, K, t6; then σ05 rr k + σ05 ss k σ05 rs k = constant for r s. Matrices satisfying this condition result in orthonormally transformed variables with spherical covariance matrices; for this reason, the assumption is sometimes referred to as the sphericity assumption. A matrix that has the property of compound symmetry (that is, identical diagonal elements and identical offdiagonal elements) automatically satisfies this assumption. Tests on the Between-Subjects Effects Procedure The procedure for testing the hypothesis of no between-subjects effects uses the following steps: 1. Compute M = Ic M11 ; L Mt; 1 where M k;1 is the first column of the contrast matrix M k of the kth within-subjects factors. Note that M is an r c matrix.. For each of the between-subjects effects including the intercept, get the L matrix, according to the specified type of sum of squares. 3. Compute S H = ( LBM $ ) 0LGL 5 ( LBM $ ) and S E = M SM. Both are c c matrices. 4. Compute the four multivariate test statistics: Wilks lambda, Pillai s trace, Hotelling-Lawley trace, Roy s largest root, and the corresponding significance levels. Also compute the individual univariate F statistics. 5. Repeat steps to 4 until all between-subjects effects have been tested. Multivariate Tests on the Within-Subjects Effects Procedure The procedure for testing the hypothesis of no within-subjects effects uses the following steps:
GLM Repeated Measures 3 1. For the kth within-subjects factor, compute M = Ic A1 L Atwhere Ak = Mk;: r k which is the second-to-last column of M k when the kth withinsubjects factor is involved in the effect. Otherwise, Ak = Mk;1. Note that M is an r cd matrix, where d is the number of columns in the Kronecker product A1... At. In general, d > 1.. For each of the between-subjects effects, get the L matrix, according to the specified type of sum of squares. 3. Compute S H = ( LBM $ ) 0LGL 5 ( LBM $ ) and S E = M SM. Both are cd cd matrices. 4. Compute the four multivariate test statistics: Wilks lambda, Pillai s trace, Hotelling-Lawley trace, Roy s largest root, and the corresponding significance levels. Also compute the individual univariate F statistics. 5. Repeat steps to 4 for the next between-subjects effect. When all the betweensubjects effects are used, go to step 6. 6. Repeat steps 1 to 5 until all within-subjects effects have been tested. Averaged Tests on the Within-Subjects Effects Procedure The procedure for the averaged test of the hypothesis of no within-subjects effects uses the following steps: 1. Take M k (k = 1, K, t) the equally spaced polynomial contrast matrix.. Compute M = Ic A1 L At where Ak = Mk;: r k which is the nd to last column of M k when the kth within-subjects factor is involved in the effect. Otherwise, Ak = 1r r k k. Note that M is an r cd matrix, where d is the number of columns in the Kronecker product A1... At. In general, d > 1. 3. For each of the between-subjects effects, get the L matrix, according to the specified type of sum of squares. 4. Compute S H = ( LBM $ ) 0LGL 5 ( LBM $ ) and S E = M SM. Both are cd cd matrices. 5. Partition S H into c block matrices each of dimension d d. The (k,l)th block, denoted as S H;k,l, (k = 1, K, c and l = 1, K, c), is a sub-matrix of S H from row k 1 d+ 1 1 + 1 0 5 to row kd, and from column 0l 5 d to column ld. Form the
4 GLM Repeated Measures c c matrix, denoted bys H, whose (k, l)th element is the trace of S H;k,l. The matrix S E is obtained similarly. 6. Use S H and S E for computing the four multivariate test statistics: Wilks lambda, Pillai s trace, Hotelling-Lawley trace, Roy s largest root, and the corresponding significance levels. Note: Set the degrees of freedom for S H (same as the row dimension of L in the test procedure) equal to dr L and that for S E equal to d1n r X 6 in the computations. Also compute the individual univariate F statistics and their significance levels. 7. Repeat steps 3 to 6 for each between-subjects effect. When all the betweensubjects effects are used, go to step 8. 8. Repeat steps to 7 until all within-subjects effects have been tested. Adjustments to Degrees of Freedom of the F Statistics The adjustments to degrees of freedom of the univariate F test statistics are the Greenhouse-Geisser epsilon, the Huynh-Feldt epsilon, and the lower-bound epsilon. Greenhouse-Geisser epsilon 1 6 trace( SE ) ε GG = d trace( SE ) Huynh-Feldt epsilon ε HF = min ndε GG, 1 dn ( rx ) d ε GG Lower bound epsilon ε LB = 1 d For any of the above three epsilons, the adjusted significance level is 1 6 L X7 1 CDF.F F, εdr, εd n r where ε is one of the above three epsilons.
GLM Repeated Measures 5 Mauchly s Test of Sphericity Mauchly s test of sphericity is displayed for every repeated measures model. Hypotheses In Mauchly s test of sphericity the null hypothesis is H o : M ΣM= σ I m, versus the alternative hypothesis H1: M ΣM σ Im, where σ > 0 is unspecified, I is an m m identity matrix, and M is the r m orthonormal matrix associated with a within-subjects effect. M is generated using equally spaced polynomial contrasts applied to the within-subjects factors (see the descriptions in the section Averaged Tests on the Within-Subjects Effects on p. 3). Mauchly s W Statistic W = % Ξ & K1trace05 Ξ 6 SYSMIS ' K m m 0 5 if trace Ξ > 0 05 if trace Ξ 0 4 9 4 9 is the r r where Ξ= MAM and A = Y XB$ W Y XB$ sums of squares and cross products. matrix of residual Chi-Square Approximation When n is large and under the null hypothesis that for n r X 1 and m, 1 6 7 4 9 4 4 9 4 99 4 9 3 Pr ρ n rx logw c = Pr χ f c + ω Pr χ f+ 4 c Pr χ f c + O n where
6 GLM Repeated Measures 0 5 f = m m+ 1 1 7 4 9 1 6 ρ = 1 m + m+ 6m n rx 3 0m+ 50m 150 m 54m + 6m + 3m+ 9 ω = 88m n r ρ 1 X6 Chi-Square Statistic % & ' n rx W W c = ρ1 6log if > 0 SYSMIS otherwise Degrees of Freedom 0 5 f = mm+ 1 1 Significance 0 5 0 5 0 5 7 1 CDF. CHISQ c, f ω CDF. CHISQ c, f + 4 CDF. CHISQ c, f References where CDF.CHISQ is the SPSS function for cumulative chi-square distribution. The significance will be reset to zero in case the computed value is less than zero due to floating point imprecision. Greenhouse, S. W., and Geisser, S. 1959. On methods in the analysis of profile data. Psychometrika, 4(): 95 111. Huynh, H., and Feldt, L. S. 1970. Conditions under which mean square ratios in repeated measurements designs have exact F-distributions. Journal of the American Statistical Association, 65: 158 1589. Huynh, H., and Feldt, L. S. 1976. Estimation of the Box correction for degrees of freedom from sample data in the randomized block and split plot designs. Journal of Educational Statistics, 1: 69 8.
GLM Repeated Measures 7 Mauchly, John W. 1940. Significance test for sphericity of a normal n-variate distribution. Annals of Mathematical Statistics, 11: 04 09. Searle, S. R. 198. Matrix algebra useful for statistics. New Work: John Wiley & Sons, Inc.