Multivariate Linear Models
|
|
- Dorcas Bishop
- 5 years ago
- Views:
Transcription
1 Multivariate Linear Models Stanley Sawyer Washington University November 7, Introduction. Suppose that we have n observations, each of which has d components. For example, we may have d measurements on each of n different individual humans or squirrels, or d test scores for n students on d different days. The observations take the form of a n d matrix Y = Y 11 Y Y 1d Y 21 Y Y 2d Y n1 Y n2... Y nd The i th row corresponds to the i th observation. We assume that we also observe r covariates whose structure is the same for all components. These might be income level, sex, altitude, etc., as in an ANOVA design. In matrix notation, the basic multivariate regression or MANOVA model is Y = Xβ + e (1.1 Here X is an n r matrix (as in the univariate case that relate the observations to the covariates and β = β 11 β β 1d β 21 β β 2d β r1 β r2... β rd is an r d matrix of parameter values. The j th column of β relates to the j th component of the observations Y i. The n d error matrix e in (1.1 is assumed to be joint Gaussian. The rows of e are independent, corresponding to independent vector-valued observations Y i R d. The columns of e corresponding to the components of the observations Y i have a fixed covariance matrix Σ. Thus the errors e satisfy E(e ia = 0 and Cov(e ia, e jb = 0, i = j Cov(e ia, e ib = Σ ab, 1 a, b d (1.2 In general, if A = { a ij } is an r s matrix and B is m n, the tensor product or Kronecker product of the matrices A and B is a matrix C = A B that
2 Multivariate Linear Models can be written in partitioned form C = a 11B a 12B... a 1sB a 21 B a 22 B... a 2s B a r1 B a r2 B... a rs B Note that C is then rm sn. In this notation, the covariate matrix of the error terms can be written Cov(e = I n Σ (1.3 This assumes that we identify the pair ia in e ia with the single integer I = (i 1m + a. The matrix Cov(e can then be written in block diagonal form with the matrix Σ down the diagonal. 2. The MLE of the matrix β. Since the errors e are jointly normal, the likelihood function of the observations Y in (1.1 can be written as L(Y, β = (2π nd/2 exp( S/2 where S = n d d i=1 a=1 b=1 ( Yia (Xβ ia Σ 1 ab ( Yib (Xβ ib Finding the matrix MLE β is equivalent to minimizing the triple sum S. Setting ( / β kc S = 0 implies n d i=1 b=1 X ik Σ 1 cb ( Yib (Xβ ib = d b=1 Σ 1 ( cb (X Y kb (X Xβ kb = 0 This is Σ 1 (X Xβ X Y = 0 in matrix form. Premultiplying by Σ leads to X Xβ = X Y. If the r r matrix X X is invertible, the matrix-valued MLE is then β = (X X 1 X Y (2.1 This is the same formula as in the univariate case (d = 1, but now β and β are r d matrices and Y is n d. That is, the d columns of the multivariate MLE β are found by applying the same r n matrix (X X 1 X as in the univariate case to each of the d columns of Y.
3 Multivariate Linear Models For Q = (X X 1 X, (2.1 and Y = Xβ + e imply β ia = β ia + Thus by (1.2 n Q iu e ua, Q = (X X 1 X u=1 Cov( β ia, β jb = = n n Q iu Q jv Cov(e ua, e vb u=1 v=1 n Q iu Q ju Σ ab = ( (X X 1 ij Σ ab (2.2 u=1 since QQ = (X X 1 X X(X X 1 = (X X 1. Thus, with the same conventions on indices in Kronecker products as before, Cov( β = (X X 1 Σ 3. Hypothesis Testing. For the regression model Y = Xβ+e, the natural hypothesis H 0 : β ia = 0, 1 a d (3.1a says that the i th covariate does not influence any of the components of Y. A generalization of (3.1a is H 0 : Lβ = 0 (3.1b where L is a q r matrix, so that Lβ is q d. Thus (3.1b is shorthand for q different relations of the form (3.1a. An example would be that a particular MANOVA effect with q degrees of freedom is equal to zero. The analog of the univariate SS term in a multivariate MANOVA is the d d matrix H L = (L β ( L(X X 1 L 1 L β The analog of the residual sum of squares SSE is the d d matrix E = (Y X β (Y X β = Y Y β X X β (3.2a (3.2b The sums in (3.2 are sometimes called SS&CP matrices to emphasize that they are matrices instead of numbers. (SS&CP stands for Sums of Squares and Cross Products. The usual one-dimensional ANOVA test for (3.1 is based on H L /E. Indeed, in the univariate case if H 0 : Lβ = 0 with rank(l = q, then F =
4 Multivariate Linear Models (1/qH L /S for S = E/(n r has an F distribution with (q, n r degrees of freedom. The multivariate analog has to be more complicated, since H L and E are d d matrices and the three matrices E 1 H L H L E 1 E 1/2 H L E 1/2 (3.3 are generally different. However, the eigenvalues of the three matrices (3.3 are the same. This is because all three matrices in (3.3 have the same characteristic polynomial f(λ = det(e 1 H L λi = det(h L λe/ det(e Several common tests of the hypothesis (3.1 are based on the eigenvalues of the matrices (3.3. Since the third matrix in (3.3 is symmetric and nonnegative definite, the eigenvalues λ 1, λ 2,..., λ d are real and nonnegative. The most common tests, and the functions of the λ i that the tests are based upon, are 1. Wilk s Lambda (test: Λ = d 2. Pillai s Trace: S 1 = d i=1 λ i λ i +1 i=1 1 λ i Hotelling-Lawley Trace: S 2 = d i=1 λ i 4. Roy s Greatest Root: S 3 = λ 1 where λ 1 λ 2... λ d 0. The last test is named after the Indian statistician S. N. Roy, so that Roy is not a first name. The likelihood ratio test for the general hypothesis (3.1 is essentially Wilk s Lambda. In general, the four tests give different P-values, and are designed to test H 0 : Lβ = 0 against four slightly different alternatives. However, there is an important special case in which the three matrices (3.3 have a single nonzero eigenvalue and the four tests give identical results. First, we need a regression into a multivariate version of the Student t distribution that is called the Hotelling T The Hotelling T 2 Statistic. Suppose that we have two samples Y 11, Y 12,..., Y 1n1, Y ij R d Y 21, Y 22,..., Y 2n2 The vector-valued observations Y ij are independent N(µ i, Σ. That is, the two samples have the same covariance function Σ but possibly different (vector-valued sample means µ 1 and µ 2.
5 Multivariate Linear Models The standard test of H 0 : µ 1 = µ 2 in the univariate case (d = 1 is based on the statistic t = n1 n 2 n 1 + n 2 ( Y 1 Y 2 / s 2 where s 2 = 1 n 1 + n n i (Y ij Y i 2 i=1 j=1 Given H 0, the statistic t has a Student s t distribution with n 1 + n 2 2 degrees of freedom and can be used as a test of H 0. Hotelling s generalization of t is T 2 = n 1n 2 n 1 + n 2 (Y 1 Y 2 S 1 (Y 1 Y 2 S = 1 n 1 + n 2 2 where 2 n i (Y ij Y i (Y ij Y i (4.1 i=1 j=1 The statistic T 2 is called the Hotelling T 2 -statistic. Most books on multivariate statistics show that (i the likelihood ratio test of H 0 : µ 1 = µ 2 is based on T 2 and (ii given H 0, the distribution of T 2 is given by T 2 d(n 1 + n 2 2 n 1 + n 2 d 1 F d, n 1 +n 2 d 1 That is, the distribution of T 2 is that constant times an F distribution with those numbers of degrees of freedom. 5. Hotelling and Wishart Distributions. In general, a random variable X is said to have a Hotelling s T 2 distribution with parameters (d, m (abbreviated T 2 (d, m if X has the same distribution as Y S 1 Y, where S = (1/m m a=1 Z az a and Y, Z 1,..., Z m are m + 1 independent random vectors with the same distribution N(0, Σ. The form of Y S 1 Y indicates that the distribution of X is independent of Σ. It can then be shown that T 2 in (4.1 has a T 2 (d, n 1 + n 2 2 distribution and that X T 2 (d, m dm m d + 1 F d, m d+1 (5.1 A random variable W is said to have a Wishart distribution with parameters Σ and m (abbreviated W (Σ, m if W has the same distribution as m a=1 Z az a, where Z 1,..., Z m are independent normally-distributed
6 Multivariate Linear Models random vectors with distribution N(0, Σ. This is a multivariate generalization of the chi-square distribution, but also depends on a matrix Σ. Since N a = Σ 1/2 Z a are independent N(0, I d, we have that W (Σ, m Σ 1/2 W (I d, mσ 1/2. Thus X has a Hotelling T 2 (d, m distribution if and only if X is distributed as Y W 1 Y, where Y is N(0, Σ, W has a Wishart distribution W (Σ, m, and Y and W are independent. 6. Hypotheses of Rank One. Suppose in the hypothesis H 0 : Lβ = 0 (6.1 for a q r matrix L that rank(l = q = 1. Section 3. Then L = h is a row vector and (See equations (3.1ab in H L = H h = (h β ( h (X X 1 h 1 h β = ch x h x h (6.2 where x h = (h β = β h and c h = 1/ ( h (X X 1 h If follows from (2.2 that Cov((x h a, (x h b = = r r h i h j Cov( β ia, β jb i=1 j=1 r i=1 j=1 r ( h i h j (X X 1 Σ ij ab = Σ ab /c h Thus, given H 0 : h β = 0, Z 0 = x h ch is N(0, Σ and H h = Z 0 Z 0. Similarly A = E 1 H L = E 1 H h = ( c h E 1 x h xh = u h x h (6.3 where E = (Y X β(y X β as in (3.2b. In general, the three matrices (3.3 all have rank one. In general, if A = cuv is a rank-one matrix where c is a nonzero constant, then Az = λz implies cu(v z = λz. If λ = 0, then v z = 0. If λ 0, then v z 0 and z is proportional to u. Then Au = λu implies cu(v u = λu. This implies that if A = cuv is rank one, then the sole nonzero eigenvalue is λ = cv u.
7 Multivariate Linear Models For A in (6.3 or for any of the matrices (3.3, the only nonzero eigenvalue is λ 1 (H, E = c h y h y h = c h (E 1/2 x h E 1/2 x h = c h x he 1 x h = h βe 1 β h / h (X X 1 h (6.4 Recall that x h ch = Z 0 has a N(0, Σ distribution if h β = 0. As in the univariate case, it can be shown that β is independent of the residuals Y X β and that the residual matrix E = (Y X β(y X β has a Wishart distribution W (Σ, n r. It follows from (6.4 that, if rank(h = 1, then (n rλ 1 (H, E has a Hotelling T 2 (d, n r distribution, which by (5.1 has an F distribution within a normalizing constant. Indeed, by (5.1, if h β = 0, λ 1 (H, E T 2 (d, n r n r d n r d + 1 F d, n r d+1 (6.5 It follows from the same argument that the eigenvector of A = E 1 H L for the eigenvalue λ 1 is y h = c h E 1 x h = c h E 1 β h. Example 1. For H 0 : β ia = 0 (1 a d, the nonzero eigenvalue is λ 1 (H, E = β i E 1 β i / ( (X X 1 ii If E is replaced by S = E/(n r and d = 1, this reduces to the square of the usual t statistic. If d > 1, it has a Hotelling T 2 (d, n r distribution. Example 2. Suppose that we have two samples Y 11, Y 12,..., Y 1n1 and Y 21, Y 22,..., Y 2n2 where Y ij R d and have the distribution N(µ i, Σ. This is equivalent to Y = Xβ + e where Y R n for n = n 1 + n 2, β = (µ 1 µ 2, and X is the n 2 matrix with X ij = 1 for j = 1, 1 i n 1 or j = 2, n i n 2, with otherwise X ij = 0. In particular ( X n1 0 X = 0 n 2 ( β = (X X 1 X Y Y = 1 Y 2 (6.6 Thus the normalized residual matrix S = E/(n 2 in (3.2b is the same as the pooled two-sample covariance matrix S in (4.1. Since β = (µ 1 µ 2,
8 Multivariate Linear Models the hypothesis H 0 : µ 1 = µ 2 is the same as H 0 : h β = 0 for h = (1 1. By (6.6, c 1 h = h (X X 1 h = (1/n 1 + (1/n 2 and Thus λ 1 (H, E = h βe 1 β h / h (X X 1 h = 1 n 1 + n 2 2 n 1 n 2 n 1 + n 2 (Y 1 Y 2 S 1 (Y 1 Y 2 (n 1 + n 2 2λ 1 (H, E = n 1n 2 n 1 + n 2 (Y 1 Y 2 S 1 (Y 1 Y 2 T 2 (2, n 1 + n 2 2 Hence any test based on λ 1 (H, E is equivalent to the Hotelling T 2 -test of Section 4. Recall that the eigenvector for λ 1 (E, H for A = E 1 H is x h = c h E 1 β h = (n 1 n 2 /(n 1 + n 2 E 1 (Y 1 Y 2. Within a constant, this is the same vector as appears in the discriminant function l(x = l 0 + X S 1 (Y 1 Y 2, S = E/(n 1 + n 2 2 in linear discriminant analysis (LDA; q.v..
Chapter 7, continued: MANOVA
Chapter 7, continued: MANOVA The Multivariate Analysis of Variance (MANOVA) technique extends Hotelling T 2 test that compares two mean vectors to the setting in which there are m 2 groups. We wish to
More informationMultivariate Regression (Chapter 10)
Multivariate Regression (Chapter 10) This week we ll cover multivariate regression and maybe a bit of canonical correlation. Today we ll mostly review univariate multivariate regression. With multivariate
More informationI L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Comparisons of Several Multivariate Populations Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS
More informationMultilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2
Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do
More informationz = β βσβ Statistical Analysis of MV Data Example : µ=0 (Σ known) consider Y = β X~ N 1 (β µ, β Σβ) test statistic for H 0β is
Example X~N p (µ,σ); H 0 : µ=0 (Σ known) consider Y = β X~ N 1 (β µ, β Σβ) H 0β : β µ = 0 test statistic for H 0β is y z = β βσβ /n And reject H 0β if z β > c [suitable critical value] 301 Reject H 0 if
More informationSTAT 730 Chapter 5: Hypothesis Testing
STAT 730 Chapter 5: Hypothesis Testing Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 28 Likelihood ratio test def n: Data X depend on θ. The
More informationApplied Multivariate Statistical Modeling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur
Applied Multivariate Statistical Modeling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur Lecture - 29 Multivariate Linear Regression- Model
More informationComparisons of Several Multivariate Populations
Comparisons of Several Multivariate Populations Edps/Soc 584, Psych 594 Carolyn J Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees,
More informationMultivariate analysis of variance and covariance
Introduction Multivariate analysis of variance and covariance Univariate ANOVA: have observations from several groups, numerical dependent variable. Ask whether dependent variable has same mean for each
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 17 for Applied Multivariate Analysis Outline Multivariate Analysis of Variance 1 Multivariate Analysis of Variance The hypotheses:
More informationMULTIVARIATE ANALYSIS OF VARIANCE
MULTIVARIATE ANALYSIS OF VARIANCE RAJENDER PARSAD AND L.M. BHAR Indian Agricultural Statistics Research Institute Library Avenue, New Delhi - 0 0 lmb@iasri.res.in. Introduction In many agricultural experiments,
More informationMANOVA MANOVA,$/,,# ANOVA ##$%'*!# 1. $!;' *$,$!;' (''
14 3! "#!$%# $# $&'('$)!! (Analysis of Variance : ANOVA) *& & "#!# +, ANOVA -& $ $ (+,$ ''$) *$#'$)!!#! (Multivariate Analysis of Variance : MANOVA).*& ANOVA *+,'$)$/*! $#/#-, $(,!0'%1)!', #($!#$ # *&,
More informationAnalysis of variance, multivariate (MANOVA)
Analysis of variance, multivariate (MANOVA) Abstract: A designed experiment is set up in which the system studied is under the control of an investigator. The individuals, the treatments, the variables
More informationExample 1 describes the results from analyzing these data for three groups and two variables contained in test file manova1.tf3.
Simfit Tutorials and worked examples for simulation, curve fitting, statistical analysis, and plotting. http://www.simfit.org.uk MANOVA examples From the main SimFIT menu choose [Statistcs], [Multivariate],
More informationRandom Matrices and Multivariate Statistical Analysis
Random Matrices and Multivariate Statistical Analysis Iain Johnstone, Statistics, Stanford imj@stanford.edu SEA 06@MIT p.1 Agenda Classical multivariate techniques Principal Component Analysis Canonical
More informationEconomics 620, Lecture 5: exp
1 Economics 620, Lecture 5: The K-Variable Linear Model II Third assumption (Normality): y; q(x; 2 I N ) 1 ) p(y) = (2 2 ) exp (N=2) 1 2 2(y X)0 (y X) where N is the sample size. The log likelihood function
More informationGLM Repeated Measures
GLM Repeated Measures Notation The GLM (general linear model) procedure provides analysis of variance when the same measurement or measurements are made several times on each subject or case (repeated
More informationAn Introduction to Multivariate Statistical Analysis
An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents
More informationMANOVA is an extension of the univariate ANOVA as it involves more than one Dependent Variable (DV). The following are assumptions for using MANOVA:
MULTIVARIATE ANALYSIS OF VARIANCE MANOVA is an extension of the univariate ANOVA as it involves more than one Dependent Variable (DV). The following are assumptions for using MANOVA: 1. Cell sizes : o
More informationMultivariate Analysis of Variance
Chapter 15 Multivariate Analysis of Variance Jolicouer and Mosimann studied the relationship between the size and shape of painted turtles. The table below gives the length, width, and height (all in mm)
More informationLecture 5: Hypothesis tests for more than one sample
1/23 Lecture 5: Hypothesis tests for more than one sample Måns Thulin Department of Mathematics, Uppsala University thulin@math.uu.se Multivariate Methods 8/4 2011 2/23 Outline Paired comparisons Repeated
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationTHE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay
THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay Lecture 5: Multivariate Multiple Linear Regression The model is Y n m = Z n (r+1) β (r+1) m + ɛ
More informationRepeated Measures Part 2: Cartoon data
Repeated Measures Part 2: Cartoon data /*********************** cartoonglm.sas ******************/ options linesize=79 noovp formdlim='_'; title 'Cartoon Data: STA442/1008 F 2005'; proc format; /* value
More informationApplied Multivariate and Longitudinal Data Analysis
Applied Multivariate and Longitudinal Data Analysis Chapter 2: Inference about the mean vector(s) II Ana-Maria Staicu SAS Hall 5220; 919-515-0644; astaicu@ncsu.edu 1 1 Compare Means from More Than Two
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationANOVA Longitudinal Models for the Practice Effects Data: via GLM
Psyc 943 Lecture 25 page 1 ANOVA Longitudinal Models for the Practice Effects Data: via GLM Model 1. Saturated Means Model for Session, E-only Variances Model (BP) Variances Model: NO correlation, EQUAL
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationOther hypotheses of interest (cont d)
Other hypotheses of interest (cont d) In addition to the simple null hypothesis of no treatment effects, we might wish to test other hypothesis of the general form (examples follow): H 0 : C k g β g p
More information4.1 Computing section Example: Bivariate measurements on plants Post hoc analysis... 7
Master of Applied Statistics ST116: Chemometrics and Multivariate Statistical data Analysis Per Bruun Brockhoff Module 4: Computing 4.1 Computing section.................................. 1 4.1.1 Example:
More informationWITHIN-PARTICIPANT EXPERIMENTAL DESIGNS
1 WITHIN-PARTICIPANT EXPERIMENTAL DESIGNS I. Single-factor designs: the model is: yij i j ij ij where: yij score for person j under treatment level i (i = 1,..., I; j = 1,..., n) overall mean βi treatment
More informationMultivariate Linear Regression Models
Multivariate Linear Regression Models Regression analysis is used to predict the value of one or more responses from a set of predictors. It can also be used to estimate the linear association between
More informationStevens 2. Aufl. S Multivariate Tests c
Stevens 2. Aufl. S. 200 General Linear Model Between-Subjects Factors 1,00 2,00 3,00 N 11 11 11 Effect a. Exact statistic Pillai's Trace Wilks' Lambda Hotelling's Trace Roy's Largest Root Pillai's Trace
More information5 Inferences about a Mean Vector
5 Inferences about a Mean Vector In this chapter we use the results from Chapter 2 through Chapter 4 to develop techniques for analyzing data. A large part of any analysis is concerned with inference that
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationApplication of Ghosh, Grizzle and Sen s Nonparametric Methods in. Longitudinal Studies Using SAS PROC GLM
Application of Ghosh, Grizzle and Sen s Nonparametric Methods in Longitudinal Studies Using SAS PROC GLM Chan Zeng and Gary O. Zerbe Department of Preventive Medicine and Biometrics University of Colorado
More informationAnalysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED. Maribeth Johnson Medical College of Georgia Augusta, GA
Analysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED Maribeth Johnson Medical College of Georgia Augusta, GA Overview Introduction to longitudinal data Describe the data for examples
More informationTHE IMPACT OF HIGH SCHOOLS OUTPUTS ON STUDENTS ACHIEVEMENT IN THE COLLEGE OF SCIENCES AT AL-HUSSEIN BIN TALAL UNIVERSITY
International Journal of Applied Mathematics Volume 26 No. 1 2013, 19-38 ISSN: 1311-1728 (printed version); ISSN: 1314-8060 (on-line version) doi: http://dx.doi.org/10.12732/ijam.v26i1.3 THE IMPACT OF
More informationNeuendorf MANOVA /MANCOVA. Model: X1 (Factor A) X2 (Factor B) X1 x X2 (Interaction) Y4. Like ANOVA/ANCOVA:
1 Neuendorf MANOVA /MANCOVA Model: X1 (Factor A) X2 (Factor B) X1 x X2 (Interaction) Y1 Y2 Y3 Y4 Like ANOVA/ANCOVA: 1. Assumes equal variance (equal covariance matrices) across cells (groups defined by
More informationJordan Canonical Form
Jordan Canonical Form Massoud Malek Jordan normal form or Jordan canonical form (named in honor of Camille Jordan) shows that by changing the basis, a given square matrix M can be transformed into a certain
More informationThe Multivariate Gaussian Distribution
The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance
More informationCanonical Correlation Analysis of Longitudinal Data
Biometrics Section JSM 2008 Canonical Correlation Analysis of Longitudinal Data Jayesh Srivastava Dayanand N Naik Abstract Studying the relationship between two sets of variables is an important multivariate
More informationANOVA: Analysis of Variance - Part I
ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.
More informationNeuendorf MANOVA /MANCOVA. Model: MAIN EFFECTS: X1 (Factor A) X2 (Factor B) INTERACTIONS : X1 x X2 (A x B Interaction) Y4. Like ANOVA/ANCOVA:
1 Neuendorf MANOVA /MANCOVA Model: MAIN EFFECTS: X1 (Factor A) X2 (Factor B) Y1 Y2 INTERACTIONS : Y3 X1 x X2 (A x B Interaction) Y4 Like ANOVA/ANCOVA: 1. Assumes equal variance (equal covariance matrices)
More informationCovariance Structure Approach to Within-Cases
Covariance Structure Approach to Within-Cases Remember how the data file grapefruit1.data looks: Store sales1 sales2 sales3 1 62.1 61.3 60.8 2 58.2 57.9 55.1 3 51.6 49.2 46.2 4 53.7 51.5 48.3 5 61.4 58.7
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationRepeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models
Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models EPSY 905: Multivariate Analysis Spring 2016 Lecture #12 April 20, 2016 EPSY 905: RM ANOVA, MANOVA, and Mixed Models
More information1. Density and properties Brief outline 2. Sampling from multivariate normal and MLE 3. Sampling distribution and large sample behavior of X and S 4.
Multivariate normal distribution Reading: AMSA: pages 149-200 Multivariate Analysis, Spring 2016 Institute of Statistics, National Chiao Tung University March 1, 2016 1. Density and properties Brief outline
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationSTAT 501 EXAM I NAME Spring 1999
STAT 501 EXAM I NAME Spring 1999 Instructions: You may use only your calculator and the attached tables and formula sheet. You can detach the tables and formula sheet from the rest of this exam. Show your
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationSTA442/2101: Assignment 5
STA442/2101: Assignment 5 Craig Burkett Quiz on: Oct 23 rd, 2015 The questions are practice for the quiz next week, and are not to be handed in. I would like you to bring in all of the code you used to
More informationNeuendorf MANOVA /MANCOVA. Model: X1 (Factor A) X2 (Factor B) X1 x X2 (Interaction) Y4. Like ANOVA/ANCOVA:
1 Neuendorf MANOVA /MANCOVA Model: X1 (Factor A) X2 (Factor B) X1 x X2 (Interaction) Y1 Y2 Y3 Y4 Like ANOVA/ANCOVA: 1. Assumes equal variance (equal covariance matrices) across cells (groups defined by
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationLeast Squares Estimation
Least Squares Estimation Using the least squares estimator for β we can obtain predicted values and compute residuals: Ŷ = Z ˆβ = Z(Z Z) 1 Z Y ˆɛ = Y Ŷ = Y Z(Z Z) 1 Z Y = [I Z(Z Z) 1 Z ]Y. The usual decomposition
More informationTest Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics
Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests
More informationMULTIVARIATE POPULATIONS
CHAPTER 5 MULTIVARIATE POPULATIONS 5. INTRODUCTION In the following chapters we will be dealing with a variety of problems concerning multivariate populations. The purpose of this chapter is to provide
More informationGroup comparison test for independent samples
Group comparison test for independent samples The purpose of the Analysis of Variance (ANOVA) is to test for significant differences between means. Supposing that: samples come from normal populations
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationT. Mark Beasley One-Way Repeated Measures ANOVA handout
T. Mark Beasley One-Way Repeated Measures ANOVA handout Profile Analysis Example In the One-Way Repeated Measures ANOVA, two factors represent separate sources of variance. Their interaction presents an
More informationApplied Multivariate Analysis
Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Discriminant Analysis Background 1 Discriminant analysis Background General Setup for the Discriminant Analysis Descriptive
More informationMultivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation
Multivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation Daniel B Rowe Division of Biostatistics Medical College of Wisconsin Technical Report 40 November 00 Division of Biostatistics
More informationPrincipal component analysis
Principal component analysis Motivation i for PCA came from major-axis regression. Strong assumption: single homogeneous sample. Free of assumptions when used for exploration. Classical tests of significance
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationChapter 2 Multivariate Normal Distribution
Chapter Multivariate Normal Distribution In this chapter, we define the univariate and multivariate normal distribution density functions and then we discuss the tests of differences of means for multiple
More informationLecture 34: Properties of the LSE
Lecture 34: Properties of the LSE The following results explain why the LSE is popular. Gauss-Markov Theorem Assume a general linear model previously described: Y = Xβ + E with assumption A2, i.e., Var(E
More informationLinear Algebra Review
Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and
More informationChapter 9. Multivariate and Within-cases Analysis. 9.1 Multivariate Analysis of Variance
Chapter 9 Multivariate and Within-cases Analysis 9.1 Multivariate Analysis of Variance Multivariate means more than one response variable at once. Why do it? Primarily because if you do parallel analyses
More informationMatrix Approach to Simple Linear Regression: An Overview
Matrix Approach to Simple Linear Regression: An Overview Aspects of matrices that you should know: Definition of a matrix Addition/subtraction/multiplication of matrices Symmetric/diagonal/identity matrix
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationLecture 2: Linear and Mixed Models
Lecture 2: Linear and Mixed Models Bruce Walsh lecture notes Introduction to Mixed Models SISG, Seattle 18 20 July 2018 1 Quick Review of the Major Points The general linear model can be written as y =
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationVectors and Matrices Statistics with Vectors and Matrices
Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc
More informationRandom Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30
Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationGaussian Models (9/9/13)
STA561: Probabilistic machine learning Gaussian Models (9/9/13) Lecturer: Barbara Engelhardt Scribes: Xi He, Jiangwei Pan, Ali Razeen, Animesh Srivastava 1 Multivariate Normal Distribution The multivariate
More informationInferences about a Mean Vector
Inferences about a Mean Vector Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University
More informationStudy Notes on Matrices & Determinants for GATE 2017
Study Notes on Matrices & Determinants for GATE 2017 Matrices and Determinates are undoubtedly one of the most scoring and high yielding topics in GATE. At least 3-4 questions are always anticipated from
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationIntroduction to Matrix Algebra
Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p
More informationSTT 843 Key to Homework 1 Spring 2018
STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ
More informationPrepared by: Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Universiti
Prepared by: Prof. Dr Bahaman Abu Samah Department of Professional Development and Continuing Education Faculty of Educational Studies Universiti Putra Malaysia Serdang Use in experiment, quasi-experiment
More informationNext is material on matrix rank. Please see the handout
B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0
More informationChapter 5 Matrix Approach to Simple Linear Regression
STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:
More informationConfidence Intervals, Testing and ANOVA Summary
Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0
More informationData Analyses in Multivariate Regression Chii-Dean Joey Lin, SDSU, San Diego, CA
Data Analyses in Multivariate Regression Chii-Dean Joey Lin, SDSU, San Diego, CA ABSTRACT Regression analysis is one of the most used statistical methodologies. It can be used to describe or predict causal
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationThe SAS System 18:28 Saturday, March 10, Plot of Canonical Variables Identified by Cluster
The SAS System 18:28 Saturday, March 10, 2018 1 The FASTCLUS Procedure Replace=FULL Radius=0 Maxclusters=2 Maxiter=10 Converge=0.02 Initial Seeds Cluster SepalLength SepalWidth PetalLength PetalWidth 1
More informationChap 2. Linear Classifiers (FTH, ) Yongdai Kim Seoul National University
Chap 2. Linear Classifiers (FTH, 4.1-4.4) Yongdai Kim Seoul National University Linear methods for classification 1. Linear classifiers For simplicity, we only consider two-class classification problems
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationand let s calculate the image of some vectors under the transformation T.
Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =
More informationLecture 6: Single-classification multivariate ANOVA (k-group( MANOVA)
Lecture 6: Single-classification multivariate ANOVA (k-group( MANOVA) Rationale and MANOVA test statistics underlying principles MANOVA assumptions Univariate ANOVA Planned and unplanned Multivariate ANOVA
More informationChapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations
Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2
More informationEcon 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )
Econ 60 Matrix Differentiation Let a and x are k vectors and A is an k k matrix. a x a x = a = a x Ax =A + A x Ax x =A + A x Ax = xx A We don t want to prove the claim rigorously. But a x = k a i x i i=
More informationPOWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE
POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE Supported by Patrick Adebayo 1 and Ahmed Ibrahim 1 Department of Statistics, University of Ilorin, Kwara State, Nigeria Department
More informationSTATS306B STATS306B. Discriminant Analysis. Jonathan Taylor Department of Statistics Stanford University. June 3, 2010
STATS306B Discriminant Analysis Jonathan Taylor Department of Statistics Stanford University June 3, 2010 Spring 2010 Classification Given K classes in R p, represented as densities f i (x), 1 i K classify
More information