Lecture 13: Simple Linear Regression in Matrix Format

Similar documents
Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices

Linear Models Review

Lecture 9 SLR in Matrix Form

MATH 320, WEEK 7: Matrices, Matrix Operations

Basic Linear Algebra in MATLAB

Linear Algebra Review

A primer on matrices

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

22 Approximations - the method of least squares (1)

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

A primer on matrices

Math (P)Review Part II:

The Hilbert Space of Random Variables

. a m1 a mn. a 1 a 2 a = a n

Chapter 5 Matrix Approach to Simple Linear Regression

Math 308 Midterm Answers and Comments July 18, Part A. Short answer questions

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

M. Matrices and Linear Algebra

Approximations - the method of least squares (1)

Dot Products, Transposes, and Orthogonal Projections

Math 291-2: Lecture Notes Northwestern University, Winter 2016

The Multivariate Gaussian Distribution [DRAFT]

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Coordinate systems and vectors in three spatial dimensions

Regression. Oscar García

Stat 206: Sampling theory, sample moments, mahalanobis

Stat 206: Linear algebra

Business Statistics. Lecture 9: Simple Regression

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Review from Bootcamp: Linear Algebra

REVIEW OF DIFFERENTIAL CALCULUS

Sometimes the domains X and Z will be the same, so this might be written:

Math 123, Week 2: Matrix Operations, Inverses

POL 213: Research Methods

Lecture 10: F -Tests, ANOVA and R 2

Partial Fractions. June 27, In this section, we will learn to integrate another class of functions: the rational functions.

3 Fields, Elementary Matrices and Calculating Inverses

STAT Homework 8 - Solutions

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Linear Algebra for Beginners Open Doors to Great Careers. Richard Han

Finite Mathematics : A Business Approach

1 Matrices and matrix algebra

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

Lecture 6 Positive Definite Matrices

Linear Algebra, Summer 2011, pt. 2

Lecture 24: Weighted and Generalized Least Squares

Answers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3

a11 a A = : a 21 a 22

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

Statistics 910, #5 1. Regression Methods

Notes on Linear Algebra I. # 1

Tangent spaces, normals and extrema

[y i α βx i ] 2 (2) Q = i=1

MATH 310, REVIEW SHEET 2

Example: 2x y + 3z = 1 5y 6z = 0 x + 4z = 7. Definition: Elementary Row Operations. Example: Type I swap rows 1 and 3

Getting Started with Communications Engineering

1 GSW Sets of Systems

STA 2101/442 Assignment Four 1

STAT 350: Geometry of Least Squares

Reference: Davidson and MacKinnon Ch 2. In particular page

Designing Information Devices and Systems I Fall 2018 Lecture Notes Note Positioning Sytems: Trilateration and Correlation

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

1 Review of the dot product

Image Registration Lecture 2: Vectors and Matrices

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

1 Least Squares Estimation - multiple regression.

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1

Lecture Notes 1: Vector spaces

Math 4310 Solutions to homework 1 Due 9/1/16

STEP Support Programme. STEP 2 Matrices Topic Notes

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

Vector calculus background

Topics in linear algebra

1 Matrices and Systems of Linear Equations. a 1n a 2n

Linear Methods for Prediction

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

Lecture 9: Elementary Matrices

Ordinary Least Squares Linear Regression

Some Notes on Linear Algebra

Gaussian elimination

Topic 15 Notes Jeremy Orloff

The Matrix Algebra of Sample Statistics

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Linear Regression 9/23/17. Simple linear regression. Advertising sales: Variance changes based on # of TVs. Advertising sales: Normal error?

Notes on Random Variables, Expectations, Probability Densities, and Martingales

Vectors and Matrices Statistics with Vectors and Matrices

Mathematics for Graphics and Vision

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Precalculus idea: A picture is worth 1,000 words

Vector, Matrix, and Tensor Derivatives

Factoring. there exists some 1 i < j l such that x i x j (mod p). (1) p gcd(x i x j, n).

Solving Quadratic Equations

5 More on Linear Algebra

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

The General Linear Model in Functional MRI

EXAM 2 REVIEW DAVID SEAL

Stat751 / CSI771 Midterm October 15, 2015 Solutions, Comments. f(x) = 0 otherwise

Transcription:

See updates and corrections at http://www.stat.cmu.edu/~cshalizi/mreg/ Lecture 13: Simple Linear Regression in Matrix Format 36-401, Section B, Fall 2015 13 October 2015 Contents 1 Least Squares in Matrix Form 2 1.1 The Basic Matrices.......................... 2 1.2 Mean Squared Error......................... 3 1.3 Minimizing the MSE......................... 4 2 Fitted Values and Residuals 5 2.1 Residuals............................... 7 2.2 Expectations and Covariances.................... 7 3 Sampling Distribution of Estimators 8 4 Derivatives with Respect to Vectors 9 4.1 Second Derivatives.......................... 11 4.2 Maxima and Minima......................... 11 5 Expectations and Variances with Vectors and Matrices 12 6 Further Reading 13 1

2 So far, we have not used any notions, or notation, that goes beyond basic algebra and calculus (and probability). This has forced us to do a fair amount of book-keeping, as it were by hand. This is just about tolerable for the simple linear model, with one predictor variable. It will get intolerable if we have multiple predictor variables. Fortunately, a little application of linear algebra will let us abstract away from a lot of the book-keeping details, and make multiple linear regression hardly more complicated than the simple version 1. These notes will not remind you of how matrix algebra works. However, they will review some results about calculus with matrices, and about expectations and variances with vectors and matrices. Throughout, bold-faced letters will denote matrices, as a as opposed to a scalar a. 1 Least Squares in Matrix Form Our data consists of n paired observations of the predictor variable X and the response variable Y, i.e., (x 1, y 1 ),... (x n, y n ). We wish to fit the model Y = β 0 + β 1 X + ɛ (1) where E [ɛ X = x] = 0, Var [ɛ X = x] = σ 2, and ɛ is uncorrelated across measurements 2. 1.1 The Basic Matrices Group all of the observations of the response into a single column (n 1) matrix y, y 1 y 2 y = (2). y n Similarly, we group both the coefficients into a single vector (i.e., a 2 1 matrix) [ ] β0 β = (3) β 1 We d also like to group the observations of the predictor variable together, but we need something which looks a little unusual at first: 1 x 1 1 x 2 x =. (4). 1 x n 1 Historically, linear models with multiple predictors evolved before the use of matrix algebra for regression. You may imagine the resulting drudgery. 2 When I need to also assume that ɛ is Gaussian, and strengthen uncorrelated to independent, I ll say so.

3 1.2 Mean Squared Error This is an n 2 matrix, where the first column is always 1, and the second column contains the actual observations of X. We have this apparently redundant first column because of what it does for us when we multiply x by β: xβ = β 0 + β 1 x 1 β 0 + β 1 x 2. β 0 + β 1 x n That is, xβ is the n 1 matrix which contains the point predictions. The matrix x is sometimes called the design matrix. 1.2 Mean Squared Error At each data point, using the coefficients β results in some error of prediction, so we have n prediction errors. These form a vector: (5) e(β) = y xβ (6) (You can check that this subtracts an n 1 matrix from an n 1 matrix.) When we derived the least squares estimator, we used the mean squared error, MSE(β) = 1 n e 2 i (β) (7) n How might we express this in terms of our matrices? I claim that the correct form is MSE(β) = 1 n et e (8) i=1 To see this, look at what the matrix multiplication really involves: e 1 e 2 [e 1 e 2... e n ]. e n (9) This, clearly equals i e2 i, so the MSE has the claimed form. Let us expand this a little for further use. MSE(β) = 1 n et e (10) = 1 n (y xβ)t (y xβ) (11) = 1 n (yt β T x T )(y xβ) (12) = 1 n ( y T y y T xβ β T x T y + β T x T xβ ) (13)

4 1.3 Minimizing the MSE Notice that (y T xβ) T = β T x T y. Further notice that this is a 1 1 matrix, so y T xβ = β T x T y. Thus MSE(β) = 1 ( y T y 2β T x T y + β T x T xβ ) (14) n 1.3 Minimizing the MSE First, we find the gradient of the MSE with respect to β: MSE(β = 1 ( y T y 2 β T x T y + β T x T xβ ) n (15) = 1 ( 0 2x T y + 2x T xβ ) n (16) = 2 ( x T xβ x T y ) n (17) We now set this to zero at the optimum, β: x T x β x T y = 0 (18) This equation, for the two-dimensional vector β, corresponds to our pair of normal or estimating equations for ˆβ 0 and ˆβ 1. Thus, it, too, is called an estimating equation. Solving, β = (x T x) 1 x T y (19) That is, we ve got one matrix equation which gives us both coefficient estimates. If this is right, the equation we ve got above should in fact reproduce the least-squares estimates we ve already derived, which are of course and ˆβ 1 = c XY s 2 X = xy xȳ x 2 x 2 (20) ˆβ 0 = y ˆβ 1 x (21) Let s see if that s right. As a first step, let s introduce normalizing factors of 1/n into both the matrix products: β = (n 1 x T x) 1 (n 1 x T y) (22) Now let s look at the two factors in parentheses separately, from right to left. y 1 y 2 1 n xt y = 1 [ ] 1 1... 1 n x 1 x 2... x n = 1 [ i y ] i n i x iy i [ ] y = xy. y n (23) (24) (25)

5 Similarly for the other factor: 1 n xt x = 1 [ n n i x i [ ] 1 x = x x 2 i x i i x2 i ] (26) (27) Now we need to take the inverse: ( ) 1 [ 1 1 x 2 x n xt x = x 2 x 2 x 1 = 1 [ ] x 2 x s 2 X x 1 ] (28) (29) Let s multiply together the pieces. (x T x) 1 x T y = 1 [ ] [ ] x 2 x y s 2 X x 1 xy = 1 [ ] x2 y xxy s 2 X xy + xy = 1 [ ] (s 2 X + x 2 )y x(c XY + xȳ) s 2 c X XY = 1 [ ] s 2 x y + x 2 y xc XY x 2 ȳ = s 2 X [ ] y c XY x s 2 X c XY s 2 X c XY (30) (31) (32) (33) (34) which is what it should be. So: n 1 x T y is keeping track of y and xy, and n 1 x T x keeps track of x and x 2. The matrix inversion and multiplication then handles all the bookkeeping to put these pieces together to get the appropriate (sample) variances, covariance, and intercepts. We don t have to remember that any more; we can just remember the one matrix equation, and then trust the linear algebra to take care of the details. 2 Fitted Values and Residuals Remember that when the coefficient vector is β, the point predictions for each data point are xβ. Thus the vector of fitted values, m(x), or m for short, is Using our equation for β, m = x β (35) m = x(x T x) 1 x T y (36)

6 Notice that the fitted values are linear in y. The matrix H x(x T x) 1 x T (37) does not depend on y at all, but does control the fitted values: m = Hy (38) If we repeat our experiment (survey, observation... ) many times at the same x, we get different y every time. But H does not change. The properties of the fitted values are thus largely determined by the properties of H. It thus deserves a name; it s usually called the hat matrix, for obvious reasons, or, if we want to sound more respectable, the influence matrix. Let s look at some of the properties of the hat matrix. 1. Influence Since H is not a function of y, we can easily verify that m i / y j = H ij. Thus, H ij is the rate at which the i th fitted value changes as we vary the j th observation, the influence that observation has on that fitted value. 2. Symmetry It s easy to see that H T = H. 3. Idempotency A square matrix a is called idempotent 3 when a 2 = a (and so a k = a for any higher power k). Again, by writing out the multiplication, H 2 = H, so it s idempotent. Idemopotency, Projection, Geometry Idempotency seems like the most obscure of these properties, but it s actually one of the more important. y and m are n-dimensional vectors. If we project a vector u on to the line in the direction of the length-one vector v, we get vv T u (39) (Check the dimensions: u and v are both n 1, so v T 1 1.) If we group the first two terms together, like so, is 1 n, and v T u is (vv T )u (40) where vv T is the n n project matrix or projection operator for that line. Since v is a unit vector, v T v = 1, and (vv T )(vv T ) = vv T (41) so the projection operator for the line is idempotent. The geometric meaning of idempotency here is that once we ve projected u on to the line, projecting its image on to the same line doesn t change anything. Extending this same reasoning, for any linear subspace of the n-dimensional space, there is always some n n matrix which projects vectors in arbitrary 3 From the Latin idem, same, and potens, power.

7 2.1 Residuals position down into the subspace, and this projection matrix is always idempotent. It is a bit more convoluted to prove that any idempotent matrix is the projection matrix for some subspace, but that s also true. We will see later how to read off the dimension of the subspace from the properties of its projection matrix. 2.1 Residuals The vector of residuals, e, is just Using the hat matrix, Here are some properties of I H: 1. Influence e i / y j = (I H) ij. 2. Symmetry (I H) T = I H. e y x β (42) e = y Hy = (I H)y (43) 3. Idempotency (I H) 2 = (I H)(I H) = I H H + H 2. But, since H is idempotent, H 2 = H, and thus (I H) 2 = (I H). Thus, simplifies to MSE( β) = 1 n yt (I H) T (I H)y (44) MSE( β) = 1 n yt (I H)y (45) 2.2 Expectations and Covariances We can of course consider the vector of random variables Y. By our modeling assumptions, Y = xβ + ɛ (46) where ɛ is an n 1 matrix of random variables, with mean vector 0, and variancecovariance matrix σ 2 I. What can we deduce from this? First, the expectation of the fitted values: E [HY] = HE [Y] (47) = Hxβ + HE [ɛ] (48) = x(x T x) 1 x T xβ + 0 (49) = xβ (50) which is as it should be, since the fitted values are unbiased.

8 Next, the variance-covariance of the fitted values: Var [HY] = Var [H(xβ + ɛ)] (51) using, again, the symmetry and idempotency of H. Similarly, the expected residual vector is zero: = Var [Hɛ] (52) = HVar [ɛ] H T (53) = σ 2 HIH (54) = σ 2 H (55) E [e] = (I H)(xβ + E [ɛ]) = xβ xβ = 0 (56) The variance-covariance matrix of the residuals: Var [e] = Var [(I H)(xβ + ɛ)] (57) = Var [(I H)ɛ] (58) = (I H)Var [ɛ] (I H)) T (59) = σ 2 (I H)(I H) T (60) = σ 2 (I H) (61) Thus, the variance of each residual is not quite σ 2, nor (unless H is diagonal) are the residuals exactly uncorrelated with each other. Finally, the expected MSE is [ ] 1 E n et e (62) which is 1 n E [ ɛ T (I H)ɛ ] (63) We know (because we proved it in the exam) that this must be (n 2)σ 2 /n; we ll see next time how to show this. 3 Sampling Distribution of Estimators Let s now turn on the Gaussian-noise assumption, so the noise terms ɛ i all have the distribution N(0, σ 2 ), and are independent of each other and of X. The vector of all n noise terms, ɛ, is an n 1 matrix. Its distribution is a multivariate Gaussian or multivariate normal 4, with mean vector 0, and 4 Some people write this distribution as MV N, and others also as N. I will stick to the former.

9 variance-covariance matrix σ 2 I. We may use this to get the sampling distribution of the estimator β: β = (x T x) 1 x T Y (64) = (x T x) 1 x T (xβ + ɛ) (65) = (x T x) 1 x T xβ + (x T x) 1 x T ɛ (66) = β + (x T x) 1 x T ɛ (67) Since ɛ is Gaussian and is being multiplied by a non-random matrix, (x T x) 1 x T ɛ is also Gaussian. Its mean vector is while its variance matrix is E [ (x T x) 1 x T ɛ ] = (x T x) 1 x T E [ɛ] = 0 (68) Var [ (x T x) 1 x T ɛ ] = (x T x) 1 x T Var [ɛ] ( (x T x) 1 x T ) T (69) = (x T x) 1 x T σ 2 Ix(x T x) 1 (70) = σ 2 (x T x) 1 x T x(x T x) 1 (71) = σ 2 (x T x) 1 (72) ] Since Var [ β = Var [ (x T x) 1 x T ɛ ] (why?), we conclude that Re-writing slightly, β MV N(β, σ 2 (x T x) 1 ) (73) β MV N(β, σ2 n (n 1 x T x) 1 ) (74) will make it easier to prove to yourself that, according [ ] to this, β 0 and β 1 are both unbiased (which we know is right), that Var ˆβ1 = σ2 n s2 X (which we know [ ] is right) and that Var ˆβ0 = σ2 n (1 + x2 /s 2 X ) (which we know is right). This will [ also give us Cov beta ˆ 0, ˆβ 1 ], which otherwise would be tedious to calculate. I will leave you to show, in a similar way, that the fitted values Hy are multivariate Gaussian, as are the residuals e, and to find both their mean vectors and their variance matrices. 4 Derivatives with Respect to Vectors This is a brief review, not intended as a full course in vector calculus. Consider some scalar function of a vector, say f(x), where x is represented as a p 1 matrix. (Here x is just being used as a place-holder or generic variable; it s not necessarily the design matrix of a regression.) We would like to think about the derivatives of f with respect to x.

10 Derivatives are rates of change; they tell us how rapidly the function changes in response to minute changes in its arguments. Since x is a p 1 matrix, we could also write f(x) = f(x 1, x 1, x p ) (75) This makes it clear that f will have a partial derivative with respect to each component of x. How much does f change when we vary the components? We can find this out by using a Taylor expansion. If we pick some base value of the matrix x 0 and expand around it, f(x) f(x 0 ) + p (x x 0 f ) i x i i=1 x 0 (76) = f(x 0 ) + (x x 0 ) T f(x 0 ) (77) where we define the gradient, f, to be the vector of partial derivatives, f f x 1 f x 2. f x p (78) Notice that this defines f to be a one-column matrix, just as x was taken to be. You may sometimes encounter people who want it to be a one-row matrix; it comes to the same thing, but you may have to track a lot of transposes to make use of their math. All of the properties of the gradient can be proved using those of partial derivatives. Here are some basic ones we ll need. 1. Linearity (af(x) + bg(x)) = a f(x) + b g(x) (79) Proof: Directly from the linearity of partial derivatives. 2. Linear forms If f(x) = x T a, with a not a function of x, then (x T a) = a (80) Proof: f(x) = i x ia i, so f/ x i = a i. Notice that a was already a p 1 matrix, so we don t have to transpose anything to get the derivative. 3. Linear forms the other way If f(x) = bx, with b not a function of x, then (bx) = b T (81) Proof: Once again, f/ x i = b i, but now remember that b was a 1 p matrix, and f is p 1, so we need to transpose.

11 4.1 Second Derivatives 4. Quadratic forms Let c be a p p matrix which is not a function of x, and consider the quadratic form x T cx. (You can check that this is scalar.) The gradient is (x T cx) = (c + c T )x (82) Proof: First, write out the matrix multiplications as explicit sums: x T cx = p j=1 x j p c jk x k = k=1 Now take the derivative with respect to x i. f x i = p p j=1 k=1 p j=1 k=1 p x j c jk x k (83) x j c jk x k x i (84) If j = k = i, the term in the inner sum is 2c ii x i. If j = i but k i, the term in the inner sum is c ik x k. If j i but k = i, we get x j c ji. Finally, if j i and k i, we get zero. The j = i terms add up to (cx) i. The k = i terms add up to (c T x) i. (This splits the 2c ii x i evenly between them.) Thus and f x i = ((c + c T x) i (85) f = (c + c T )x (86) (You can, and should, double check that this has the right dimensions.) 5. Symmetric quadratic forms If c = c T, then 4.1 Second Derivatives x T cx = 2cx (87) The p p matrix of second partial derivatives is called the Hessian. I won t step through its properties, except to note that they, too, follow from the basic rules for partial derivatives. 4.2 Maxima and Minima We need all the partial derivatives to be equal to zero at a minimum or maximum. This means that the gradient must be zero there. At a minimum, the Hessian must be positive-definite (so that moves away from the minimum always increase the function); at a maximum, the Hessian must be negative definite (so moves away always decrease the function). If the Hessian is neither positive nor negative definite, the point is neither a minimum nor a maximum, but a saddle (since moving in some directions increases the function but moving in others decreases it, as though one were at the center of a horse s saddle).

12 5 Expectations and Variances with Vectors and Matrices If we have p random variables, Z 1, Z 2,... Z p, we can grow them into a random vector Z = [Z 1 Z 2... Z p ] T. (That is, the random vector is an n 1 matrix of random variables.) This has an expected value: E [Z] zp(z)dz (88) and a little thought shows E [Z] = E [Z 1 ] E [Z 2 ]. E [Z p ] (89) Since expectations of random scalars are linear, so are expectations of random vectors: when a and b are non-random scalars, If a is a non-random matrix, E [az + bw] = ae [Z] + be [W] (90) E [az] = ae [Z] (91) Every coordinate of a random vector has some covariance with every other coordinate. The variance-covariance matrix of Z is the p p matrix which stores these: Var [Z] Var [Z 1 ] Cov [Z 1, Z 2 ]... Cov [Z 1, Z p ] Cov [Z 2, Z 1 ] Var [Z 2 ]... Cov [Z 2, Z p ].... Cov [Z p, Z 1 ] Cov [Z p, Z 2 ]... Var [Z p ] (92) This inherits properties of ordinary variances and covariances. Just Var [Z] = E [ Z 2] (E [Z]) 2, Var [Z] = E [ ZZ T ] E [Z] (E [Z]) T (93) For a non-random vector a and a non-random scalar b, For a non-random matrix c, Var [a + bz] = b 2 Var [Z] (94) Var [cz] = cvar [Z] c T (95)

13 (Check that the dimensions all conform here: if c is q p, Var [cz] should be q q, and so is the right-hand side.) For a quadratic form, Z T cz, with non-random c, the expectation value is E [ Z T cz ] = E [Z] T ce [Z] + tr cvar [Z] (96) where tr is of course the trace of a matrix, the sum of its diagonal entries. To see this, notice that Z T cz = tr Z T cz (97) because it s a 1 1 matrix. But the trace of a matrix product doesn t change when we cyclic permute the matrices, so Therefore Z T cz = tr czz T (98) E [ Z T cz ] = E [ tr czz T ] (99) = tr E [ czz T ] (100) = tr ce [ ZZ T ] (101) = tr c(var [Z] + E [Z] E [Z] T ) (102) = tr cvar [Z] + tr ce [Z] E [Z] T ) (103) = tr cvar [Z] + tr E [Z] T ce [Z] (104) = tr cvar [Z] + E [Z] T ce [Z] (105) using the fact that tr is a linear operation so it commutes with taking expectations; the decomposition of Var [Z]; the cyclic permutation trick again; and finally dropping tr from a scalar. Unfortunately, there is generally no simple formula for the variance of a quadratic form, unless the random vector is Gaussian. 6 Further Reading Linear algebra is a pre-requisite for this class; I strongly urge you to go back to your textbook and notes for review, if any of this is rusty. If you desire additional resources, I recommend Axler (1996) as a concise but thorough course. Petersen and Pedersen (2012), while not an introduction or even really a review, is an extremely handy compendium of matrix, and matrix-calculus, identities. References Axler, Sheldon (1996). Linear Algebra Done Right. Undergraduate Texts in Mathematics. Berlin: Springer-Verlag.

14 REFERENCES Petersen, Kaare Brandt and Michael Syskind Pedersen (2012). The Matrix Cookbook. Tech. rep., Technical University of Denmark, Intelligent Signal Processing Group. URL http://www2.imm.dtu.dk/pubdb/views/ publication_details.php?id=3274.