Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution

Size: px
Start display at page:

Download "Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution"

Transcription

1 Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution Relevant textbook passages: Larsen Marx [8]: Section Pitman [9]: Section 6.5. If you just need a quick refresher, I recommend my on-line notes [4, 5]. In what follows, vectors in R n are usually considered to be n 1 column matrices, and denotes transposition. To save space I may write a vector a x = (x 1,..., x n ), but you should still think of it as a column vector The geometry of the Euclidean inner product Vectors x and y in R n are orthogonal if x y = 0, written as x y. More generally, for nonzero vectors x and y in a Euclidean space, x y = x y = x y cos θ, where θ is the angle between x and y. To see this, orthogonally project y on the space spanned by x. That is, write y = αx + z where z x = 0. Thus z x = (y αx) x = y x αx x = 0 = α = x y/x x. Referring to Figure 22.1 we see that cos θ = α x / y = x y/ ( x y ). y z θ αx x Figure Dot product and angles: cos θ = α x / y = x y/ ( x y ) Reminder of matrix operations Here is a quick reminder of some of the properties of matrix operations that we will use often. (AB) = B A. 22 1

2 KC Border Linear Algebra and Multivariate Normal 22 2 A(B + C)D = ABD + ACD. Operations can proceed blockwise: [ ] [ ] A B E F = C D G H [ ] AE + BG AF + BH CE + DG CF + DH For an n 1 column matrix (an n-vector) the matrix xx is the n n matrix whose i, j- element is x i x j Orthogonal matrices Definition Let A be an n n square matrix. We say that A is an orthogonal matrix if its transpose is its inverse, A = A 1. This may seem like an odd property to study, but the following theorem explains why it is so useful. Essentially an orthogonal matrix rotates (or reflects) the coordinate system without distorting distances or angles Proposition For an n n square matrix A, the following are equivalent. 1. A is orthogonal. That is, A A = I. 2. A preserves norms. That is, for all x, 3. A preserves inner products, that is, for every x, y R n, Ax = x. (1) (Ax) (Ay) = x y Proof : (1) = (2) Assume A = A 1. For any vector, x 2 = x x, so Ax 2 = (Ax) (Ax) = x A Ax = x Ix = x x = x 2. (2) = (3) Assume A preserves norms. By Lemma below, (Ax) (Ay) = = = Ax + Ay Ax Ay 4 A(x + y) A(x y) 4 x + y x y 4 = x y. (3) = (1) Assume A preserves inner products. Pick an arbitrary x and z, and let y = Az. Then y Ax = z A Ax = z x since A preserves inner products. That is, z A Ax = z x for every z, which implies that A Ax = x. But x is arbitrary, so A A = I, that is, A is orthogonal Lemma x + y 2 x y 2 = 4x y. v ::14.45 KC Border

3 KC Border Linear Algebra and Multivariate Normal 22 3 Proof : x + y 2 x y 2 = (x + y) (x + y) (x y) (x y) = x x + 2x y + y y ( x x 2x y + y y ) = 4x y Quadratic forms An expression of the form x Ax, where x is an n 1 column vector and A is an n n matrix, is called a quadratic form 1 in x, and x Ax = a ij x i x j. j=1 If A and B are n n and x, y are n-vectors, then x (A + B)y = x Ay + x By and (x + y) A(x + y) = x Ax + 2x Ay + y Ay. The quadratic form, or the matrix A, is called positive definite if and positive semidefinite if x Ax > 0 whenever x 0, x Ax 0 whenever x 0. Letting x be the i th unit coordinate vector, we have x Ax = a ii. As a consequence, if A is positive definite, then the diagonal elements satisfy a ii > 0. If A is only positive semidefinite, then a ii 0. The level surfaces of a positive definite quadratic form are ellipsoids in R n. If A is of the form A = B B, then A is necessarily positive semidefinite, since x Ax = x B Bx = (Bx) (Bx) = Bx 2. This also implies that if A is positive semidefinite and nonsingular, then it is in fact positive definite. There is also a converse. If A is positive semidefinite, then there is a square matrix B such that A = B B (sort of like a square root of A). 1 For decades I was baffled by the term form. I once asked Tom Apostol at a faculty cocktail party what it meant. He professed not to know (it was a cocktail party, so that is excusable), but suggested that I should ask John Todd. He hypothesized that mathematicians don t know the difference between form and function, a clever reference to modern architectural philosophy. Professor Todd was quite imperious looking and had an English accent so I was too intimidated to ask him, but I subsequently learned (where, I can t recall) that form refers to a polynomial function in several variables where each term in the polynomial has the same degree. (The degree of the term is the sum of the exponents. For example, in the expression xyz + x 2 y + xz + z, the first two terms have degree three, the third term has degree two and the last one has degree one. It is thus not a form.) This is most often encountered in the phrases linear form (each term has degree one) or quadratic form (each term has degree two). KC Border v ::14.45

4 KC Border Linear Algebra and Multivariate Normal Gradients of linear and quadratic forms A function f given by f(x) = a Ax, where a is m 1, A is m n and x is n 1 is a linear function of x so the partial derivative f(x)/ x j is just the coefficient on x j, which is a A j. Thus, the gradient is f(x) = f(x) x 1. f(x) x n = A a. For a symmetric square n n matrix, the quadratic form Q(x) = x Ax = j=1 a ij x i x j satisfies Q(x)/ x k = n a ikx i + n j=1 a kjx j = 2 n l=1 a klx l, where the last equality is due to the symmetry of A. Thus Q(x) x 1 Q(x) =. = 2Ax. Q(x) x n 22.4 Eigenthingies and quadratic forms If A is an n n symmetric matrix, if there is a nonzero vector x and real number λ so that Ax = λx, then x is called an eigenvector of A and λ is its corresponding eigenvalue. For a given eigenvalue λ, the set of corresponding eigenvectors together with zero form a linear subspace of R n called the eigenspace of λ. The dimension of this subspace is the multiplicity of λ. The eigenspaces of distinct eigenvalues are orthogonal. The sum of the multiplicities of the eigenvalues of A is n, or equivalently, there is an orthogonal basis for R n consisting of eigenvectors of A. The Principal Axis Theorem may also be known as the Spectral Theorem, or the Diagonalization Theorem. It says that for a symmetric matrix, we can rotate the coordinate system so that the transformation x Ax simply multiplies each new coordinate j of x by some λ j Principal Axis Theorem Let A be an n n symmetric matrix. Let x 1,..., x n be an orthonormal basis for R n made up of eigenvectors of A, with corresponding eigenvalues λ 1,..., λ n. Set Λ = λ λ n and let C be the matrix whose columns are x 1,..., x n., v ::14.45 KC Border

5 KC Border Linear Algebra and Multivariate Normal 22 5 Then and C is orthogonal, that is, A = CΛC 1, Λ = C 1 AC, C 1 = C. We can use this to diagonalize the quadratic form A. Let C and Λ be as in the Principal Axis Theorem. Given x, let y = C x = C 1 x, so x = Cy. Then x Ax = (Cy) A(Cy) = y C ACy = y Λy = 22.5 Orthogonal projection onto a subspace λ i yi 2. Let M be a linear subspace of R n. The set of vectors in R n that are orthogonal to every x M is a linear subspace of R n, and it is called the orthogonal complement of M, denoted M Orthogonal Complement Theorem For each x R n we can write x in a unique way as x = x M + x, where x M M and x M. The vector x M is called the orthogonal projection of x onto M. In fact, let y 1,..., y m be an orthonormal basis for M. Put z i = (x y i )y i for i = 1,..., m. Then x M = m z i. A key property of the orthogonal projection of x on M is that x M is the point in M closest to x. Formally, we have: Proposition Let M be a linear subspace of R n. Let y R n. Then y y M y x for all x M. Proof : This is really just the Pythagorean Theorem. Let x M. Then y x 2 = (y y M ) + (y M x) 2 ( = (y y M ) + (y M x) ) ( ) (y y M ) + (y M x) = (y y M ) (y y M ) + 2(y y M ) (y M x) + (y M x) (y M x) but y y M = y M, and y M x M, so (y y M ) (y M x) = 0, so = y y M 2 + y M x 2 y y M 2. Note that x, y M, and y form a right triangle, so this is the Pythagorean Theorem. That is, x = y M minimizes y x over M Proposition (Linearity of Projection) The orthogonal projection satisfies (x + z) M = x M + z M and (αx) M = αx M. KC Border v ::14.45

6 KC Border Linear Algebra and Multivariate Normal 22 6 y y b 1 y M 0 b 2 Figure Orthogonal projection onto M = span{b 2, b 2 }. Proof : Let y 1,..., y k be an orthonormal basis for M. Use x M = k j=1 (x y j)y j and z m = k j=1 (x z j)z j. Then Use linearity of the dot product. (x + z) M = k (x + z y j )y j. j=1 Since orthogonal projection onto M is a linear transformation, there is a matrix P such that for every x, x M = P x. What does it look like? Let M R n be a k dimensional subspace. Let x 1,..., x k be a basis for M, and let X be the n k matrix whose columns are x 1,..., x k, and set P = X(X X) 1 X. Then for any y R n, y M = P y = X(X X) 1 X y. There are a number of points worth mentioning. Note that P is symmetric and idempotent, that is, P P = P. This implies that X X has an inverse. To see why, suppose X Xz = 0. Then z X Xz = 0, but z X Xz = Xz 2 so Xz = 0. But Xz is a linear combination of the columns of X, which are independent by hypothesis, so z = 0. We have just shown that X Xz = 0 = z = 0, so X X is invertible. Suppose y belongs to M. Then y is a linear combination of x 1,..., x k so there is a k-vector b with y = Xb. Then P y = X(X X) 1 X Xb = Xb = y, so P acts as the identity on M and every nonzero vector in M is an eigenvector of P corresponding to an eigenvalue of 1. Now suppose that y is orthogonal to M. Then y X = 0, which implies X y = 0, so P y = X(X X) 1 X y = 0. Thus every nonzero vector in M is an eigenvector of P corresponding to an eigenvalue 0. v ::14.45 KC Border

7 KC Border Linear Algebra and Multivariate Normal 22 7 We now show that y P y is orthogonal to every basis vector x j. That is, we want to show that (y P y) X = 0. Now (y P y) X = y X y P X = y X y X (X X) 1 X X = y X y X = 0. By the uniqueness of the orthogonal decomposition this shows that P is the orthogonal projection onto M. Thus P y is always a linear combination x 1,..., x k, that is, P y = Xb for some k-vector b, namely b = (X X) 1 X y. In fact, we have just shown that any symmetric idempotent matrix is a projection on its range Review of random vectors Let X = X 1.. be a random vector. Define Define the covariance matrix of X by Let X n E X 1 µ = E X =.. E X n Var X = [ Cov X i X j ] = [ E(Xi µ i )(X j µ j ) ] = [ σ ij ] = E ( (X µ)(x µ) ). be an m n matrix of constants, and let Then, since expectation is a linear operator, Moreover a a 1n A =.. a m1... a mn Y = AX E Y = Aµ. Var Y = A(Var X)A since Var Y = E ( (AX Aµ)(AX Aµ) ) = E ( A(X µ)(x µ) A ) = A(Var X)A. The covariance matrix Σ of a random vector Y = (Y 1,..., Y n ) is always positive semidefinite, since for any vector w of weights, w Σw is the variance of the random variable w Y, and variances are always nonnegative The Normal density Recall that the Normal N(µ, σ 2 ) has a density of the form f(x) = although I don t normally write it in that form. 1 2πσ 2 e 1 2 (x µ) 1 σ 2 (x µ) (2) KC Border v ::14.45

8 KC Border Linear Algebra and Multivariate Normal The Multivariate Normal distribution I am going to use a not quite standard definition of the multivariate normal that makes life simpler. I learned this approach from Dave Grether who told me that he learned it from the late statistician Edwin James George Pitman, who happens to be the father of Jim Pitman, the author of the probability textbook [9] for the course. I have also found this definition is the one used by Jacod and Protter [7, Definition 16.1, p. 126] and by Rao [10, p. 518] Definition Extend the notion of a Normal random variable to include constants as N(µ, 0) zero-variance random variables. A random vector X = (X 1,..., X n ) R n has a multivariate Normal distribution or a jointly Normal distribution if for every constant vector T R n the linear combination T X = n T ix i has a Normal(µ T, σ 2 T ) distribution. Recall Proposition : Proposition If X and Y are independent normals with X N(µ, σ 2 ) and Y N(λ, τ 2 ), then X + Y N(µ + λ, σ 2 + τ 2 ). It has the following Corollary: Corollary If X 1,..., X n are independent Normal random variables, then the random vector X = (X 1,..., X n ) has a multivariate Normal distribution. Proof : We need to show that for any constant vector T, the linear combination T X = n T ix i has a Normal distribution. But since the X i s are independent Normals, the T i X i s are also independent Normals, so by the Proposition, their sum is a Normal random variable. So now we know that multivariate Normal random vectors do exist Proposition If X is an n-dimensional multivariate Normal random vector, and A is an m m constant matrix, then Y = AX is an m-dimensional multivariate Normal random vector. Proof : For a constant 1 m-vector T, the linear combination T Y is just the linear combination (T A)X, which by hypothesis is Normal Proposition If X = (X 1,..., X n ) has a multivariate Normal distribution, then Each component X i has a Normal distribution. Every subvector of X has a multivariate Normal distribution Definition Let µ = E X and let Σ = E ((X µ)(x µ) ), that is, σ i,j := Σ i,j = Cov(X i, X j ) = E(X i E X i )(X j E X j ) v ::14.45 KC Border

9 KC Border Linear Algebra and Multivariate Normal Theorem If X N(µ, Σ), then Y = CX N(Cµ, CΣC ) Corollary Uncorrelated jointly normal random variables are in fact independent!! Proof : If the random variables are uncorrelated, then Σ is diagonal. In that case the quadratic form (x µ) Σ 1 (x µ) reduces to a sum of squares, so the density factors into the product of the marginal densities, which implies independence. Let C be diagonal. Then CX is a linear combination c 1 X c n X n of the components and has a (univariate) normal N(Cµ, CΣC ) distribution. Σ is positive semi-definite. To see this, let a R n. Then Var(a X) = a Σa, but variances are always nonnegative. The following theorem may be found in Jacod and Protter [7, Theorem 16.1, p ]. It is used by Anderson [1, 2.4] as the definition of a multivariate normal Proposition A random vector X = (X 1,..., X n ) with nonsingular covariance matrix Σ has a multivariate normal distribution if and only if its density is ( ) n 1 1 f(x) = e 1 2 (x µ) Σ 1 (x µ). (3) 2π det Σ Note that this agrees with (2) when n = 1. The multivariate Normal density is constant on ellipsoids of the form (x µ) Σ 1 (x µ) = constant. Singular covariance matrices When a jointly Normal distribution has a singular covariance matrix, then the density does not exist. But if the matrix has rank k, there is a k-dimensional flat 2 on which it is possible to define a density of the form given in (3), but we will not make use of that here Not jointly normal vectors If X 1,..., X n are independent Normals, then the vector (X 1,..., X n ) is a jointly Normal vector. But if the components are not independent, then the vector may not be jointly Normal. That is, there may be a linear combination of them that does not have a Normal distribution. Here are a couple of examples, the details of which I leave as an exercise Example (Example 2, [7, p. 134]) Let Z be a standard Normal and let a > 0. Define { Z if Z a X = Z if Z > a. Then you can show that X has a standard Normal distribution, X + Z is not constant, but P (X + Z > 2a) = 0. Thus (X, Z) is not jointly Normal. 2 A flat is a translate of a linear subspace. That is, if M is a linear subspace, then x + M = {x + y : y M} is a flat. KC Border v ::14.45

10 KC Border Linear Algebra and Multivariate Normal Example (Exercise 4.47, [6, p. 200]) Let X and Y be independent standard normal random variables. Define the random variable Z by { X if XY > 0 Z = X if XY < 0. Then Z is a standard normal random variable, but the vector (Y, Z) is not jointly normal. I leave the verification as an exercise. (Hint for the second claim: Y and Z always have the same sign.) See Casella and Berger [6, Exercises , pp ] for more peculiar examples of pairs of normals that are not jointly normal. Behboodian [3] describes a pair of uncorrelated normal random variables that are not independent they aren t jointly normal either Multivariate Central Limit Theorem The following vector version of the Central Limit Theorem may be found in Jacod and Protter [7, Theorem 21.3, p. 183] Multivariate Central Limit Theorem Let X 1,..., X i... be independent and identically distributed random k-vectors with common mean µ and covariance matrix Σ. Let S n = X X n. Then S n µ D N(0, Σ). n Note that this result asserts that the limiting distribution is jointly Normal, which is stronger than saying that the distribution of each component is Normal Multivariate Normal and Chi-square Proposition Let X N(0, I n ). Then X AX χ 2 (k) if and only if A is symmetric, idempotent, and has rank k. Proof : I ll prove only one half. Assume A is symmetric, idempotent, and has rank k. (Then it is orthogonal projection onto a k-dimensional subspace.) Its eigenvalues are 0 and 1, so it is positive semidefinite. So by the Principal Axis Theorem, there is an orthogonal matrix C such that C AC = Λ = λ λ n where the λ i s are the eigenvalues, k of them are 1 and n k are zero. Setting Y = C X, we see that Y N(0, Λ). This means that the components of Y are independent. Moreover the Principal Axis Theorem also implies X AX = Y ΛY = Yi 2, which has a χ 2 (k) distribution. i:λ i = Corollary Let X N(µ, σ 2 I n ). Then ( ) ( ) X µ X µ A χ 2 (k) σ σ if and only if A is symmetric, idempotent, and has rank k. v ::14.45 KC Border

11 KC Border Linear Algebra and Multivariate Normal Proof : Note that X µ σ N(0, I). The following theorem is useful in deriving the distribution of certain test statistics. You can find a proof in C. Radakrishna Rao [10, Item (v), p. 187] or Henri Theil [11, pp ] Theorem Let X N(0, σ 2 I) and let A 1 and A 2 be symmetric idempotent matrices that satisfy A 1 A 2 = A 2 A 1 = 0. Then X A 1 X and X A 2 X are independent Independence of Sample Mean and Variance Statistics Let X 1,..., X n be independent and identically distributed Normal N(µ, σ 2 ) random variables. Let D i = X i X Theorem The sample average X and the random vector (D 1,..., D n ) are stochastically independent. [N.B. This does not say that each D i and D j are independent of each other, rather, they are jointly independent of X.] Proof : The random vector X = (X 1,..., X n ) has a multivariate Normal distribution (Corollary ). Therefore the random vector 1 1 D 1 n 1 n 1 n 1. n 1 1 n 1 n 1 n X 1. = D n. X 1 n 1 n 1 1 n X n n n n is also a multivariate Normal random vector (Theorem ). But by the Covariance Menagerie , Cov(D i, X) = 0. But for multivariate Normal vectors, this means that S and (D 1,..., D n ) are stochastically independent Corollary If X 1,..., X n are independent and identically distributed Normal N(µ, σ 2 ) random variables. Define n S 2 = (X i X) 2. n 1 1. X N(µ, σ 2 /n). 2. X and S 2 are independent. 3. (n 1)S 2 σ 2 = 1 σ 2 (X i X) 2 χ 2 (n 1) Proof : (1). This is old news. (2). By Theorem , X is independent of (D1,..., D n ). But S 2 = i D2 i /(n 1) is a function of (D 1,..., D n ), so it too is independent of X. (3). Define the standardized version of X i, Y i = X i µ, (i = 1,..., n), Ȳ = σ Y i /n = ( X µ)/σ. KC Border v ::14.45

12 KC Border Linear Algebra and Multivariate Normal Note that for each i, Y i Ȳ = (X i X)/σ, each Y i is a standard Normal random variable, and the Y i s are independent. So Y = (Y 1,..., Y n ) multivariate Normal with covariance matrix I. Let ( ) 1 1 v = n,...,. n Note that v v = 1. Now create an orthogonal matrix B that has v as its last row. We can always do this. (Think of the Gram Schmidt procedure.) Define the transformed variables Z = BY. Since Y is multivariate Normal, therefore so is Z. By Proposition , the covariance matrices satisfy Var Z = B(Var Y )B = BB = BB 1 = I, so Z is a vector of independent standard Normal random variables. By Proposition , multiplication by B preserves inner products, so Zi 2 = But Z n is the dot product of the last row of B with Y, or So combining (4) and (5), we have On the other hand, we can write Combining (6) and (7) implies Z n = v Y = 1 n Now rewrite this in terms of the X i s: n Y 2 i. (4) Y i = nȳ. (5) n 1 Yi 2 = Zi 2 + nȳ 2. (6) Y 2 i = (Y i Ȳ )2 + nȳ 2, (7) n 1 (Y i Ȳ )2 = Zi 2. (8) (Y i Ȳ )2 = 1 n σ 2 (X i X) 2. (9) Combining (8) and (9) shows that 1 σ 2 (X i X) n 1 2 = where Z i are independent standard Normals. In other words, 1 σ 2 n (X i X) 2 has a χ 2 (n 1) distribution. Z 2 i v ::14.45 KC Border

13 KC Border Linear Algebra and Multivariate Normal Bibliography [1] T. W. Anderson An introduction to multivariate statistical analysis. A Wiley Publication in Mathematical Statistics. New York: Wiley. [2] S. Axler Linear algebra done right, 2d. ed. Undergraduate Texts in Mathematics. New York: Springer. [3] J. Behboodian Examples of uncorrelated dependent random variables using a bivariate mixture. American Statistician 44(3): [4] K. C. Border. More than you wanted to know about quadratic forms. [5]. Quick review of matrix and linear algebra. [6] G. Casella and R. L. Berger Statistical inference, 2d. ed. Pacific Grove, California: Wadsworth. [7] J. Jacod and P. Protter Probability essentials, 2d. ed. Berlin and Heidelberg: Springer. [8] R. J. Larsen and M. L. Marx An introduction to mathematical statistics and its applications, fifth ed. Boston: Prentice Hall. [9] J. Pitman Probability. Springer Texts in Statistics. New York, Berlin, and Heidelberg: Springer. [10] C. R. Rao Linear statistical inference and its applications, 2d. ed. Wiley Series in Probability and Mathematical Statistics. New York: Wiley. [11] H. Theil Principles of econometrics. New York: Wiley. KC Border v ::14.45

14

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

Linear Models Review

Linear Models Review Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Exercises * on Principal Component Analysis

Exercises * on Principal Component Analysis Exercises * on Principal Component Analysis Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 207 Contents Intuition 3. Problem statement..........................................

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

Symmetric matrices and dot products

Symmetric matrices and dot products Symmetric matrices and dot products Proposition An n n matrix A is symmetric iff, for all x, y in R n, (Ax) y = x (Ay). Proof. If A is symmetric, then (Ax) y = x T A T y = x T Ay = x (Ay). If equality

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................

More information

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality. 88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Gaussian random variables inr n

Gaussian random variables inr n Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b

More information

Chapter 6: Orthogonality

Chapter 6: Orthogonality Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

Chapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix

Chapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix Probability Theory Linear transformations A transformation is said to be linear if every single function in the transformation is a linear combination. Chapter 5 The multivariate normal distribution When

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

The Multivariate Gaussian Distribution

The Multivariate Gaussian Distribution The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Math 291-2: Lecture Notes Northwestern University, Winter 2016 Math 291-2: Lecture Notes Northwestern University, Winter 2016 Written by Santiago Cañez These are lecture notes for Math 291-2, the second quarter of MENU: Intensive Linear Algebra and Multivariable Calculus,

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

Chapter 3. Matrices. 3.1 Matrices

Chapter 3. Matrices. 3.1 Matrices 40 Chapter 3 Matrices 3.1 Matrices Definition 3.1 Matrix) A matrix A is a rectangular array of m n real numbers {a ij } written as a 11 a 12 a 1n a 21 a 22 a 2n A =.... a m1 a m2 a mn The array has m rows

More information

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares MATH 167: APPLIED LINEAR ALGEBRA Least-Squares October 30, 2014 Least Squares We do a series of experiments, collecting data. We wish to see patterns!! We expect the output b to be a linear function of

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

Algebra II. Paulius Drungilas and Jonas Jankauskas

Algebra II. Paulius Drungilas and Jonas Jankauskas Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive

More information

Introducing the Normal Distribution

Introducing the Normal Distribution Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 10: Introducing the Normal Distribution Relevant textbook passages: Pitman [5]: Sections 1.2,

More information

Chapter Two Elements of Linear Algebra

Chapter Two Elements of Linear Algebra Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to

More information

Introducing the Normal Distribution

Introducing the Normal Distribution Department of Mathematics Ma 3/13 KC Border Introduction to Probability and Statistics Winter 219 Lecture 1: Introducing the Normal Distribution Relevant textbook passages: Pitman [5]: Sections 1.2, 2.2,

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Nonlinear Programming Algorithms Handout

Nonlinear Programming Algorithms Handout Nonlinear Programming Algorithms Handout Michael C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 5376 September 9 1 Eigenvalues The eigenvalues of a matrix A C n n are

More information

MTH 2032 SemesterII

MTH 2032 SemesterII MTH 202 SemesterII 2010-11 Linear Algebra Worked Examples Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education December 28, 2011 ii Contents Table of Contents

More information

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT 204 - FALL 2006 PRINCETON UNIVERSITY ALFONSO SORRENTINO 1 Adjoint of a linear operator Note: In these notes, V will denote a n-dimensional euclidean vector

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Typical Problem: Compute.

Typical Problem: Compute. Math 2040 Chapter 6 Orhtogonality and Least Squares 6.1 and some of 6.7: Inner Product, Length and Orthogonality. Definition: If x, y R n, then x y = x 1 y 1 +... + x n y n is the dot product of x and

More information

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes Written by Santiago Cañez These are notes which provide a basic summary of each lecture for Math 290-2, the second

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

Stat 206: Linear algebra

Stat 206: Linear algebra Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two

More information

Linear Algebra. Session 12

Linear Algebra. Session 12 Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)

More information

Department of Mathematics

Department of Mathematics Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Supplement 2: Review Your Distributions Relevant textbook passages: Pitman [10]: pages 476 487. Larsen

More information

The Standard Linear Model: Hypothesis Testing

The Standard Linear Model: Hypothesis Testing Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 25: The Standard Linear Model: Hypothesis Testing Relevant textbook passages: Larsen Marx [4]:

More information

MATH 369 Linear Algebra

MATH 369 Linear Algebra Assignment # Problem # A father and his two sons are together 00 years old. The father is twice as old as his older son and 30 years older than his younger son. How old is each person? Problem # 2 Determine

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Week Quadratic forms. Principal axes theorem. Text reference: this material corresponds to parts of sections 5.5, 8.2,

Week Quadratic forms. Principal axes theorem. Text reference: this material corresponds to parts of sections 5.5, 8.2, Math 051 W008 Margo Kondratieva Week 10-11 Quadratic forms Principal axes theorem Text reference: this material corresponds to parts of sections 55, 8, 83 89 Section 41 Motivation and introduction Consider

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

5 More on Linear Algebra

5 More on Linear Algebra 14.102, Math for Economists Fall 2004 Lecture Notes, 9/23/2004 These notes are primarily based on those written by George Marios Angeletos for the Harvard Math Camp in 1999 and 2000, and updated by Stavros

More information

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM 33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM (UPDATED MARCH 17, 2018) The final exam will be cumulative, with a bit more weight on more recent material. This outline covers the what we ve done since the

More information

Further Mathematical Methods (Linear Algebra)

Further Mathematical Methods (Linear Algebra) Further Mathematical Methods (Linear Algebra) Solutions For The 2 Examination Question (a) For a non-empty subset W of V to be a subspace of V we require that for all vectors x y W and all scalars α R:

More information

Lecture Note 1: Probability Theory and Statistics

Lecture Note 1: Probability Theory and Statistics Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

Linear Algebra: Characteristic Value Problem

Linear Algebra: Characteristic Value Problem Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number

More information

Cheat Sheet for MATH461

Cheat Sheet for MATH461 Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A

More information

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as Reading [SB], Ch. 16.1-16.3, p. 375-393 1 Quadratic Forms A quadratic function f : R R has the form f(x) = a x. Generalization of this notion to two variables is the quadratic form Q(x 1, x ) = a 11 x

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work. Assignment 1 Math 5341 Linear Algebra Review Give complete answers to each of the following questions Show all of your work Note: You might struggle with some of these questions, either because it has

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

Vector Spaces, Orthogonality, and Linear Least Squares

Vector Spaces, Orthogonality, and Linear Least Squares Week Vector Spaces, Orthogonality, and Linear Least Squares. Opening Remarks.. Visualizing Planes, Lines, and Solutions Consider the following system of linear equations from the opener for Week 9: χ χ

More information

Department of Mathematics

Department of Mathematics Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2018 Supplement 2: Review Your Distributions Relevant textbook passages: Pitman [10]: pages 476 487. Larsen

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Random Vectors and Multivariate Normal Distributions

Random Vectors and Multivariate Normal Distributions Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random 75 variables. For instance, X = X 1 X 2., where each

More information

Lecture 7 Spectral methods

Lecture 7 Spectral methods CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional

More information

Seminar on Linear Algebra

Seminar on Linear Algebra Supplement Seminar on Linear Algebra Projection, Singular Value Decomposition, Pseudoinverse Kenichi Kanatani Kyoritsu Shuppan Co., Ltd. Contents 1 Linear Space and Projection 1 1.1 Expression of Linear

More information

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010 Diagonalization of Symmetric Real Matrices (from Handout Definition 1 Let δ ij = { 1 if i = j 0 if i j A basis V = {v 1,..., v n } of R n

More information

MATH 22A: LINEAR ALGEBRA Chapter 4

MATH 22A: LINEAR ALGEBRA Chapter 4 MATH 22A: LINEAR ALGEBRA Chapter 4 Jesús De Loera, UC Davis November 30, 2012 Orthogonality and Least Squares Approximation QUESTION: Suppose Ax = b has no solution!! Then what to do? Can we find an Approximate

More information

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013 Multivariate Gaussian Distribution Auxiliary notes for Time Series Analysis SF2943 Spring 203 Timo Koski Department of Mathematics KTH Royal Institute of Technology, Stockholm 2 Chapter Gaussian Vectors.

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Expectation is a positive linear operator

Expectation is a positive linear operator Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 6: Expectation is a positive linear operator Relevant textbook passages: Pitman [3]: Chapter

More information

P = A(A T A) 1 A T. A Om (m n)

P = A(A T A) 1 A T. A Om (m n) Chapter 4: Orthogonality 4.. Projections Proposition. Let A be a matrix. Then N(A T A) N(A). Proof. If Ax, then of course A T Ax. Conversely, if A T Ax, then so Ax also. x (A T Ax) x T A T Ax (Ax) T Ax

More information

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2 Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,

More information

There are two things that are particularly nice about the first basis

There are two things that are particularly nice about the first basis Orthogonality and the Gram-Schmidt Process In Chapter 4, we spent a great deal of time studying the problem of finding a basis for a vector space We know that a basis for a vector space can potentially

More information

STAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017

STAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017 STAT 151A: Lab 1 Billy Fang 2 September 2017 1 Logistics Billy Fang (blfang@berkeley.edu) Office hours: Monday 9am-11am, Wednesday 10am-12pm, Evans 428 (room changes will be written on the chalkboard)

More information

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,

More information

Exercises * on Linear Algebra

Exercises * on Linear Algebra Exercises * on Linear Algebra Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 7 Contents Vector spaces 4. Definition...............................................

More information

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix

More information

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix. Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric

More information

Stable Process. 2. Multivariate Stable Distributions. July, 2006

Stable Process. 2. Multivariate Stable Distributions. July, 2006 Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.

More information

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix: Joint Distributions Joint Distributions A bivariate normal distribution generalizes the concept of normal distribution to bivariate random variables It requires a matrix formulation of quadratic forms,

More information

INNER PRODUCT SPACE. Definition 1

INNER PRODUCT SPACE. Definition 1 INNER PRODUCT SPACE Definition 1 Suppose u, v and w are all vectors in vector space V and c is any scalar. An inner product space on the vectors space V is a function that associates with each pair of

More information

MA 265 FINAL EXAM Fall 2012

MA 265 FINAL EXAM Fall 2012 MA 265 FINAL EXAM Fall 22 NAME: INSTRUCTOR S NAME:. There are a total of 25 problems. You should show work on the exam sheet, and pencil in the correct answer on the scantron. 2. No books, notes, or calculators

More information

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of CHAPTER III APPLICATIONS The eigenvalues are λ =, An orthonormal basis of eigenvectors consists of, The eigenvalues are λ =, A basis of eigenvectors consists of, 4 which are not perpendicular However,

More information