Eigenvalues, Eigenvectors and the Jordan Form

Size: px
Start display at page:

Download "Eigenvalues, Eigenvectors and the Jordan Form"

Transcription

1 EE/ME 701: Advanced Linear Systems Eigenvalues, Eigenvectors and the Jordan Form Contents 1 Introduction Review of basic facts about eigenvectors and eigenvalues Looking at eigenvalues and eigenvectors in relation to the null space of (A λ k I) Eigenvector example Bay section 4.1, A-Invariant Subspaces Properties of the Eigensystem Finding eigenvalues and eigenvectors Interpreting complex eigenvalues / eigenvectors Example: 3D Rotation The eigenvectors diagonalize A The eigensystem of symmetric (Hermitian) matrices Using the eigensystem to find the matrix exponential Defining the matrix exponential The matrix exponential provides the homogeneous solution to the matrix differential equation Using the eigensystem to solve for e At Conclusion on e At Repeated eigenvalues Analysis of the structure of the eigensystem of a matrix Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 1

2 EE/ME 701: Advanced Linear Systems 4 The Jordan-form Nature of the Jordan-form An example where the Jordan form arises Constructing the Jordan Form Regular and Generalized Eigenvectors of A First Jordan form example More on Jordan blocks The Jordan form, a second example One more twist, freedom to choose the regular eigenvector Example where regular E-vecs do not lie in the column space of (A λ k I) Summary of the Jordan Form Why Matlab does not have a numeric Jordan command Conclusions 60 6 Review questions and skills 61 Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 2

3 EE/ME 701: Advanced Linear Systems Section Introduction We have seen the basic case of eigenvalues and eigenvectors, eigenvectors are special vectors that are not rotated by the action of matrix A the Av k = λ k v k. (1) 1.1 Review of basic facts about eigenvectors and eigenvalues Only square matrices have eigensystems Every n n matrix has n eigenvalues, λ 1...λ n The eigenvector satisfies the relationship Av k = λ k v k, which leads to the eigenvector being the solution to (A λ k I) v k = 0 (2) or, said another way, the eigenvector is a vector in the null space of the matrix (A λ k I). Notes: 1. Any vector in the null space of (A λ k I) is an eigenvector. For example, if the null space is 2 dimensional, then any vec. in this 2D subspace is an E-vec. 2. Since the determinant of any matrix with a non-empty null space is zero, we have: det(a λ k I) = 0, k = 1..n (3) which gives the characteristic equation of matrix A. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 3

4 EE/ME 701: Advanced Linear Systems Section Looking at eigenvalues and eigenvectors in relation to the null space of (A λ k I) Starting from (A λ k I) v k = 0 (4) The eigenvalues are values of λ k such that (A λ k I) has a non-trivial null space. The eigenvectors are any vectors in that null space! Since det(a λ k I) = 0 (5) We know the null space is at least 1 dimensional. Theorem: for each distinct eigenvalue, there is at least one independent eigenvector. Proof: The proof follows directly from Eqns (4) and (5). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 4

5 EE/ME 701: Advanced Linear Systems Section Eigenvector example Considering matrix A A = >> [V, U] = eig(a) V = U = So the eigenvectors are: v 1 = , v 2 = , v 3 = And the eigenvalues are λ 1 = , λ 2 = , λ 2 = We can scale eigenvectors by any non-zero constant, if v 1 is an eigenvector, αv 1 is an eigenvector. For example, scaling each eigenvector so that the first element is 1.0: >> V = [ V(:,1)/V(1,1), V(:,2)/V(1,2), V(:,3)/V(1,3)] V = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 5

6 EE/ME 701: Advanced Linear Systems Section Bay section 4.1, A-Invariant Subspaces Consider a linear mapping from R n to R n A : R n R n (6) A subspace W is an invariant subspace with respect to operator A if the action of A on any vector in W returns a vector in W. This can be stated as w W = Aw W (7) The eigenvectors of a square matrix A are a suitable choice to be basis vectors for a subspace. The subspaces so given are A-invariant subspaces. The A-invariance property of subspaces spanned by eigenvectors is essential for our study of state-space systems. Since linear system dynamics can be written x(t 1 ) = Ax(t 0 ) (8) the A-invariance property says that if the state of a system lies within a subspace spanned by a subset of eigenvectors at a moment in time t 0, the state of the system will remain in that subspace. The property will permit us to break down the response of a system into modes, corresponding to eigenvalues and vector subspaces of state space given by the corresponding eigenvectors. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 6

7 EE/ME 701: Advanced Linear Systems Section Properties of the Eigensystem First we ll cover properties of the eigenvalues and regular eigenvectors. Example (Bay Example 4.1): Electric Fields In an isotropic dielectric medium, the electric field follows the relation D = εe where E is the electric field vector, D is the electric flux density (also called the displacement vector) and ε is the dielectric constant. Some materials, however, are anisotropic, governed by D 1 D 2 = ε 11 ε 12 ε 13 ε 21 ε 22 ε 23 E 1 E 2 D 3 ε 31 ε 32 ε 33 E 3 Find the directions, if any, in which the E-field and flux density are collinear. Solution: For the E-field and flux density to be collinear they must satisfy D = λe, Which is to say, the anisotropic directions are the eigenvectors of the dielectric tensor. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 7

8 EE/ME 701: Advanced Linear Systems Section Finding eigenvalues and eigenvectors For (A λi) v = 0 to have a solution, (A λi) must have a non-trivial null space. This is equivalent to saying det(a λi) = 0 (9) We have seen that Eqn (9) gives an n th order polynomial in λ This is more important for understanding than as a solution method Use: >> [V, U] = eig(a) There is no closed-form solution for the eigenvectors and values of a matrix larger than 4x4. And so eigen-solvers employ iterative numerical techniques. Matlab s eig() command is actually a front-end for a family of numerical techniques. Routine eig(a) starts by analyzing properties of matrix A, then selects the preferred routine. A good example is symmetric matrices. The eigensystem of symmetric matrices has the properties that all eigenvalues are real and that the eigenvectors are real and orthogonal. These special properties lead to a special eigensystem solver that applies only to symmetric matrices. Details of eigenvalue / eigenvector algorithms are beyond the scope of EE/ME 701 Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 8

9 EE/ME 701: Advanced Linear Systems Section Interpreting complex eigenvalues / eigenvectors Real eigenvalues correspond to scaling the eigenvector Complex eigenvalues lead to complex eigenvectors, and correspond to rotations and come in complex conjugate pairs. Recall that a complex number can be written λ k = a+ j b = M e j θ (10) Examples: Consider the basic rotation matrix: R = C θ S θ S θ C θ (11) Forming (λi R): (λi R) = λ C θ S θ S θ λ C θ (12) which gives the characteristic equation det(λi R) = λ 2 2C θ λ+c 2 θ + S2 θ = λ2 2C θ λ+1 = 0 (13) which solves to give λ = 2C θ ± 4Cθ 2 4 = 2C θ ± 2 4S 2 θ 2 = C θ ± j S θ (14) The eigenvalues of R are a complex conjugate pair given by the ± solutions of the quadratic equation. λ k = 1.0 ± θ Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 9

10 EE/ME 701: Advanced Linear Systems Section Example: 3D Rotation We ve seen the rotation matrix c tr = C κ C ϕ +C ω S κ C κ S ϕ S ω S ω S κ C ω C κ S ϕ S κ C ϕ S κ S ϕ S ω +C κ C ω C ω S ϕ S κ C κ S ω (15) S ϕ C ϕ S ω C ϕ C ω For example, the rotation matrix R = 2/3 1/3 2/3 2/3 2/3 1/3 1/3 2/3 2/3 (16) Corresponds to ω = 135 o, κ = 135 o, ϕ = o Like the general 3D rotation matrix, this matrix has one real eigenvalue and a complex conjugate pair >> [V, U] = eig(r) V = i i i i U = i i Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 10

11 EE/ME 701: Advanced Linear Systems Section The only real R-invariant subspace is given by the real eigenvector. v 3 is the axis of rotation! v v 3 = y w Rw z x Figure 1: Illustration of the action of rotation matrix R on a vector w to give vector Rw. Every vector w v 3 contains a projection onto subspaces spanned by the complex eigenvectors, and will be rotated by R. To solve differential equations such as ẋ(t) = R x(t), we will be interested in solutions of the form x(t) = V e U t V 1 x(0) (17) In which case complex eigenvalue and eigenvector pair combines to form a single 2-D R-invariant subspace spanned by w 1 = 1 2 (v 1 + v 2 ) (w 1 from real part) (18) w 2 = j 1 2 (v 1 v 2 ) (w 2 from imaginary part) (19) With solutions to the differential equation written x(t) = a 1 e αt cos(ωt + θ)w 1 + a 2 e αt sin(ωt + θ)w 2 (20) Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 11

12 EE/ME 701: Advanced Linear Systems Section The cos and sin terms in Eqn (20) show rotation in the 2D R-invariant subspace spanned by w 1 and w 2. (We treat matrix differential equations in up-coming sections). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 12

13 EE/ME 701: Advanced Linear Systems Section The eigenvectors diagonalize A Given a matrix A with eigensystem V, U. When A has a complete set of eigenvectors, A is related to its eigenvectors and eigenvalues by: A = V U V 1 (21) Example (from above), A = >> [V, U] = eig(a) V = U = >> V*U*inv(V) ans = Writing Ax = V U V 1 x (22) Eqn (22) can be understood as: V 1 (x) ( ) U V 1 x ( ) V U V 1 x representation of x on V basis vectors, V x scaling of elements of V x by the eigenvalues transformation back onto the s basis (The difference with the SVD is that the columns of V are not orthonormal, we get no vector-by-vector decomposition). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 13

14 EE/ME 701: Advanced Linear Systems Section The eigensystem of symmetric (Hermitian) matrices The eigensystem of a Hermitian matrix Q, (symmetric matrix, if real) have special, and very helpful properties. Notation: use A to be the complex-conjugate transpose (equivalent to A in Matlab). Property 1: If A = A, then for all complex vectors x, x Ax is real. Proof: Define y = (x Ax). Applying the transpose to the product, y = (x Ax) = x A x = x A x (23) Since A = A, y = y. For a number to equal its complex conjugate, it must be real. Property 2: The eigenvalues of a Hermitian matrix must be real. Proof: Suppose the λ is an eigenvalue of Q, with v a corresponding eigenvector, then Qv = λv (24) Now multiply on the left and right by v, v Qv = v λv = λv v = λ v 2 (25) By property 1, v Qv must be real, and v 2 must be real, there for λ = v Qv v 2 (26) must be real. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 14

15 EE/ME 701: Advanced Linear Systems Section Property 3: The eigenvectors of a Hermitian matrix, if they correspond to distinct eigenvalues, must be orthogonal. Proof: Starting with the given information, eigenvalue values λ 1 λ 2, and corresponding eigenvectors v 1 and v 2 Qv 1 = λ 1 v 1 (27) Qv 2 = λ 2 v 2 (28) Forming the complex-conjugate transpose of Eqn (27) v 1 Q = λ 1 v 1 = λ 1 v 1 (29) where we can drop the complex conjugate on λ 1, because we know λ 1 is real. Now multiplying on the right by v 2 gives (v 1 Q ) v 2 = λ 1 v 1 v 2 = λ 1 v 1 v 2 (30) But the multiplication also gives: v 1 (Q v 2 ) = v 1 λ 2 v 2 = λ 2 v 1 v 2 (31) So we find λ 1 v 1 v 2 = λ 2 v 1 v 2 (32) If λ 1 λ 2, Eqn (32) is only possible if v 1 v 2 = 0, which is to say that v 1 and v 2 are orthogonal. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 15

16 EE/ME 701: Advanced Linear Systems Section Property 4: A Hermitian matrix has a complete set of orthogonal eigenvectors. Proof: Any square matrix Q has a Schur decomposition T = V 1 QV (33) where V is an orthonormal matrix and T is an upper-triangular matrix. Since T is upper-triangular, T will be lower-triangular. However, V is orthonormal and Q is Hermitian, so V = V 1, Q = Q, and ( V 1) = (V ) = V So T = V Q ( V 1) = V 1 QV = T (34) Since T = T, T must be both upper-triangular and lower-triangular. For T to be both upper-triangular and lower-triangular, it must be diagonal. Let U = T be a diagonal matrix. Multiplying Eqn (33) on the left by V and on the right by V 1 gives Q = V T V 1 = V U V 1 (35) Which is precisely the form of the eigen-decomposition of Q, where Diagonal matrix U holds the eigenvalues of Q, and Orthogonal matrix V holds the eigenvectors of Q. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 16

17 EE/ME 701: Advanced Linear Systems Section Remarks: Since T = T, the diagonal elements (the eigenvalues) must be real (see property 2). When Q is real, V will be real. Significance: When a matrix Q is Hermitian, it has a complete set of orthogonal eigenvectors. This is the property that assures that the SVD exists for any matrix A. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 17

18 EE/ME 701: Advanced Linear Systems Section Using the eigensystem to find the matrix exponential Defining the matrix exponential Given a square matrix A, the matrix exponential is defined in a way analogous to the scalar exponential e at e At = I + At + 1 2! A2 t ! A3 t ! A4 t ! A5 t 5 + (36) For example, A = t = >> T5 = eye(2) + A * t + (1/2)*A*A*t^2 + (1/6)*A*A*A*t^3 + (1/24)*A*A*A*A*t^4 + (1/120)*A*A*A*A*A*t^5 T5 = >> B = expm(a*t) B = Note: Matlab function exp(a*t) computes the element-wise exponential. For >> B = exp(a*t) gives A = a 11 a 12 a 21 a 22 B = ea 11 t e a 21 t e a 12 t e a 22 t Matlab function expm(a*t) evaluates the matrix exponential, Eqn (36). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 18

19 EE/ME 701: Advanced Linear Systems Section The matrix exponential provides the homogeneous solution to the matrix differential equation Consider the homogeneous matrix differential equation (no forcing function for now) The solution is where x 0 is x(t) at t = 0 (the initial condition). d x(t) = Ax(t) (37) dt x(t) = e At x 0 (38) To show that Eqn (38) is the solution to (37), start with the definition of the matrix exponential e At = I+At+ 1 2! A2 t ! A3 t ! A4 t ! A5 t 5 + (36, repeated) And take the derivative w.r.t. time, d e At dt = 0+AI ! A2 t ! A3 t ! A4 t ! A5 t 4 + (39) Adjusting the factorial for the integers 2, 3,... from the differentiation, and factoring out one power of A d e At dt = AI + A At + A 1 2! A2 t 2 + A 1 3! A3 t 3 + A 1 4! A4 t 4 + ( = A I + At + 1 2! A2 t ! A3 t ) 4! A4 t 4 + = Ae At (40) Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 19

20 EE/ME 701: Advanced Linear Systems Section So just as in the scalar case, where d e at dt = ae at For the matrix case, we have d e At dt = Ae At Choosing x(t) = e At x 0, and plugging into differential equation (37) gives d dt x(t) = d dt eat x 0 = Ae At x 0 = A x(t) or d x(t) = A x(t) (41) dt and so the matrix exponential gives the solution to the homogeneous matrix differential equation. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 20

21 EE/ME 701: Advanced Linear Systems Section Using the eigensystem to solve for e At To solve for e At, in principle one could evaluate the series expansion e At = I+At+ 1 2! A2 t ! A3 t ! A4 t ! A5 t 5 + (36, repeated) However, for large A or t the series will converge slowly. For general matrices A, there is no simple expression for the elements of A p However, for a diagonal matrix U, U 2... U p are easily expressed: U = λ λ 2, then U 2 = λ λ 2 2,, U p = λp λ p 2 Given a diagonal matrix U = λ 1 0 λ λ n The matrix exponential is given by: e U t = I +U t + 1 2! U2 t ! U3 t ! U4 t λ 1 t 0 λ 2 1 t2 0 1 λ = + 2 t + 1 λ 2 2 t ! λ n t 0 λ 2 nt 2 Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 21 +

22 EE/ME 701: Advanced Linear Systems Section Summing the terms e U t = 1+λ 1 t + 1 2! λ2 1 t λ 2 t + 1 2! λ2 2 t λ n t + 1 2! λ2 n t2 + And so e U t = e λ 1 t 0 e λ 2t... 0 e λ nt (42) The matrix exponential of a diagonal matrix, is the matrix of element-wise exponentials Now consider again the matrix exponential of A e At = I + At + 1 2! A2 t ! A3 t 3 + When V is full rank, expanding A with its eigen-decomposition, A = V U V 1, and replacing I with V V 1 gives: e At = V V 1 +V U V 1 t + 1 2! V U V 1 V U V 1 t ! V U V 1 V U V 1 V U V 1 t 3 + Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 22

23 EE/ME 701: Advanced Linear Systems Section Factoring out V on the left, and V 1 on the right gives e At = V ( I +U t + 1 2! U V 1 V U t ) 3! U V 1 V U V 1 V U t 3 + V 1 The terms V 1 V cancel, and the inner term is e U t e At = V ( I +U t + 1 2! U2 t ) 3! U3 t 3 + V 1 = V e U t V 1 (43) Since U is diagonal, e U t we can solve for, e At = V e U t V 1 = V e λ 1 t 0 e λ 2t... 0 e λ nt V 1 (44) When V is full rank, equation (44) solves for e At, and is computed using n scalar exponentials. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 23

24 EE/ME 701: Advanced Linear Systems Section Example: A = t = >> [V, U] = eig(a) V = U = The matrix exponential e U t >> B = [ exp(u(1,1)*t) 0; B = e At >> C = V * B * inv(v) 0, exp(u(2,2)*t) ] C = And directly computing: >> D = expm(a*t) D = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 24

25 EE/ME 701: Advanced Linear Systems Section Conclusion on e At Since e At is part of the solution to ẋ(t) = Ax(t)+Bu(t) the eigenvectors and eigenvalues are needed to solve the matrix differential equation. Note that a complete set of eigenvectors is required to form V 1 for Eqn (44). To find e At we need a complete set of eigenvectors (this is where generalized eigenvectors will come in). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 25

26 EE/ME 701: Advanced Linear Systems Section Repeated eigenvalues Matrices can have repeated eigenvalues, for example: >> A = [ 2 1; 0 2] A = >> [V,U] = eig(a) V = U = When there are repeated eigenvalues: 1. We are assured to have at least 1 independent eigenvector. 2. There may fewer independent eigenvectors than eigenvalues Definitions: The algebraic multiplicity of an eigenvalue is the number of times the eigenvalue is repeated. The geometric multiplicity is the number of independent eigenvectors corresponding to the eigenvalue. (dim null (A λi)) Consider the example above: >> A = [ 2 1; 0 2] A = >> [V,U] = eig(a) V = U = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 26

27 EE/ME 701: Advanced Linear Systems Section Eigenvalue: 2.0 Algebraic multiplicity: 2 Geometric Multiplicity: 1 Number of missing eigenvectors: 1 Recall the eigen-decomposition of a matrix: A = V U V 1 The eigen-decomposition only exists if V is invertible. That is if there is a complete set of independent eigenvectors. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 27

28 EE/ME 701: Advanced Linear Systems Section Analysis of the structure of the eigensystem of a matrix Analysis of the eigensystem of a matrix proceeds by completing table Group the λ into k sets of repeated eigenvalues (one set for each unique value of λ) The number of λ k in the k th set is called the algebraic multiplicity, and is given by m k 2. Determine the number of independent eigenvectors corresponding to λ k by evaluating q(a λ k I) = dim null(a λ k I). This is called the geometric multiplicity, and is given by g k. If m k 2, it is possible that there are fewer independent eigenvectors than eigenvalues. 1 g k m k (45) 3. If m k > g k for any k, the Jordan form and generalized eigenvectors are required. k λ k m k g k # Needed Generalized Evecs m k g k 1 2. Table 1: Analysis of the structure of the eigensystem of A. Example, >> A = [ ; ; ] A = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 28

29 EE/ME 701: Advanced Linear Systems Section Recall: for triangular and diagonal matrices, the eigenvalues are the diagonal elements >> [V, U ] = eig(a) >> RoundByRatCommand(V) V = ans = U = Find g k : (A λ 1 I) = (A 2I) = (46) dim null (A λ 1 I) = g 1 = 1 (47) The analysis of the structure of the eigensystem of matrix A is seen in table 2. We see that there is one eigenvalue that is triply repeated. dim null (A λ k I) = 1, there is one eigenvector. The Jordan form will be required. k λ k m k g k # Needed Generalized Evecs m k g k Table 2: Analysis of the structure of the eigensystem of A. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 29

30 EE/ME 701: Advanced Linear Systems Section The Jordan-form 4.1 Nature of the Jordan-form Recall that the matrix exponential utilizes the eigen-decomposition, which as the form A = V U V 1 (48) But Eqn (48) is only valid when A has a complete set of eigenvectors. When A does not have a complete set of eigenvectors, we complete the set with generalized eigenvectors, giving the Jordan form decomposition of A: A = M J M 1 (49) where J is a block-diagonal matrix and M contains the regular eigenvectors of A, as well as one or more generalized eigenvectors, to make a complete set and assure that M 1 exists. When A has repeated eigenvalues it may have missing regular eigenvectors (g k < m k ), Analysis of e At requires decomposing matrix A using the Jordan form. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 30

31 EE/ME 701: Advanced Linear Systems Section An example where the Jordan form arises With scalar differential equations, we know that equations with repeated roots give solutions of the form y(t) = c 1 e λ 1 t + c 2 t e λ 1 t. For example, ÿ(t)+6ẏ(t)+9y(t) = 0 (50) has the characteristic equation s 2 + 6s+9 = 0 which as the roots s = { 3, 3}. The solution to Eqn (50) is: y(t) = c 1 e 3t + c 2 t e 3t. (51) But Eqn (??) for e At has no terms of the form t e 3t. And yet Eqn (50) is simply represented in state space with: ẋ(t) = Ax(t) with: x(t) = ẏ(t) y(t), ẋ(t) = d dt ẏ(t) y(t) = ẏ(t) y(t) (52) Yet the solution to Eqn (52) is (as always for ẋ(t) = Ax(t)) x(t) = e At x(0). So how can e At have a term of the form t e 3t? Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 31

32 EE/ME 701: Advanced Linear Systems Section Consider e J t for this matrix J: J = λ 1 0 λ (53) The expression for e J t is e J t = I + J t + J2 t 2 2! + + Jk t k + (54) k! = t λ 1 + t2 λ2 2λ + t3 λ3 3λ ! 0 λ 2! 0 λ 2 3! 0 λ 3 + tk k! λk kλ k 1 0 λ k + (55) The (1,1) and (2,2) elements give rise to summations of the form 1+ and the (2,1) element of e J t is 0. k=1 1 k! λk t k = e λt (56) Now consider the summation for the (1,2) element, which is given as 0+ t 1! + k=2 1 k! kλk 1 t k = t ( 1+ k=2 ) ( 1 (k 1)! λk 1 t k 1 = t 1+ k=1 ) 1 k! λk t k = t e λt So if J has the form of Eqn (53), then e J t = expm λ 1 t = 0 λ eλt t e λt 0 e λt! (57) Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 32

33 EE/ME 701: Advanced Linear Systems Section By the argument of Eqns (53)-(57) above, expm λ λ λ t = e λt t e λt 1 2 t2 e λt 0 e λt t e λt 0 0 e λt (58) which gives terms of the form t 2 e λt, expm λ λ λ λ t (59) gives terms of the form t 3 e λt, etc. For our specific example, A decomposes according to with A = M J M 1 (60) A = then M = , J = And e At = M e J t M 1 = M e 3t t e 3t 0 e 3t M 1 (61) So the solution x(t) = e At x(0) will have terms of the form e 3t and t e 3t, as need! Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 33

34 EE/ME 701: Advanced Linear Systems Section Constructing the Jordan Form Matrix A has been transformed to the Jordan form (sometimes called the Jordan canonical form) when J = M 1 A M (62) The columns of M are regular eigenvectors and generalized eigenvectors J is a block-diagonal matrix composed of m k m k Jordan blocks along the main diagonal. Each block on the diagonal of J is called a Jordan block. Eqns (53), (58) and (59) give examples of 2x2, 3x3 and 4x4 Jordan blocks Regular and Generalized Eigenvectors of A The regular eigenvectors are the eigenvectors we have considered all along, they satisfy the relationship Av = λ k v or (A λ k I) v = 0 (63) From Eqn (63), a set of independent regular eigenvectors is given by the null space of (A λ k I). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 34

35 EE/ME 701: Advanced Linear Systems Section The generalized eigenvectors form chains starting with a regular eigenvector. The generalized eigenvectors satisfy the relationship A V l+1 k, j = λ k V l+1 k, j +V l k, j (64) Or, rearranging (A λ k I) V l+1 k, j = V l k, j (65) Where V l+1 k, j is the the next generalized eigenvector in a chain (see Bay Eqn (4.14)). Each chain of generalized eigenvectors is anchored by a regular eigenvector. In this notation, V 1 k, j (66) is the first element of a chain; it is a regular eigenvector\. It is the j th regular eigenvector of the k th distinct eigenvalue. The l = 1 designates that Vk, 1 j is the first eigenvector in a chain, so it must be a regular eigenvector. Eqn (65) is an example of a recursive relationship, it is an equation that is applied repeatedly to get all elements of the chain. It has the solution V l+1 k, j = (A λ k I) # V l k, j (67) The method presented here to determine the Jordan form is the bottom-up method presented in Bay, section Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 35

36 EE/ME 701: Advanced Linear Systems Section First Jordan form example Consider e At, with A given as: >> A = [ ; ; ] A = First look at the eigenvalues >> U = eig(a) U = %% This command rounds-off values to nearby rational numbers %% which may be integers >> U = RoundByRatCommand(U) U = A has a repeated eigenvalue, we can make a table analyzing the structure of the eigensystem of A k λ k m k g k # Needed Gen. Evecs m k g k or 2 0 or Table 3: Analysis of the structure of the eigensystem of A. Table 3 shows that A has two distinct eigenvalues, and we don t yet know if λ 1 has 1 or 2 independent eigenvectors. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 36

37 EE/ME 701: Advanced Linear Systems Section Evaluate the geometric multiplicity of λ 1 >> lambda1=3; lambda2=6; I = eye(3); >> v1 = null(a-lambda1*i); v1 = v1/v1(1) v1 = 1 %% Eigenvector, scaled so the -1 1 %% first element is an integer The geometric multiplicity is the dimension of the null space in which the eigenvectors lie. For λ 1, g 1 = 1. Putting this information into the table k λ k m k g k # Needed Generalized Evecs m k g k Table 4: Analysis of the structure of the eigensystem of A. The total number of eigenvectors (regular+generalized) needed for λ k is m k. The number of regular eigenvectors is g k. The regular eigenvectors get the notation: where j = 1...g k. V 1 k,1, V 1 k,2,..., V 1 k,g k The number of needed generalized eigenvectors, corresponding to λ k, is m k g k. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 37

38 EE/ME 701: Advanced Linear Systems Section In the example, we need 1 generalized eigenvector for the k = 1 eigenvalue. In this case for k = 1 we have only one regular eigenvector, so it must serve as the first element, or anchor, of the chain of generalized eigenvectors regular eigenvector: V 1 1,1 solves (A λ 1 I)V 1 1,1 = 0 (68) generalized eigenvector: V 2 1,1 solves (A λ 1 I)V 2 1,1 = V 1 1,1 In Matlab >> lambda1=1; lambda2=2; I = eye(3); %% Find the first regular eigenvector, >> V111 = null(a-lambda1*i); V111=V111/V111(1) V111 = %% Find the generalized eigenvector by solving Eqn (68) >> V112 = pinv(a-lambda1*i)*v111 V112 = %% Find the regular eigenvector for lambda2 >> V211 = null(a-lambda2*i); V211=V211/V211(2) V211 = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 38

39 EE/ME 701: Advanced Linear Systems Section Put the eigenvectors (regular and generalized) together in the M matrix. The regular and generalized eigenvectors of a chain must be put in order. For each k and j, put the vectors Vk, l j going to the end of the chain. into M, starting with l = 1, and Put in the chain corresponding to for each regular eigenvector ( j) for each distinct eigenvalue (k). Chains may have a length of 1. For the example, [ M = V 1 1,1 V 2 1,1 V 1 2,1 ] >> M = [V111, V112, V211] M = J has 2 Jordan blocks J = >> J = inv(m) * A * M >> J = RoundByRatCommand(J) J = , e At = M e J t M 1 (69) For a system governed by ẋ(t) = Ax(t), and considering the J matrix, the output of the system will have solutions of the form y(t) = c 1 e 3t + c 2 t e 3t + c 3 e 6t (70) where the first two terms correspond to the first Jordan block, and the last term to the second Jordan block. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 39

40 EE/ME 701: Advanced Linear Systems Section k=1 j=1 λ κ =3 k=2 j=1 λ κ =6 V 1 1,1 V 2 1,1 V 1 2,1 Figure 2: Illustration of chains of eigenvectors. Three eigenvectors, two chains, one chain has two elements More on Jordan blocks A matrix in Jordan canonical form has a block diagonal structure, with Eigenvalues on the main diagonal Ones on the super diagonal within each block. A 2x2 block has 1 one, a 3x3 block has 2 ones, etc. One Jordan block corresponds to each regular eigenvector If the regular eigenvector has no generalized eigenvectors, then it creates a 1x1 block. If the regular eigenvector anchors a chain with one generalized eigenvector, then it creates a 2x2 block, etc. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 40

41 EE/ME 701: Advanced Linear Systems Section Each Jordan block corresponds to: 1x1 block: a regular eigenvector n n block, n 2: a chain anchored by a regular eigenvector, with n 1 generalized eigenvectors Using the Vk, l j notation, if we look at the structure of the M matrix, we can determine the layout of Jordan blocks. For example, [ M = V 1 1,1 V 1 1,2 V 2 1,2 V 1 2,1 V 2 2,1 ] Then the blocks of J are arranged: J = 1x1 2x2 2x2 Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 41

42 EE/ME 701: Advanced Linear Systems Section The Jordan form, a second example A = λ 1 = 0, λ 2 = 2 is repeated 5 times. >> U = eig(a) U = >> lambda1=0; lambda2=2; I = eye(6); >> [Row, Col, Null, LNull] = spaces(a-lambda2*i); >> g2=rank(null); g2 = 2 (A λ 2 I) has a 2-dimensional null space, so there are 2 independent regular eigenvectors. Null = For convenience, scale the eigenvectors to get integer values >> V211 = Null(:,1)/Null(3,1) V211 = >> V221 = Null(:,2)/Null(1,2) V221 = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 42

43 EE/ME 701: Advanced Linear Systems Section Check that the regular eigenvectors are actually eigenvectors for λ 2 = 2 >> %% Check that the eigenvectors are indeed eigenvectors >> %% These norms come out to zero, very small would be sufficient >> NDiff1 = norm( A*V211 - lambda2*v211 ) NDiff1 = 0 >> NDiff2 = norm( A*V221 - lambda2*v221 ) NDiff2 = 0 Note: All vectors from null(a λ 2 I) are eigenvectors. For example, >> x = 0.3*V *V221 x = 0.4 >> NDiffx = norm( A*x - lambda2*x ) 0.4 NDiffx = Make a table of the structure of the eigensystem. k λ k m k g k # Needed Gen. Evecs m k g k Table 5: Structure of the eigensystem of A. We need 3 generalized eigenvectors to have a complete set. These three will be in chains anchored by one or both of the regular eigenvectors of λ 2. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 43

44 EE/ME 701: Advanced Linear Systems Section The equation giving generalized eigenvectors is (A λ k I) V l+1 k, j = V l k, j (71) This is simply the relation Bx = y (72) with B = (A λ k I), y = Vk, l l+1 j and x = Vk, j. We know some things about the solutions of Eqn (72) 1. For an exact solution to exist V l k, j must lie in the column space of (A λ k I) 2. If we find a solution V l+1 k, j, it is not unique. We can add any component from the null space of (A λ k I), and it will still be a solution. Considering again the example problem, Check that V211 and V221 lie in the column space of (A λ 2 I) by checking that the projection of each onto the column space is equal to the original vector >> [Row, Col, Null, LNull] = spaces(a-lambda2*i); >> NIsInColumnSpaceV211 = norm( Col*Col *V211-V211 ) NIsInColumnSpaceV211 = e-16 %% V211 is in the column space of (A-lambda2*I) >> NIsInColumnSpaceV2 = norm( Col*Col *V221-V221 ) NIsInColumnSpaceV2 = e-15 %% V221 is in the column space of (A-lambda2*I) Both vectors lie in the column space of (A λ 2 I), so each will have at least one generalized eigenvector. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 44

45 EE/ME 701: Advanced Linear Systems Section Find V2,1 2 >> V212 = pinv(a-lambda2*i)*v211 V212 = %% Check that V212 is a generalized eigenvector >> NDiffV212 = norm( (A-lambda2*I)*V212 - V211 ) NDiffV212 = e-16 Yes, V 2 2,1 is a generalized eigenvector. Test to see if V 2 2,1 is in the column space of (A λ 2 I) >> norm( Col*Col *V212-V212 ) ans = No, so there is no V 3 2,1 Now evaluate V2,2 2 >> V222 = pinv(a-lambda2*i)*v221 V222 = >> %% Check that V222 is a gen. eigenvector >> NDiffV222 = norm( (A-lambda2*I)*V222 - V221 ) NDiffV222 = e Yes, V2,2 2 is a generalized eigenvector Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 45

46 EE/ME 701: Advanced Linear Systems Section Now check to see that V2,2 2 is in the column space of (A λ 2 I) >> NIsInColumnSpaceV222 = norm( Col*Col *V222-V222 ) NIsInColumnSpaceV222 = e-16 Yes, so there is a V 3 2,2. This will be the third generalized eigenvector >> V223 = pinv(a-lambda2*i)*v222, V223 = The chains of eigenvectors are seen in figure 3. k=1 j=1 λ κ =0 k=2 j=1 λ κ =2 k=2 j=2 λ κ =2 V 1,1 1 V 2 1,1 V 2 1,2 V 2 2,1 V 2 2,2 V 2 3,2 Figure 3: Illustration of chains of eigenvectors for this 6x6 example. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 46

47 EE/ME 701: Advanced Linear Systems Section Now build the M matrix First we need the regular eigenvalue corresponding to λ 1 >> V111 = null(a-lambda1*i); V111 = 0 >> V111 = V111/ V111(5) Put in the chains of E-vecs, starting each chain with its regular E-vec >> M = [ [V111] [V211 V212] [V221 V222 V223] ] >> M = RoundByRatCommand(M) M = Now find J >> J = inv(m)*a*m; >> J = RoundByRatCommand(J) J = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 47

48 EE/ME 701: Advanced Linear Systems Section Interpreting the result J has 3 Jordan blocks J = Correspondingly, M has 3 chains of eigenvectors M = The first eigenvector in each chain is a regular eigenvector. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 48

49 EE/ME 701: Advanced Linear Systems Section One more twist, freedom to choose the regular eigenvector Fact: If a matrix A has repeated eigenvalues, with g k > 1 independent eigenvector, the g k eigenvectors form a vector subspace X k = null (A λ k I) (73) Any vector from X k is an eigenvector. When m k 3, it is possible that g k 2, and we need to find a generalized eigenvector. In this case, dim null(a λ k I) = g k 2 and any vector from the 2-dimensional (or larger) null space of (A λ k I) is an eigenvector. Consider the generating equation for the generalized eigenvector (A λ k I) V 2 k, j = V 1 k, j (74) The anchor V 1 k, j must also lie in the column space of (A λ ki) A regular eigenvector that anchors a chain of generalized eigenvectors must lie in 2 spaces at once: The null space of (A λ k I) To be a regular e-vec of A. The column space of (A λ k I)... To generate a generalized e-vec of A. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 49

50 EE/ME 701: Advanced Linear Systems Section When g k 2, we have the freedom to choose the anchor for the chain of generalized eigenvectors Not just from a list, V 1 k,1, V 1 k,2,... It may be that we have valid eigenvectors, Vk,1 1, V k,2 1 which lies in the column space of (A λ k I)!, neither one of We can choose the regular eigenvectors as any vector from the null space of (A λ k I). By forming the intersection of the null and column spaces of (A λ k I), we can find the needed regular eigenvector. W = col(a λ k I) null(a λ k I), V 1 k, j W (75) Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 50

51 EE/ME 701: Advanced Linear Systems Section Example where regular E-vecs do not lie in the column space of (A λ k I) Consider the matrix A = (76) Analyzing the structure of its eigensystem >> RoundByRatCommand ( eig(a) ) ans = >> I = eye(4) lambda1=2, >> [Row, Col, Null, Lnull] = spaces(a - lambda1*i) Col = Null = So the structure of the eigensystem is given in table 6 k λ k m k g k # Needed Gen. Evecs m k g k Table 6: Analysis of the structure of the eigensystem of A. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 51

52 EE/ME 701: Advanced Linear Systems Section A first choice for eigenvectors are the two basis vectors of the null space of (A λ k I) >> v1 = RoundByRatCommand( Null(:,1) / Null(3,1) ) >> v2 = RoundByRatCommand( Null(:,2) ) v1 = 0 v2 = Determine if the eigenvectors lie in the column space of (A λ k I) >> NIsInColumnSpaceV1 = norm( Col*Col *v1-v1 ) NIsInColumnSpaceV1 = >> NIsInColumnSpaceV2 = norm( Col*Col *v2-v2 ) NIsInColumnSpaceV2 = No, neither eigenvector lies in the column space of (A λ k I) But what about the possibility that there exists another eigenvector which lies in the null space and column space of (A λ k I): x 1 = a 1 v 1 + a 2 v 2 First, consider the possibilities The universe is R 4, or 4D, with dim col(a λ k I) = 2, and dim null(a λ k I) = 2. So there are 3 possibilities: 1. Two 2D spaces can fit in a 4D universe and not intersect, so it is possible that col (A λ k I) null (A λ k I) = 0 2. It is possible that the intersection is 1 D 3. It is possible that the intersection is 2 D Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 52

53 EE/ME 701: Advanced Linear Systems Section Previously, we have seen how to form the intersection of two subspaces, Given sets of basis vectors U = [u 1, u 2,...,u nu ] and V = [v 1, v 2,...,v nv ], vectors in the intersection Are solutions to W = U V (77) w i = a 1 u 1 + a 2 u 2 + +a nu u nu = b 1 v 1 + b 2 v 2 + +b nv v nv (78) [ for some non-zero values a 1 a nu, b 1 b nv ]. Where U and V are the sets of basis vectors on spaces U and V, [ a 1 a nu, b 1 b nv ] must solve [U, V] a 1. b nv = 0 (79) [ ] The coefficient vector must lie in the null space of Col, Null, where [Col] and [Null] are sets of basis vectors on the column and null spaces of (A λ k I). >> CoeffVec = null([col, -Null]) CoeffVec = [ ] Since the null space of Col, Null is one dimensional, the intersection of the column and null spaces is 1D. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 53

54 EE/ME 701: Advanced Linear Systems Section Now find w 1, a vector in both the column and null spaces of (A λ k I) >> w1 = Col*CoeffVec(1:2) >> w1 = RoundByRatCommand( w1/w1(1)), Following the principle of check everything, w1 = verify that w 1 is a regular eigenvector and also lies in the needed Col. space >> NIsEigenvectorW1 = norm( A*w1 - lambda1 * w1) NIsEigenvectorW1 = 0 w 1 is constructed in the column space of (A λ k I), and the fact that it is an eigenvector shows it is also in the null space. Because the intersection of column and null spaces is 1-dimensional, (A λ k I) has exactly one regular eigenvector that can anchor a chain, = the chain must have length 3 (2 generalized E-vecs are needed). Compute a candidate for the first generalized eigenvector, V1,1 2 as solution to (A λ k I) V1,1 2 = V 1,1 1 >> V111 = w1, >> v3 = pinv(a - lambda1*i) * V111, v3 = Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 54

55 EE/ME 701: Advanced Linear Systems Section To find the remaining generalized eigenvector, V 3 1,1, solve (A λ k I) V 3 1,1 = V 2 1,1 (80) which requires that V1,1 2 be in the column space of (A λ k I) >> NIsInColumnSpaceV112 = norm( Col*Col *v3 - v3) NIsInColumnSpaceV112 = It is not! Vector v 3 is a candidate generalized eigenvector, but we can not use it for V1,1 2, because it does not lead to V 1,1 3. Going back to the generating Eqn, V1,1 2 must simultaneously be part of two solutions: (A λ k I) V1,1 2 = V 1,1 1 (81) and (A λ k I) V 3 1,1 = V 2 1,1 (82) Vector v 3 is a particular solution to Eqn (81), but it is not the only solution. Any vector V1,1 2 = v 3 + b 1 n 1 + b 2 n 2 (83) is an exact solution to (81), where n 1 and n 2 are basis vectors for the null space of (A λ k I). We must find a vector V1,1 2 that is given by Eqn (83), but also lies in the columns space of (A λ k I) to provide a solution to Eqn (82). Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 55

56 EE/ME 701: Advanced Linear Systems Section To find a value for V1,1 2 that is in the column space of (A λ k I), we need a solution to V1,1 2 = v 3 + b 1 n 1 + b 2 n 2 = a 1 c 1 + a 2 c 2 (84) where c 1 and c 2 are the basis vectors on the column space of (A λ k I). The solution to Eqn (84) is found by writing [ ] c 1 c 2 n 1 n 2 a 1 a 2 b 1 = v 3 (85) b 2 1. Find the coefficient vector >> CoeffVec2 = pinv( [Col, -Null]) * v3 CoeffVec2 = (a) Determine the candidate vector >> v3b = v3 + Null * CoeffVec2(3:4), v3b = (b) Check to be sure the new candidate is in the column space of (A λ k I) >> NIsInColumnSpaceV112b = norm( Col*Col *v3b - v3b) NIsInColumnSpaceV112b = e-16 Yes! Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 56

57 EE/ME 701: Advanced Linear Systems Section Set V1,1 2 and compute V 1,1 3 >> V112 = v3b >> V113 = pinv(a - lambda1*i) * V112 V112 = V113 = Build the M matrix, V1,2 1 is any independent regular eigenvector. Compute J. >> M = [ V111, V112, V113, V121 ] M = >> J = RoundByRatCommand( inv(m) * A * M ) J = J has a 3x3 block, and a 1x1 block. Done! Recapping k=1 j=1 λ κ =2 V 1 1,1 k=1 j=2 λ κ =2 V 1 1,2 m k = 4 and g k = 2, 1. Regular E-vec V1,1 1 had to be a regular E-vec, but also had to satisfy Eqn (74) to lead to V1, Not any V1,1 2 would do, it had to be the one that could lead to V1,1 3 V 1 2,1 V 1 3,1 Figure 4: Illustration of chains of eigenvectors for this 4x4 example. All of our work on vector spaces, leading up to the four fundamental spaces, was required to find the solution ẋ(t) = Ax(t) with this A matrix! Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 57

58 EE/ME 701: Advanced Linear Systems Section Summary of the Jordan Form Square matrices always have n eigenvalues, λ i The regular eigenvectors are given as the null space of (A λ i I) For a repeated eigenvalue λ k The algebraic multiplicity, m k, is the number of times λ k is repeated The geometric multiplicity, g k, is the dimension of null(a λ k I) When eigenvalues are repeated, we may not have enough independent regular eigenvectors (g k < m k ), in which case the Jordan form is required. The Jordan form corresponds to scalar differential equations with repeated roots and solutions of the form y(t) = a 1 e λ 1 t + a 2 t e λ 1 t + a 3 t 2 e λ 1 t... For repeated eigenvalues, regular eigenvectors give rise to chains of generalized eigenvectors. The generalized eigenvectors are solutions to (A λ i I)v l+1 k, j = v l k, j (86) where Eqn (86) is repeated as needed to obtain m k eigenvectors. Examples have shown several characteristics of eigensystems with repeated roots. Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 58

59 EE/ME 701: Advanced Linear Systems Section Why Matlab does not have a numeric Jordan command Strikingly, Matlab has no numerical routine to find the generalized eigenvectors or Jordan form (standard Matlab no jordan() routine!) This is because the Jordan form calculation is numerically very sensitive, a small perturbation in A produces a large change in the chains of eigenvectors This sensitivity is true of the differential equations themselves, ÿ(t)+6ẏ(t) y(t) = 0 has two distinct roots! Consider the stages where a decision must be made When there are two eigenvalues with λ a λ b, are they repeated or distinct? What is the dimension of null(a λi)? Does v l k, j lie in the column space of (A λi), or does it not? Is v l+1 k, j independent of the existing eigenvectors? There is no known numerical routine to find the Jordan form that is sufficiently numerically robust to be included by Mathworks in Matlab. The Matlab symbolic algebra package does have a jordan() routine. It runs symbolically on rational numbers to operate without round-off error, for example: 21/107 52/12 3/2 A = 0 119/120 8/5 1/1 11/12 13/14 Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 59

60 EE/ME 701: Advanced Linear Systems Section Conclusions To solve The solution includes ẋ(t) = Ax(t)+Bu(t) (87) x(t) = e At x(0) (88) When we have a complete set of eigenvectors, we can diagonalize A and e At = V e U t V 1 (89) with U = V 1 AV (90) When we lack a complete set of eigenvectors, the Jordan form is used, with e At = M e J t M 1 (91) with J = M 1 AM (92) Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 60

61 EE/ME 701: Advanced Linear Systems Section Review questions and skills 1. In what fundamental space do the regular eigenvectors lie. 2. Given the eigenvalues of a matrix, analyze the structure of the eigensystem (a) Determine the number of required generalized eigenvectors 3. Indicate the generating equation for the generalized eigenvectors. 4. Indicate in what fundamental space the vectors of the generating equations must lie. 5. When g k 2, and no regular eigenvector lies in the column space of (A λ k I), what steps can be taken? 6. When additional generalized eigenvectors are needed, and v, a candidate generalized eigenvector does not lie in the column space of (A λ k I), what steps can be taken? Part 4b: Eigenvalues and Eigenvectors (Revised: Oct 22, 2016) Page 61

21 Linear State-Space Representations

21 Linear State-Space Representations ME 132, Spring 25, UC Berkeley, A Packard 187 21 Linear State-Space Representations First, let s describe the most general type of dynamic system that we will consider/encounter in this class Systems may

More information

2. Every linear system with the same number of equations as unknowns has a unique solution.

2. Every linear system with the same number of equations as unknowns has a unique solution. 1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015 Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See

More information

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017 Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

. = V c = V [x]v (5.1) c 1. c k

. = V c = V [x]v (5.1) c 1. c k Chapter 5 Linear Algebra It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix Because they form the foundation on which we later work,

More information

Symmetric and anti symmetric matrices

Symmetric and anti symmetric matrices Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal

More information

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015 Final Review Written by Victoria Kala vtkala@mathucsbedu SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015 Summary This review contains notes on sections 44 47, 51 53, 61, 62, 65 For your final,

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

Chapter 4 & 5: Vector Spaces & Linear Transformations

Chapter 4 & 5: Vector Spaces & Linear Transformations Chapter 4 & 5: Vector Spaces & Linear Transformations Philip Gressman University of Pennsylvania Philip Gressman Math 240 002 2014C: Chapters 4 & 5 1 / 40 Objective The purpose of Chapter 4 is to think

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors /88 Chia-Ping Chen Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Eigenvalue Problem /88 Eigenvalue Equation By definition, the eigenvalue equation for matrix

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Chapter 5 Eigenvalues and Eigenvectors

Chapter 5 Eigenvalues and Eigenvectors Chapter 5 Eigenvalues and Eigenvectors Outline 5.1 Eigenvalues and Eigenvectors 5.2 Diagonalization 5.3 Complex Vector Spaces 2 5.1 Eigenvalues and Eigenvectors Eigenvalue and Eigenvector If A is a n n

More information

Chapter Two Elements of Linear Algebra

Chapter Two Elements of Linear Algebra Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Linear Algebra and ODEs review

Linear Algebra and ODEs review Linear Algebra and ODEs review Ania A Baetica September 9, 015 1 Linear Algebra 11 Eigenvalues and eigenvectors Consider the square matrix A R n n (v, λ are an (eigenvector, eigenvalue pair of matrix A

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

Section 8.2 : Homogeneous Linear Systems

Section 8.2 : Homogeneous Linear Systems Section 8.2 : Homogeneous Linear Systems Review: Eigenvalues and Eigenvectors Let A be an n n matrix with constant real components a ij. An eigenvector of A is a nonzero n 1 column vector v such that Av

More information

Notes on Linear Algebra

Notes on Linear Algebra 1 Notes on Linear Algebra Jean Walrand August 2005 I INTRODUCTION Linear Algebra is the theory of linear transformations Applications abound in estimation control and Markov chains You should be familiar

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible. MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:

More information

Announcements Monday, October 29

Announcements Monday, October 29 Announcements Monday, October 29 WeBWorK on determinents due on Wednesday at :59pm. The quiz on Friday covers 5., 5.2, 5.3. My office is Skiles 244 and Rabinoffice hours are: Mondays, 2 pm; Wednesdays,

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Examples and MatLab. Vector and Matrix Material. Matrix Addition R = A + B. Matrix Equality A = B. Matrix Multiplication R = A * B.

Examples and MatLab. Vector and Matrix Material. Matrix Addition R = A + B. Matrix Equality A = B. Matrix Multiplication R = A * B. Vector and Matrix Material Examples and MatLab Matrix = Rectangular array of numbers, complex numbers or functions If r rows & c columns, r*c elements, r*c is order of matrix. A is n * m matrix Square

More information

MIT Final Exam Solutions, Spring 2017

MIT Final Exam Solutions, Spring 2017 MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

More information

Eigenvalues and Eigenvectors. Review: Invertibility. Eigenvalues and Eigenvectors. The Finite Dimensional Case. January 18, 2018

Eigenvalues and Eigenvectors. Review: Invertibility. Eigenvalues and Eigenvectors. The Finite Dimensional Case. January 18, 2018 January 18, 2018 Contents 1 2 3 4 Review 1 We looked at general determinant functions proved that they are all multiples of a special one, called det f (A) = f (I n ) det A. Review 1 We looked at general

More information

6 Inner Product Spaces

6 Inner Product Spaces Lectures 16,17,18 6 Inner Product Spaces 6.1 Basic Definition Parallelogram law, the ability to measure angle between two vectors and in particular, the concept of perpendicularity make the euclidean space

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

LINEAR ALGEBRA QUESTION BANK

LINEAR ALGEBRA QUESTION BANK LINEAR ALGEBRA QUESTION BANK () ( points total) Circle True or False: TRUE / FALSE: If A is any n n matrix, and I n is the n n identity matrix, then I n A = AI n = A. TRUE / FALSE: If A, B are n n matrices,

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Solutions to Final Exam

Solutions to Final Exam Solutions to Final Exam. Let A be a 3 5 matrix. Let b be a nonzero 5-vector. Assume that the nullity of A is. (a) What is the rank of A? 3 (b) Are the rows of A linearly independent? (c) Are the columns

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Reduction to the associated homogeneous system via a particular solution

Reduction to the associated homogeneous system via a particular solution June PURDUE UNIVERSITY Study Guide for the Credit Exam in (MA 5) Linear Algebra This study guide describes briefly the course materials to be covered in MA 5. In order to be qualified for the credit, one

More information

Solving a system by back-substitution, checking consistency of a system (no rows of the form

Solving a system by back-substitution, checking consistency of a system (no rows of the form MATH 520 LEARNING OBJECTIVES SPRING 2017 BROWN UNIVERSITY SAMUEL S. WATSON Week 1 (23 Jan through 27 Jan) Definition of a system of linear equations, definition of a solution of a linear system, elementary

More information

Lecture 1: Review of linear algebra

Lecture 1: Review of linear algebra Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers. Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

ECE504: Lecture 6. D. Richard Brown III. Worcester Polytechnic Institute. 07-Oct-2008

ECE504: Lecture 6. D. Richard Brown III. Worcester Polytechnic Institute. 07-Oct-2008 D. Richard Brown III Worcester Polytechnic Institute 07-Oct-2008 Worcester Polytechnic Institute D. Richard Brown III 07-Oct-2008 1 / 35 Lecture 6 Major Topics We are still in Part II of ECE504: Quantitative

More information

LINEAR ALGEBRA SUMMARY SHEET.

LINEAR ALGEBRA SUMMARY SHEET. LINEAR ALGEBRA SUMMARY SHEET RADON ROSBOROUGH https://intuitiveexplanationscom/linear-algebra-summary-sheet/ This document is a concise collection of many of the important theorems of linear algebra, organized

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Study Guide for Linear Algebra Exam 2

Study Guide for Linear Algebra Exam 2 Study Guide for Linear Algebra Exam 2 Term Vector Space Definition A Vector Space is a nonempty set V of objects, on which are defined two operations, called addition and multiplication by scalars (real

More information

Basic Elements of Linear Algebra

Basic Elements of Linear Algebra A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A =

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A = AMS1 HW Solutions All credit is given for effort. (- pts for any missing sections) Problem 1 ( pts) Consider the following matrix 1 1 9 a. Calculate the eigenvalues of A. Eigenvalues are 1 1.1, 9.81,.1

More information

Math 110 Linear Algebra Midterm 2 Review October 28, 2017

Math 110 Linear Algebra Midterm 2 Review October 28, 2017 Math 11 Linear Algebra Midterm Review October 8, 17 Material Material covered on the midterm includes: All lectures from Thursday, Sept. 1st to Tuesday, Oct. 4th Homeworks 9 to 17 Quizzes 5 to 9 Sections

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems. Chapter 3 Linear Algebra In this Chapter we provide a review of some basic concepts from Linear Algebra which will be required in order to compute solutions of LTI systems in state space form, discuss

More information

Linear Algebra Lecture Notes-II

Linear Algebra Lecture Notes-II Linear Algebra Lecture Notes-II Vikas Bist Department of Mathematics Panjab University, Chandigarh-64 email: bistvikas@gmail.com Last revised on March 5, 8 This text is based on the lectures delivered

More information

ELE/MCE 503 Linear Algebra Facts Fall 2018

ELE/MCE 503 Linear Algebra Facts Fall 2018 ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2

More information

2. Review of Linear Algebra

2. Review of Linear Algebra 2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear

More information

and let s calculate the image of some vectors under the transformation T.

and let s calculate the image of some vectors under the transformation T. Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =

More information

Generalized eigenspaces

Generalized eigenspaces Generalized eigenspaces November 30, 2012 Contents 1 Introduction 1 2 Polynomials 2 3 Calculating the characteristic polynomial 5 4 Projections 7 5 Generalized eigenvalues 10 6 Eigenpolynomials 15 1 Introduction

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

Linear algebra II Tutorial solutions #1 A = x 1

Linear algebra II Tutorial solutions #1 A = x 1 Linear algebra II Tutorial solutions #. Find the eigenvalues and the eigenvectors of the matrix [ ] 5 2 A =. 4 3 Since tra = 8 and deta = 5 8 = 7, the characteristic polynomial is f(λ) = λ 2 (tra)λ+deta

More information

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the

More information

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar

More information

Linear Algebra. Min Yan

Linear Algebra. Min Yan Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................

More information

Chapter 5. Eigenvalues and Eigenvectors

Chapter 5. Eigenvalues and Eigenvectors Chapter 5 Eigenvalues and Eigenvectors Section 5. Eigenvectors and Eigenvalues Motivation: Difference equations A Biology Question How to predict a population of rabbits with given dynamics:. half of the

More information

Linear System Theory

Linear System Theory Linear System Theory Wonhee Kim Lecture 4 Apr. 4, 2018 1 / 40 Recap Vector space, linear space, linear vector space Subspace Linearly independence and dependence Dimension, Basis, Change of Basis 2 / 40

More information

m We can similarly replace any pair of complex conjugate eigenvalues with 2 2 real blocks. = R

m We can similarly replace any pair of complex conjugate eigenvalues with 2 2 real blocks. = R 1 RODICA D. COSTIN Suppose that some eigenvalues are not real. Then the matrices S and are not real either, and the diagonalization of M must be done in C n. Suppose that we want to work in R n only. Recall

More information

Singular Value Decomposition (SVD)

Singular Value Decomposition (SVD) School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Summer Session Practice Final Exam

Summer Session Practice Final Exam Math 2F Summer Session 25 Practice Final Exam Time Limit: Hours Name (Print): Teaching Assistant This exam contains pages (including this cover page) and 9 problems. Check to see if any pages are missing.

More information

LINEAR ALGEBRA REVIEW

LINEAR ALGEBRA REVIEW LINEAR ALGEBRA REVIEW SPENCER BECKER-KAHN Basic Definitions Domain and Codomain. Let f : X Y be any function. This notation means that X is the domain of f and Y is the codomain of f. This means that for

More information

Calculating determinants for larger matrices

Calculating determinants for larger matrices Day 26 Calculating determinants for larger matrices We now proceed to define det A for n n matrices A As before, we are looking for a function of A that satisfies the product formula det(ab) = det A det

More information

ft-uiowa-math2550 Assignment OptionalFinalExamReviewMultChoiceMEDIUMlengthForm due 12/31/2014 at 10:36pm CST

ft-uiowa-math2550 Assignment OptionalFinalExamReviewMultChoiceMEDIUMlengthForm due 12/31/2014 at 10:36pm CST me me ft-uiowa-math255 Assignment OptionalFinalExamReviewMultChoiceMEDIUMlengthForm due 2/3/2 at :3pm CST. ( pt) Library/TCNJ/TCNJ LinearSystems/problem3.pg Give a geometric description of the following

More information

Linear Algebra Fundamentals

Linear Algebra Fundamentals Linear Algebra Fundamentals It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix. Because they form the foundation on which we later

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors LECTURE 3 Eigenvalues and Eigenvectors Definition 3.. Let A be an n n matrix. The eigenvalue-eigenvector problem for A is the problem of finding numbers λ and vectors v R 3 such that Av = λv. If λ, v are

More information

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that

More information

π 1 = tr(a), π n = ( 1) n det(a). In particular, when n = 2 one has

π 1 = tr(a), π n = ( 1) n det(a). In particular, when n = 2 one has Eigen Methods Math 246, Spring 2009, Professor David Levermore Eigenpairs Let A be a real n n matrix A number λ possibly complex is an eigenvalue of A if there exists a nonzero vector v possibly complex

More information

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses

More information

Generalized Eigenvectors and Jordan Form

Generalized Eigenvectors and Jordan Form Generalized Eigenvectors and Jordan Form We have seen that an n n matrix A is diagonalizable precisely when the dimensions of its eigenspaces sum to n. So if A is not diagonalizable, there is at least

More information

Lec 2: Mathematical Economics

Lec 2: Mathematical Economics Lec 2: Mathematical Economics to Spectral Theory Sugata Bag Delhi School of Economics 24th August 2012 [SB] (Delhi School of Economics) Introductory Math Econ 24th August 2012 1 / 17 Definition: Eigen

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2 Contents Preface for the Instructor xi Preface for the Student xv Acknowledgments xvii 1 Vector Spaces 1 1.A R n and C n 2 Complex Numbers 2 Lists 5 F n 6 Digression on Fields 10 Exercises 1.A 11 1.B Definition

More information

4. Linear transformations as a vector space 17

4. Linear transformations as a vector space 17 4 Linear transformations as a vector space 17 d) 1 2 0 0 1 2 0 0 1 0 0 0 1 2 3 4 32 Let a linear transformation in R 2 be the reflection in the line = x 2 Find its matrix 33 For each linear transformation

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

MATHEMATICS 217 NOTES

MATHEMATICS 217 NOTES MATHEMATICS 27 NOTES PART I THE JORDAN CANONICAL FORM The characteristic polynomial of an n n matrix A is the polynomial χ A (λ) = det(λi A), a monic polynomial of degree n; a monic polynomial in the variable

More information

Announcements Wednesday, November 01

Announcements Wednesday, November 01 Announcements Wednesday, November 01 WeBWorK 3.1, 3.2 are due today at 11:59pm. The quiz on Friday covers 3.1, 3.2. My office is Skiles 244. Rabinoffice hours are Monday, 1 3pm and Tuesday, 9 11am. Section

More information

12. Homogeneous Linear Systems of ODEs

12. Homogeneous Linear Systems of ODEs 2. Homogeneous Linear Systems of ODEs A system of first order differential equations is linear, homogeneous, and has constant coefficients if it has the specific form x = a x + a 2 x 2 + +a n x n, x 2

More information

Review of Some Concepts from Linear Algebra: Part 2

Review of Some Concepts from Linear Algebra: Part 2 Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers.. EIGENVALUE PROBLEMS Background on eigenvalues/ eigenvectors / decompositions Perturbation analysis, condition numbers.. Power method The QR algorithm Practical QR algorithms: use of Hessenberg form and

More information

Notes on the matrix exponential

Notes on the matrix exponential Notes on the matrix exponential Erik Wahlén erik.wahlen@math.lu.se February 14, 212 1 Introduction The purpose of these notes is to describe how one can compute the matrix exponential e A when A is not

More information

Designing Information Devices and Systems II

Designing Information Devices and Systems II EECS 16B Fall 2016 Designing Information Devices and Systems II Linear Algebra Notes Introduction In this set of notes, we will derive the linear least squares equation, study the properties symmetric

More information