The Cayley-Hamilton Theorem and Minimal Polynomials

Similar documents
The minimal polynomial

COMMUTATIVE RINGS. Definition 3: A domain is a commutative ring R that satisfies the cancellation law for multiplication:

Solutions for Math 225 Assignment #4 1

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

Linear Algebra Lecture Notes-II

2a 2 4ac), provided there is an element r in our

and let s calculate the image of some vectors under the transformation T.

Eigenvalues and Eigenvectors: An Introduction

MTH310 EXAM 2 REVIEW

0.1 Rational Canonical Forms

Chapter 4. Remember: F will always stand for a field.

Basic Math Review for CS4830

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Linear Algebra III Lecture 11

6x 3 12x 2 7x 2 +16x 7x 2 +14x 2x 4

Polynomials. Chapter 4

Questionnaire for CSET Mathematics subset 1

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Math 375 First Exam, Page a) Give an example of a non-abelian group of innite order and one of nite order. (Say which is

Math 4377/6308 Advanced Linear Algebra

Basic Math Review for CS1340

(Inv) Computing Invariant Factors Math 683L (Summer 2003)

88 CHAPTER 3. SYMMETRIES

More Polynomial Equations Section 6.4

Homework #2 solutions Due: June 15, 2012

MATH Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product.

Homework 8 Solutions to Selected Problems

Matrix Algebra 2.1 MATRIX OPERATIONS Pearson Education, Inc.

Rings. Chapter 1. Definition 1.2. A commutative ring R is a ring in which multiplication is commutative. That is, ab = ba for all a, b R.

Lecture 7: Polynomial rings

Q N id β. 2. Let I and J be ideals in a commutative ring A. Give a simple description of

Midterm 3 Review. Terms. Formulas and Rules to Use. Math 1010, Fall 2011 Instructor: Marina Gresham. Odd Root ( n x where n is odd) Exponent

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics

[ Here 21 is the dot product of (3, 1, 2, 5) with (2, 3, 1, 2), and 31 is the dot product of

Topics in linear algebra

Math 113 Homework 5. Bowei Liu, Chao Li. Fall 2013

Chapter 2 Notes, Linear Algebra 5e Lay

CHAPTER 10: POLYNOMIALS (DRAFT)

Minimum Polynomials of Linear Transformations

Determinants Chapter 3 of Lay

Dividing Polynomials: Remainder and Factor Theorems

Linear Equations in Linear Algebra

Lecture 6: Finite Fields

Math Matrix Algebra

Homework 9 Solutions to Selected Problems

ENGR-1100 Introduction to Engineering Analysis. Lecture 21. Lecture outline

Systems of Second Order Differential Equations Cayley-Hamilton-Ziebur

Math 4310 Solutions to homework 7 Due 10/27/16

Definition For a set F, a polynomial over F with variable x is of the form

Name: MATH 3195 :: Fall 2011 :: Exam 2. No document, no calculator, 1h00. Explanations and justifications are expected for full credit.

Letting be a field, e.g., of the real numbers, the complex numbers, the rational numbers, the rational functions W(s) of a complex variable s, etc.

A connection between number theory and linear algebra

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017

ENGR-1100 Introduction to Engineering Analysis. Lecture 21

Rings If R is a commutative ring, a zero divisor is a nonzero element x such that xy = 0 for some nonzero element y R.

2.1 Matrices. 3 5 Solve for the variables in the following matrix equation.

THE MINIMAL POLYNOMIAL AND SOME APPLICATIONS

Lecture 2: Groups. Rajat Mittal. IIT Kanpur

MATH 115, SUMMER 2012 LECTURE 12

Section 19 Integral domains

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

MATH 235. Final ANSWERS May 5, 2015

MATRICES The numbers or letters in any given matrix are called its entries or elements

1. Select the unique answer (choice) for each problem. Write only the answer.

Polynomial Rings. i=0

Math 309 Notes and Homework for Days 4-6

Notes 6: Polynomials in One Variable

Review for Exam Find all a for which the following linear system has no solutions, one solution, and infinitely many solutions.

Evaluating Determinants by Row Reduction

Minimal Polynomials and Jordan Normal Forms

5.1 Introduction to Matrices

Math 370 Spring 2016 Sample Midterm with Solutions

3. Coding theory 3.1. Basic concepts

Chapter 3. Determinants and Eigenvalues

The Cyclic Decomposition Theorem

1. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. The augmented matrix of this linear system is

Matrices. In this chapter: matrices, determinants. inverse matrix

MS 2001: Test 1 B Solutions

MATHEMATICS 217: HONORS LINEAR ALGEBRA Spring Term, First Week, February 4 8

Linear algebra and applications to graphs Part 1

Chapter 1 Vector Spaces

Rings. EE 387, Notes 7, Handout #10

Undergraduate Mathematical Economics Lecture 1

Multiplication of Polynomials

Matrix Algebra Determinant, Inverse matrix. Matrices. A. Fabretti. Mathematics 2 A.Y. 2015/2016. A. Fabretti Matrices

Lecture 7.4: Divisibility and factorization

[06.1] Given a 3-by-3 matrix M with integer entries, find A, B integer 3-by-3 matrices with determinant ±1 such that AMB is diagonal.

MATRICES AND MATRIX OPERATIONS

Elementary maths for GMT

Vector Spaces ปร ภ ม เวกเตอร

x y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2

Chapter 2:Determinants. Section 2.1: Determinants by cofactor expansion

Econ Slides from Lecture 7

MATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:

(a + b)c = ac + bc and a(b + c) = ab + ac.

Advanced Engineering Mathematics Prof. Pratima Panigrahi Department of Mathematics Indian Institute of Technology, Kharagpur

Chapter 3: Theory Review: Solutions Math 308 F Spring 2015

February 11, 2019 LECTURE 5: THE RULES OF DIFFERENTIATION.

Self-Directed Course: Transitional Math Module 4: Algebra

Transcription:

Math 5327 Spring 208 The Cayley-Hamilton Theorem and Minimal Polynomials Here are some notes on the Cayley-Hamilton Theorem, with a few extras thrown in. I will start with a proof of the Cayley-Hamilton theorem, that the characteristic polynomial is an annihilating polynomial for its n n matrix A, along with a 3 3 example of the various aspects of the proof. In class, I only used a 2 2 example, which seems a little too simple. The proof starts this way: given a matrix, A, we consider the formula (xi A)adj(xI A) = det(xi A)I = c A (x)i. It would be nice to just plug A into this equation for x, and say that A A = 0, so the result follows. This does not work because adj(xi A) has entries which are polynomials in x so we would be dealing with a matrix where some of its entries are scalars and others are matrices. For example, if 2 x 2 4x + 3 x x A = 2, then adj(xi A) = x x 2 4x + 3 x. 2 x x x 2 4x + 3 Also, (xi A)adj(xI A) = (x 3 6x 2 + 9x 4)I. What we do is write adj(xi A) as a polynomial with matrix coefficients. In this case, 4 3 adj(xi A) = x 2 I + x 4 + 3. 4 3 More generally, adj(xi A) = x n B n + x n 2 B n 2 + + xb + B 0 for some collection of matrices. At this point, we have 2 4 3 xi 2 x 2 I + x 4 + 3 = (x 3 6x 2 +9x 4)I, 2 4 3 and it is legal to replace x by a matrix. This does not mean it is valid to do so, just that the expressions can make mathematical sense at this point. For example, if we were to replace 0 0 x by, say, C = 0 0 0, then on the left hand side of the expression above we have 0 0 0 0 0 2 0 0 4 3 0 0 0 2 0I + 0 0 0 4 + 3 0 0 0 2 0 0 0 4 3 2 0 4 0 5 7 3 = 2 3 = 5 4. 2 3 0

On the other hand, 0 0 0 0 4 0 9 (x 3 6x 2 + 9x 4)I = 0 + 0 + 9 0 0 0 4 0 0 = 0 4 0, 0 0 0 0 0 0 0 4 a very different answer. Thus, (xi A)(x n B n + x n 2 B n 2 + + xb + B 0 ) = c A (x)i is correct for scalars x but not for matrices. Now if x is a scalar, we can multiply the left hand side out, and collect powers of x to get x n I + x n (B n 2 AB n ) + + x(b 0 AB ) AB 0. If we denote c A (x) by x n + a n x n + + xa + a 0 then x n B n + x n (B n 2 AB n ) + + x(b 0 AB ) AB 0 () = x n I + a n x n I + + xa I + a 0 I. Two polynomials are equal if and only if they have the same coefficients, and this is true even if those coefficients are matrices. Consequently, B n = I, B n 2 AB n = a n I,..., B 0 ab = a I, AB 0 = a 0 I. In fact, for our running example, we see that B 2 = I, B AB 2 = 3 2 4 6I, B 0 AB = 3 2 4 = 9I and AB 0 = 4I. 3 2 4 At this point, it is possible to replace x by any 3 3 matrix, and everything will work. The problem is that for general matrices x, it need not be the case that (xi A)(x n B n + x n 2 B n 2 + + xb + B 0 ) = x n B n + x n (B n 2 AB n ) + + x(b 0 AB ) AB 0. That is, this equation is true for all scalars x but not all matrices. What does it take to make it true for matrices? We must pay attention to order in matrix multiplication. A careful multiplication shows To make (xi A)(x n B n + x n 2 B n 2 + + xb + B 0 ) = x n B n + (x n B n 2 Ax n B n ) + + (xb 0 AxB ) AB 0. x n B n + (x n B n 2 Ax n B n ) + + (xb 0 AxB ) AB 0 look like x n B n + x n (B n 2 AB n ) + + x(b 0 AB ) AB 0 Page 2

we need x to be able to commute with all powers of A. This was the problem with the matrix C we plugged in for x, that AC CA. But A obviously commutes with all powers of A so it is legal to replace x with A. When we do, we get c A (A) = A n B n + A n (B n 2 AB n ) + + A(B 0 AB ) AB 0, and you can see by the cancellation that this will be 0. This completes the proof. Here is an extension of the Cayley-Hamilton Theorem. It uses adj(xi A) to calculate the minimal polynomial of A. Suppose that the greatest common divisor of all the entries in adj(xi A) is g(x). Then the minimal polynomial is m A (x) = c A(x). The proof is essentially g(x) the same as the proof of the Cayley-Hamilton theorem. If every entry in adj(xi A) is divisible by g(x) then we may factor g(x) outside the matrix, and then write the remaining matrix as a polynomial in x. This would give adj(xi A) = g(x)(x m C m + x m C m + + xc + C 0 ) for some matrices C 0,..., C m. This means or c A (x)i = g(x)(xi A)(x m C m + x m C m + + xc + C 0 ), [c A (x)/g(x)]i = (xi A)(x m C m + x m C m + + xc + C 0 ). We now follow the rest of the proof of the Cayley-Hamilton theorem to justify that this equation remains valid when x is replaced by A. Next, if f(x) = x n C n + + xc + C 0 is any polynomial with matrix coefficients, then if x is a variable that commutes with A we can divide by xi A to get f(x) = (xi A)q(x) + R, for some polynomial q(x) with matrix coefficients and some matrix R. If f(x) annihilates A then 0 = f(a) = (A A)q(A) + R forces R = 0. In particular, if m A (x) is the minimal polynomial of A then we can write m A (x) = (xi A)q(x) for some polynomial with matrix coefficients. If c A (x) = m A (x)h(x) then we have (xi A)adj(xI A) = c A (x)i = m A (x)h(x)i = (xi A)h(x)q(x). Comparing the start and the end, adj(xi A) = h(x)q(x) = h(x)q(x) where Q(x) is a single matrix (not a polynomial) but with polynomial entries. This means that every entry in the adjoint of xi A must be divisible by h(x), and this finishes the proof. For example, with the matrix A we have been using as an example, if you look back at the adjoint of xi A and note that x 2 4x+3 = (x )(x 3) then we see that x is the greatest common divisor of the entries of the adjoint. The characteristic polynomial is x 3 6x 2 +9x 4 = (x ) 2 (x 4). Page 3

This means the minimal polynomial of A should be (x )2 (x 4) = (x )(x 4), which x is easy to check. 3 For another example, let A = 3 0 0 3. In this case, the adjoint is 0 0 3 (x 2)(x 3)(x 4) (x 2)(x 4) (x 2) 2 (x 2) 2 (x 2)(x 4) (x 2)(x 3)(x 4) (x 2) 2 (x 2) 2 0 0 (x 2)(x 3)(x 4) (x 2)(x 4) 0 0 (x 2)(x 4) (x 2)(x 3)(x 4) As we can see, the gcd of the terms in the adjoint is x 2. You should check that the characteristic polynomial is (x 2) 2 (x 4) 2 and that the minimal polynomial is (x 2)(x 4) 2. One final note on this material. A matrix with integer entries or polynomial entries can be put into a special form, called the Smith-Normal form: Given A there are matrices P c c 2 and Q each with determinant ± for which P AQ =.., a diagonal matrix.. In this matrix, each entry should divide subsequent entries. That is, c n is divisible by c n is divisible by c n 2, and so on. Moreover, det(a) = ±c c 2 c n. If we find the Smith-Normal form for xi A then we can make all the c s be monic polynomials, their product will be the characteristic polynomial, and c n will be the minimal polynomial. Let s show this on the matrix A above. We do row and column operations on xi A, but if we multiply or divide a row by anything other than or - we must do the reverse later. The method is to write the greatest common divisor of all the entries as a combination of the entries, use row/column operations to produce that gcd, interchange rows and columns to put that element in the (, ) position, and use it to get rid of everything in its row and column. Then work on the (n ) (n ) submatrix that s left, etc. In our case, the calculations look like this: xi A = x 3 x 3 0 0 x 3 0 0 x 3 c n x 3 x 3 0 0 x 3 0 0 x 3 x 3 0 x 2 6x + 8 x + 2 x + 2 0 0 x 3 0 x 2 6x + 8 x + 2 x + 2 0 0 x 3. 0 0 x 3 0 0 x 3 In the lower right 3 3 submatrix, the greatest common divisor of the entries is still. We could bring the (4, 3) entry to the (2, 2) entry by row/column operations. Doing this and. Page 4

continue the reduction: 0 x 2 6x + 8 x + 2 x + 2 0 0 x 3 0 0 x 3 0 0 x 3 0 0 x 3 0 x 2 6x + 8 x + 2 x + 2 0 0 x 3 0 x 3 0 0 0 x 3 0 0 0 x 2 6x + 8 0 x + 2 x 2 6x + 8 x + 2 0 0 x 2 6x + 8 (x 2 4x + 4) 0 0 0 0 0 0 x 2 6x + 8. 0 0 x 2 6x + 8 (x 2 4x + 4) Now things get more complicated. Each of the for lower right entries is divisible by x 2, since x 2 6x + 8 = (x 2)(x 4) and x 2 4x + 4 = (x 2) 2. If we interchange the third and fourth rows, and then the third and fourth columns, and then add the bottom row from to third, we get 0 0 0 0 0 0 0 0 x 2 + 4x 4 x 2 6x + 8 0 0 x 2 6x + 8 0 0 0 2x + 4 x 2 6x + 8 0 0 x 2 6x + 8 0 Fractions arise when using the (3, 3) entry: We multiply the third row by (x 4) and add 2 to the bottom: 0 0 0 0 0 2x + 4 x 2 6x + 8 0 0 0 (x 2 4)(x2 6x + 8) 0 0 0 0 0 2x + 4 0 0 0 0 (x 2 4)(x2 6x + 8) Finally, we make the diagonal polynomials monic by dividing the (3, 3) entry by -2, multiplying the (4, 4) entry by 2. Factoring, we have the Smith-Normal form: 0 0 0 0 0 x 2 0. 0 0 0 (x 2)(x 4) 2 Thus, the minimal polynomial is (x 2)(x 4) 2 and the characteristic polynomial polynomial is the product of the diagonals: (x 2) 2 (x 4) 2.. Page 5