Matrices A(t) depending on a Parameter t. Jerry L. Kazdan

Similar documents
Math 312 Final Exam Jerry L. Kazdan May 5, :00 2:00

DIFFERENTIABLE PERTURBATION OF UNBOUNDED OPERATORS. Andreas Kriegl, Peter W. Michor

Math Ordinary Differential Equations

Math 108b: Notes on the Spectral Theorem

ft-uiowa-math2550 Assignment NOTRequiredJustHWformatOfQuizReviewForExam3part2 due 12/31/2014 at 07:10pm CST

MATH 115A: SAMPLE FINAL SOLUTIONS

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 250B Midterm II Information Spring 2019 SOLUTIONS TO PRACTICE PROBLEMS

Properties of Linear Transformations from R n to R m

Solutions for Math 225 Assignment #5 1

Problem Set 9 Due: In class Tuesday, Nov. 27 Late papers will be accepted until 12:00 on Thursday (at the beginning of class).

BASIC MATRIX PERTURBATION THEORY

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Diagonalization of Matrix

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

Econ Slides from Lecture 7

Reduction to the associated homogeneous system via a particular solution

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Exercise Set 7.2. Skills

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Final Examination 201-NYC-05 - Linear Algebra I December 8 th, and b = 4. Find the value(s) of a for which the equation Ax = b

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Linear Algebra II Lecture 13

Lecture Summaries for Linear Algebra M51A

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Generalized eigenspaces

Math Abstract Linear Algebra Fall 2011, section E1 Practice Final. This is a (long) practice exam. The real exam will consist of 6 problems.

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

Math 315: Linear Algebra Solutions to Assignment 7

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Linear Differential Equations. Problems

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

k is a product of elementary matrices.

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

MATH 583A REVIEW SESSION #1

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Solutions for Math 225 Assignment #4 1

Diagonalization. P. Danziger. u B = A 1. B u S.

Math 113 Final Exam: Solutions

SPECTRAL THEORY EVAN JENKINS

Homework 2 Solutions

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Final Exam. Linear Algebra Summer 2011 Math S2010X (3) Corrin Clarkson. August 10th, Solutions

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

4. Linear transformations as a vector space 17

MATH 235. Final ANSWERS May 5, 2015

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

Solutions to Final Exam

Eigenvalues and Eigenvectors

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision?

MATH 24 EXAM 3 SOLUTIONS

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A =

Linear Algebra Exercises

Linear Algebra (Math-324) Lecture Notes

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Linear Algebra- Final Exam Review

REVIEW FOR EXAM II. The exam covers sections , the part of 3.7 on Markov chains, and

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

1 Inner Product and Orthogonality

Notes on Special Functions

Definition (T -invariant subspace) Example. Example

Maths for Signals and Systems Linear Algebra in Engineering

Math 3191 Applied Linear Algebra

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

Answer Keys For Math 225 Final Review Problem

A note on linear differential equations with periodic coefficients.

Linear Algebra Primer

JUST THE MATHS SLIDES NUMBER 9.6. MATRICES 6 (Eigenvalues and eigenvectors) A.J.Hobson

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

MAC Module 12 Eigenvalues and Eigenvectors

6 Linear Equation. 6.1 Equation with constant coefficients

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

Lecture 15, 16: Diagonalization

235 Final exam review questions

1. General Vector Spaces

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Remarks on Definitions

MATH 369 Linear Algebra

COMP 558 lecture 18 Nov. 15, 2010

Linear Algebra (MATH ) Spring 2011 Final Exam Practice Problem Solutions

A Brief Outline of Math 355

Math Linear Algebra II. 1. Inner Products and Norms

LINEAR ALGEBRA QUESTION BANK

Section 9.8 Higher Order Linear Equations

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear algebra and differential equations (Math 54): Lecture 10

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

2015 School Year. Graduate School Entrance Examination Problem Booklet. Mathematics

CSL361 Problem set 4: Basic linear algebra

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Columbus State Community College Mathematics Department Public Syllabus

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam June 10, 2011

Transcription:

Matrices A(t depending on a Parameter t Jerry L. Kazdan If a square matrix A(t depends smoothly on a parameter t are its eigenvalues and eigenvectors also smooth functions of t? The answer is yes most of the time but not always. This story while old is interesting and elementary and deserves to be better known. One can also ask the same question for objects such as the Schrödinger operator whose potential depends on a parameter. Much of our current understanding arose from this fundamental example. Warm-up Exercise Given a polynomial p(xt = x n + a n 1 (tx n 1 + + a 1 (tx + a 0 (t whose coefficients depend smoothly on a parameter t. Assume at t = 0 the number x = c is a simple root of this polynomial p(c0 = 0. Show that for all t sufficiently near 0 there is a unique root x(t with x(0 = c that depends smoothly on t. Moreover if p(xt is a real analytic function of t that is it has a convergent power series expansion in t near t = 0 then so does x(t. SOLUTION: Given that p(c0 = 0 we want to solve p(xt = 0 for x(t with x(0 = c. The assertions are immediate from the implicit function theorem. Since x(0 = c is a simple zero of p(x0 = 0 then p(x0 = (x cg(x where g(c 0. Thus the derivative p x (c0 0. The example p(xt := x 3 t = 0 so x(t = t 1/3 shows x(t may not be a smooth function at a multiple root. In this case the best one can get is a Puiseux expansion in fractional powers of t (see [Kn 15]. The Generic Case: a simple eigenvalue In the following let λ be an eigenvalue and X a corresponding eigenvector of a matrix A. We say λ is a simple eigenvalue if λ is a simple root of the characteristic polynomial. We will use the equivalent version: if (A λ V = 0 then V = cx for some constant c. [The proof of the equivalence is for instance an immediate consequence of the fact that every matrix is similar to an upper triangular matrix.] The point is to eliminate matrices such as the zero matrix A = 0 where λ = 0 is a double eigenvalue and any vector V 0 is an eigenvector as well as the more complicated matrix A = ( 0 1 0 0 which has λ = 0 as an eigenvalue with geometric multiplicity one but algebraic multiplicity two. Theorem Given a square matrix A(t whose elements depend smoothly on a real parameter t if λ = λ 0 is a simple eigenvalue at t = 0 with a corresponding unit eigenvector X 0 then for all t near 0 there is a corresponding eigenvalue and unique (normalized eigenvector that depend smoothly on t. Also if the elements of A(t are real analytic function of t then so are the eigenvalue and eigenvector. 1

The only books I know of that have proofs of this theorem are [Evans p.?] and [Lax p. 131]. The proofs there are different. PROOF: Although we won t use it the eigenvalue part is immediate from the warm-up exercise above applied to the characteristic polynomial. It is the eigenvector aspect that is takes a bit more work. Given A(0X 0 = λ 0 X 0 for some vector X 0 with X 0 = 1 we want a function λ(t and a vector X(t that depend smoothly on t with the properties A(tX(t = λ(tx(t X 0 X(t = 1 and λ(0 = λ 0 X(0 = X 0. Here X Y is the standard inner product. Of course we could also have used a different normalization such as X(t = 1. SOME BACKGROUND ON THE IMPLICIT FUNCTION THEOREM. If H : R N R R N say we want to solve the equations H(Zt = 0 for Z = Z(t. These are N equations for the N unknowns Z(t. Assume that Z = Z 0 is a solution at t = 0 so H(Z 0 0 = 0. Expanding H in a Taylor series in the variable Z near Z = Z 0 we get H(Zt = H(Z 0 t+h Z (Z 0 t(z Z 0 + where H Z is the derivative matrix and represent higher order terms. If these higher order terms were missing then the solution of H(Zt = 0 would be simply Z Z 0 = [H Z (Z 0 t] 1 H(Z 0 t that is Z = Z 0 [H Z (Z 0 t] 1 H(Z 0 t. This assumes that the first derivative matrix H Z (Z 0 0 is invertible (since it is then invertible for all t near zero. The implicit function theorem says that this is still true even if there are higher order terms. The key assumption is that the first derivative matrix H Z (Z 0 0 is invertible. Although we may think of the special case where t R this works without change if the parameter t R k is a vector. (CONTINUATION OF THE PROOF We may assume that λ 0 = 0. Write our equations as f(xλt A(tX λx F(Xλt := := g(x λt X 0 X 1 where we have written f(xλt := A(tX λx and g(xλt := X 0 X 1. We wish to solve: F(Xλt = 0 for both X(t and λ(t near t = 0. In the notation of the previous paragraph Z = (Xλ and H(Zt = F(Xλt. Thus the derivative matrix H Z involves differentiation with respect to both X and λ. The derivative with respect to the parameters X and λ is the partitioned matrix F fx f (Xλt = λ = g X g λ Here we used X 0 X = X0 T X where X 0 T at t = 0 F (X 0 λ 0 0 = This was worked out at the blackboard with Dennis DeTurck. ( A(t λi X X0 T 0. is the transpose of the column vector X 0. Thus A(0 λ0 I X 0 X0 T. 0

For the implicit function theorem we check that the matrix on the right is invertible. It is enough to show its kernel is zero. Thus say F (X 0 λ 0 0W = 0 where W = V r 0. Then A(0V = λ 0 V + rx 0 and X 0 V = 0. Assume V 0. Then by the last condition X 0 and V are linearly independent. Write A(0 as a matrix in a basis whose first vector is X 0 and second is V. Since AX 0 = λ 0 X 0 and A(0V = rx 0 + λ 0 V in this basis λ 0 r 0 λ 0 A(0 = 0 0... 0 0 where can be anything. Thus λ 0 is at least a double root of the charachteristic polynomial which contradicts the assumption that λ 0 is a simple eigenvalue of A(0. Thus V = 0 and hence r = 0 so W = 0. Consequently the derivative matrix F (X 0 λ 0 0 is invertible. By the implicit function theorem the equation F(X λt = 0 has the desired smooth solution X = X(t λ = λ(t near t = 0. If F is real analytic in t near t = 0 then so is this solution. The General Case If the eigenvalue λ(t is not simple then the situation is more complicated except if A(t is self-adjoint and depends analytically on t. Then both the eigenvalue and corresponding eigenvectors are analytic functions of t (see [Ka] and [Re]. However it is obviously false that the largest eigenvalue is an analytic function of t as one sees from the simple example ( t 0 0 0 for t near 0. We conclude with three examples showing that if either the self-adjoint or analyticity assumptions are deleted the eigenvalue and/or eigenvector may not depend smoothly on t. 0 1 EXAMPLE 1 At t = 0 the matrix A(t = has 0 as a double eigenvalue. Since the t 0 characteristic polynomial is p(t := λ t the eigenvalues are not smooth functions of t for t near 0. EXAMPLE a This is a symmetric matrix depending smoothly (but not analytically on t. Near t = 0 the eigenvectors are not even continuous functions of t. This is from Rellich s nice book [Re page 41]. Let B(t = ( a(t 0 0 a(t ( cosϕ(t sinϕ(t and R(t = sin ϕ(t cos ϕ(t where a(0 = 0. For t 0 we let a(t := exp( 1/t and ϕ(t := 1/t. The desired symmetric matrix is A(t = R(tB(tR 1 (t. It is similar to B(t but the new basis determined by the orthogonal matrix R(t is spinning quickly near t = 0. We find cosϕ(t sinϕ(t A(t = a(t sin ϕ(t cos ϕ(t 3

4 Its eigenvalues are λ ± = ±a(t. For t 0 the orthonormal eigenvectors are cosϕ(t sinϕ(t V + (t = and V sin ϕ(t (t =. cos ϕ(t Since a(t goes to 0 so quickly near t = 0 even though ϕ(t = 1/t the matrix A(t is a C function of t. However the eigenvectors keep spinning and don t even converge as t 0. EXAMPLE b Another example of the same phenomenon. Let M + and M be symmetric matrices with different orthonormal eigenevctors V 1 V and W 1 W respectively. For instance M + = 0 1 1 0 and M = 1 0 0 so we can let V 1 = ( 1 1 V = ( 1 1 and W 1 = (10 W (01. With a(t := exp( 1/t as above let A(t = { a(tm + for t > 0 a(tm for t < 0 with A(0 = 0. This matrix A(t depends smoothly on t and has the eigenvectors V 1 and V for t > 0 but W 1 and W for t < 0. The eigenvectors are not continuous in a neighborhood of t = 0. EXAMPLE 3 In the previous example the eigenvalues were still smooth functions but this was lucky. There are symmetric matrices depending smoothly on a parameter t whose eigenvalues are not C functions of t. Since the eigenvalues are roots of the characteristic polynomial this is just the situation of a polynomial whose coefficients depend on a parameeter and asking how smoothly the roots depend on the parameter. One instance is x f(t = 0 where f(t 0 is smooth. The key observation (certainly known by Rellich in [Re] is the perhaps surprising fact that this f(t may not have a smooth square root. This has been rediscovered many times. One example is f(t = sin (1/te 1/t + e /t for t > 0 while f(t = 0 for t 0. For a recent discussion with additional details and references see [AKML]. References [AKML] Alekseevsky D. Kriegl A. Michor P. Losik M. Choosing Roots of Polynomials Smoothly II Israel Journal of Mathematics 139 004 183 188. [Evans] Evans L. C. Partial Differential Equations (n d. ed. Grad. Stud. Math. vol. 19 Amer. Math. Soc. Providence RI 010. [Ka] [Kn] Kato Tosio Perturbation theory for linear operators Grundlehren 13 1976 Springer- Verlag Berlin New York. Knopp K. Theory of Functions Vol. II 1947 Dover New York (translated from the German. [Lax] Lax Peter D. Linear Algebra and Its Applications nd. ed. Wiley-Interscience 007

5 [Re] Rellich Franz Perturbation Theory of Eigenvalue Problems 1953 New York University Lecture Notes reprinted by Gordon and Breach 1968. Jerry L. Kazdan Department of Mathematics University of Pennsylvania Philadelphia PA 19104-6395 (U.S.A. email: kazdan@math.upenn.edu