Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices

Similar documents
Perron Frobenius Theory

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?

MATH36001 Perron Frobenius Theory 2015

Nonnegative and spectral matrix theory Lecture notes

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Random Walks on Graphs, and the Laplacian

LinGloss. A glossary of linear algebra

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

MAA704, Perron-Frobenius theory and Markov chains.

The Perron-Frobenius Theorem and its Application to Population Dynamics

Section 1.7: Properties of the Leslie Matrix

Detailed Proof of The PerronFrobenius Theorem

Notes on Linear Algebra and Matrix Theory

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

Spectral radius, symmetric and positive matrices

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Perron-Frobenius theorem for nonnegative multilinear forms and extensions

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Markov Chains and Stochastic Sampling

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Eigenvalues in Applications

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Convex Optimization CMU-10725

Linear Algebra: Matrix Eigenvalue Problems

Some bounds for the spectral radius of the Hadamard product of matrices

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Maths for Signals and Systems Linear Algebra in Engineering

On the simultaneous diagonal stability of a pair of positive linear systems

Quantum Computing Lecture 2. Review of Linear Algebra

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Properties of Matrices and Operations on Matrices

Review of Some Concepts from Linear Algebra: Part 2

A Brief Outline of Math 355

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

1 Linear Algebra Problems

T.8. Perron-Frobenius theory of positive matrices From: H.R. Thieme, Mathematics in Population Biology, Princeton University Press, Princeton 2003

Krylov Subspace Methods to Calculate PageRank

First, we review some important facts on the location of eigenvalues of matrices.

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013

Matrix functions that preserve the strong Perron- Frobenius property

Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

MAT 1302B Mathematical Methods II

Moore-Penrose Conditions & SVD

City Suburbs. : population distribution after m years

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini

Linear Algebra in Actuarial Science: Slides to the lecture

Geometric Mapping Properties of Semipositive Matrices

Chapter 7 Special matrices

Characterization of half-radial matrices

Synopsis of Numerical Linear Algebra

Diagonalization. P. Danziger. u B = A 1. B u S.

Example: Filter output power. maximization. Definition. Eigenvalues, eigenvectors and similarity. Example: Stability of linear systems.

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Recall : Eigenvalues and Eigenvectors

and let s calculate the image of some vectors under the transformation T.

Math 3191 Applied Linear Algebra

Lecture 8 : Eigenvalues and Eigenvectors

1 9/5 Matrices, vectors, and their applications

Mathematical foundations - linear algebra

On the spectra of striped sign patterns

REPRESENTATION THEORY WEEK 7

ECE 275A Homework # 3 Due Thursday 10/27/2016

Linear algebra and applications to graphs Part 1

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Computational math: Assignment 1

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Numerical Methods - Numerical Linear Algebra

7. Symmetric Matrices and Quadratic Forms

EE5120 Linear Algebra: Tutorial 6, July-Dec Covers sec 4.2, 5.1, 5.2 of GS

Calculating Web Page Authority Using the PageRank Algorithm

Lecture 3: graph theory

Lecture 2: September 8

Z-Pencils. November 20, Abstract

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

On the convergence of weighted-average consensus

Lecture 10 - Eigenvalues problem

Linear Algebra Primer

Diagonalization of Matrix

NOTES ON THE PERRON-FROBENIUS THEORY OF NONNEGATIVE MATRICES

Notes on Eigenvalues, Singular Values and QR

Discrete Math Class 19 ( )

An estimation of the spectral radius of a product of block matrices

Symmetric and anti symmetric matrices

Computational Linear Algebra

Lecture notes: Applied linear algebra Part 1. Version 2

UCSD ECE269 Handout #18 Prof. Young-Han Kim Monday, March 19, Final Examination (Total: 130 points)

From nonnegative matrices to nonnegative tensors

Wavelets and Linear Algebra

AN ITERATION. In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b

Lecture 15, 16: Diagonalization

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Math Homework 8 (selected problems)

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Transcription:

Geršgorin circles Lecture 8: Outline Chapter 6 + Appendix D: Location and perturbation of eigenvalues Some other results on perturbed eigenvalue problems Chapter 8: Nonnegative matrices Geršgorin s Thm: Let A = D +B, where D = diag(d 1,...,d n ), and B = [b ij ] M n has zeros on the diagonal. Define r i(b) = b ij C i (D,B) = {z C : z d i r i(b)} Then, all eigenvalues of A are located in n λ k (A) G(A) = C i (D,B) i=1 The C i (D,B) are called Geršgorin circles. If G(A) contains a region of k circles that are disjoint from the rest, then there are k eigenvalues in that region. k KTH Signal Processing 1 Magnus Jansson/Mats Bengtsson KTH Signal Processing 3 Magnus Jansson/Mats Bengtsson Geršgorin, Improvements Eigenvalue Perturbation Results, Motivation We know from a previous lecture that ρ(a) A for any matrix norm. That is, we know that all eigenvalues are in a circular disk with radius upper bounded by any matrix norm. Better results? What can be said about the eigenvalues and eigenvectors of A+ǫB when ǫ is small? Since A T has the same eigenvalues as A, we can do the same but summing over columns instead of rows. We conclude that λ i (A) G(A) G(A T ) Since S 1 AS has the same eigenvalues as A, the above can be improved by i λ i (A) G(S 1 AS) G ( (S 1 AS) T) for any choice of S. For it to be useful, S should be simple, e.g., diagonal (see Corollary 6.1.6). i KTH Signal Processing 2 Magnus Jansson/Mats Bengtsson KTH Signal Processing 4 Magnus Jansson/Mats Bengtsson

Invertibility and stability If A M n is strictly diagonally dominant such that then 1. A is invertible. a ii > a ij i 2. If all main diagonal elements are real and positive then all eigenvalues are in the right half plane. 3. If A is Hermitian with all diagonal elements positive, then all eigenvalues are real and positive. Irreducibly diagonally dominant If A M n is called irreducibly diagonally dominant if i) A is irreducible (= A has the SC property). ii) A is diagonally dominant, iii) For at least one row, i, a ii a ij i a ii > a ij KTH Signal Processing 5 Magnus Jansson/Mats Bengtsson KTH Signal Processing 7 Magnus Jansson/Mats Bengtsson Reducible matrices A matrix A M n is called reducible if n = 1 and A = 0 or n 2 and there is a permutation matrix P M n such that P T AP = B C } r 0 D } n r }{{}}{{} r n r for some integer 1 r n 1. Invertibility and stability, stronger result If A M n is irreducibly diagonally dominant, then 1. A is invertible. 2. If all main diagonal elements are real and positive then all eigenvalues are in the right half plane. 3. If A is Hermitian with all diagonal elements positive, then all eigenvalues are real and positive. A matrix A M n that is not reducible is called irreducible. A matrix is irreducible iff it describes a strongly connected directed graph, A has the SC property. KTH Signal Processing 6 Magnus Jansson/Mats Bengtsson KTH Signal Processing 8 Magnus Jansson/Mats Bengtsson

Perturbation theorems Thm: Let A,E M n and let A be diagonalizable, A = SΛS 1. Further, let ˆλ be an eigenvalue of A+E. Then there is some eigenvalue λ i of A such that ˆλ λ i S S 1 E = κ(s) E for some particular matrix norms (e.g., 1, 2, ). Cor: If A is a normal matrix, S is unitary = S 2 = S 1 2 = 1. This gives ˆλ λ i E 2 indicating that normal matrices are perfectly conditioned for eigenvalue computations. Perturbation of a simple eigenvalue Let λ be a simple eigenvalue of A M n and let y and x be the corresponding left and right eigenvectors. Then y x 0. Thm: Let A(t) M n be differentiable at t = 0 and assume λ is a simple eigenvalue of A(0) with left and right eigenvectors y and x. If λ(t) is an eigenvalue of A(t) for small t such that λ(0) = λ then λ (0) = y A (0)x y x Example: A(t) = A+tE gives λ (0) = y Ex y x. KTH Signal Processing 9 Magnus Jansson/Mats Bengtsson KTH Signal Processing 11 Magnus Jansson/Mats Bengtsson Perturbation cont d If both A and E are Hermitian, we can use Weyl s theorem (here we assume the eigenvalues are indexed in non-decreasing order): λ 1 (E) λ k (A+E) λ k (A) λ n (E) We also have for this case [ n ] 1/2 λ k (A+E) λ k (A) 2 E 2 k=1 where 2 is the Frobenius norm. k Perturbation of eigenvalues cont d Errors in eigenvalues may also be related to the residual r = Aˆx ˆλˆx. Assume for example that A is diagonalizable A = SΛS 1 and let ˆx and ˆλ be a given complex vector and scalar, respectively. Then there is some eigenvalue of A such that ˆλ λ i κ(s) r ˆx (for details and conditions see book). We conclude that a small residual implies a good approximation of the eigenvalue. KTH Signal Processing 10 Magnus Jansson/Mats Bengtsson KTH Signal Processing 12 Magnus Jansson/Mats Bengtsson

Literature with perturbation results J. R. Magnus and H. Neudecker. Matrix Differential Calculus with Applications in Statistics and Econometrics. John Wiley & Sons Ltd., 1988. H. Krim and P. Forster. Projections on unstructured subspaces. IEEE Trans. SP, 44(10):2634 2637, Oct. 1996. J. Moro, J. V. Burke, and M. L. Overton. On the Lidskii-Vishik- Lyusternik perturbation theory for eigenvalues of matrices with arbitrary Jordan structure. SIAM Journal on Matrix Analysis and Applications, 18(4):793 817, 1997. F. Rellich. Perturbation Theory of Eigenvalue Problems. Gordon & Breach, 1969. J. Wilkinson. The Algebraic Eigenvalue Problem. Clarendon Press, 1965. Perturbation of eigenvectors with simple eigenvalues: The real symmetric case Assume that A M n (R) is real symmetric matrix with normalized eigenvectors x i and eigenvalues λ i. Further assume that λ 1 is a simple distinct eigenvalue. Let  = A+ǫB where ǫ is a small scalar, B is real symmetric and let ˆx 1 be an eigenvector of  that approaches x 1 as ǫ 0. Then a first order approximation (in ǫ) is ˆx 1 x 1 = ǫ x T k Bx 1 x k λ 1 λ k k=2 Warning: Non-unique derivative in the complex valued case! Warning, Warning Warning: No extension to multiple eigenvalues! KTH Signal Processing 13 Magnus Jansson/Mats Bengtsson KTH Signal Processing 15 Magnus Jansson/Mats Bengtsson Perturbation of eigenvectors with simple eigenvalues Thm: Let A(t) M n be differentiable at t = 0 and assume λ 0 is a simple eigenvalue of A(0) with left and right eigenvectors y 0 and x 0. If λ(t) is an eigenvalue of A(t), it has a right eigenvector x(t) for small t normalized such that x 0x(t) = 1 with derivative ( x (0) = (λ 0 I A(0)) I x 0y0 ) y0 x A (0)x 0 0 B denotes the Moore-Penrose pseudo inverse of a matrix B. (See, e.g., J. R. Magnus and H. Neudecker. Matrix Differential Calculus with Applications in Statistics and Econometrics. John Wiley & Sons Ltd., 1988, rev. 1999) Chapter 8: Nonnegative matrices Def: A matrix A = [a ij ] M n,r is nonnegative if a ij 0 for all i,j, and we write this as A 0. (Note that this should not be confused with the matrix being nonnegative definite!) If a ij > 0 for all i,j, we say that A is positive and write this as A > 0. (We write A > B to mean A B > 0 etc.) We also define A = [ a ij ]. Typical applications where nonnegative or positive matrices occur are problems in which we have matrices where the elements correspond to probabilities (e.g., Markov chains) power levels or power gain factors (e.g., in power control for wireless systems). any other application where only nonnegative quantities appear. KTH Signal Processing 14 Magnus Jansson/Mats Bengtsson KTH Signal Processing 16 Magnus Jansson/Mats Bengtsson

Nonnegative matrices: Some properties Let A,B M n and x C n. Then Ax A x AB A B If A 0, then A m 0; if A > 0, then A m > 0. If A 0, x > 0, and Ax = 0 then A = 0. Nonnegative matrices: Spectral radius Thm: Let A M n and A 0. Then min i min j i=1 a ij ρ(a) max i a ij ρ(a) max j Thm: Let A M n and A 0. If Ax = λx and x > 0, then λ = ρ(a). i=1 a ij a ij If A B, then A B, for any absolute norm ; that is, a norm for which A = A. KTH Signal Processing 17 Magnus Jansson/Mats Bengtsson KTH Signal Processing 19 Magnus Jansson/Mats Bengtsson Nonnegative matrices: Spectral radius Positive matrices For positive matrices we can say a little more. Lemma: If A M n, A 0, and if the row sums of A are constant, then ρ(a) = A. If the column sums are constant, then ρ(a) = A 1. The following theorem can be used to give upper and lower bounds on the spectral radius of arbitrary matrices. Thm: Let A,B M n. If A B, then ρ(a) ρ( A ) ρ(b). Perron s theorem: If A M n and A > 0, then 1. ρ(a) > 0 2. ρ(a) is an eigenvalue of A 3. There is an x R n with x > 0 such that Ax = ρ(a)x 4. ρ(a) is an algebraically (and geometrically) simple eigenvalue of A 5. λ < ρ(a) for every eigenvalue λ ρ(a) of A 6. [A/ρ(A)] m L as m, where L = xy T, Ax = ρ(a)x, y T A = ρ(a)y T, x > 0, y > 0, and x T y = 1. The root ρ(a) is sometimes called a Perron root and the vector x = [x i ] a Perron vector if it is scaled such that n i=1 x i = 1. KTH Signal Processing 18 Magnus Jansson/Mats Bengtsson KTH Signal Processing 20 Magnus Jansson/Mats Bengtsson

Nonnegative matrices Generalization of Perron s theorem to general non-negative matrices? Thm: If A M n and A 0, then 1. ρ(a) is an eigenvalue of A 2. There is a non-zero x R n with x 0 such that Ax = ρ(a)x For stronger results, we need a stronger assumption on A. Irreducible matrices Frobenius theorem: If A M n, A 0 is irreducible, then 1. ρ(a) > 0 2. ρ(a) is an eigenvalue of A 3. There is an x R n with x > 0 such that Ax = ρ(a)x 4. ρ(a) is an algebraically (and geometrically) simple eigenvalue of A 5. If there are exactly k eigenvalues with λ p = ρ(a), p = 1,...,k, then λ p = ρ(a)e i2πp/k, p = 0,1,...,k 1 (suitably ordered) If λ is any eigenvalue of A, then λe i2πp/k is also an eigenvalue of A for all p = 0,1,...,k 1 diag[a m ] 0 for all m that are not multiples of k (e.g. m = 1). KTH Signal Processing 21 Magnus Jansson/Mats Bengtsson KTH Signal Processing 23 Magnus Jansson/Mats Bengtsson Irreducible matrices Reminder: A matrix A M n, n 2 is called reducible if there is a permutation matrix P M n such that P T AP = B C } r 0 D } n r }{{}}{{} r n r for some integer 1 r n 1. A matrix A M n that is not reducible is called irreducible. Thm: A matrix A M n with A 0 is irreducible iff (I +A) n 1 > 0 Primitive matrices A matrix A M n, A 0 is called primitive if A is irreducible ρ(a) is the only eigenvalue with λ p = ρ(a). Thm: If A M n, A 0 is primitive, then lim m [A/ρ(A)]m = L where L = xy T, Ax = ρ(a)x, y T A = ρ(a)y T, x > 0, y > 0, and x T y = 1. Thm: If A M n, A 0, then it is primitive iff A m > 0 for some m 1. KTH Signal Processing 22 Magnus Jansson/Mats Bengtsson KTH Signal Processing 24 Magnus Jansson/Mats Bengtsson

Stochastic matrices A nonnegative matrix with all its row sums equal to 1 is called a (row) stochastic matrix. A column stochastic matrix is the transpose of a row stochastic matrix. If a matrix is both row and column stochastic it is called doubly stochastic. Example, Markov processes Consider a discrete stochastic process that at each time instant is in one of the states S 1,...,S n. Let p ij be the probability to change from state S i to state S j. Note that the transition matrix P = [p ij ], is a stochastic matrix. Let µ i (t) denote the probability of being in state S i at time t and µ(t) = [µ 1 (t),...,µ n (t)], then µ(t+1) = µ(t)p contains the corresponding probabilities for time t+1. If P is primitive (other terms are used in the statistics literature), then µ(t) µ as t where µ = µ P, no matter what µ(0) is. µ is called the stationary distribution. Nice article: The Perron Frobenius Theorem: Some of its applications, S. U. Pillai, T. Suel, S. Cha, IEEE Signal Processing Magazine, Mar. 2005. KTH Signal Processing 25 Magnus Jansson/Mats Bengtsson KTH Signal Processing 27 Magnus Jansson/Mats Bengtsson Stochastic matrices cont d The set of stochastic matrices in M n is a compact convex set. Let 1 = [1,1,...,1] T. A matrix is stochastic if and only if A1 = 1 = 1 is an eigenvector with eigenvalue +1 of all stochastic matrices. An example of a doubly stochastic matrix is A = [ u ij 2 ] where U = [u ij ] is a unitary matrix. Also, notice that all permutation matrices are doubly stochastic. Thm: A matrix is doubly stochastic if and only if it can be written as a convex combination of a finite number of permutation matrices. Corr: The maximum of a convex function on the set of doubly stochastic matrices is attained at a permutation matrix! Other books contain more results. Further results In Matrix Theory, vol. II by Gantmacher, for example, you can find results such as: Thm: If A M n, A 0 is irreducible, then for all α > ρ(a). (αi A) 1 > 0 (Useful, for example, in connection with power control of wireless systems). KTH Signal Processing 26 Magnus Jansson/Mats Bengtsson KTH Signal Processing 28 Magnus Jansson/Mats Bengtsson