T.8. Perron-Frobenius theory of positive matrices From: H.R. Thieme, Mathematics in Population Biology, Princeton University Press, Princeton 2003

Size: px
Start display at page:

Download "T.8. Perron-Frobenius theory of positive matrices From: H.R. Thieme, Mathematics in Population Biology, Princeton University Press, Princeton 2003"

Transcription

1 T.8. Perron-Frobenius theory of positive matrices From: H.R. Thieme, Mathematics in Population Biology, Princeton University Press, Princeton 2003 A vector x R n is called positive, symbolically x > 0, if all components are nonnegative and at least one is positive. It is called strictly positive, x 0, if all components are positive. A square matrix is called positive if all entries are non-negative numbers and the matrix is not the zero matrix. It is called quasi-positive if it is not the zero matrix and all off-diagonal entries are non-negative numbers. It is called strictly positive if all entries are strictly positive. If n 2, an n n matrix A = (a ik ) is called irreducible if if the following holds: For any proper non-empty subset P of {1,, m} there are k P, j P such that a jk 0. A 1 1 matrix is called irreducible if it is not the 0 matrix. Equivalently A is irreducible if and only if, for all i, k = 1,..., n, there exist numbers j 1,..., j r {1,..., n} such that i = j 1, k = j r and a jl j l+1 0 for all l = 1,... r 1. A non-negative matrix A is irreducible if and only if the matrix exponential e A is strictly positive. A non-negative square matrix A is called primitive if one of its powers, A k, has strictly positive entries. It is easily seen that a non-negative matrix is primitive if it is irreducible and all entries in its main diagonal are strictly positive (Exercise 2). If A is a complex square matrix, a complex number λ is called a spectral value of A if the matrix λ A is singular. The set of spectral values of A is called the spectrum of A and is denoted by σ(a). For a matrix, a spectral value is an eigenvalue and vice versa, i.e., there exists a non-zero vector x, called eigenvector of A, such that Ax = λx. The spectral radius of the matrix A, r(a), is defined as r(a) = max{ λ ; λ σ(a)}, while the spectral bound of the matrix A, s(a) is defined as s(a) = max{rλ; λ σ(a)}. Obviously, s(a) r(a). Theorem 8.1. Let A be a positive matrix. Then its spectral radius, r(a), is an eigenvalue associated both with a positive eigenvector of A and a positive eigenvector of the transposed matrix A. In particular, s(a) = r(a). 29

2 For a proof see Schaefer (1974), Prop. I.2.3. Theorem 8.2. Let A be a positive matrix and x > 0 be a vector and µ 0 such that A q z µz for some natural number q and some vector z = z z with z, z R m +, z R m +. Then the spectral radius of A satisfied r(a) µ 1/q. Proof: This is the finite dimensional special case of Theorem 2.5 by Krasnosel skii (1964). Theorem 8.3. Let A be a quasi-positive matrix. Then its spectral bound (modulus of stability), s(a), is an eigenvalue of A associated both with a positive eigenvector of A and a positive eigenvector of the transposed matrix A. Moreover if x > 0 is a vector and µ R such that Ax µx, there exists some vector z > 0 and some scalar λ µ such that Az = λz and in particular s(a) µ. Proof: Since all off-diagonal elements of A are non-negative, then the matrix A + νi is non-negative for some (and then all) sufficiently large ν > 0. Let λ C be an eigenvalue of A such that Rλ = s(a). Then there exists a (possibly complex) vector x 0 such that Ax = λx. So (A + ν)x = (λ + ν)x. Let x = ( x 1,..., x n ) be the modulus (or absolute value) of the vector x. Since A+νI is a positive matrix, ν +λ x = (A+ν)x (A+ν) x. By Corollary 8.2 and Theorem 8.1, there exists some r ν+λ ν+s(a) and some vector z > 0 such that (A + ν)z = rz. So Az = (r ν)z. By definition of s(a), r ν s(a). Together with our previous inequality, r ν = s(a) and s(a) is an eigenvalue of A associated with a non-negative eigenvector. Since s(a) = s(a ) and A is a quasi-positive matrix, we can conclude that s(a) is also associated with a positive eigenvector of A. Now let Ax µx for some vector x > 0 and some µ R. Then (A + ν)x (ν + µ)x. Since A + νi is a positive matrix, by Corollary 8.2 and Theorem 8.1, there exists some r (ν + µ) and some vector z > 0 such that (A + ν)z = rz. Obviously Az = (r ν)z and r ν µ. So we choose λ = r ν. By definition of s(a), λ s(a) and so µ s(a). Theorem 8.4. Let A and D be positive matrices, D diagonal with all diagonal elements being positive. Then s(a D) and r(d 1 A) 1 have the same sign, i.e., these numbers are simultaneously positive, zero, or negative. 30

3 Proof: Let λ = s(a D). Since the off-diagonal elements of A D are non-negative, by Theorem 8.3 there exists some vector x > 0 such that (A D)x = λx. Reorganizing terms, Ax = Dx + λx. Since D is an invertible matrix, D 1 Ax = x + λd 1 x (1 + λɛ)x with ɛ being the reciprocal of the largest of the diagonal elements in D. So r(d 1 A) 1 + λɛ. Now let r = r(d 1 A) 1. By Theorem 1 there exist some vector x > 0 such that D 1 Ax = rx. Reorganizing terms, Ax = rdx. So (A D)x = (r 1)Dx (r 1)dx with d being the smallest diagonal element of D. By Theorem 8.3, s(a D) (r 1)d. The continuous and discrete dynamical systems associated with irreducible quasipositive or even primitive matrices have a strikingly simple large-time behavior. In the following x, y = m k=1 x ky k is the canonical scalar (or inner) product on R m. Theorem 8.5. Let A be a quasi-positive irreducible matrix. Then s = s(a) is an eigenvalue of both A and A with strictly positive eigenvectors v and v and s(a) is larger than the real parts of all other eigenvalues of A. Further any non-negative solution x of the differential equation x = Ax which is not identical 0 satisfies e s(t r) x(t) x(r), v v, v v, t, r > 0. Proof: If A is a quasi-positive irreducible matrix, then A+ν is a positive irreducible matrix for a sufficiently large ν > 0. So all matrices e ta = e νt e t(a+ν) are strictly positive and so form an irreducible uniformly continuous semigroup of compact operators on the Banach lattice R m. The claim now follows from Theorem 9.11 in Heijmans, de Pagter (1987). Remarks. Theorem 8.5 has significant side effects for an irreducible quasi-positive matrix A: (a) Every subspace that is forward invariant under A and contains a positive vector also contains the eigenvector v associated with s(a). In particular (b) Eigenvalues of A different from s(a) have no positive eigenvector or positive generalized eigenvector. (c) There are no generalized eigenvectors associated with s(a) and the eigenspace associated with s(a) is one-dimensional. In other words, s(a) is a simple eigenvalue. 31

4 (d) s(a) > Rλ for all eigenvalues λ of A that are different from s(a). Proof of (a): Let Y be a subspace of X and x 0 Y positive. Consider the solution x = Ax with x(0) = x 0. Then x(t) = e ta x 0 Y for all t 0 and so is v = lim t v, v x 0, v e s(a)t x(t). (d) Let λ be an eigenvalue of A that is different from s(a), but with Rλ = s(a). Let x be the solution of x = Ax with x(0) = v being the eigenvector associated with λ. Then e s(a)t x(t) = e ı(iλ)t v does not converge, contradicting the statement in Theorem 8.5. Lemma 8.6. Let A be a quasi-positive irreducible matrix and Ax λx or A x λx with λ R and x being a positive vector. Then s(a) λ. Proof: We consider Ax λx, the other case is done similarly. By Theorem 8.5, there exists a strictly positive vector v such that A v = sv with s = s(a). Then λ x, v = λx, v Ax, v = x, A v = x, sv = s x, v. Since the vector x is positive and the vector v strictly positive, x, v > 0 and λ s = s(a) follows by division. Remark. Choosing x with x j = 1 for j = 1,..., m in Lemma 8.6 provides the estimates s(a) max 1 j m k=1 m a jk, s(a) max 1 k m j=1 m a jk. Since the vector x is strictly positive, it is actually sufficient that A is quasi-positive. Theorem 8.7. Let A be a primitive matrix with spectral radius r = r(a) = s(a). Then r k A k x x, v v, v v, k where v and v are strictly positive eigenvectors of A and A associated with the eigenvalue r according to Theorem 8.5. A perhaps more intuitive equivalent formulation is the following one. 32

5 Corollary 8.8. Let A be a primitive matrix with spectral radius r = r(a) = s(a). Then, for every positive vector x and for every norm on R m, A k x A k x v v, k, where v is a positive eigenvector of A associated with the eigenvalue r according to Theorem 8.5. Proof: Corollary 8.8 obviously follows from Theorem 8.7 by the continuity of the norm. The converse follows by choosing the norm x = x, v where x is the vector ( x 1,..., x n ) and v a strictly positive eigenvector of A associated with r. Theorem 8.7 is only valid for primitive matrices. Actually, for non-negative matrices, primitivity is equivalent to the ergodicity statement in Theorem 8.7 (Schaefer, 1974, I.Proposition 7.3). But mean ergodicity still holds for irreducible matrices, which means that the convergence in Theorem 8.7 holds in average (Schaefer, 1974, end of Section I.6). Notice that the convergence statement in Theorem 8.7 implies the convergence statement in Theorem 8.9 (Exercise 2). Theorem 8.9. Let A be an irreducible positive matrix with spectral radius r = r(a). Then 1 k + 1 k j=0 r j A j x x, v v, v v, k where v and v are the strictly positive eigenvectors of A and A associated with the eigenvalue r according to Theorem 8.5. Noticing that A j x, v = r j x, v, Theorem 8.8 can be reformulated as follows. Theorem Let A be an irreducible positive matrix with spectral radius r = r(a) and x a positive vector. Then 1 k + 1 k j=0 A j x A j x, v v v, v, k where v and v are the strictly positive eigenvectors of A and A associated with the eigenvalue r according to Theorem

6 Exercises 1. Show that an irreducible non-negative matrix is primitive if all entries in the main diagonal are strictly positive. 2. Let (z(l)) be a convergent sequence of vectors in a normed vector space. Show that the averages converge to the same limit as l. 1 l + 1 l z(l) l=0 34

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Notes on Linear Algebra and Matrix Theory

Notes on Linear Algebra and Matrix Theory Massimo Franceschet featuring Enrico Bozzo Scalar product The scalar product (a.k.a. dot product or inner product) of two real vectors x = (x 1,..., x n ) and y = (y 1,..., y n ) is not a vector but a

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

NOTES ON THE PERRON-FROBENIUS THEORY OF NONNEGATIVE MATRICES

NOTES ON THE PERRON-FROBENIUS THEORY OF NONNEGATIVE MATRICES NOTES ON THE PERRON-FROBENIUS THEORY OF NONNEGATIVE MATRICES MIKE BOYLE. Introduction By a nonnegative matrix we mean a matrix whose entries are nonnegative real numbers. By positive matrix we mean a matrix

More information

Section 3.9. Matrix Norm

Section 3.9. Matrix Norm 3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

Section 1.7: Properties of the Leslie Matrix

Section 1.7: Properties of the Leslie Matrix Section 1.7: Properties of the Leslie Matrix Definition: A matrix A whose entries are nonnegative (positive) is called a nonnegative (positive) matrix, denoted as A 0 (A > 0). Definition: A square m m

More information

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 Introduction Square matrices whose entries are all nonnegative have special properties. This was mentioned briefly in Section

More information

Review of Some Concepts from Linear Algebra: Part 2

Review of Some Concepts from Linear Algebra: Part 2 Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set

More information

5 Compact linear operators

5 Compact linear operators 5 Compact linear operators One of the most important results of Linear Algebra is that for every selfadjoint linear map A on a finite-dimensional space, there exists a basis consisting of eigenvectors.

More information

Geometric Mapping Properties of Semipositive Matrices

Geometric Mapping Properties of Semipositive Matrices Geometric Mapping Properties of Semipositive Matrices M. J. Tsatsomeros Mathematics Department Washington State University Pullman, WA 99164 (tsat@wsu.edu) July 14, 2015 Abstract Semipositive matrices

More information

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2 Contents Preface for the Instructor xi Preface for the Student xv Acknowledgments xvii 1 Vector Spaces 1 1.A R n and C n 2 Complex Numbers 2 Lists 5 F n 6 Digression on Fields 10 Exercises 1.A 11 1.B Definition

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

WEIGHTED LIMITS FOR POINCARÉ DIFFERENCE EQUATIONS. Rotchana Chieocan

WEIGHTED LIMITS FOR POINCARÉ DIFFERENCE EQUATIONS. Rotchana Chieocan WEIGHTED LIMITS FOR POINCARÉ DIFFERENCE EQUATIONS Rotchana Chieocan Department of Mathematics, Faculty of Science, Khon Kaen University, 40002 Khon Kaen, Thail rotchana@kku.ac.th Mihály Pituk* Department

More information

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS G. RAMESH Contents Introduction 1 1. Bounded Operators 1 1.3. Examples 3 2. Compact Operators 5 2.1. Properties 6 3. The Spectral Theorem 9 3.3. Self-adjoint

More information

Spectral theory for compact operators on Banach spaces

Spectral theory for compact operators on Banach spaces 68 Chapter 9 Spectral theory for compact operators on Banach spaces Recall that a subset S of a metric space X is precompact if its closure is compact, or equivalently every sequence contains a Cauchy

More information

Z-Pencils. November 20, Abstract

Z-Pencils. November 20, Abstract Z-Pencils J. J. McDonald D. D. Olesky H. Schneider M. J. Tsatsomeros P. van den Driessche November 20, 2006 Abstract The matrix pencil (A, B) = {tb A t C} is considered under the assumptions that A is

More information

18.S34 linear algebra problems (2007)

18.S34 linear algebra problems (2007) 18.S34 linear algebra problems (2007) Useful ideas for evaluating determinants 1. Row reduction, expanding by minors, or combinations thereof; sometimes these are useful in combination with an induction

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Detailed Proof of The PerronFrobenius Theorem

Detailed Proof of The PerronFrobenius Theorem Detailed Proof of The PerronFrobenius Theorem Arseny M Shur Ural Federal University October 30, 2016 1 Introduction This famous theorem has numerous applications, but to apply it you should understand

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Recall : Eigenvalues and Eigenvectors

Recall : Eigenvalues and Eigenvectors Recall : Eigenvalues and Eigenvectors Let A be an n n matrix. If a nonzero vector x in R n satisfies Ax λx for a scalar λ, then : The scalar λ is called an eigenvalue of A. The vector x is called an eigenvector

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why? KTH ROYAL INSTITUTE OF TECHNOLOGY Norms for vectors and matrices Why? Lecture 5 Ch. 5, Norms for vectors and matrices Emil Björnson/Magnus Jansson/Mats Bengtsson April 27, 2016 Problem: Measure size of

More information

Functional Analysis Review

Functional Analysis Review Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all

More information

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012. Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

On the simultaneous diagonal stability of a pair of positive linear systems

On the simultaneous diagonal stability of a pair of positive linear systems On the simultaneous diagonal stability of a pair of positive linear systems Oliver Mason Hamilton Institute NUI Maynooth Ireland Robert Shorten Hamilton Institute NUI Maynooth Ireland Abstract In this

More information

Matrix Theory, Math6304 Lecture Notes from Sept 11, 2012 taken by Tristan Whalen

Matrix Theory, Math6304 Lecture Notes from Sept 11, 2012 taken by Tristan Whalen Matrix Theory, Math6304 Lecture Notes from Sept 11, 2012 taken by Tristan Whalen 1 Further Review continued Warm-up Let A, B M n and suppose det(a) = 0.Defineamatrixvaluedfunctionasfollows: F (t) =(A +

More information

Lecture 15 Perron-Frobenius Theory

Lecture 15 Perron-Frobenius Theory EE363 Winter 2005-06 Lecture 15 Perron-Frobenius Theory Positive and nonnegative matrices and vectors Perron-Frobenius theorems Markov chains Economic growth Population dynamics Max-min and min-max characterization

More information

Symmetric and self-adjoint matrices

Symmetric and self-adjoint matrices Symmetric and self-adjoint matrices A matrix A in M n (F) is called symmetric if A T = A, ie A ij = A ji for each i, j; and self-adjoint if A = A, ie A ij = A ji or each i, j Note for A in M n (R) that

More information

Cayley-Hamilton Theorem

Cayley-Hamilton Theorem Cayley-Hamilton Theorem Massoud Malek In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n Let A be an n n matrix Although det (λ I n A

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

Spectral radius, symmetric and positive matrices

Spectral radius, symmetric and positive matrices Spectral radius, symmetric and positive matrices Zdeněk Dvořák April 28, 2016 1 Spectral radius Definition 1. The spectral radius of a square matrix A is ρ(a) = max{ λ : λ is an eigenvalue of A}. For an

More information

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS SPRING 006 PRELIMINARY EXAMINATION SOLUTIONS 1A. Let G be the subgroup of the free abelian group Z 4 consisting of all integer vectors (x, y, z, w) such that x + 3y + 5z + 7w = 0. (a) Determine a linearly

More information

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A). Math 344 Lecture #19 3.5 Normed Linear Spaces Definition 3.5.1. A seminorm on a vector space V over F is a map : V R that for all x, y V and for all α F satisfies (i) x 0 (positivity), (ii) αx = α x (scale

More information

Basic Elements of Linear Algebra

Basic Elements of Linear Algebra A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,

More information

Analysis Preliminary Exam Workshop: Hilbert Spaces

Analysis Preliminary Exam Workshop: Hilbert Spaces Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H

More information

Lax Solution Part 4. October 27, 2016

Lax Solution Part 4.   October 27, 2016 Lax Solution Part 4 www.mathtuition88.com October 27, 2016 Textbook: Functional Analysis by Peter D. Lax Exercises: Ch 16: Q2 4. Ch 21: Q1, 2, 9, 10. Ch 28: 1, 5, 9, 10. 1 Chapter 16 Exercise 2 Let h =

More information

Math 421, Homework #7 Solutions. We can then us the triangle inequality to find for k N that (x k + y k ) (L + M) = (x k L) + (y k M) x k L + y k M

Math 421, Homework #7 Solutions. We can then us the triangle inequality to find for k N that (x k + y k ) (L + M) = (x k L) + (y k M) x k L + y k M Math 421, Homework #7 Solutions (1) Let {x k } and {y k } be convergent sequences in R n, and assume that lim k x k = L and that lim k y k = M. Prove directly from definition 9.1 (i.e. don t use Theorem

More information

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT 204 - FALL 2006 PRINCETON UNIVERSITY ALFONSO SORRENTINO 1 Adjoint of a linear operator Note: In these notes, V will denote a n-dimensional euclidean vector

More information

Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices

Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices Geršgorin circles Lecture 8: Outline Chapter 6 + Appendix D: Location and perturbation of eigenvalues Some other results on perturbed eigenvalue problems Chapter 8: Nonnegative matrices Geršgorin s Thm:

More information

Componentwise perturbation analysis for matrix inversion or the solution of linear systems leads to the Bauer-Skeel condition number ([2], [13])

Componentwise perturbation analysis for matrix inversion or the solution of linear systems leads to the Bauer-Skeel condition number ([2], [13]) SIAM Review 4():02 2, 999 ILL-CONDITIONED MATRICES ARE COMPONENTWISE NEAR TO SINGULARITY SIEGFRIED M. RUMP Abstract. For a square matrix normed to, the normwise distance to singularity is well known to

More information

Definition (T -invariant subspace) Example. Example

Definition (T -invariant subspace) Example. Example Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin

More information

INVARIANT PROBABILITIES ON PROJECTIVE SPACES. 1. Introduction

INVARIANT PROBABILITIES ON PROJECTIVE SPACES. 1. Introduction INVARIANT PROBABILITIES ON PROJECTIVE SPACES YVES DE CORNULIER Abstract. Let K be a local field. We classify the linear groups G GL(V ) that preserve an probability on the Borel subsets of the projective

More information

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational

More information

Asymptotic Stability by Linearization

Asymptotic Stability by Linearization Dynamical Systems Prof. J. Rauch Asymptotic Stability by Linearization Summary. Sufficient and nearly sharp sufficient conditions for asymptotic stability of equiiibria of differential equations, fixed

More information

A Brief Introduction to Functional Analysis

A Brief Introduction to Functional Analysis A Brief Introduction to Functional Analysis Sungwook Lee Department of Mathematics University of Southern Mississippi sunglee@usm.edu July 5, 2007 Definition 1. An algebra A is a vector space over C with

More information

E2 212: Matrix Theory (Fall 2010) Solutions to Test - 1

E2 212: Matrix Theory (Fall 2010) Solutions to Test - 1 E2 212: Matrix Theory (Fall 2010) s to Test - 1 1. Let X = [x 1, x 2,..., x n ] R m n be a tall matrix. Let S R(X), and let P be an orthogonal projector onto S. (a) If X is full rank, show that P can be

More information

0.1 Rational Canonical Forms

0.1 Rational Canonical Forms We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Nonlinear Programming Algorithms Handout

Nonlinear Programming Algorithms Handout Nonlinear Programming Algorithms Handout Michael C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 5376 September 9 1 Eigenvalues The eigenvalues of a matrix A C n n are

More information

Linear Algebra. Session 12

Linear Algebra. Session 12 Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Eigenvectors. Prop-Defn

Eigenvectors. Prop-Defn Eigenvectors Aim lecture: The simplest T -invariant subspaces are 1-dim & these give rise to the theory of eigenvectors. To compute these we introduce the similarity invariant, the characteristic polynomial.

More information

Wavelets and Linear Algebra

Wavelets and Linear Algebra Wavelets and Linear Algebra 4(1) (2017) 43-51 Wavelets and Linear Algebra Vali-e-Asr University of Rafsanan http://walavruacir Some results on the block numerical range Mostafa Zangiabadi a,, Hamid Reza

More information

Krein-Rutman Theorem on the Spectrum of Compact Positive Operators on Ordered Banach Spaces

Krein-Rutman Theorem on the Spectrum of Compact Positive Operators on Ordered Banach Spaces D I P L O M A R B E I T Krein-Rutman Theorem on the Spectrum of Compact Positive Operators on Ordered Banach Spaces ausgeführt am Institut für Analysis und Scientific Computing der Technischen Universität

More information

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra Foundations of Numerics from Advanced Mathematics Linear Algebra Linear Algebra, October 23, 22 Linear Algebra Mathematical Structures a mathematical structure consists of one or several sets and one or

More information

Linear Compartmental Systems-The Basics

Linear Compartmental Systems-The Basics Linear Compartmental Systems-The Basics Hal Smith October 3, 26.1 The Model Consider material flowing among n compartments labeled 1, 2,, n. Assume to begin that no material flows into any compartment

More information

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T 1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.

More information

Stabilization of Distributed Parameter Systems by State Feedback with Positivity Constraints

Stabilization of Distributed Parameter Systems by State Feedback with Positivity Constraints Stabilization of Distributed Parameter Systems by State Feedback with Positivity Constraints Joseph Winkin Namur Center of Complex Systems (naxys) and Dept. of Mathematics, University of Namur, Belgium

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Nonnegative and spectral matrix theory Lecture notes

Nonnegative and spectral matrix theory Lecture notes Nonnegative and spectral matrix theory Lecture notes Dario Fasino, University of Udine (Italy) Lecture notes for the first part of the course Nonnegative and spectral matrix theory with applications to

More information

Positive Stabilization of Infinite-Dimensional Linear Systems

Positive Stabilization of Infinite-Dimensional Linear Systems Positive Stabilization of Infinite-Dimensional Linear Systems Joseph Winkin Namur Center of Complex Systems (NaXys) and Department of Mathematics, University of Namur, Belgium Joint work with Bouchra Abouzaid

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Eventual Cone Invariance

Eventual Cone Invariance Electronic Journal of Linear Algebra Volume 32 Volume 32 (2017) Article 17 2017 Eventual Cone Invariance Michael Kasigwa Washington State University, kasigwam@gmail.com Michael Tsatsomeros Washington State

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,

More information

On the mathematical background of Google PageRank algorithm

On the mathematical background of Google PageRank algorithm Working Paper Series Department of Economics University of Verona On the mathematical background of Google PageRank algorithm Alberto Peretti, Alberto Roveda WP Number: 25 December 2014 ISSN: 2036-2919

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,

More information

SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT

SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT SPECTRAL THEOREM FOR SYMMETRIC OPERATORS WITH COMPACT RESOLVENT Abstract. These are the letcure notes prepared for the workshop on Functional Analysis and Operator Algebras to be held at NIT-Karnataka,

More information

LECTURE 7. k=1 (, v k)u k. Moreover r

LECTURE 7. k=1 (, v k)u k. Moreover r LECTURE 7 Finite rank operators Definition. T is said to be of rank r (r < ) if dim T(H) = r. The class of operators of rank r is denoted by K r and K := r K r. Theorem 1. T K r iff T K r. Proof. Let T

More information

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra D. R. Wilkins Contents 3 Topics in Commutative Algebra 2 3.1 Rings and Fields......................... 2 3.2 Ideals...............................

More information

MATH36001 Perron Frobenius Theory 2015

MATH36001 Perron Frobenius Theory 2015 MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,

More information

Linear Algebra Review

Linear Algebra Review January 29, 2013 Table of contents Metrics Metric Given a space X, then d : X X R + 0 and z in X if: d(x, y) = 0 is equivalent to x = y d(x, y) = d(y, x) d(x, y) d(x, z) + d(z, y) is a metric is for all

More information

Lecture notes: Applied linear algebra Part 2. Version 1

Lecture notes: Applied linear algebra Part 2. Version 1 Lecture notes: Applied linear algebra Part 2. Version 1 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 First, some exercises: xercise 0.1 (2 Points) Another least

More information

Matrix functions that preserve the strong Perron- Frobenius property

Matrix functions that preserve the strong Perron- Frobenius property Electronic Journal of Linear Algebra Volume 30 Volume 30 (2015) Article 18 2015 Matrix functions that preserve the strong Perron- Frobenius property Pietro Paparella University of Washington, pietrop@uw.edu

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. Math 5327 Fall 2018 Homework 7 1. For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. 3 1 0 (a) A = 1 2 0 1 1 0 x 3 1 0 Solution: 1 x 2 0

More information

ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS

ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS Giorgio Picci with T. Taylor, ASU Tempe AZ. Wofgang Runggaldier s Birthday, Brixen July 2007 1 CONSENSUS FOR RANDOM GOSSIP ALGORITHMS Consider a finite

More information

Eventual Positivity of Operator Semigroups

Eventual Positivity of Operator Semigroups Eventual Positivity of Operator Semigroups Jochen Glück Ulm University Positivity IX, 17 May 21 May 2017 Joint work with Daniel Daners (University of Sydney) and James B. Kennedy (University of Lisbon)

More information

Balanced Truncation 1

Balanced Truncation 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors week -2 Fall 26 Eigenvalues and eigenvectors The most simple linear transformation from R n to R n may be the transformation of the form: T (x,,, x n ) (λ x, λ 2,, λ n x n

More information

Fiedler s Theorems on Nodal Domains

Fiedler s Theorems on Nodal Domains Spectral Graph Theory Lecture 7 Fiedler s Theorems on Nodal Domains Daniel A. Spielman September 19, 2018 7.1 Overview In today s lecture we will justify some of the behavior we observed when using eigenvectors

More information

The Spectral Theorem for normal linear maps

The Spectral Theorem for normal linear maps MAT067 University of California, Davis Winter 2007 The Spectral Theorem for normal linear maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 14, 2007) In this section we come back to the question

More information

UCSD ECE269 Handout #18 Prof. Young-Han Kim Monday, March 19, Final Examination (Total: 130 points)

UCSD ECE269 Handout #18 Prof. Young-Han Kim Monday, March 19, Final Examination (Total: 130 points) UCSD ECE269 Handout #8 Prof Young-Han Kim Monday, March 9, 208 Final Examination (Total: 30 points) There are 5 problems, each with multiple parts Your answer should be as clear and succinct as possible

More information

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

In English, this means that if we travel on a straight line between any two points in C, then we never leave C. Convex sets In this section, we will be introduced to some of the mathematical fundamentals of convex sets. In order to motivate some of the definitions, we will look at the closest point problem from

More information

Diagonalization of Matrix

Diagonalization of Matrix of Matrix King Saud University August 29, 2018 of Matrix Table of contents 1 2 of Matrix Definition If A M n (R) and λ R. We say that λ is an eigenvalue of the matrix A if there is X R n \ {0} such that

More information

MATRIX ANALYSIS HOMEWORK

MATRIX ANALYSIS HOMEWORK MATRIX ANALYSIS HOMEWORK VERN I. PAULSEN 1. Due 9/25 We let S n denote the group of all permutations of {1,..., n}. 1. Compute the determinants of the elementary matrices: U(k, l), D(k, λ), and S(k, l;

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information