A Note on Eigenvalues of Perturbed Hermitian Matrices


 Allyson Glenn
 11 months ago
 Views:
Transcription
1 A Note on Eigenvalues of Perturbed Hermitian Matrices ChiKwong Li RenCang Li July 2004 Let ( H1 E A = E H 2 Abstract and Ã = ( H1 H 2 be Hermitian matrices with eigenvalues λ 1 λ k and λ 1 λ k, respectively. Denote by E the spectral norm of the matrix E, and η the spectral gap between the spectra of H 1 and H 2. It is shown that λ i λ i 2 E 2 η + η E 2, which improves all the existing results. Similar bounds are obtained for singular values of matrices under block perturbations. AMS Classifications: 15A42, 15A18, 65F15. Keywords: Hermitian matrix, eigenvalue, singular value. 1 Introduction Consider a partitioned Hermitian matrix A = ( m n m H 1 E n E H 2, (1.1 where E is E s complex conjugate transpose. At various situations (typically when E is small, one is interested in knowing the impact of removing E and E on the eigenvalues Department of Mathematics, The College of William and Mary, Williamsburg, Virginia Research supported in part by the NSF grant DMS Department of Mathematics, University of Kentucky, Lexington, KY This work was supported in part by the National Science Foundation CAREER award under Grant No. CCR
2 of A. More specifically, one would like to obtain bounds for the differences between that eigenvalues of A and those of its perturbed matrix Ã = ( m n m H 1 n H 2. (1.2 Let λ(x be the spectrum of the square matrix X, and let Y be the spectral norm of a matrix Y, i.e., the largest singular value of Y. There are two kinds of bounds for the eigenvalues λ 1 λ m+n and λ 1 λ m+n of A and Ã, respectively: 1. [1, 7, 8] λ i λ i E. ( [1, 2, 3, 5, 7, 8] If the spectra of H 1 and H 2 are disjoint, then where and λ(h i is the spectrum of H i. λ i λ i E 2 /η, (1.4 η def = min µ 1 λ(h 1, µ 2 λ(h 2 µ 1 µ 2, The bounds of the first kind do not use information of the spectral distribution of the H 1 and H 2, which will give (much weaker bounds when η is not so small; while the bounds of the second kind may blow up whenever H 1 and H 2 have a common eigenvalue. Thus both kinds have their own drawbacks, and it would be advantageous to have bounds that are always no bigger than E, of ( E as η 0, and at the same time behave like ( E 2 /η for not so small η. To further motivate our study, let us look at the following 2 2 example. Example 1 Consider the 2 2 Hermitian matrix ( α ɛ A =. (1.5 ɛ β Interesting cases are when ɛ is small, and thus α and β are approximate eigenvalues of A. We shall analyze by how much the eigenvalues of A differ from α and β. Without loss of generality, assume α > β. The eigenvalues of A, denoted by λ ±, satisfy λ 2 (α + βλ + αβ ɛ 2 = 0; and thus λ ± = α + β ± (α + β 2 4(αβ ɛ 2 2 = α + β ± (α β 2 + 4ɛ
3 Now { λ+ α 0 < β λ } = (α β + (α β 2 + 4ɛ 2 2 = 2ɛ 2 (α β + (α β 2 + 4ɛ 2 (1.6 which provides a difference that enjoys the following properties: 2ɛ 2 ɛ always, (α β + ɛ as α β +, (α β 2 + 4ɛ 2 ɛ 2 /(α β. The purpose of this note is to extend this 2 2 example and obtain bounds which improve both (1.3 and (1.4. Such results are not only of theoretical interest but also important in the computations of eigenvalues of Hermitian matrices [4, 6, 9]. As an application, similar bounds are presented for the singular value problem. 2 Main Result Theorem 2 Let A = ( m n m H 1 E n E H 2 and Ã = ( m n m H 1 n H 2 be Hermitian matrices with eigenvalues λ 1 λ 2 λ m+n and λ1 λ 2 λ m+n, (2.1 respectively. Define η i η def = min λ i µ 2, if λ i λ(h 1, µ 2 λ(h 2 min λ i µ 1, if λ i λ(h 2, µ 1 λ(h 1 (2.2 def = min η i = min µ 1 µ 2. (2.3 1 i m+n µ 1 λ(h 1, µ 2 λ(h 2 Then for i = 1, 2,, m + n, we have λ i λ i 2 E 2 (2.4 η i + η 2i + 4 E 2 2 E 2 η + η E 2. (2.5 3
4 Proof. Suppose U H 1 U and V H 2 V are in diagonal form with diagonal entries arranged in descending order. We may assume that U = I m and V = I n. therwise, replace A by (U V A(U V. We may perturb the diagonal of A so that all entries are distinct, and apply continuity argument for the general case. We prove the result by induction on m + n. If m + n = 2, the result is clear (from our Example. Assume that m + n > 2, and the result is true for Hermitian matrices of size m + n 1. First, refining an argument of Mathias [5], we show that (2.4 holds for i = 1. Assume that the (1, 1 entry of H 1 equals λ 1. By the minmax principle [1, 7, 8], we have λ 1 e 1Ae 1 = λ 1, where e 1 is the first column of the identity matrix. Let ( I X = m 0 (H 2 µ 1 I n 1. E I n Then where ( X H1 (λ (A λ 1 IX = H 2 λ 1 I n, H 1 (λ 1 = H 1 λ 1 I m E (H 2 λ 1 I n 1 E. Since A and X AX have the same inertia, we see that H 1 (λ 1 has zero as the largest eigenvalue. Notice that the largest eigenvalue of H 1 λ 1 I is λ 1 λ 1 0. Thus, for δ 1 = λ 1 λ 1 = λ λ 1, we have (see [7, (10.9] λ 1 λ 1 + E 2 2/(δ 1 + η 1, and hence Consequently, δ 1 E 2 /(δ 1 + η 1. 2 E δ 1 η 1 + η E 2 as asserted. Similarly, we can prove the result if the (1, 1 entry of H 2 equals λ 1. In this case, we will apply the inertia arguments to A and Y AY with ( I Y = m 0 E(H 1 λ 1 I m 1. I n Applying the result of the last paragraph to A, we see that (2.2 holds for i = m + n. 4
5 Now, suppose 1 < i < m + n. The result trivially holds if λ i = λ i. Suppose λ i λ i. We may assume that λ i > λ i. therwise, replace (A, Ã, i by ( A, Ã, m + n i + 1. Delete the row and column of A that contain the diagonal entry λ n. Suppose the resulting matrix Â has eigenvalues ν 1 ν m+n 1. By the interlacing inequalities [7, Section 10.1], we have λ i ν i and hence λi λ i λ i νi. (2.6 Note that λ i is the ith largest diagonal entries in Â. Let η i be the minimum distance between λ i and the diagonal entries in the diagonal block Ĥj in Â not containing λ i ; here j {1, 2}. Then η i η i because Ĥj may have one fewer diagonal entries than H j. Let Ê be the offdiagonal block of Â. Then Ê E. Thus, λ i λ i = λ i λ i because λ i > λ i λ i ν i by (2.6 2 Ê 2 by induction assumption η i + η 2i + 4 Ê 2 2 Ê 2 because η i η i η i + η 2i + 4 Ê 2 = 1 ηi Ê 2 η i 1 ηi E 2 η i because Ê E = 2 E 2 η i + ηi E 2 as asserted. 3 Application to Singular Value Problem In this section, we apply the result in Section 2 to study singular values of matrices. For notational convenience in connection to our discussion, we define the sequence of singular values of a complex p q matrix X by σ(x = (σ 1 (X,..., σ k (X, where k = max{p, q} and σ 1 (X σ k (X are the nonnegative square roots of the eigenvalues of the matrix XX or X X depending on which one has a larger size. Note that the nonzero eigenvalues of XX and X X are the same, and they give rise to the nonzero singular values of X which are of importance. We have the following result concerning the nonzero singular values of perturbed matrices. 5
6 Theorem 3 Let B = ( k l m G 1 E 1 n E 2 G 2 ( m G and B = 1 n G 2 k l be complex matrices with singular values σ 1 σ 2 σ max{m+n,k+l} and σ 1 σ 2 σ max{m+n,k+l}, (3.1 respectively, so that G 1 and G 2 are nontrivial. Define ɛ = max{ E 1, E 2 }, and min σ i µ 2, if σ i σ(g 1, µ 2 σ(g 2 η i η def = min σ i µ 1, if σ i σ(g 2, µ 1 σ(g 1 (3.2 def = min η i = min µ 1 µ 2. (3.3 1 i m+n µ 1 σ(g 1, µ 2 σ(g 2 Then for i = 1, 2,, min{m + n, k + l}, we have σ i σ i 2ɛ 2 (3.4 η i + η 2i + 4ɛ2 2ɛ 2 η + η 2 + 4ɛ 2, (3.5 and σ i = σ i = 0 for i > min{m + n, k + l}. Proof: By JordanWielandt Theorem [8, Theorem I.4.2], the eigenvalues of ( B B are ±σ i and possibly some zeros adding up to m+n+k+l eigenvalues. A similar statement holds for B. Permuting the rows and columns appropriately, we see that ( G 1 E 1 B B is similar to G 1 E2 E 2 G 2, E1 G 2 and ( B B is similar to G 1 G 1 G 2 G 2. 6
7 Applying Theorem 2 with ( Gi H i = G i ( E2 and E =, E 1 we get the result. ne can also apply the above proof to the degenerate cases when G 1 or G 2 in the matrix B is trivial, i.e., one of the parameters m, n, k, l is zero. These cases are useful in applications. We state one of them, and one can easily extend it to other cases. Theorem 4 Suppose B = ( G E and B = ( G are p q matrices with singular values σ 1 σ max{p,q} and σ 1... σ max{p,q}, respectively. Then for i = 1,..., min{p, q}, σ i σ i 2 E. 2 σ i + σ 2i + 4 E 2 Acknowledgment This paper evolved from an early note of the second author in which Theorem 2 was proven for special cases. The authors wish to thank Editor Volker Mehrmann for making this collaboration possible. 7
8 References [1] R. Bhatia. Matrix Analysis. Graduate Texts in Mathematics, vol Springer, New York, [2] J. Demmel. Applied Numerical Linear Algebra. SIAM, Philadelphia, [3] G. H. Golub and C. F. van Loan. Matrix Computations. Johns Hopkins University Press, Baltimore, Maryland, 2nd edition, [4] W. Kahan. When to neglect offdiagonal elements of symmetric tridiagonal matrices. Technical Report CS42, Computer Science Department, Stanford University, [5] R. Mathias. Quadratic residual bounds for the hermitian eigenvalue problem. SIAM Journal on Matrix Analysis and Applications, 19: , [6] C. C. Paige. Eigenvalues of perturbed Hermitian matrices. Linear Algebra and its Applications, 8:1 10, [7] B. N. Parlett. The Symmetric Eigenvalue Problem. SIAM, Philadelphia, This SIAM edition is an unabridged, corrected reproduction of the work first published by PrenticeHall, Inc., Englewood Cliffs, New Jersey, [8] G. W. Stewart and J.G. Sun. Matrix Perturbation Theory. Academic Press, Boston, [9] J. H. Wilkinson. The Algebraic Eigenvalue Problem. Clarendon Press, xford, England,
216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are
Numer. Math. 68: 215{223 (1994) Numerische Mathemati c SringerVerlag 1994 Electronic Edition Bacward errors for eigenvalue and singular value decomositions? S. Chandrasearan??, I.C.F. Isen??? Deartment
More informationEIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems
EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues
More informationarxiv: v4 [quantph] 8 Mar 2013
Decomposition of unitary matrices and quantum gates ChiKwong Li, Rebecca Roberts Department of Mathematics, College of William and Mary, Williamsburg, VA 2387, USA. (Email: ckli@math.wm.edu, rlroberts@email.wm.edu)
More informationThe Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment
he Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment William Glunt 1, homas L. Hayden 2 and Robert Reams 2 1 Department of Mathematics and Computer Science, Austin Peay State
More informationConjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
More informationfor all A M m and B M n Key words. Complex matrix, linear preserver, spectral norm, Ky Fan knorm, Schatten pnorm, tensor product.
LINEAR MAPS PRESERVING KY FAN NORMS AND SCHATTEN NORMS OF TENSOR PRODUCTS OF MATRICES AJDA FOŠNER, ZEJUN HUANG, CHIKWONG LI, AND NUNGSING SZE Abstract. For a positive integer n, let M n be the set of
More informationComplementary bases in symplectic matrices and a proof that their determinant is one
Linear Algebra and its Applications 419 (2006) 772 778 www.elsevier.com/locate/laa Complementary bases in symplectic matrices and a proof that their determinant is one Froilán M. Dopico a,, Charles R.
More informationBlock companion matrices, discretetime block diagonal stability and polynomial matrices
Block companion matrices, discretetime block diagonal stability and polynomial matrices Harald K. Wimmer Mathematisches Institut Universität Würzburg D97074 Würzburg Germany October 25, 2008 Abstract
More informationRankone LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the wellknown Lyapunov's matrix inequality about
Rankone LMIs and Lyapunov's Inequality Didier Henrion 1;; Gjerrit Meinsma Abstract We describe a new proof of the wellknown Lyapunov's matrix inequality about the location of the eigenvalues of a matrix
More informationLine segments on the boundary of the numerical ranges of some tridiagonal matrices
Electronic Journal of Linear Algebra Volume 30 Volume 30 (2015) Article 46 2015 Line segments on the boundary of the numerical ranges of some tridiagonal matrices Ilya M. Spitkovsky NYUAD and College of
More informationHOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)
HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationDistance Preserving Maps on Matrices. ChiKwong Li Department of Mathematics College of William and Mary Williamsburg, Virginia
Distance Preserving Maps on Matrices ChiKwong Li Department of Mathematics College of William and Mary Williamsburg, Virginia 231878795 General problem Suppose is a norm on a matrix space M. Characterize
More informationRanks and determinants of the sum of matrices from unitary orbits
Ranks and determinants of the sum of matrices from unitary orbits ChiKwong Li, YiuTung Poon and NungSing Sze Dedicated to Professor YikHoi AuYeung on the occasion of his 70th birthday. Abstract The
More informationSecondorder approximation of dynamic models without the use of tensors
Secondorder approximation of dynamic models without the use of tensors Paul Klein a, a University of Western Ontario First draft: May 17, 2005 This version: January 24, 2006 Abstract Several approaches
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150  Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationI = i 0,
Special Types of Matrices Certain matrices, such as the identity matrix 0 0 0 0 0 0 I = 0 0 0, 0 0 0 have a special shape, which endows the matrix with helpful properties The identity matrix is an example
More informationKEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio
COMPUTING THE SVD OF A GENERAL MATRIX PRODUCT/QUOTIENT GENE GOLUB Computer Science Department Stanford University Stanford, CA USA golub@sccm.stanford.edu KNUT SLNA SCCM Stanford University Stanford,
More informationMatching Polynomials of Graphs
Spectral Graph Theory Lecture 25 Matching Polynomials of Graphs Daniel A Spielman December 7, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened in class The notes
More informationAppendix C Vector and matrix algebra
Appendix C Vector and matrix algebra Concepts Scalars Vectors, rows and columns, matrices Adding and subtracting vectors and matrices Multiplying them by scalars Products of vectors and matrices, scalar
More informationw T 1 w T 2. w T n 0 if i j 1 if i = j
Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible 
More informationNONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction
NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques
More informationRemarks on the Cayley Representation of Orthogonal Matrices and on Perturbing the Diagonal of a Matrix to Make it Invertible
Remarks on the Cayley Representation of Orthogonal Matrices and on Perturbing the Diagonal of a Matrix to Make it Invertible Jean Gallier Department of Computer and Information Science University of Pennsylvania
More information9.1 Eigenanalysis I Eigenanalysis II Advanced Topics in Linear Algebra Kepler s laws
Chapter 9 Eigenanalysis Contents 9. Eigenanalysis I.................. 49 9.2 Eigenanalysis II................. 5 9.3 Advanced Topics in Linear Algebra..... 522 9.4 Kepler s laws................... 537
More informationBaltzer Journals April 24, Bounds for the Trace of the Inverse and the Determinant. of Symmetric Positive Denite Matrices
Baltzer Journals April 4, 996 Bounds for the Trace of the Inverse and the Determinant of Symmetric Positive Denite Matrices haojun Bai ; and Gene H. Golub ; y Department of Mathematics, University of Kentucky,
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationON PERTURBATIONS OF MATRIX PENCILS WITH REAL SPECTRA. II
MATEMATICS OF COMPUTATION Volume 65, Number 14 April 1996, Pages 637 645 ON PERTURBATIONS OF MATRIX PENCILS WIT REAL SPECTRA. II RAJENDRA BATIA AND RENCANG LI Abstract. A wellknown result on spectral
More informationRecursive Solutions of the Matrix Equations X + A T X 1 A = Q and X A T X 1 A = Q
Applied Mathematical Sciences, Vol. 2, 2008, no. 38, 18551872 Recursive Solutions of the Matrix Equations X + A T X 1 A = Q and X A T X 1 A = Q Nicholas Assimakis Department of Electronics, Technological
More informationLeast squares Solution of Homogeneous Equations
Least squares Solution of Homogeneous Equations supportive text for teaching purposes Revision: 1.2, dated: December 15, 2005 Tomáš Svoboda Czech Technical University, Faculty of Electrical Engineering
More informationMATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian.
MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian. Spanning set Let S be a subset of a vector space V. Definition. The span of the set S is the smallest subspace W V that contains S. If
More informationarxiv:math/ v2 [math.ag] 21 Sep 2003
EIGENVALUES, SINGULAR VALUES, AND LITTLEWOODRICHARDSON COEFFICIENTS arxiv:math/0301307v2 [math.ag] 21 Sep 2003 SERGEY FOMIN, WILLIAM FULTON, CHIKWONG LI, AND YIUTUNG POON Abstract. We characterize the
More informationComparison of perturbation bounds for the stationary distribution of a Markov chain
Linear Algebra and its Applications 335 (00) 37 50 www.elsevier.com/locate/laa Comparison of perturbation bounds for the stationary distribution of a Markov chain Grace E. Cho a, Carl D. Meyer b,, a Mathematics
More informationLinear Algebra. Carleton DeTar February 27, 2017
Linear Algebra Carleton DeTar detar@physics.utah.edu February 27, 2017 This document provides some background for various course topics in linear algebra: solving linear systems, determinants, and finding
More informationInstitute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{95{93 TR{3535 On the Perturbation of LU and Cholesky Factors G. W. Stewart y October, 1995
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More information1.10 Matrix Representation of Graphs
42 Basic Concepts of Graphs 1.10 Matrix Representation of Graphs Definitions: In this section, we introduce two kinds of matrix representations of a graph, that is, the adjacency matrix and incidence matrix
More informationOn the mathematical background of Google PageRank algorithm
Working Paper Series Department of Economics University of Verona On the mathematical background of Google PageRank algorithm Alberto Peretti, Alberto Roveda WP Number: 25 December 2014 ISSN: 20362919
More informationResearch Article A Note on Kantorovich Inequality for Hermite Matrices
Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 0, Article ID 5767, 6 pages doi:0.55/0/5767 Research Article A Note on Kantorovich Inequality for Hermite Matrices Zhibing
More informationSpectrally arbitrary star sign patterns
Linear Algebra and its Applications 400 (2005) 99 119 wwwelseviercom/locate/laa Spectrally arbitrary star sign patterns G MacGillivray, RM Tifenbach, P van den Driessche Department of Mathematics and Statistics,
More information1 Determinants. 1.1 Determinant
1 Determinants [SB], Chapter 9, p.188196. [SB], Chapter 26, p.719739. Bellow w ll study the central question: which additional conditions must satisfy a quadratic matrix A to be invertible, that is to
More informationAn Observation on the Positive Real Lemma
Journal of Mathematical Analysis and Applications 255, 48 49 (21) doi:1.16/jmaa.2.7241, available online at http://www.idealibrary.com on An Observation on the Positive Real Lemma Luciano Pandolfi Dipartimento
More informationSENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*
SIAM J Matrix Anal Appl c 1994 Society for Industrial and Applied Mathematics Vol 15, No 3, pp 715728, July, 1994 001 SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN* CARL D MEYER Abstract
More informationc c c c c c c c c c a 3x3 matrix C= has a determinant determined by
Linear Algebra Determinants and Eigenvalues Introduction: Many important geometric and algebraic properties of square matrices are associated with a single real number revealed by what s known as the determinant.
More informationAn SVDLike Matrix Decomposition and Its Applications
An SVDLike Matrix Decomposition and Its Applications Hongguo Xu Abstract [ A matrix S C m m is symplectic if SJS = J, where J = I m I m. Symplectic matrices play an important role in the analysis and
More informationMath 2331 Linear Algebra
5. Eigenvectors & Eigenvalues Math 233 Linear Algebra 5. Eigenvectors & Eigenvalues ShangHuan Chiu Department of Mathematics, University of Houston schiu@math.uh.edu math.uh.edu/ schiu/ ShangHuan Chiu,
More informationMath 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.
Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses
More informationc 2002 Society for Industrial and Applied Mathematics
SIAM J. MATRIX ANAL. APPL. Vol. 24, No. 1, pp. 150 164 c 2002 Society for Industrial and Applied Mathematics VARIANTS OF THE GREVILLE FORMULA WITH APPLICATIONS TO EXACT RECURSIVE LEAST SQUARES JIE ZHOU,
More informationScaling of Symmetric Matrices by Positive Diagonal Congruence
Scaling of Symmetric Matrices by Positive Diagonal Congruence Abstract Charles R. Johnson a and Robert Reams b a Department of Mathematics, College of William and Mary, P.O. Box 8795, Williamsburg, VA
More informationGATE Engineering Mathematics SAMPLE STUDY MATERIAL. Postal Correspondence Course GATE. Engineering. Mathematics GATE ENGINEERING MATHEMATICS
SAMPLE STUDY MATERIAL Postal Correspondence Course GATE Engineering Mathematics GATE ENGINEERING MATHEMATICS ENGINEERING MATHEMATICS GATE Syllabus CIVIL ENGINEERING CE CHEMICAL ENGINEERING CH MECHANICAL
More informationMatrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...
Chapter Matrices We review the basic matrix operations What is a Matrix? An array of numbers a a n A = a m a mn with m rows and n columns is a m n matrix Element a ij in located in position (i, j The elements
More informationSandwich Theorem and Calculation of the Theta Function for Several Graphs
Brigham Young University BYU ScholarsArchive All Theses and Dissertations 2003037 Sandwich Theorem and Calculation of the Theta Function for Several Graphs Marcia Ling Riddle Brigham Young University
More informationCounting Matrices Over a Finite Field With All Eigenvalues in the Field
Counting Matrices Over a Finite Field With All Eigenvalues in the Field Lisa Kaylor David Offner Department of Mathematics and Computer Science Westminster College, Pennsylvania, USA kaylorlm@wclive.westminster.edu
More informationDeterminants. Beifang Chen
Determinants Beifang Chen 1 Motivation Determinant is a function that each square real matrix A is assigned a real number, denoted det A, satisfying certain properties If A is a 3 3 matrix, writing A [u,
More informationHW #3 Solutions: M552 Spring (c)
HW #3 Solutions: M55 Spring 6. (4.Trefethen & Bau: parts (a), (c) and (e)) Determine the SVDs of the following matrices (by hand calculation): (a) [ 3 (c) (e) [ ANS: In each case we seek A UΣV. The general
More informationWHAT IS THE MINIMUM FUNCTION OBSERVER ORDER
WHAT IS THE MINIMUM FUNCTION OBSERVER ORDER ChiaChi Tsui 743 Clove Road, NY 10310, USA, ctsui@ny.devry.edu poles (or the eigenvalues of F) of Keywords:function observer order, much lower & lowest possible
More informationEigensolving via reduction to DPR1 matrices
Computers and Mathematics with Applications 56 (2008) 166 171 www.elsevier.com/locate/camwa Eigensolving via reduction to DPR1 matrices V.Y. Pan a,, B. Murphy a, R.E. Rosholt a, Y. Tang b, X. Wang c,
More informationThe passage time distribution for a birthanddeath chain: Strong stationary duality gives a first stochastic proof
The passage time distribution for a birthanddeath chain: Strong stationary duality gives a first stochastic proof James Allen Fill 1 Department of Applied Mathematics and Statistics The Johns Hopkins
More informationRelation of Pure Minimum Cost Flow Model to Linear Programming
Appendix A Page 1 Relation of Pure Minimum Cost Flow Model to Linear Programming The Network Model The network pure minimum cost flow model has m nodes. The external flows given by the vector b with m
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationMaths for Signals and Systems Linear Algebra for Engineering Applications
Maths for Signals and Systems Linear Algebra for Engineering Applications Lectures 12, Tuesday 11 th October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON
More informationProblem Set # 1 Solution, 18.06
Problem Set # 1 Solution, 1.06 For grading: Each problem worths 10 points, and there is points of extra credit in problem. The total maximum is 100. 1. (10pts) In Lecture 1, Prof. Strang drew the cone
More informationThe Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points. Sorin Costiner and Shlomo Ta'asan
The Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points Sorin Costiner and Shlomo Ta'asan Department of Applied Mathematics and Computer Science The Weizmann
More informationSingular Value Decomposition
Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our
More informationBoolean InnerProduct Spaces and Boolean Matrices
Boolean InnerProduct Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver
More informationAN INEXACT INVERSE ITERATION FOR COMPUTING THE SMALLEST EIGENVALUE OF AN IRREDUCIBLE MMATRIX. 1. Introduction. We consider the eigenvalue problem
AN INEXACT INVERSE ITERATION FOR COMPUTING THE SMALLEST EIGENVALUE OF AN IRREDUCIBLE MMATRIX MICHIEL E. HOCHSTENBACH, WENWEI LIN, AND CHINGSUNG LIU Abstract. In this paper, we present an inexact inverse
More informationand let s calculate the image of some vectors under the transformation T.
Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =
More informationarxiv:quantph/ v1 29 Mar 2003
FiniteDimensional PT Symmetric Hamiltonians arxiv:quantph/0303174v1 29 Mar 2003 Carl M. Bender, Peter N. Meisinger, and Qinghai Wang Department of Physics, Washington University, St. Louis, MO 63130,
More informationYinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method
The Simplex Method Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye (LY, Chapters 2.32.5, 3.13.4) 1 Geometry of Linear
More informationDimension reduction based on centroids and least squares for efficient processing of text data
Dimension reduction based on centroids and least squares for efficient processing of text data M. Jeon,H.Park, and J.B. Rosen Abstract Dimension reduction in today s vector space based information retrieval
More informationJordan Normal Form. Chapter Minimal Polynomials
Chapter 8 Jordan Normal Form 81 Minimal Polynomials Recall p A (x) =det(xi A) is called the characteristic polynomial of the matrix A Theorem 811 Let A M n Then there exists a unique monic polynomial q
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2
MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2 1 Ridge Regression Ridge regression and the Lasso are two forms of regularized
More informationDifferential Equations and Linear Algebra Supplementary Notes. Simon J.A. Malham. Department of Mathematics, HeriotWatt University
Differential Equations and Linear Algebra Supplementary Notes Simon J.A. Malham Department of Mathematics, HeriotWatt University Contents Chapter 1. Linear algebraic equations 5 1.1. Gaussian elimination
More informationWelsh s problem on the number of bases of matroids
Welsh s problem on the number of bases of matroids Edward S. T. Fan 1 and Tony W. H. Wong 2 1 Department of Mathematics, California Institute of Technology 2 Department of Mathematics, Kutztown University
More informationCHAPTER 6. Direct Methods for Solving Linear Systems
CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to
More informationToday: eigenvalue sensitivity, eigenvalue algorithms Reminder: midterm starts today
AM 205: lecture 22 Today: eigenvalue sensitivity, eigenvalue algorithms Reminder: midterm starts today Posted online at 5 PM on Thursday 13th Deadline at 5 PM on Friday 14th Covers material up to and including
More informationMath 121 Homework 5: Notes on Selected Problems
Math 121 Homework 5: Notes on Selected Problems 12.1.2. Let M be a module over the integral domain R. (a) Assume that M has rank n and that x 1,..., x n is any maximal set of linearly independent elements
More informationGaussian Elimination without/with Pivoting and Cholesky Decomposition
Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination WITHOUT pivoting Notation: For a matrix A R n n we define for k {,,n} the leading principal submatrix a a k A
More information8.1 Bifurcations of Equilibria
1 81 Bifurcations of Equilibria Bifurcation theory studies qualitative changes in solutions as a parameter varies In general one could study the bifurcation theory of ODEs PDEs integrodifferential equations
More informationSkewSymmetric Matrix Polynomials and their Smith Forms
SkewSymmetric Matrix Polynomials and their Smith Forms D. Steven Mackey Niloufer Mackey Christian Mehl Volker Mehrmann March 23, 2013 Abstract Two canonical forms for skewsymmetric matrix polynomials
More informationThe NewtonRaphson Algorithm
The NewtonRaphson Algorithm David Allen University of Kentucky January 31, 2013 1 The NewtonRaphson Algorithm The NewtonRaphson algorithm, also called Newton s method, is a method for finding the minimum
More information18.06 Problem Set 2 Solution
18.06 Problem Set 2 Solution Total: 100 points Section 2.5. Problem 24: Use GaussJordan elimination on [U I] to find the upper triangular U 1 : 1 a b 1 0 UU 1 = I 0 1 c x 1 x 2 x 3 = 0 1 0. 0 0 1 0 0
More informationLinear Algebra: Linear Systems and Matrices  Quadratic Forms and Deniteness  Eigenvalues and Markov Chains
Linear Algebra: Linear Systems and Matrices  Quadratic Forms and Deniteness  Eigenvalues and Markov Chains Joshua Wilde, revised by Isabel Tecu, Takeshi Suzuki and María José Boccardi August 3, 3 Systems
More informationSection Signed Measures: The Hahn and Jordan Decompositions
17.2. Signed Measures 1 Section 17.2. Signed Measures: The Hahn and Jordan Decompositions Note. If measure space (X, M) admits measures µ 1 and µ 2, then for any α,β R where α 0,β 0, µ 3 = αµ 1 + βµ 2
More informationarxiv: v1 [math.co] 3 Nov 2014
SPARSE MATRICES DESCRIBING ITERATIONS OF INTEGERVALUED FUNCTIONS BERND C. KELLNER arxiv:1411.0590v1 [math.co] 3 Nov 014 Abstract. We consider iterations of integervalued functions φ, which have no fixed
More informationLecture 2: The Simplex method. 1. Repetition of the geometrical simplex method. 2. Linear programming problems on standard form.
Lecture 2: The Simplex method. Repetition of the geometrical simplex method. 2. Linear programming problems on standard form. 3. The Simplex algorithm. 4. How to find an initial basic solution. Lecture
More informationEigenvalues and Eigenvectors
Eigenvalues and Eigenvectors Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Eigenvalues and Eigenvectors Fall 2015 1 / 14 Introduction We define eigenvalues and eigenvectors. We discuss how to
More informationUpdating the QR decomposition of block tridiagonal and block Hessenberg matrices
Applied Numerical Mathematics 58 (2008) 871 883 www.elsevier.com/locate/apnum Updating the QR decomposition of block tridiagonal and block Hessenberg matrices Martin H. Gutknecht a,, Thomas Schmelzer b
More informationInverse Iteration on Defective Matrices*
MATHEMATICS OF COMPUTATION, VOLUME 31, NUMBER 139 JULY 1977, PAGES 726732 Inverse Iteration on Defective Matrices* By Naifu Chen Abstract. Very often, inverse iteration is used with shifts to accelerate
More informationSPRING OF 2008 D. DETERMINANTS
18024 SPRING OF 2008 D DETERMINANTS In many applications of linear algebra to calculus and geometry, the concept of a determinant plays an important role This chapter studies the basic properties of determinants
More informationc 1997 Society for Industrial and Applied Mathematics Vol. 39, No. 2, pp , June
SIAM REV. c 997 Society for Industrial and Applied Mathematics Vol. 39, No. 2, pp. 254 29, June 997 003 COMPUTING AN EIGENVECTOR WITH INVERSE ITERATION ILSE C. F. IPSEN Abstract. The purpose of this paper
More informationFINITEDIMENSIONAL LINEAR ALGEBRA
DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITEDIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup
More informationRegular Sparse Crossbar Concentrators
Regular Sparse Crossbar Concentrators Weiming Guo and A. Yavuz Oruç Electrical Engineering Department College Park, MD 20742 ABSTRACT A bipartite concentrator is a single stage sparse crossbar switching
More informationarxiv:math/ v2 [math.oa] 8 Oct 2006
arxiv:math/0507166v2 math.oa] 8 Oct 2006 ADVANCES ON THE BESSIS MOUSSAVILLANI TRACE CONJECTURE CHRISTOPHER J. HILLAR Abstract. A longstanding conjecture asserts that the polynomial p(t = Tr(A + tb m
More informationI.C.F. Ipsen small magnitude. Our goal is to give some intuition for what the bounds mean and why they hold. Suppose you have to compute an eigenvalue
Acta Numerica (998), pp. { Relative Perturbation Results for Matrix Eigenvalues and Singular Values Ilse C.F. Ipsen Department of Mathematics North Carolina State University Raleigh, NC 769585, USA ipsen@math.ncsu.edu
More informationConjugate Gradient Tutorial
Conjugate Gradient Tutorial Prof. ChungKuan Cheng Computer Science and Engineering Department University of California, San Diego ckcheng@ucsd.edu December 1, 2015 Prof. ChungKuan Cheng (UC San Diego)
More informationRecovery Based on Kolmogorov Complexity in Underdetermined Systems of Linear Equations
Recovery Based on Kolmogorov Complexity in Underdetermined Systems of Linear Equations David Donoho Department of Statistics Stanford University Email: donoho@stanfordedu Hossein Kakavand, James Mammen
More informationTypes of convergence of matrices. Olga Pryporova. A dissertation submitted to the graduate faculty
Types of convergence of matrices by Olga Pryporova A dissertation submitted to the graduate faculty in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Maor: Mathematics Program
More informationYOUNG TABLEAUX AND THE REPRESENTATIONS OF THE SYMMETRIC GROUP
YOUNG TABLEAUX AND THE REPRESENTATIONS OF THE SYMMETRIC GROUP YUFEI ZHAO ABSTRACT We explore an intimate connection between Young tableaux and representations of the symmetric group We describe the construction
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis Volume 38, pp 7530, 011 Copyright 011, ISSN 10689613 ETNA PERTURBATION ANALYSIS FOR COMPLEX SYMMETRIC, SKEW SYMMETRIC, EVEN AND ODD MATRIX POLYNOMIALS SK
More informationA fullnewton step infeasible interiorpoint algorithm for linear programming based on a kernel function
A fullnewton step infeasible interiorpoint algorithm for linear programming based on a kernel function Zhongyi Liu, Wenyu Sun Abstract This paper proposes an infeasible interiorpoint algorithm with
More information