Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

Similar documents
Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of

UMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart.

Institute for Advanced Computer Studies. Department of Computer Science. On the Adjoint Matrix. G. W. Stewart y ABSTRACT

On the Perturbation of the Q-factor of the QR Factorization

Institute for Advanced Computer Studies. Department of Computer Science. On Markov Chains with Sluggish Transients. G. W. Stewart y.

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.

Institute for Advanced Computer Studies. Department of Computer Science. On the Convergence of. Multipoint Iterations. G. W. Stewart y.

then kaxk 1 = j a ij x j j ja ij jjx j j: Changing the order of summation, we can separate the summands, kaxk 1 ja ij jjx j j: let then c = max 1jn ja

Department of Mathematics Technical Report May 2000 ABSTRACT. for any matrix norm that is reduced by a pinching. In addition to known

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices

Scientific Computing

AMS526: Numerical Analysis I (Numerical Linear Algebra)

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Gaussian Elimination without/with Pivoting and Cholesky Decomposition

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

This can be accomplished by left matrix multiplication as follows: I

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Multiplicative Perturbation Analysis for QR Factorizations

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Numerical Linear Algebra

14.2 QR Factorization with Column Pivoting

Matrix decompositions

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

lecture 2 and 3: algorithms for linear algebra

Perturbation Analyses for the Cholesky Factorization with Backward Rounding Errors

Algebra C Numerical Linear Algebra Sample Exam Problems

Numerical Methods - Numerical Linear Algebra

c 2010 Society for Industrial and Applied Mathematics

Introduction to Mathematical Programming

Numerical Linear Algebra And Its Applications

MULTIPLICATIVE PERTURBATION ANALYSIS FOR QR FACTORIZATIONS. Xiao-Wen Chang. Ren-Cang Li. (Communicated by Wenyu Sun)

An algorithm for symmetric generalized inverse eigenvalue problems

Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

MODULE 7. where A is an m n real (or complex) matrix. 2) Let K(t, s) be a function of two variables which is continuous on the square [0, 1] [0, 1].

Computational Linear Algebra

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Matrix Factorizations

On the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

Numerical Methods I: Eigenvalues and eigenvectors

y b where U. matrix inverse A 1 ( L. 1 U 1. L 1 U 13 U 23 U 33 U 13 2 U 12 1

Review Questions REVIEW QUESTIONS 71

Linear Algebra Review

An exploration of matrix equilibration

WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

Important Matrix Factorizations

ESTIMATION OF ERROR. r = b Abx a quantity called the residual for bx. Then

Numerical Methods I: Numerical linear algebra

lecture 3 and 4: algorithms for linear algebra

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Lecture Note 2: The Gaussian Elimination and LU Decomposition

Matrix decompositions

The Matrix Sign Function Method and the. Computation of Invariant Subspaces. November Abstract

The System of Linear Equations. Direct Methods. Xiaozhou Li.

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

I-v k e k. (I-e k h kt ) = Stability of Gauss-Huard Elimination for Solving Linear Systems. 1 x 1 x x x x

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

Linear Algebraic Equations

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Numerical Methods in Matrix Computations

Numerical Linear Algebra

the Unitary Polar Factor æ Ren-Cang Li Department of Mathematics University of California at Berkeley Berkeley, California 94720

Matrix Multiplication Chapter IV Special Linear Systems

the Unitary Polar Factor æ Ren-Cang Li P.O. Box 2008, Bldg 6012

Lecture 9. Errors in solving Linear Systems. J. Chaudhry (Zeb) Department of Mathematics and Statistics University of New Mexico

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

Computational Methods. Least Squares Approximation/Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Matrix decompositions

1 GSW Sets of Systems

ON THE QR ITERATIONS OF REAL MATRICES

Section 5.6. LU and LDU Factorizations

A Parallel Implementation of the. Yuan-Jye Jason Wu y. September 2, Abstract. The GTH algorithm is a very accurate direct method for nding

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Gaussian Elimination and Back Substitution

Applied Linear Algebra in Geoscience Using MATLAB

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables,

Draft. Lecture 12 Gaussian Elimination and LU Factorization. MATH 562 Numerical Analysis II. Songting Luo

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are

LU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b

Part IB Numerical Analysis

Dense LU factorization and its error analysis

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

or H = UU = nx i=1 i u i u i ; where H is a non-singular Hermitian matrix of order n, = diag( i ) is a diagonal matrix whose diagonal elements are the

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

MATRICES WITH POSITIVE DEFINITE HERMITIAN PART: INEQUALITIES AND LINEAR SYSTEMS ROY MATHIAS

EBG # 3 Using Gaussian Elimination (Echelon Form) Gaussian Elimination: 0s below the main diagonal

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4

Analysis of Block LDL T Factorizations for Symmetric Indefinite Matrices

Transcription:

University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{95{93 TR{3535 On the Perturbation of LU and Cholesky Factors G. W. Stewart y October, 1995 ABSTRACT In a recent paper, Chang and Paige have shown that the usual perturbation bounds for Cholesky factors can systematically overestimate the errors. In this note we sharpen their results and extend them to the factors of the LU decomposition. The results are based on a new formula for the rst order terms of the error in the factors. This report is available by anonymous ftp from thales.cs.umd.edu in the directory pub/reports. y Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 074. This work was supported in part by the National Science Foundation under grant CCR 9550316.

On the Perturbation of LU and Cholesky Factors G. W. Stewart ABSTRACT In a recent paper, Chang and Paige have shown that the usual perturbation bounds for Cholesky factors can systematically overestimate the errors. In this note we sharpen their results and extend them to the factors of the LU decomposition. The results are based on a new formula for the rst order terms of the error in the factors. 1. Introduction Let A be a positive denite matrix of order n. Then A has a unique Cholesky factorization of the form A = R T R, where R is upper triangular with positive diagonal elements. Let ~ A = A + E be a perturbation of A in which E is symmetric. If E is suciently small, then ~ A also has a Cholesky factorization: A + E = (R + F R ) T (R + F R ): Several workers [1, 3, 4, 7] have given bounds on the matrix F R. The common result is essentially that kf R k F krk 1 p (A) kek F kak + O(kEk ): (1.1) Recently Chang and Paige [] have shown that (1.1) can consistently overestimate the error in the Cholesky factor and have proposed new bounds. Following [3], they note that E = R T F R + F T R R + O(kF Rk ): (1.) Consequently if one denes the linear operator T R on the space of upper triangular matrices by T R (F ) = R T F + F T R; then F R = FR T 1 R (E); 1

Perturbation of LU Factorizations and kf R k < kt 1 R kkek; where k k denotes a suitably chosen norm. By examining the matrix representation of T R, Chang and Paige were able to show that bounds based on kt 1 R k are sharper than the conventional bounds. They also derive a lower bound for kt 1 R k, and show by example that the failure of (1.1) is somehow connected with pivoting in the computation of the decomposition. The purpose of this note is to generalize and strengthen the results of Chang and Paige. We do so by exhibiting an explicit matrix representation of F R. The representation is invariant under a certain kind of diagonal scaling, and by adjusting the scaling we can improve the usual bound. Since the approach works for the more general LU decomposition, we will treat that case rst and then specialize to the Cholesky decomposition. Throughout this note k k will denote an absolute norm, such as the 1-norm, the 1-norm, or the Frobenius norm, which will also be denoted by k k F. The matrix -norm, which is not absolute, will be denoted by kk. For any nonsingular matrix X we will dene (X) = kxk kx 1 k : For more on norms see [5].. The LU Decomposition Let A be a matrix of order n whose leading principal submatrices are nonsingular. Then A can be written in the form A = LU; where L is lower triangular and U is upper triangular. The decomposition is not unique, but it can be made so by specifying the diagonal elements of L. (The conventional choice is to require them to be one.) If E is suciently small, A + E has an LU factorization: A + E = (L + F L )(U + F U ): (.1) Again, the factorization is not unique, but it can be made so, say by requiring that the diagonals of L remain unaltered.

Perturbation of LU Factorizations 3 Multiplying out the right hand side of (.1) and ignoring higher order terms, we obtain a linear matrix equation for rst order approximations F L and F U to F L and F U : L FU + U FL = E: We shall show how to solve this equation in terms of two matrix operators. Let 0 p 1, and dene L p and U p as illustrated below for a 3 3 matrix: L p (X) = 0 B @ px 11 0 0 x 1 px 0 x 31 x 3 px 33 It then follows that for any matrix X, 1 C A and U p (X) = 0 B @ px 11 x 1 x 13 0 px x 3 0 0 px 33 1 C A : X = L p (X) + U 1 p (X); (.) and Finally, if X is symmetric kl p (X)k; ku p (X)k kxk: Our basic result is the following: To see this, write ku 1 (X)k F p 1 kxk F : (.3) F L = LL p (L 1 EU 1 ) and FU = U 1 p (L 1 EU 1 )U: L[U 1 p (L 1 EU 1 )U] + [LL p (L 1 EU 1 )]U = L[U 1 p (L 1 EU 1 ) + L p (L 1 EU 1 )]U = L(L 1 EU 1 )U by (.) = E: The number p is a normalizing parameter, controling how much of the perturbation is attached to the diagonals of L and U. If p = 0, the diagonal elements of L do not change. If p = 1, the diagonal elements of U do not change. We can take norms in the expressions FL and FU to get rst order perturbation bounds for the LU decomposition. But it is possible to introduce degrees of freedom in the expressions that can later be used to reduce the bounds. Specically, for any nonsingular diagonal matrix D L, we have F L = LD L L 0 (D 1 L L 1 EU 1 ) ^LL0 (^L 1 EU 1 )

4 Perturbation of LU Factorizations Consequently or Since kak klkkuk, we have k F L k k^lkk^l 1 kku 1 kkek; k FL k klk (^L)(U)kEk : (.4) klkku k k F L k klk (^L)(U) kek kak : (.5) Similarly, if D U is a nonsingular diagonal matrix and we set then ^U = D U U; k F U k kuk (L)( ^U) kek kak : (.6) The bounds (.5) and (.6) dier from the usual bounds (e.g., see [4]) by the substitution of ^L or ^U for L or U. However, if the diagonal matrices D L and D U are chosen appropriately, (^L) and ( ^U) can be far less that (L) or (U). For example, if! U = 1 ; (.7) 0 then 1 (U) = 1=. But if we set D U = diag(1; 1=), then ( ^U) = 1. Poorly scaled but essentially well-conditioned matrices like U in (.7) occur naturally. If A is ill-conditioned and the LU decomposition of A is computed with pivoting, the ill-conditioning of A will usually reveal itself in the diagonal elements of U. In [6] the author has shown that such upper triangular matrices are articially ill conditioned in the sense that they can be made well conditioned by scaling their rows. If (^L) = 1 (it cannot be less), then the bound (.4) reduces to k F L k ku 1 kkek: (.8) It is reasonable to ask if there are problems for which we can replace ku 1 k by an even smaller number and still have inequality for all E. The answer depends on p. For example, suppose that k k is the 1-norm. Let e i be the unit coordinate

Perturbation of LU Factorizations 5 vectors, and let k be such that ke T k U 1 k = ku 1 k. Let E = e n e T k, so that kek = 1 Then it is easy to see that Hence k^llp (^L 1 EU 1 )k = kl p (e n e T k U 1 )k ku 1 kkek k FL k pku 1 kkek: (.9) Consequently, if p is near one, (.8) is essentially the smallest bound that holds uniformly for all E. The reason for the appearance of the factor p in (.9) is that the error may concentrate in the last column of L 1 EU 1, in which case it is reduced by a factor of at least p by the operator L p. This can happen, for example, when L = I and U = diag(i n 1 ; ) for small. However, if p si small, the perturbation will show up in F U, for which the factor is 1 p. The bounds (.5) and (.6) suggest a strategy for estimating the condition of the LU factorizations. Van der Sluis [8] has shown that in the -norm, the condition number is approximately minimized when the rows or columns of the matrix are scaled to have norm one. Thus the strategy is to so scale ^L and ^U and use a condition estimator to estimate the condition of L, ^L, U, ^U. In [4] it is shown how to obtain rigorous bounds for the errors in the rst order approximations FL and FU. Since the second order terms decay rapidly, the error bounds are less important than the condition that insures their existence: namely, kl 1 kku 1 kkek 1 4 : 3. The Cholesky Decomposition We now return to the Cholesky decomposition. In analyzing the perturbation of the the Cholesky factor R it is natural to take p = 1 so that symmetry is preserved. In this case the solution of the perturbation equation becomes F R = U 1 (R T ER 1 )R: Hence if ^R is dened in analogy with ^L and ^U, it follows from (.3) that k F R k F krk 1 p ( ^R) (R) kek F kak :

6 Perturbation of LU Factorizations Moreover, by a variant of the argument that lead to (.9), for any A, there is an E such that kr 1 k kek F k F R k F 1 kr 1 k kek F which shows that we cannot reduce the constant in the bound to less than 1 kr 1 k. k F R k F kek F References [1] A. Barrland. Perturbation bounds for the LDL H and the LU factorizations. BIT, 31:358{363, 1991. [] X-W. Chang and C. C. Paige. A new perturbaton analysis for the Cholesky factorization. School of Computer Science, McGill University. To appear in the IMA Journal of Numerical Analysis., 1995. [3] G. W. Stewart. Perturbation bounds for the QR factorization of a matrix. SIAM Journal on Numerical Analysis, 1977:509{518, 1977. [4] G. W. Stewart. On the perturbation of LU, Cholesky, and QR factorizations. SIAM Journal on Matrix Analysis and Applications, 14:1141{1146, 1993. [5] G. W. Stewart and J.-G. Sun. Matrix Perturbation Theory. Academic Press, Boston, 1990. [6] G. W. Stwart. The triangular matrices of Gaussian elimination and related decompositions. Technical Report CS-TR-3533 UMIACS-TR-95-91, University of Maryland, Department of Computer Science, 1995. [7] J.-G. Sun. Rounding-error and perturbation bounds for the Cholesky and LDL T factorizations. Linear Algebra and Its Applications, 173:77{98, 199. [8] A. van der Sluis. Condition numbers and equilibration of matrices. Numerische Mathematik, 14:14{3, 1969.