Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices

Similar documents
Matrix Multiplication Chapter IV Special Linear Systems

Scientific Computing

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

AMS526: Numerical Analysis I (Numerical Linear Algebra)

9. Numerical linear algebra background

Numerical Linear Algebra

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Symmetric matrices and dot products

Lecture 9: Numerical Linear Algebra Primer (February 11st)

9. Numerical linear algebra background

MS&E 318 (CME 338) Large-Scale Numerical Optimization

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

7. Symmetric Matrices and Quadratic Forms

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

Numerical Methods I Non-Square and Sparse Linear Systems

Analysis of Block LDL T Factorizations for Symmetric Indefinite Matrices

B553 Lecture 5: Matrix Algebra Review

Numerical Methods I: Numerical linear algebra

Matrix decompositions

ANONSINGULAR tridiagonal linear system of the form

Numerical Linear Algebra

ECE133A Applied Numerical Computing Additional Lecture Notes

Chapter 3 Transformations

Scientific Computing: Solving Linear Systems

1 GSW Sets of Systems

An Algorithm for Solving the Convex Feasibility Problem With Linear Matrix Inequality Constraints and an Implementation for Second-Order Cones

LINEAR SYSTEMS (11) Intensive Computation

Matrix decompositions

5.1 Banded Storage. u = temperature. The five-point difference operator. uh (x, y + h) 2u h (x, y)+u h (x, y h) uh (x + h, y) 2u h (x, y)+u h (x h, y)

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Conjugate Gradient Method

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix

Regularized Least Squares

Lecture 2 INF-MAT : , LU, symmetric LU, Positve (semi)definite, Cholesky, Semi-Cholesky

Numerical Methods I: Eigenvalues and eigenvectors

Notes on Eigenvalues, Singular Values and QR

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

arxiv:cs/ v1 [cs.ms] 13 Aug 2004

Solving linear systems (6 lectures)

Key words. numerical linear algebra, direct methods, Cholesky factorization, sparse matrices, mathematical software, matrix updates.

Review Questions REVIEW QUESTIONS 71

There are six more problems on the next two pages

Data Analysis and Manifold Learning Lecture 2: Properties of Symmetric Matrices and Examples

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Numerical Methods in Matrix Computations

6.4 Krylov Subspaces and Conjugate Gradients

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Let x be an approximate solution for Ax = b, e.g., obtained by Gaussian elimination. Let x denote the exact solution. Call. r := b A x.

A Revised Modified Cholesky Factorization Algorithm 1

Positive Definite Matrix

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

LU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b

Regularized Least Squares

Modified Cholesky algorithms: a catalog with new approaches

ETNA Kent State University

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Scientific Computing: Dense Linear Systems

Numerical Analysis Fall. Gauss Elimination

HW2 - Due 01/30. Each answer must be mathematically justified. Don t forget your name.

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

Key words. modified Cholesky factorization, optimization, Newton s method, symmetric indefinite

Lecture Notes to be used in conjunction with. 233 Computational Techniques

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

CS 542G: Conditioning, BLAS, LU Factorization

ACM106a - Homework 2 Solutions

Computational Methods. Eigenvalues and Singular Values


AN ITERATION. In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b

Incomplete Cholesky preconditioners that exploit the low-rank property

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS

The Singular Value Decomposition

Matrix decompositions

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Parallel Singular Value Decomposition. Jiaxing Tan

Course Notes: Week 1

Iterative Methods. Splitting Methods

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Linear Algebra in IPMs

Solving Large Nonlinear Sparse Systems

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

Homework 2 Foundations of Computational Math 2 Spring 2019

Linear Algebraic Equations

This ensures that we walk downhill. For fixed λ not even this may be the case.

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4

Numerical Linear Algebra

Next topics: Solving systems of linear equations

Lecture Notes to be used in conjunction with. 233 Computational Techniques

Gaussian Elimination and Back Substitution

12. Cholesky factorization

Transcription:

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as the difference of two positive semidefinite matrices. We start with simple ideas such as eigenvalue decomposition. hen we develop a simple adaptation of the Cholesky that returns a difference-of-cholesky representation of indefinite matrices. Heuristics that promote sparsity can be applied directly to this modification. 1 Introduction Let A S n be a real-valued symmetric matrix. It is always possible to write A as the difference P + P of two positive semidefinite matrices P + and P. ake for example P + A + ti and P ti for large enough t > 0. With this representation one can rewrite an indefinite quadratic form x Ax as a difference-of-convex expression x P + x x P x which can be used in various algorithms for convex-concave programming [LB14 SDGB16]. In this note we explore a number of different such representations and their properties. 1.1 Desired properties In addition to the running time of the algorithm there are several other properties that are important in typical applications. ime complexity. Clearly we want the algorithm for computing the representation to be fast. Asymptotic performance is important but in practice even with the same asymptotic complexity one method may outperform another depending on how sparsity is exploited. Memory usage. Ideally representing P + and P should not cost much more than representing P itself. In the worst case both P + and P are dense and 2n 2 floating point numbers must be stored. It is possible to save more space by exploting the sparsity pattern of A or representing the matrices P + and P implicitly in a memory-efficient form. 1

Numerical stability. Due to the round-off errors of floating point numbers the resulting representation may be numerically inaccurate. his can be particularly problematic when P + and P are stored implicitly (e.g. via Cholesky factorization) rather than explicitly. Singular or badly conditioned matrices can often introduce big round-off errors and we want the algorithm to be robust in such settings. Additional curvature. In order for convex-concave procedure to perform better we want P + and P to have as little distortion as possible i.e. we want the additional curvature measured by the nuclear norm P + + P A to be small. Since P + and P are required to be positive semidefinite this quantity is equal to r P + + r P A. 2 Simple representations Any indefinite matrix can be made positive semidefinite by adding a large enough multiple of the identity. Let λ min < 0 be the smallest eigenvalue of A. hen for any t λ min P + A + ti P ti is a pair of positive semidefinite matrices whose difference is A. If the magnitude of the maximum eigenvalue λ max > 0 is smaller than that of λ min then an alternative representation is also possible: P + ti P ti A where t λ max. his representation is relatively easy to compute as it only requires a lower bound on λ min or λ max. It also has a property that the sparsity of A is preserved as much as possible in that no new off-diagonal nonzero entries are introduced in P + or P. Its disadvantage is the additional curvature it introduces: P + + P A 2t. Another simple representation is based on the full eigenvalue decomposition of A. his representation preserves the norm of A and thus introduces no additional curvature. Let A QΛQ be the eigenvalue decomposition of A where Λ diag(λ 1... λ n ) is the eigenvalue matrix with λ 1 λ k 0 > λ k+1 λ n. hen Λ can be written as Λ Λ + Λ where Λ + diag(λ 1... λ k 0... 0) Λ diag(0... 0 λ k+1... λ n ). 2

Setting P + QΛ + Q P + QΛ Q then gives a difference of positive semidefinite matrix representation of A. In general the cost of explicitly computing this representation is very high as it not only requires the full eigenvalue decomposition but also destroys the sparsity of A even when it is very sparse. 3 Cholesky-like representations When A is positive semidefinite there exists a unique lower triangular matrix L R n n that satisfies A LL. his representation is known as the Cholesky factorization of A. In the simplest form the Cholesky algorithm can be described as a simple recursive algorithm. Formally let chol : S n R n n that computes given A S n a lower triangular matrix L chol(a) that satisfies A LL. If A is 1-by-1 then L is simply given by A 11. If A has two or more rows let [ ] A with a 0 v R n 1 M R (n 1) (n 1). If a 0 is zero then v must also equal zero for otherwise A is indefinite. In this case the Cholesky factorization is given by ([ ]) [ ] 0 0 0 chol. 0 M 0 chol(m) If a > 0 the recursion is: ([ chol ]) [ a v/ a chol(m vv /a) ]. he cost of computing a dense Cholesky factorization is (1/3)n 3 flops. In case A is sparse various pivoting heuristics can be used to exploit the sparsity structure and speed up the computation. Cholesky factorization does not exist when A has both positive and negative eigenvalues. However the LDL decomposition which is a close variant of the Cholesky factorization exists for all symmetric matrices. It also has an additional computational advantage since there is no need to take square roots. he idea is to write A as LDL where L R n n is lower triangular with ones on the main diagonal and D R n n is block diagonal consisting of 1-by-1 or 2-by-2 blocks. In general some additional computation is required to transform the LDL decomposition into the difference of two positive semidefinite matrices due to the 2-by-2 blocks in D. When D is diagonal (i.e. no 2-by-2 blocks) then one can easily separate out the negative entries of D and the corresponding columns in L to write A as A L 1 D 1 L 1 L 2 D 2 L 2 where D 1 and D 2 are diagonal matrices consisting of nonnegative entries only. 3

4 Difference-of-Cholesky representation In this section we describe a simple modification : S n R n n R n n of the Cholesky algorithm that computes given A S n a pair of lower triangular matrices L 1 and L 2 such that A L 1 L 1 L 2 L 2 where (L 1 L 2 ) (A). Let a 0 v R n 1 and M S n 1. Below is a simple recursion for computing when a > 0 is large enough: ([ ]) ([ ] [ ]) a v/ 0 (S ) (M (1/a)vv ) a S 0 ([ ]) ([ ] [ ]) 0 a 0 S v/ (S ) (M + (1/a)vv ). a In this form the algorithm is no different from computing the LDL decomposition and separating out positive and negative entries in D (assuming the LDL decomposition exists with a diagonal D). In case a is too small so that dividing by a would give big round-off errors the following recursion can be used for any value of δ > 0. Notice that with this recursion the algorithm can proceed even when a 0. δ + a δ (S ) (M v S 1 v1 + v 2 v2 ). v 1 v 2 Here v 1 and v 2 are any vectors satisfying δ + 1 δ v 2 v. his introduces an additional degree of freedom as we can freely choose v 1 or v 2. For example we can let v 1 0 or v 2 0. his way it is possible to trade off the number of nonzero elements in L 1 against L 2 which is a property that is not readily available in other representations discussed above. When we choose v 1 0 the recursion becomes δ + a δ 0 S v/ (S ) (M +(1/δ)vv ). δ For completeness we show the recursion in the case of A 11 0. δ δ + a (S ) (M v S 1 v1 +v 2 v2 ) v 1 v 2 where v 1 and v 2 are any vectors satisfying δ v 1 δ + 2 v. It is simple to modify this algorithm to return a pair of LDL factorizations to avoid computing square roots. With the additional parameter δ > 0 both factorizations become positive definite and thus there will not be any 2-by-2 diagonal block in the factorizations. Since the method is very close to the original Cholesky algorithm and thus most extensions and techniques applicable to Cholesky algorithm or LDL decomposition naturally apply to the difference-of-cholesky method. One such example is pivoting heuristics for LDL decomposition for avoiding round-off errors or ill-conditioned matrices and at the same time preserving the sparsity of the intermediate matrices as much as possible. 4

References [BP71] J. R. Bunch and B. N. Parlett. Direct methods for solving symmetric indefinite systems of linear equations. SIAM Journal on Numerical Analysis 8(4):639 655 1971. [CH98] S. H. Cheng and N. J. Higham. A modified Cholesky algorithm based on a symmetric indefinite factorization. SIAM Journal on Matrix Analysis and Applications 19(4):1097 1110 1998. [Fle76] [LB14] [PS75] R. Fletcher. Factorizing symmetric indefinite matrices. Linear Algebra and its Applications 14(3):257 272 1976.. Lipp and S. Boyd. Variations and extension of the convex concave procedure. Optimization and Engineering pages 1 25 2014. C. C. Paige and M. A. Saunders. Solution of sparse indefinite systems of linear equations. SIAM journal on numerical analysis 12(4):617 629 1975. [SDGB16] X. Shen S. Diamond Y. Gu and S. Boyd. Disciplined convex-concave programming. arxiv preprint:1604.02639 2016. 5