Aplikace matematiky. Gene Howard Golub Least squares, singular values and matrix approximations. Terms of use:

Similar documents
Aplikace matematiky. Gene Howard Golub Numerical methods for solving linear least squares problems

Aplikace matematiky. Jiří Neuberg Some limit properties of the best determined terms method. Terms of use:

Czechoslovak Mathematical Journal

Časopis pro pěstování matematiky

Bounds for Eigenvalues of Tridiagonal Symmetric Matrices Computed. by the LR Method. By Gene H. Golub

Časopis pro pěstování matematiky

σ 11 σ 22 σ pp 0 with p = min(n, m) The σ ii s are the singular values. Notation change σ ii A 1 σ 2

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u

Časopis pro pěstování matematiky

Časopis pro pěstování matematiky

Mathematica Slovaca. Antonín Dvořák; David Jedelský; Juraj Kostra The fields of degree seven over rationals with a normal basis generated by a unit

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Theorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k

Časopis pro pěstování matematiky

Orthogonalization and least squares methods

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Linear Algebra Review. Vectors

Homework 1. Yuan Yao. September 18, 2011

UNIT 6: The singular value decomposition.

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

LinGloss. A glossary of linear algebra

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

In this section again we shall assume that the matrix A is m m, real and symmetric.

Main matrix factorizations

CHARACTERIZATIONS. is pd/psd. Possible for all pd/psd matrices! Generating a pd/psd matrix: Choose any B Mn, then

Computational Linear Algebra

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Computational Methods. Eigenvalues and Singular Values

Lecture 3: QR-Factorization

3D Computer Vision - WT 2004

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Časopis pro pěstování matematiky

Mathematica Slovaca. Jaroslav Hančl; Péter Kiss On reciprocal sums of terms of linear recurrences. Terms of use:

Parallel Numerical Algorithms

EQUADIFF 1. Milan Práger; Emil Vitásek Stability of numerical processes. Terms of use:

Applied Numerical Linear Algebra. Lecture 8

Math 489AB Exercises for Chapter 2 Fall Section 2.3

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Large Scale Data Analysis Using Deep Learning

Matematický časopis. Bedřich Pondělíček A Note on Classes of Regularity in Semigroups. Terms of use: Persistent URL:

Linear Differential Transformations of the Second Order

WSAA 14. A. K. Kwaśniewski Algebraic properties of the 3-state vector Potts model. Terms of use:

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

CS 246 Review of Linear Algebra 01/17/19

Lecture 5 Singular value decomposition

Preliminary Examination, Numerical Analysis, August 2016

ELEMENTARY LINEAR ALGEBRA

Solving large scale eigenvalue problems

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Commentationes Mathematicae Universitatis Carolinae

Vector and Matrix Norms. Vector and Matrix Norms

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

The QR Factorization

Total Least Squares Approach in Regression Methods

Properties of Matrices and Operations on Matrices

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

A matrix over a field F is a rectangular array of elements from F. The symbol

Chapter 3. Matrices. 3.1 Matrices

Linear Algebra. Brigitte Bidégaray-Fesquet. MSIAM, September Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble.

Block Bidiagonal Decomposition and Least Squares Problems

Section 4.4 Reduction to Symmetric Tridiagonal Form

CS 143 Linear Algebra Review

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Linear algebra for computational statistics

Introduction. Vectors and Matrices. Vectors [1] Vectors [2]

Chapter 3 Transformations

ELEMENTARY LINEAR ALGEBRA

Linear Algebraic Equations

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

MAT 610: Numerical Linear Algebra. James V. Lambers

Math. H110 Answers for Final Examination December 21, :56 pm

WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS

Singular Value Decomposition

Positive Definite Matrix

Matrices, Moments and Quadrature, cont d

Notes on basis changes and matrix diagonalization

B553 Lecture 5: Matrix Algebra Review

Singular Value Decompsition

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

ELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n

Numerical Methods - Numerical Linear Algebra

1 Singular Value Decomposition and Principal Component

WSGP 23. Lenka Lakomá; Marek Jukl The decomposition of tensor spaces with almost complex structure

14.2 QR Factorization with Column Pivoting

Introduction to Numerical Linear Algebra II

ELEMENTARY LINEAR ALGEBRA

Phys 201. Matrices and Determinants

Notes on Householder QR Factorization

14 Singular Value Decomposition

Linear Algebra for Machine Learning. Sargur N. Srihari

Fundamentals of Engineering Analysis (650163)

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Linear Algebra, part 3 QR and SVD

Knowledge Discovery and Data Mining 1 (VO) ( )

Unit 2, Section 3: Linear Combinations, Spanning, and Linear Independence Linear Combinations, Spanning, and Linear Independence

Algebra II. Paulius Drungilas and Jonas Jankauskas

Numerical Linear Algebra

1 Linearity and Linear Systems

Linear Least squares

Transcription:

Aplikace matematiky Gene Howard Golub Least squares, singular values and matrix approximations Aplikace matematiky, Vol. 3 (968), No., 44--5 Persistent URL: http://dml.cz/dmlcz/0338 Terms of use: Institute of Mathematics AS CR, 968 Institute of Mathematics of the Academy of Sciences of the Czech Republic provides access to digitized documents strictly for personal use. Each copy of any part of this document must contain these Terms of use. This paper has been digitized, optimized for electronic delivery and stamped with digital signature within the project DML-CZ: The Czech Digital Mathematics Library http://dml.cz

SVAZEK 3 (968) APLIKACE MATEMATIKY ČÍSLO LEAST SQUARES, SINGULAR VALUES AND MATRIX APPROXIMATIONS ) GENE HOWARD GOLUB 0. Let A be a real, m x n matrix (for notational convenience we assume that m j_ n). It is well known (cf. [6]) that (0.) A - UIV T where and I = n) x n The matrix U consists of the orthonormalized eigenvectors of AA r, and the matrix V consists of the orthonormalized eigenvectors of A T A. The diagonal elements of I are the non-negative square roots of the eigenvalues of A T A; they are called singular values or principal values of A. Throughout this note, we assume 0"i = "2 = = a n = 0 Thus if rank (A) = r, er r+ = er r+2 =... = a n = 0. The decomposition (0.) is called the singular value decomposition. In this paper, we shall present a numerical algorithm for computing the singular value decomposition.. The singular value decomposition plays an important role in a number of least squares problems, and we will illustrate this with some examples. Throughout this discussion, we use the euclidean or Frobenius norm of a matrix, viz. Ml = (EM 2 ) ' 2 ) This work was in part supported by NSF and ONR. 44

A. Let tfl n be the set of all n x n orthogonal matrices. For an arbitrary n x n real matrix A, determine Q e l/ n such that A - Q ^ \\A - Kj for any X e ll n. It has been shown by FAN and HOFFMAN [2] that if A = UIV T, then Q = UV T. B. An important generalization of problem A occurs in factor analysis. For arbitrary n x n real matrices A and B, determine 0 e ll n such that A - BQ ^ A - BX\\ for any X e % n. It has been shown by GREEN [5] and by SCHONEMANN [9] that if B T A = UIV T, then Q = UV T. C. Let Ji n \] n be the set of all m x n matrices of rank k. Assume A e Ji^n- Determine B e M^n (k <; r) such that A - B g AL - X\\ for all Xe, It has been shown by ECKART and YOUNG [] that if f(k) m,n (i.l) UIV then B = UQ.V where (.2) ӣь ffi.o,...,0 0, ff 2,...,0 o, o,.. ff* Note that (.3) B = 2 - fì t (ff, 2 + +... + (т r2 ) / 2. D. An n x m matrix X is said to be the pseudo-inverse of an m x n matrix A if X satisfies the following four properties: i) AXA = A, ii) XAX = X, iii) (AX) T = AX, iv) (XA) T = XA. We denote the pseudo-inverse by A +. We wish to determine A + be shown [8] that A + can always be determined and is unique. numerically. It can 45

It is easy to verify that (.4) where Л = o-i 0, A + - VAU 7 0,..,0 г..,0 0, 0, o r n x m In recent years there have been a number of algorithms proposed for computing the pseudo-inverse of a matrix. These algorithms usually depend upon a knowledge of the rank of the matrix or upon some suitably chosen parameter. For example in the latter case, if one uses (.4) to compute the pseudo-inverse, then after one has computed the singular value decomposition numerically it is necessary to determine which of the singular values are zero by testing against some tolerance. Alternatively, suppose we know that the given matrix A can be represented as where SB is a matrix of perturbations and A = B + 8B \\8B\\ S * Now, we wish to construct a matrix B such that and -4 - B\\ = ri rank (S) = minimum. This can be accomplished with the aid of the solution to problem (C). Let as in equation (.2). if Then using (.3), B k = UQ k V T B = B and (al + al + +... + a 2 y 2 >n. 46

Since rank (B) = p by construction, B + = VQ + pu T. Thus, we take B + as our approximation to A +. E. Let A be a given matrix, and let b be a known vector. Determine a vector x such that for \\b Ax 2 = min and x 2 = min, where y 2 = CEy 2 ) /2 for any vector y. It is easy to verify that x = A + b. A norm is said to be unitarily invariant if \\AU\\ = \\VA\\ = A when U*U = I and V*V = I. FAN and HOFFMAN [2] have shown that the solution to problem (A) is the same for all unitarily invariant norms and MIRSKY [7] has proved a similar result for the solution to problem (C). 2. In [4] it was shown by GOLUB and KAHAN that it is possible to construct a sequence of orthogonal matrices {P (fc) } =i> {Q (k) } =i via Householder transformation so that poopoi-l) po) A Q<DQi2) goi-l) s pt A Q = j and J is an m x n bi-diagonal matrix of the form J =., o o-i, ß u 0,.. 0, a 2, ß 2..., 0 0, 0, 0,..., /? n _i 0, 0, o,.. 0 } (m n) x n. The singular values of J are the same as those of A. Thus if the singular value decomposition of J = XIY T then A = PXIY T Q T so that U = PX, V= QY. A number of algorithms were proposed in [4] for computing the singular value decomposition of J. We now describe a new algorithm, based on the QR algorithm of FRANCIS [3], for computing the singular value decomposition of J. Let "0 a t (x x 0 pi 0 ІC = ßl u a 2 a 2 a, a и 0 2n x 2и 47

It can be shown [4] that K is a symmetric, tri-diagonal matrix whose eigenvalues arc ± singular values of J. One of the most effective methods of computing the eigenvalues of a tri-diagonal matrix is the QR algorithm of Francis, which proceeds as follows: Begin with the given matrix K = K 0. Compute the factorization where M 0 M 0 = and R 0 is an upper triangular matrix, and then multiply the matrices in reverse order so that K x = R 0 M 0 = MlK 0 M 0. Now one treats K x in the same fashion as the matrix K 0, and a sequence of matrices is obtained by continuing ad infinitum. Thus so that K, = MiRi and K i + =.R t M, = M i+ R i+, K i+ = MjK-Mi = MjMj_... M T 0KM 0 M X... M {. The method has the advantage that K, remains tri-diagonal throughout the computation. For suitably chosen shift parameters s if we can accelerate the convergence of the QR method by computing (2.) (K; - s,7) = MiRi, RiMi + sj = K i+. Unfortunately, the shift parameter 5, may destroy the zeroes on the diagonal of K. Since the eigenvalues of K always occur in pairs, it would seem more appropriate to compute the QR decomposition of so that (K i -s l I)(K i + SiI) = K 2 i-s 2 ii MiRi = K\ - s 2 I. It has been shown by Francis that it is not necessary to compute (2.) explicitly but it is possible to perform the shift implicitly. Let {-V,} w = {M i } ka, k = U2,...,2n. (i.e., the elements of the first column of N are equal to the elements of the first column of M) and NjN ( = I. Then if 48 i) T l+ = NjKiNi, *0 -Ti+i s a tri-diagonal matrix,

iii) Ki is non-singular, iv) the sub-diagonal elements of T i+i are positive, it follows that T i + i = K J +. The calculation proceeds quite simply. Dropping the iteration counter i), let GO (P + i) (P + 2) z = cos 0 p, 0, sin p, o,, 0, sin p 0 cos Op (P) (P + l) (p + 2) 2n x 2п. Then cos t9 x is chosen so that Then the matrix {Z X (K 2 - s 2 I)} M - 0 for fc = 2,3 2n.^ Z^KZ X = "0 ai 0 d! a; o /І; o /;;. a 2 di a 2. 0 a л a 0 and T Z 2n ~2 Z Í KZ Í where Z 2,..., Z 2rj _ 2 are constructed so that Tis tri-diagonal. The product of all the orthogonal transformations which gives the singular values yields the matrix of orthogonal eigenvectors of K. For ease of notation let us write Jij-i = «j 9 I =, 2,..., n, /2j = Pj, I =,2,..., n -. Then explicitly, the calculation goes as follows: Dropping the iteration counter i, To = y\ - s 2, d 0 = ri72 49

For j = 0,,... 2/i 3, wtf + rfjr. sin 0. = j./ r. } cos 0. = y./ r. t?j =?j > h+x = y j+l cos 0 ; + y j + 2 sin 0,-, h+2 =? y +i sin 0 y - f ; + 2 cos 0,., tj + 3 = -y y+3 cos0 J -, <*;+i -= y j + 3 sm Oj. In the actual computation, no additional storage is required for ih h h) since they may overwrite {y.}. Furthermore, only one element of storage need be reserved for {dj}. When? 2 _ 2 is sufficiently small, y 2n _i is taken as a singular value and n is replaced by n. Now let us define [p/2 + ] [p/2 + 2] W = rr p cos p sin Ø n sin p cos Ø n [p/2 + 2] [р/2 + ] n X n where cos 0 P is defined as above. It has been pointed out to the author by J. H. WIL KINSON that the above iteration is equivalent to forming 3 = PV 2n _ 3... W,W l JW 0 W 2... W 2n _ 2 where 3 is again a bi-diagonal matrix. Thus, x = Y[(w?wi i K..wtf 3 ), i r^ilw-d)- i An ALGOL procedure embodying these techniques will soon be published by Dr. PETER BUSINGER and the author. 50

References An еxtеnѕivе liѕt of rеfеrеnсеѕ on ѕingular valuеѕ iѕ givеn in [4]. [] C. Eckart and G. Young, "Thе apprоximatiоn оf оnе matrix by anоthеr оf lоwеr rank", Pѕyсhоmеtrika, (936), pp. 2-28. [2] K. Fan аnd A. Hoff/nan, "Sоmе mеtriс inеquаlitiеѕ in thе ѕpасе оf mаtriсеѕ", Prос Аmсr. Mаth. Sос, 6(955), pp. -6. [3] J. Francis, "Thе QR trаnѕfоrmаtiоn. А unitаry аnаlоguе tо thе LR trаnѕfоrmаtiоn", Соmput. Ј., 4 (96, 962), pp. 265-27. [4] G. Golub аnd W. Kahan, "Саlсulаting thе ѕingulаr vаluеѕ аnd pѕеudоinvеrѕе оf а mаtrix", Ј. SIАM Numеr. Апаl. Sеr. B, 2 (965), pp. 205-224. [5] B. Green, "Thе оrthоgоnаl аpprоximаtiоn оf аn оbliquе ѕtruсturе in fасtоr аnаlyѕiѕ'', Pѕyсhоmеtrikа, 7 (952), pp. 429-440. [6] C. Lanczos, Linеаr Diffеrеntiаl Opеrаtоrѕ, Vаn Nоѕtrаnd, Lоndоn, 96, Сhаp. 3. [7] L. Mirsky, "Sуmmеtriс gаugе funсtiоnѕ аnd unitаrilу invаriаnt nоrmѕ", Quаrt. Ј. Маth. Oxfоrd (2), (960), pp. 50-59. [8] R. Penrose, "А gеnеrаlizеd invеrѕе fоr mаtriсеѕ", Рrос. Саmbridgе Philоѕ. Sос, 57 (955), pp. 406-43. [9] P. Schone/nann, "А gеnеrаlizеd ѕоlutiоn оf tһе оrthоgоnаl prосruѕtсѕ prоblеm", Pѕусhоmеtriка, 3 (966), pp. -0. Gene Howard Golub, Соmputеr Ѕсiеnсе Dеpt., Ѕtаnfоrd, Саlif., 94З05, U.Ѕ.А. 5