AMS subject classifications. 15A06, 65F10, 65F20, 65F22, 65F25, 65F35, 65F50, 93E24

Size: px
Start display at page:

Download "AMS subject classifications. 15A06, 65F10, 65F20, 65F22, 65F25, 65F35, 65F50, 93E24"

Transcription

1 : AN IERAIVE ALGORIHM FOR SPARSE LEAS-SQUARES PROBLEMS DAVID CHIN-LUNG FONG AND MICHAEL SAUNDERS Abstract. An iterative method is presented for solving linear systems Ax = b and leastsquares problems min Ax b 2, with A being sparse or a fast linear operator. is based on the Golub-Kahan bidiagonalization process. It is analytically equivalent to the MINRES method applied to the normal equation A Ax = A b, so that the quantities A r are monotonically decreasing (where r = b Ax is the residual for the current iterate x. In practice we observe that r also decreases monotonically. Compared to, for which only r is monotonic, it is safer to terminate early. Improvements for the new iterative method in the presence of extra available memory are also explored. Key words. least-squares problem, sparse matrix,, MINRES, Krylov subspace method, Golub-Kahan process, conjugate-gradient method, minimum-residual method, iterative method AMS subject classifications. 15A6, 65F1, 65F2, 65F22, 65F25, 65F35, 65F5, 93E24 DOI. xxx/xxxxxxxxx 1. Introduction. We present a numerical method called for computing a solution x to the following problems: Unsymmetric equations: solve Ax = b Linear least squares (LS: minimize Ax b 2 ( Regularized least squares: minimize A b ( x λi 2 where A R m n, b R m, and λ. he matrix A is used as an operator for which products of the form Av and A u can be computed for various v and u. hus A is normally large and sparse and need not be explicitly stored. is similar in style to the well nown method [15, 16] in being based on the Golub-Kahan bidiagonalization of A [5]. is equivalent to the conjugategradient (CG method applied to the normal equation (A A+λ 2 Ix = A b. It has the property of reducing r monotonically, where r = b Ax is the residual for the approximate solution x. (For simplicity, we are letting λ =. In contrast, is equivalent to MINRES [14] applied to the normal equation, so that the quantities A r are monotonically decreasing. In practice we observe that r also decreases monotonically, and is never very far behind the corresponding value for. Hence, although and ultimately converge to similar points, it is safer to use in situations where the solver must be terminated early. Stopping conditions are typically based on bacward error: the norm of some perturbation to A for which the current iterate x solves the perturbed problem exactly. Experiments on many sparse LS test problems show that for, a certain cheaply computable bacward error for each x is close to the optimal (smallest possible bacward error. his is an unexpected but highly desirable advantage. Version of August 16, 211. echnical Report SOL 21-2, revised March 14, for Copper Mountain Special Issue 21 icme, Stanford University (clfong@stanford.edu. Partially supported by a Stanford Graduate Fellowship. Systems Optimization Laboratory, Department of Management Science and Engineering, Stanford University, CA (saunders@stanford.edu. Partially supported by Office of Naval Research grant N and by the U.S. Army Research Laboratory, through the Army High Performance Computing Research Center, Cooperative Agreement W911NF

2 2 DAVID FONG AND MICHAEL SAUNDERS 1.1. Overview. Section 2 introduces the Golub-Kahan process and derives the basic algorithm with λ =. Section 3 derives various norms and stopping criteria. Section 4 discusses singular systems and complexity. Section 5 derives the algorithm with λ. Section 6 describes bacward error estimates. Section 7 gives numerical results on a range of overdetermined and square systems. Section 8 summarizes our findings, and Appendix A proves one of the main lemmas Notation. Matrices are denoted by A, B,..., vectors by v, w,..., and scalars by α, β,.... wo exceptions are c and s, which denote the significant components of a plane rotation matrix, with c 2 + s 2 = 1. For a vector v, v always denotes the 2-norm of v. For a matrix A, A usually denotes the Frobenius norm, and the condition number of a matrix A is defined by cond(a = A A +, where A + denotes the pseudoinverse of A. Vectors e 1 and e denote columns of an identity matrix. Items lie β and β are about to change to something similar lie β. 2. Derivation of. We begin with the Golub-Kahan process [5], an iterative procedure for transforming ( b A to upper-bidiagonal form ( β 1 e 1 B he Golub-Kahan process. 1. Set β 1 u 1 = b (shorthand for β 1 = b, u 1 = b/β 1 and α 1 v 1 = A u For = 1, 2,..., set After steps, we have β +1 u +1 = Av α u and α +1 v +1 = A u +1 β +1 v. (2.1 AV = U +1 B and A U +1 = V +1 L +1, where we define V = ( v 1 v 2... v, U = ( u 1 u 2... u, and Now consider α 1 β 2 α 2 B = β, α L +1 = ( B α +1 e +1. β +1 A AV = A U +1 B = V +1 L +1B = V +1 ( B α +1 e +1 = V +1 ( B B α +1 β +1 e his is equivalent to what would be generated by the symmetric Lanczos process with matrix A A and starting vector A b. (For this reason we define β α β below Using Golub-Kahan to solve the normal equation. Krylov subspace methods for solving linear equations form solution estimates x = V y for some y, where the columns of V are an expanding set of theoretically independent vectors. (In this case, V and also U are theoretically orthonormal. For the equation A Ax = A b, any solution x has the property of minimizing r, where r = b Ax is the corresponding residual vector. hus, in the development of it was natural to choose y to minimize r at each stage. Since r = b AV y = β 1 u 1 U +1 B y = U +1 (β 1 e 1 B y, B.

3 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 3 where U +1 is theoretically orthonormal, the subproblem min y β 1 e 1 B y easily arose. In contrast, for we wish to minimize A r. Let β α β for all. Since A r = A b A Ax = β 1 α 1 v 1 A AV y, we have ( A r = β B 1 v 1 V B +1 α +1 β +1 e and we are led to the subproblem min y A r = min y y = V +1 ( β 1 e 1 ( B B β +1 e ( β B 1 e 1 B β +1 e y, y. (2.2 Efficient solution of this LS subproblem is the heart of algorithm wo QR factorizations. As in, we form the QR factorization ρ 1 θ 2 ( R. Q +1 B =, R = ρ θ. (2.3 ρ If we define t = R y and solve R q = β +1 e, we have q = ( β +1 /ρ e = ϕ e with ρ = (R and ϕ β +1 /ρ. hen we perform a second QR factorization ρ 1 θ2 ( ( R Q β1 e 1 R z +1 ϕ e =., R = ρ.. 2 ζ+1... θ. (2.4 ρ Combining what we have with (2.2 gives ( ( min A r = min y y β R 1 e 1 R q R y = min t β R 1 e 1 ϕ e ( ( = min z R t ζ +1 he subproblem is solved by choosing t from R t = z. t t. ( Recurrence for x. Let W and W be computed by forward substitution from R W = V and R W = W. hen from x = V y, R y = t, and R t = z, we have x and x = W R y = W t = W R t = W z = x + ζ w Recurrence for W and W. If we write V = ( v 1 v 2 v, W = ( w 1 w 2 w, W = ( w 1 w 2 w, z = ( ζ 1 ζ 2 ζ, an important fact is that when increases to + 1, all quantities remain the same except for one additional term.

4 4 DAVID FONG AND MICHAEL SAUNDERS he first QR factorization proceeds as follows. At iteration we construct a plane rotation operating on rows l and l + 1: I l P l = c l s l s l c l. I l Now if Q +1 = P... P 2 P 1, we have ( B α Q +1 B +1 = Q +1 e = β +2 R θ +1 e ᾱ +1, β +2 Q +2 B +1 = P +1 R θ +1 e ᾱ +1 = R θ +1 e ρ +1 β +2 and we see that θ +1 = s α +1 = (β +1 /ρ α +1 = β +1 /ρ = ϕ. herefore we can write θ +1 instead of ϕ. For the second QR factorization, if Q +1 = P... P 2 P1 we now that and so ( R Q +1 θ +1 e = ( R R Q θ +2 e = P ( R, θ+1 e R c ρ +1 = θ +2 θ+1 e ρ +1. (2.6 By considering the last row of the matrix equation R+1 W +1 = V +1 and the last row of R +1 W +1 = W +1 we obtain equations that define w +1 and w +1 : θ +1 w + ρ +1 w +1 = v +1, θ +1 w + ρ +1 w +1 = w he two rotations. o summarize, the rotations P and P have the following effects on our computation: ( ( ( c s ᾱ ρ θ = +1 s c β +1 α +1 ᾱ +1 ( c s ( c ρ ζ ( ρ θ+1 ζ =. s c θ +1 ρ +1 c ρ +1 ζ Speeding up forward substitution. he forward substitutions for computing w and w can be made more efficient if we define h = ρ w and h = ρ ρ w. We then obtain the updates described in part 6 of the pseudo-code below Algorithm. he following summarizes the main steps of algorithm for solving Ax b, excluding the norms and stopping rules developed later. 1. (Initialize β 1 u 1 = b α 1 v 1 = A u 1 ᾱ 1 = α 1 ζ1 = α 1 β 1 ρ = 1 ρ = 1 c = 1 s = h 1 = v 1 h = x =

5 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 5 2. For = 1, 2, 3..., repeat steps (Continue the bidiagonalization β +1 u +1 = Av α u, 4. (Construct and apply rotation P 1 α +1 v +1 = A u +1 β +1 v ρ = ( ᾱ 2 + β c = ᾱ /ρ s = β +1 /ρ (2.7 θ +1 = s α +1 ᾱ +1 = c α +1 ( (Construct and apply rotation P 6. (Update h, h x θ = s ρ ρ = ( ( c ρ 2 + θ c = c ρ / ρ s = θ +1 / ρ (2.9 ζ = c ζ ζ+1 = s ζ (2.1 h = h ( θ ρ /(ρ ρ h x = x + (ζ /(ρ ρ h h +1 = v +1 (θ +1 /ρ h 3. Norms and stopping rules. Here we derive r, A r, x and estimates of A and cond(a for use within stopping rules. All quantities require O(1 computation at each iteration Computing r. We transform R to upper-bidiagonal form using a third QR factorization: R = Q R with Q = P... P 1. his amounts to one additional rotation per iteration. Now let ( t = Q Q t, b = Q +1 e 1 β 1. (3.1 1 hen r = b Ax = β 1 u 1 AV y = U +1 e 1 β 1 U +1 B y gives ( ( ( r = U +1 (e 1 β 1 Q R +1 y = U +1 e 1 β 1 Q t +1 ( ( Q = U +1 (Q Q +1 b Q t +1 1 ( ( Q = U +1 Q ( t +1 b. 1 herefore, assuming orthogonality of U +1, we have r = ( t b. (3.2 he vectors b and t can be written in the form b = ( β1 β β β+1 t = ( τ 1 τ τ. (3.3 he vector t can be computed by forward substitution from R t = z. Lemma 3.1. In (3.2 (3.3, β i = τ i for i = 1,..., 1. Proof. Appendix A proves the lemma by induction. Using this lemma we can estimate r from just the last two elements of b and the last element of t, as shown in (3.6 below. 1

6 6 DAVID FONG AND MICHAEL SAUNDERS Pseudo-code for computing r. he following summarizes how r may be obtained from quantities arising from the first and third QR factorizations. 1. (Initialize β 1 = β 1 β = ρ = 1 τ = θ = ζ = 2. For the th iteration, repeat steps (Apply rotation P ˆβ = c β β+1 = s β ( (If 2, construct and apply rotation P 1 2 ρ = ( ρ 2 + θ 2 c = ρ / ρ s = θ / ρ (3.5 θ = s ρ ρ = c ρ β = c β + s ˆβ β = s β + c ˆβ 5. (Update t by forward substitution 6. (Form r τ = (ζ θ τ / ρ τ = (ζ θ τ / ρ γ = ( β τ 2 + β 2 +1, r = γ ( Computing A r. From (2.5 we have A r = ζ +1, which by (2.1 is monotonically decreasing Computing x. From section 2.4 we have x = V R R z. From the third QR factorization Q R = R in section 3.1 and a fourth QR factorization ˆQ ( Q R = ˆR we can write x = V R R z = V R R R Q z = V R Q Q R ˆQ ẑ = V ˆQ ẑ, where z and ẑ are defined by forward substitutions R z = z and ˆR ẑ = z. Assuming orthogonality of V we arrive at the estimate x = ẑ. Since only the last diagonal of R and the bottom 2 2 part of ˆR change each iteration, this estimate of x can again be updated cheaply. he pseudo-code, omitted here, can be derived as in section Experimentally we have observed that for every iteration, x > x is either true or very nearly true Estimates of A and cond(a. It is nown that the singular values of B are interlaced by those of A and are bounded above and below by the largest and smallest nonzero singular values of A [15]. herefore we can estimate A and cond(a by B and cond(b respectively. Considering the Frobenius norm of B, we have the recurrence relation B +1 2 F = B 2 F + α2 + β2 +1. From (2.3 (2.4 and (2.6, we can show that the following QLP factorization [22] holds: ( R Q +1 B Q = θ e c ρ (the same as R except for the last diagonal. Since the singular values of B are approximated by the diagonal elements of that lower-bidiagonal matrix [22], and since the diagonals are all positive, we can estimate cond(a by the ratio of the largest and smallest values in { ρ 1,..., ρ, c ρ }. hose values can be updated cheaply.

7 : AN IERAIVE ALGORIHM FOR LEAS SQUARES Stopping criteria. With exact arithmetic, the Golub-Kahan process terminates when either α +1 = or β +1 =. For certain data b, this could happen in practice when is small (but is unliely later. We show that will have solved the problem at that point and should therefore terminate. When α +1 =, with the expression of A r from section 3.2, we have A r = ζ +1 = s ζ = θ +1 ρ ζ = s α +1 ρ ζ =, where (2.1, (2.9, (2.8 are used. hus, a least-squares solution has been obtained. When β +1 =, we have s = β +1 ρ =. (from (2.7 (3.7 β +1 = s β =. (from (3.4, (3.7 (3.8 β = c ( β s ( s ( c +1 β 1 (from (A.6 = c (from (3.7 = ρ ρ β (from (3.5 = ρ ρ τ (from Lemma 3.1 = τ. (from (A.2, (A.3 (3.9 By (3.9, (3.8, and (3.6 we conclude that r =. It follows that Ax = b Practical stopping criteria. For we use the same stopping rules as [15], involving dimensionless quantities AOL, BOL, CONLIM: S1: Stop if r BOL b + AOL A x S2: Stop if A r AOL A r S3: Stop if cond(a CONLIM S1 applies to consistent systems, allowing for uncertainty in A and b [9, heorem 7.1]. S2 applies to inconsistent systems and comes from Stewart s bacward error estimate E 2 assuming uncertainty in A; see section 6.1. S3 applies to any system. 4. Characteristics of the solution on singular systems. If A does not have full column ran, the normal equation A Ax = A b is singular but consistent. We show that and both give the minimum-norm LS solution. hat is, they both solve the optimization problem min x 2 such that A Ax = A b. Let N(A and R(A denote the nullspace and range of a matrix A. Lemma 4.1. If A R m n and p R n satisfy A Ap =, then p N(A. Proof. A Ap = p A Ap = (Ap Ap = Ap =. heorem 4.2. returns the minimum-norm solution. Proof. he final solution satisfies A Ax = A b, and any other solution x satisfies A A x = A b. With p = x x, the difference between the two normal equations gives A Ap =, so that Ap = by Lemma 4.1. From α 1 v 1 = A u 1 and α +1 v +1 = A u +1 β +1 v (2.1, we have v 1,..., v R(A. With Ap =, this implies p V =, so that x 2 2 x 2 2 = x + p 2 2 x 2 2 = p p + 2p x = p p + 2p V y = p p.

8 8 DAVID FONG AND MICHAEL SAUNDERS Corollary 4.3. returns the minimum-norm solution. Proof. At convergence, α +1 = or β +1 =. hus β +1 = α +1 β +1 =, which means equation (2.2 becomes min β 1 e 1 B B y and hence B B y = β 1 e 1, since B has full ran. his is the normal equation for min B y β 1 e 1, the same LS subproblem solved by. We conclude that at convergence, y = y x = V y = V y = x, and heorem 4.2 applies. and thus 4.1. Complexity. We compare the storage requirement and computational complexity for and on Ax b and MINRES on the normal equation A Ax = A b. In able 4.1, we list the vector storage needed (excluding storage for A and b. Recall that A is m n and for LS systems m may be considerably larger than n. Av denotes the woring storage for matrix-vector products. Wor represents the number of floating-point multiplications required at each iteration. able 4.1 Storage and computational requirements for various least-squares methods Storage Wor m n m n Av, u x, v, h, h 3 6 Av, u x, v, w 3 5 MINRES on A Ax = A b Av x, v 1, v 2, w 1, w 2, w Regularized least squares. In this section, we extend to the regularized LS problem: ( ( min A b x. (5.1 λi 2 ( ( A b If Ā = and r λi = Āx, then ( ( B Ā r = A r λ 2 ( x = V +1 β 1 e 1 B β +1 e y λ 2 y ( ( R = V +1 β 1 e 1 R β +1 e y and the rest of the main algorithm follows the same as in the unregularized case. In the last equality, R is defined by the QR factorization ( ( B R Q 2+1 =, Q λi 2+1 P ˆP... P 2 ˆP2 P 1 ˆP1, where ˆP l is a rotation operating on rows l and l he effects of ˆP 1 and P 1 are illustrated here: α 1 ˆα 1 ˆα 1 ρ 1 θ 2 β 2 α 2 ˆP 1 β 3 λ = β 2 α 2 β 3, P β 2 α 2 1 β 3 = ᾱ 2 β 3. λ λ λ λ

9 : AN IERAIVE ALGORIHM FOR LEAS SQUARES Effects on r. he introduction of regularization changes the residual norm as follows: ( ( ( ( ( ( b A u1 AV u1 U+1 B r = x λi = β 1 y λv = β 1 y λv ( ( ( U+1 B = e V 1 β 1 y λi ( ( ( U+1 = e V 1 β 1 Q R 2+1 y ( ( ( U+1 = e V 1 β 1 Q t 2+1 ( ( ( U+1 Q = Q ( t V 2+1 b 1 ( with b Q = Q 2+1 e 1 β 1, where we adopt the notation 1 b = ( β1 β β β+1 ˇβ1 ˇβ. We conclude that r 2 = ˇβ ˇβ 2 +( β τ 2 + β he effect of regularization on the rotations is summarized as ( ĉ ŝ (ᾱ β (ˆα β = ŝ ĉ λ ˇβ ( ( ( c s ˆα β ρ θ = +1 ˆβ. s c β +1 α +1 β+1 ᾱ Pseudo-code for regularized. he following summarizes algorithm for solving the regularized problem (5.1 with given λ. Our Matlab implementation is based on these steps. 1. (Initialize β 1 u 1 = b α 1 v 1 = A u 1 ᾱ 1 = α 1 ζ1 = α 1 β 1 ρ = 1 ρ = 1 c = 1 s = β1 = β 1 β = ρ = 1 τ = θ = ζ = d = h 1 = v 1 h = x = 2. For = 1, 2, 3,... repeat steps (Continue the bidiagonalization β +1 u +1 = Av α u α +1 v +1 = A u +1 β +1 v 4. (Construct rotation ˆP ˆα = ( ᾱ 2 + λ ĉ = ᾱ /ˆα ŝ = λ/ˆα 5. (Construct and apply rotation P ρ = (ˆα 2 + β c = ˆα /ρ s = β +1 /ρ θ +1 = s α +1 ᾱ +1 = c α +1

10 1 DAVID FONG AND MICHAEL SAUNDERS 6. (Construct and apply rotation P θ = s ρ ρ = ( ( c ρ 2 + θ c = c ρ / ρ ζ = c ζ s = θ +1 / ρ ζ+1 = s ζ 7. (Update h, x, h 8. (Apply rotation ˆP, P h = h ( θ ρ /(ρ ρ h x = x + (ζ /(ρ ρ h h +1 = v +1 (θ +1 /ρ h β = ĉ β ˇβ = ŝ β ˆβ = c β β+1 = s β 9. (If 2, construct and apply rotation P ρ = ( ρ 2 + θ c = ρ / ρ θ = s ρ β = c β + s ˆβ s = θ / ρ ρ = c ρ β = s β + c ˆβ 1. (Update t by forward substitution τ = (ζ θ τ / ρ τ = (ζ θ τ / ρ 11. (Compute r d = d + ˇβ 2 γ = d + ( β τ 2 + β 2 +1 r = γ 12. (Compute A r, x, estimate Ā, cond(ā, and test for termination A r = ζ +1 (section 3.2 Compute x (section 3.3 Estimate σ max (B, σ min (B and hence Ā, cond(ā (section 3.4 erminate if any of the stopping criteria are satisfied (section Bacward errors. For inconsistent problems with uncertainty in A (but not b, let x be any approximate solution. he normwise bacward error for x measures the perturbation to A that would mae x an exact LS solution: µ(x min E E s.t. (A + E (A + Ex = (A + E b. (6.1 It is nown to be the smallest singular value of a certain m (n + m matrix C; see Waldén et al. [25] and Higham [9, pp ]: [ ( ] µ(x = σ min (C, C A r x I rr r. 2 Since it is generally too expensive to evaluate µ(x, we need to find approximations.

11 : AN IERAIVE ALGORIHM FOR LEAS SQUARES Approximate bacward errors E 1 and E 2. In 1975, Stewart [2] discussed a particular bacward error estimate that we will call E 1. Let x and r = b A x be the exact LS solution and residual. Stewart showed that an approximate solution x with residual r = b Ax is the exact LS solution of the perturbed problem min b (A + E 1 x, where E 1 is the ran-one matrix E 1 = ex x 2, E 1 = e, e r r, (6.2 x with r 2 = r 2 + e 2. Soon after, Stewart [21] gave a further important result that can be used within any LS solver. he approximate x and a certain vector r = b (A + E 2 x are the exact solution and residual of the perturbed LS problem min b (A + E 2 x, where E 2 = rr A r 2, E 2 = A r, r = b Ax. (6.3 r and both compute E 2 for each iterate x because the current r and A r can be accurately estimated at almost no cost. An added feature is that for both solvers, r = b (A + E 2 x = r because E 2 x = (assuming orthogonality of V. hat is, x and r are theoretically exact for the perturbed LS problem (A + E 2 x b. Stopping rule S2 (section 3.6 requires E 2 AOL A. Hence the following property gives an advantage over for stopping early. heorem 6.1. E2 E 2. Proof. his follows from A r A r and r r Approximate optimal bacward error µ(x. Various authors have derived expressions for a quantity µ(x that has proved to be a very accurate approximation to µ(x in (6.1 when x is at least moderately close to the exact solution x. Grcar, Saunders, and Su [23, 7] show that µ(x can be obtained from a full-ran LS problem as follows: K = [ A r x I ], v = [ ] r, min Ky v, µ(x = Ky / x, (6.4 y and give the following Matlab script for computing the economy size sparse QR factorization K = QR and c Q v (for which c = Ky and thence µ(x: [m,n] = size(a; r = b - A*x; normx = norm(x; eta = norm(r/normx; p = colamd(a; K = [A(:,p; eta*speye(n]; v = [ r ; zeros(n,1]; [c,r] = qr(k,v,; mutilde = norm(c/normx; In our experiments we use this script to compute µ(x for each and iterate x. We refer to this as the optimal bacward error for x because it is provably very close to the true µ(x [6].

12 12 DAVID FONG AND MICHAEL SAUNDERS 6.3. Related wor. More precise stopping rules have been derived recently by Arioli and Gratton [1] and itley-péloquin et al. [3, 12, 24]. he rules allow for uncertainty in both A and b, and may prove to be useful for,, and least-squares methods in general. However, we would lie to emphasize that rule S2 already terminates significantly sooner than on most of our inconsistent test cases; see heorem 6.1, Fig. 7.2 (left, and Fig. 7.3 (top left. 7. Numerical results. For test examples, we have drawn from the University of Florida Sparse Matrix Collection (Davis [4]. We discuss overdetermined systems first, and then some square examples Least-squares problems. he LPnetlib group provides data for 138 linear programming problems of widely varying origin, structure, and size. he constraint matrix and objective function may be used to define a sparse LS problem min Ax b. Each example was downloaded in Matlab format, and a sparse matrix A and dense vector b were extracted from the data structure via A = (Problem.A and b = Problem.c (where denotes transpose. Five examples had b =, and a further six gave A b =. he remaining 127 problems had up to 243 rows, 1 columns, and 1.4M nonzeros in A. Diagonal scaling was applied to the columns of [ A b ] to give a scaled problem min Ax b in which the columns of A (and also b have unit 2-norm. and were run on each of the 127 scaled problems with stopping tolerance AOL = 1, generating sequences of approximate solutions {x } and {x }. he iteration indices are omitted below. he associated residual vectors are denoted by r without ambiguity, and x is the solution to the LS problem, or the minimum-norm solution to the LS problem if the system is singular. As expected, the optimal residual is nonzero in all cases. We record some general observations. 1. r is monotonic by design. r seems to be monotonic (no counterexamples were found and nearly as small as r for all iterations on almost all problems. Figure 7.1 shows a typical example and a rare case. 1 Name:lp greenbeb, Dim:5598x2392, nnz:317, id=631 1 Name:lp woodw, Dim:8418x198, nnz:37487, id= r.7 r Fig For most iterations, r appears to be monotonic and nearly as small as r. Left: A typical case (problem lp greenbeb. Right: A rare case (problem lp woodw. s residual norm is significantly larger than s during early iterations.

13 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 13 Name:lp pilot ja, Dim:2267x94, nnz:14977, id=657 Name:lp sc25, Dim:317x25, nnz:665, id=665 log(e2 log(e Fig For most iterations, E 2 appears to be monotonic (but E 2 is not. Left: A typical case (problem lp pilot ja. is liely to terminate much sooner than (see heorem 6.1. Right: Sole exception (problem lp sc25 at iterations he exception remains even if U and/or V are reorthogonalized. 2. x is nearly monotonic for and even more closely monotonic for. With r monotonic for and essentially so for, E 1 in (6.2 is liely to appear monotonic for both solvers. Although E 1 is not normally available for each iteration, it provides a benchmar for E E 2 is not monotonic, but E 2 appears monotonic almost always. Figure 7.2 shows a typical case. he sole exception for this observation is also shown. 4. Note that Benbow [2] has given numerical results comparing a generalized form of with application of MINRES to the corresponding normal equation. he curves in [2, Figure 3] show the irregular and smooth behavior of and MINRES respectively in terms of A r. hose curves are effectively a preview of the left-hand plots in Figure 7.2 (where serves as our more reliable implementation of MINRES. 5. E 1 E 2 often, but not so for. Some examples are shown on Figure 7.3 along with µ(x, the accurate estimate (6.4 of the optimal bacward error for each point x. 6. E2 µ(x almost always. Figure 7.4 shows a typical example and a rare case. In all such rare cases, E1 µ(x instead! 7. µ(x is not always monotonic. µ(x does seem to be monotonic. Figure 7.5 gives examples. 8. µ(x µ(x almost always. Figure 7.6 gives examples. 9. he errors x x and x x seem to decrease monotonically, with the error typically smaller than for. Figure 7.7 gives examples. his is one property for which seems more desirable (and it has been suggested [17] that for LS problems, could be terminated when rule S2 would terminate.

14 14 DAVID FONG AND MICHAEL SAUNDERS Name:lp cre a, Dim:7248x3516, nnz:18168, id=69 Name:lp pilot, Dim:486x1441, nnz:44375, id=654 E2 E1 Optimal E2 E1 Optimal log(bacward Error for log(bacward Error for Name:lp cre a, Dim:7248x3516, nnz:18168, id=69 E2 E1 Optimal Name:lp pilot, Dim:486x1441, nnz:44375, id=654 E2 E1 Optimal log(bacward Error for log(bacward Error for Fig E 1, E 2, and µ(x for (top figures and (bottom figures. op left: A typical case. E 1 is close to the optimal bacward error, but the computable E 2 is not. op right: A rare case in which E 2 is close to optimal. Bottom left: E 1 and E 2 are often both close to the optimal bacward error. Bottom right: E 1 is far from optimal, but the computable E 2 is almost always close (too close to distinguish in the plot!. Problems lp cre a (left and lp pilot (right. Name:lp en 11, Dim:21349x14694, nnz:4958, id=638 E2 E1 Optimal Name:lp ship12l, Dim:5533x1151, nnz:16276, id=688 E2 E1 Optimal log(bacward Error for log(bacward Error for Fig Again, E 2 µ(x almost always (the computable bacward error estimate is essentially optimal. Left: A typical case (problem lp en 11. Right: A rare case (problem lp ship12l. Here, E 1 µ(x!

15 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 15.5 Name:lp maros, Dim:1966x846, nnz:1137, id=642 Name:lp cre c, Dim:6411x368, nnz:15977, id=611 log(optimal Bacward Error log(optimal Bacward Error Fig µ(x seems to be always monotonic, but µ(x is usually not. Left: A typical case for both and (problem lp maros. Right: A rare case for, typical for (problem lp cre c. Name:lp pilot, Dim:486x1441, nnz:44375, id=654 Name:lp standgub, Dim:1383x361, nnz:3338, id=693 log(optimal Bacward Error log(optimal Bacward Error Fig µ(x µ(x almost always. Left: A typical case (problem lp pilot. Right: A rare case (problem lp standgub. 2 1 Name:lp ship12l, Dim:5533x1151, nnz:16276, id=688 1 Name:lp pds 2, Dim:7716x2953, nnz:16571, id=649 log x x * log x x * Fig he errors x x and x x seem to decrease monotonically, with s errors smaller than for. Left: A nonsingular LS system (problem lp ship12l. Right: A singular system (problem lp pds 2. and both converge to the minimum-norm LS solution.

16 16 DAVID FONG AND MICHAEL SAUNDERS Name:Hamm.hcircuit, Dim:15676x15676, nnz:51372, id=542 Name:Hamm.hcircuit, Dim:15676x15676, nnz:51372, id=542 log r log A r x 1 4 Name:IBM EDA.trans5, Dim:116835x116835, nnz:7498, id= x 1 4 Name:IBM EDA.trans5, Dim:116835x116835, nnz:7498, id=1324 log r log A r Fig and solving two square nonsingular systems Ax = b: problems Hamm/hcircuit (top and IBM EDA/trans5 (bottom. Left: log 1 r for both solvers, with prolonged plateaus for. Right: log 1 A r (preferable for Square systems. Since and are applicable to consistent systems, it is of interest to compare them on an unbiased test set. We used the search facility of Davis [4] to select a set of square real linear systems Ax = b. With index = UFget, the criteria ids = find(index.nrows > 1 & index.nrows < 2 &... index.nrows == index.ncols & index.isreal == 1 &... index.posdef == & index.numerical_symmetry < 1; returned a list of 42 examples. esting isfield(ufget(id, b left 26 cases for which [ ] b was supplied. For each, diagonal scaling was first applied to the rows of A b and then to its columns to give a scaled problem Ax = b in which the columns of [ A b ] have unit 2-norm. In spite of the scaling, most examples required more than n iterations of or to reduce r satisfactorily (rule S1 in section 3.6 with AOL = BOL = 1. o simulate better preconditioning, we chose two cases that required about n/5 and n/1 iterations. Figure 7.8 (left shows both solvers reducing r monotonically but with plateaus that are prolonged for. With loose stopping tolerances, could terminate somewhat sooner. Figure 7.8 (right shows A r for each solver. he plateaus for correspond to gaining ground with r, but falling significantly bacward by the A r measure.

17 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 17 Name:lp ship12l, Dim:5533x1151, nnz:16276, id=688 Name:lpi gran, Dim:2525x2658, nnz:2111, id=717 NoOrtho OrthoU OrthoV OrthoUV NoOrtho OrthoU OrthoV OrthoUV log(e2 log(e Fig with and without reorthogonalization of V and/or U. Left: An easy case (problem lp ship12l. Right: A helpful case (problem lp gran Reorthogonalization. It is well nown that Krylov-subspace methods can tae arbitrarily many iterations because of loss of orthogonality. For the Golub-Kahan bidiagonalization, we have two sets of vectors U and V. As an experiment, we implemented the following options in : 1. No reorthogonalization. 2. Reorthogonalize V (that is, reorthogonalize v with respect to V. 3. Reorthogonalize U (that is, reorthogonalize u with respect to U. 4. Both 2 and 3. Each option was tested on all of the over-determined test problems with fewer than 16K nonzeros. Figure 7.9 shows an easy case in which all options converge equally well (convergence before significant loss of orthogonality, and an extreme case in which reorthogonalization maes a large difference. Unexpectedly, options 2, 3, and 4 proved to be indistinguishable in all cases. o loo closer, we forced to tae n iterations. Option 2 (with V orthonormal to machine precision ɛ was found to be eeping U orthonormal to at least O( ɛ. Option 3 (with U orthonormal was not quite as effective but it ept V orthonormal to at least O( ɛ up to the point where would terminate when AOL = ɛ. Note that for square or rectangular A with exact arithmetic, is equivalent to MINRES on the normal equation (and hence to the conjugate-residual method [11] and GMRES [19] on the same equation. Reorthogonalization maes the equivalence essentially true in practice. We now focus on reorthogonalizing V but not U. Other authors have presented numerical results involving reorthogonalization. For example, on some randomly generated LS problems of increasing condition number, Hayami et al. [8] compare their BA-GMRES method with an implementation of CGLS (equivalent to [15] in which V is reorthogonalized, and find that the methods require essentially the same number of iterations. he preconditioner chosen for BA- GMRES made that method equivalent to GMRES on A Ax = A b. hus, GMRES without reorthogonalization was seen to converge essentially as well as CGLS or with reorthogonalization of V (option 2 above. his coincides with the analysis by Paige et al. [13], who conclude that MGS-GMRES does not need reorthogonalization of the Arnoldi vectors V.

18 18 DAVID FONG AND MICHAEL SAUNDERS Name:lp maros, Dim:1966x846, nnz:1137, id=642 Name:lp cre c, Dim:6411x368, nnz:15977, id=611 NoOrtho Restart5 Restart1 Restart5 FullOrtho NoOrtho Restart5 Restart1 Restart5 FullOrtho log(e2 log(e Fig with reorthogonalized V and restarting. Restart(l with l = 5, 1, 5 is slower than standard with or without reorthogonalization. Problems lp maros and lp cre c. Name:lp fit1p, Dim:1677x627, nnz:9868, id=625 NoOrtho Local5 Local1 Local5 FullOrtho 1 Name:lp bnl2, Dim:4486x2324, nnz:14996, id=65 NoOrtho Local5 Local1 Local5 FullOrtho log(e2 log(e Fig with local reorthogonalization of V. Local(l with l = 5, 1, 5 illustrates reduced iterations as l increases. Problems lp fit1p and lp bnl Restarting. o conserve storage, a simple approach is to restart the algorithm every l steps, as with GMRES(l [19]. Figure 7.1 shows that restarting even with full reorthogonalization (of V may lead to stagnation. In general, convergence with restarting is much slower than without reorthogonalization Local reorthogonalization. Here we reorthogonalize each new v with respect to the previous l vectors, where l is a specified parameter. Figure 7.11 shows that l = 5 has little effect, but partial speedup was achieved with l = 1 and 5 in the two chosen cases. here is evidence of a useful storage-time tradeoff. he potential speedup depends strongly on the computational cost of Av and A u Partial reorthogonalization. Larsen [18] uses partial reorthogonalization of both V and U within his PROPACK software for computing a set of singular values and vectors for a sparse rectangular matrix A. Similar techniques might prove helpful within. We leave this for future research.

19 : AN IERAIVE ALGORIHM FOR LEAS SQUARES Summary. We have presented, an iterative algorithm for square or rectangular systems, along with details of its implementation and experimental results to suggest that it has advantages over the widely adopted algorithm. As in, theoretical and practical stopping criteria are provided for solving Ax = b and min Ax b with optional ihonov regularization, using estimates of r, A r, x, A and cond(a that are cheaply computable. For LS problems, the Stewart bacward error estimate E 2 (6.3 seems experimentally to be very close to the optimal bacward error µ(x at each iterate x (section 6.2. his often allows to terminate significantly sooner than. Experiments with full reorthogonalization have shown that the Golub-Kahan process retains high accuracy if the columns of either V or U are reorthogonalized. here is no need to reorthogonalize both. his discovery could be helpful for other uses of the Golub-Kahan process. Matlab, Python, and Fortran 9 implementations of are available from [1]. hey all allow local reorthogonalization of V. Acnowledgements. We are grateful to Chris Paige for his helpful comments on reorthogonalization and other aspects of this wor. We are also grateful to two referees for their extremely helpful and perceptive reviews. Further thans go to Martin van Gijzen and Mie Botchev for their help with testing on square systems arising from convection-diffusion problems, and to Victor Pereyra for proposing that be used to terminate if a smaller final error x x is important. Appendix A. Proof of Lemma 3.1. he effects of the rotations P and P can be summarized as ( ( ( ρ 1 θ2 c s β ˆβ =, R = ρ θ, s c β +1 ( ( c s ρ β = ρ s c θ ρ ˆβ ρ β ( ρ θ β where β 1 = β 1, ρ 1 = ρ 1, β1 = ˆβ 1 and where c, s are defined in section 2.6. We define s ( = s 1... s and s ( = s 1... s. hen from (3.3 and (2.4 we have R t = z = ( I Q+1 e +1 β1. Expanding this and (3.1 gives c c 1 1 R t s 1 c 2 ( =. β s 1 c 2 Q 1, b = 1. β 1, ( +1 s ( c ( +1 s ( c ( +2 s ( and we see that τ 1 = ρ 1 c 1 β 1 (A.1 τ = ρ (( s ( c β1 θ τ, (A.2 τ = ρ ((+1 s ( c β1 θ τ. (A.3 β 1 = ˆβ 1 = c 1 β 1 (A.4 β = s β + c ( s ( c β 1 (A.5 β = c β + s ( s ( c +1 β 1. (A.6

20 2 DAVID FONG AND MICHAEL SAUNDERS We want to show by induction that τ i = β i for all i. When i = 1, β 1 = c 1 c 1 β 1 s 1 s 1 c 2 β 1 = β 1 ρ 1 (c 1 ρ 1 θ 2 s 1 c 2 = β 1 ρ 1 α 1 ρ 1 ρ 2 1 ρ 1 = β 1 ρ 1 ρ 1 ρ 1 = β 1 ρ 1 c 1 = τ 1 where the third equality follows from the two lines below: c 1 ρ 1 θ 2 s 1 c 2 = c 1 ρ 1 θ 2 s 1 c 1 α 2 ρ 2 = ρ 1 θ 2 s 1 α 2 ρ 2 = α 1 ρ 1 ( ρ 1 1 ρ 2 θ2 s 1 α 2 ρ 1 1 ρ 2 θ2 s 1 α 2 = ρ 1 1 ρ 2 ( s 1 ρ 2 θ 2 = ρ 1 θ 2 ρ 1 θ 2 = ρ2 1 θ 2 2 ρ 1 = ρ2 1 + θ 2 2 θ 2 2 ρ 1. Suppose τ = β. We consider the expression s ( c ρ c2 ρ 2 β 1 = c ρ (s ( c c ρ β 1 ρ = c θ 2 θ α 1 ρ 1 ρ ρ 1 ρ ρ 1 ρ ρ β 1 = c θ 2 ρ 1 θ ρ β1 = c s 1 s β1 = c s ( β1. (A.7 Applying the induction hypothesis on τ = ρ (( +1 s ( c β1 θ τ gives τ = ρ (( +1 s ( c β1 θ ( c β + s ( s ( c β 1 = ρ θ c β + ( +1 ρ ( s ( c β1 θ s s ( c β 1 = ρ ( ρ s c β + ( +1 ρ ( s( β 1 ρ c c θ +1 s c +1 = c s β + ( +1 s ( β 1 ( c c c s s c +1 = c ( s β + c ( +1 s ( c β 1 + s ( +1 s ( c +1 β 1 = c β + s ( +1 s ( c +1 β 1 = β with the second equality obtained by the induction hypothesis, and the fourth from s ( c β1 θ s s ( c β 1 = s ( c ρ c2 ρ 2 β 1 ( s ρ s s ( c β 1 = s ( c 2 β 1 ( c ρ ρ 2 s 2 ρ 2 = s ( β 1 ( ρ c c θ +1 s c +1, where the first equality follows from (A.7 and the last from c 2 ρ 2 s 2 ρ 2 = ( ρ 2 θ+1 2 s 2 ρ 2 = ρ 2 (1 s 2 θ+1 2 = ρ 2 c 2 θ+1, 2 c ρ 2 ρ c 2 = ρ c 2 c = ρ c c, c θ+1 2 = θ +1 θ +1 c = θ +1ρ +1 c s α +1 = ρ ρ ρ ρ θ +1 s c herefore by induction, we now that τ i = β i for i = 1, 2,.... From (3.3, we see that at iteration, the first 1 elements of b and t are equal.

21 : AN IERAIVE ALGORIHM FOR LEAS SQUARES 21 REFERENCES [1] M. Arioli and S. Gratton, Least-squares problems, normal equations, and stopping criteria for the conjugate gradient method, echnical Report RAL-R-28-8, Rutherford Appleton Laboratory, Oxfordshire, UK, 28. [2] S. J. Benbow, Solving generalized least-squares problems with, SIAM J. Matrix Anal. Appl., 21 (1999, pp [3] X.-W. Chang, C. C. Paige, and D. itley-péloquin, Stopping criteria for the iterative solution of linear least squares problems, SIAM J. Matrix Anal. Appl., 31 (29, pp [4]. A. Davis, University of Florida Sparse Matrix Collection. research/sparse/matrices. [5] G. H. Golub and W. Kahan, Calculating the singular values and pseudo-inverse of a matrix, J. of the Society for Industrial and Applied Mathematics, Series B: Numerical Analysis, 2 (1965, pp [6] S. Gratton, P. Jiráne, and D. itley-péloquin, On the accuracy of the Karlson-Waldén estimate of the bacward error for linear least squares problems. Manuscript, Feb 22, 211. [7] J. F. Grcar, M. A. Saunders, and Z. Su, Estimates of optimal bacward perturbations for linear least squares problems, Report SOL 27-1, Department of Management Science and Engineering, Stanford University, Stanford, CA, pp. [8] K. Hayami, J.-F. Yin, and. Ito, GMRES methods for least squares problems, SIAM J. Matrix Anal. Appl., 31 (21, pp [9] N. J. Higham, Accuracy and Stability of Numerical Algorithms, SIAM, Philadelphia, second ed., 22. [1] software for linear systems and least squares. software.html. [11] D. G. Luenberger, he conjugate residual method for constrained minimization problems, SIAM J. Numer. Anal., 7 (197, pp [12] P. Jiráne and D. itley-péloquin, Estimating the bacward error in, SIAM J. Matrix Anal. Appl., 31 (21, pp [13] C. C. Paige, M. Rozlozni, and Z. Straos, Modified Gram-Schmidt (MGS, least squares, and bacward stability of MGS-GMRES, SIAM J. Matrix Anal. Appl., 28 (26, pp [14] C. C. Paige and M. A. Saunders, Solution of sparse indefinite systems of linear equations, SIAM J. Numer. Anal., 12 (1975, pp [15], : An algorithm for sparse linear equations and sparse least squares, ACM rans. Math. Softw., 8 (1982, pp [16], Algorithm 583; : Sparse linear equations and least-squares problems, ACM rans. Math. Softw., 8 (1982, pp [17] V. Pereyra, 21. Private communication. [18] PROPACK software for SVD of sparse matrices. [19] Y. Saad and M. H. Schultz, GMRES: a generalized minimum residual algorithm for solving nonsymmetric linear systems, SIAM J. Sci. and Statist. Comput., 7 (1986, pp [2] G. W. Stewart, An inverse perturbation theorem for the linear least squares problem, SIGNUM Newsletter, 1 (1975, pp [21], Research, development and LINPACK, in Mathematical Software III, J. R. Rice, ed., Academic Press, New Yor, 1977, pp [22], he QLP approximation to the singular value decomposition, SIAM J. Sci. Comput., 2 (1999, pp [23] Z. Su, Computational Methods for Least Squares Problems and Clinical rials, PhD thesis, SCCM, Stanford University, 25. [24] D. itley-péloquin, Bacward Perturbation Analysis of Least Squares Problems, PhD thesis, School of Computer Science, McGill University, 21. [25] B. Waldén, R. Karlson, and J.-G. Sun, Optimal bacward perturbation bounds for the linear least squares problem, Numerical Linear Algebra with Applications, 2 (1995, pp

LSMR: An iterative algorithm for least-squares problems

LSMR: An iterative algorithm for least-squares problems LSMR: An iterative algorithm for least-squares problems David Fong Michael Saunders Institute for Computational and Mathematical Engineering (icme) Stanford University Copper Mountain Conference on Iterative

More information

CG and MINRES: An empirical comparison

CG and MINRES: An empirical comparison School of Engineering CG and MINRES: An empirical comparison Prequel to LSQR and LSMR: Two least-squares solvers David Fong and Michael Saunders Institute for Computational and Mathematical Engineering

More information

CG versus MINRES on Positive Definite Systems

CG versus MINRES on Positive Definite Systems School of Engineering CG versus MINRES on Positive Definite Systems Michael Saunders SOL and ICME, Stanford University Joint work with David Fong Recent Advances on Optimization In honor of Philippe Toint

More information

CG VERSUS MINRES: AN EMPIRICAL COMPARISON

CG VERSUS MINRES: AN EMPIRICAL COMPARISON VERSUS : AN EMPIRICAL COMPARISON DAVID CHIN-LUNG FONG AND MICHAEL SAUNDERS Abstract. For iterative solution of symmetric systems Ax = b, the conjugate gradient method () is commonly used when A is positive

More information

CG VERSUS MINRES: AN EMPIRICAL COMPARISON

CG VERSUS MINRES: AN EMPIRICAL COMPARISON VERSUS : AN EMPIRICAL COMPARISON DAVID CHIN-LUNG FONG AND MICHAEL SAUNDERS Abstract. For iterative solution of symmetric systems Ax = b, the conjugate gradient method () is commonly used when A is positive

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Experiments with iterative computation of search directions within interior methods for constrained optimization

Experiments with iterative computation of search directions within interior methods for constrained optimization Experiments with iterative computation of search directions within interior methods for constrained optimization Santiago Akle and Michael Saunders Institute for Computational Mathematics and Engineering

More information

Generalized MINRES or Generalized LSQR?

Generalized MINRES or Generalized LSQR? Generalized MINRES or Generalized LSQR? Michael Saunders Systems Optimization Laboratory (SOL) Institute for Computational Mathematics and Engineering (ICME) Stanford University New Frontiers in Numerical

More information

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME MS&E 38 (CME 338 Large-Scale Numerical Optimization Course description Instructor: Michael Saunders Spring 28 Notes : Review The course teaches

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

F as a function of the approximate solution x

F as a function of the approximate solution x ESTIMATES OF OPTIMAL BACKWARD PERTURBATIONS FOR LINEAR LEAST SQUARES PROBLEMS JOSEPH F. GRCAR, MICHAEL A. SAUNDERS, AND ZHENG SU Abstract. Numerical tests are used to validate a practical estimate for

More information

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

On the loss of orthogonality in the Gram-Schmidt orthogonalization process CERFACS Technical Report No. TR/PA/03/25 Luc Giraud Julien Langou Miroslav Rozložník On the loss of orthogonality in the Gram-Schmidt orthogonalization process Abstract. In this paper we study numerical

More information

Math 310 Final Project Recycled subspaces in the LSMR Krylov method

Math 310 Final Project Recycled subspaces in the LSMR Krylov method Math 310 Final Project Recycled subspaces in the LSMR Krylov method Victor Minden November 25, 2013 1 Introduction In many applications, it is the case that we are interested not only in solving a single

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Solving large sparse Ax = b.

Solving large sparse Ax = b. Bob-05 p.1/69 Solving large sparse Ax = b. Stopping criteria, & backward stability of MGS-GMRES. Chris Paige (McGill University); Miroslav Rozložník & Zdeněk Strakoš (Academy of Sciences of the Czech Republic)..pdf

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems Abstract and Applied Analysis Article ID 237808 pages http://dxdoiorg/055/204/237808 Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite

More information

AN ITERATIVE METHOD WITH ERROR ESTIMATORS

AN ITERATIVE METHOD WITH ERROR ESTIMATORS AN ITERATIVE METHOD WITH ERROR ESTIMATORS D. CALVETTI, S. MORIGI, L. REICHEL, AND F. SGALLARI Abstract. Iterative methods for the solution of linear systems of equations produce a sequence of approximate

More information

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS HOMER F. WALKER Abstract. Recent results on residual smoothing are reviewed, and it is observed that certain of these are equivalent

More information

Journal of Computational and Applied Mathematics. New matrix iterative methods for constraint solutions of the matrix

Journal of Computational and Applied Mathematics. New matrix iterative methods for constraint solutions of the matrix Journal of Computational and Applied Mathematics 35 (010 76 735 Contents lists aailable at ScienceDirect Journal of Computational and Applied Mathematics journal homepage: www.elseier.com/locate/cam New

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition AM 205: lecture 8 Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition QR Factorization A matrix A R m n, m n, can be factorized

More information

REORTHOGONALIZATION FOR THE GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART I SINGULAR VALUES

REORTHOGONALIZATION FOR THE GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART I SINGULAR VALUES REORTHOGONALIZATION FOR THE GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART I SINGULAR VALUES JESSE L. BARLOW Department of Computer Science and Engineering, The Pennsylvania State University, University

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

THE RELATIONSHIPS BETWEEN CG, BFGS, AND TWO LIMITED-MEMORY ALGORITHMS

THE RELATIONSHIPS BETWEEN CG, BFGS, AND TWO LIMITED-MEMORY ALGORITHMS Furman University Electronic Journal of Undergraduate Mathematics Volume 12, 5 20, 2007 HE RELAIONSHIPS BEWEEN CG, BFGS, AND WO LIMIED-MEMORY ALGORIHMS ZHIWEI (ONY) QIN Abstract. For the solution of linear

More information

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR POLYNOMIAL PRECONDITIONED BICGSTAB AND IDR JENNIFER A. LOE AND RONALD B. MORGAN Abstract. Polynomial preconditioning is applied to the nonsymmetric Lanczos methods BiCGStab and IDR for solving large nonsymmetric

More information

Preconditioned conjugate gradient algorithms with column scaling

Preconditioned conjugate gradient algorithms with column scaling Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 28 Preconditioned conjugate gradient algorithms with column scaling R. Pytla Institute of Automatic Control and

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization

More information

Non-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error

Non-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error on-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error Divya Anand Subba and Murugesan Venatapathi* Supercomputer Education and Research

More information

REORTHOGONALIZATION FOR GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART II SINGULAR VECTORS

REORTHOGONALIZATION FOR GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART II SINGULAR VECTORS REORTHOGONALIZATION FOR GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART II SINGULAR VECTORS JESSE L. BARLOW Department of Computer Science and Engineering, The Pennsylvania State University, University

More information

Tikhonov Regularization of Large Symmetric Problems

Tikhonov Regularization of Large Symmetric Problems NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi

More information

Iterative Linear Solvers

Iterative Linear Solvers Chapter 10 Iterative Linear Solvers In the previous two chapters, we developed strategies for solving a new class of problems involving minimizing a function f ( x) with or without constraints on x. In

More information

Block Lanczos Tridiagonalization of Complex Symmetric Matrices

Block Lanczos Tridiagonalization of Complex Symmetric Matrices Block Lanczos Tridiagonalization of Complex Symmetric Matrices Sanzheng Qiao, Guohong Liu, Wei Xu Department of Computing and Software, McMaster University, Hamilton, Ontario L8S 4L7 ABSTRACT The classic

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated. Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces

More information

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725 Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Numerical behavior of inexact linear solvers

Numerical behavior of inexact linear solvers Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

LU Preconditioning for Overdetermined Sparse Least Squares Problems

LU Preconditioning for Overdetermined Sparse Least Squares Problems LU Preconditioning for Overdetermined Sparse Least Squares Problems Gary W. Howell 1 and Marc Baboulin 2 1 North Carolina State University, USA gwhowell@ncsu.edu 2 Université Paris-Sud and Inria, France

More information

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1. J. Appl. Math. & Computing Vol. 15(2004), No. 1, pp. 299-312 BILUS: A BLOCK VERSION OF ILUS FACTORIZATION DAVOD KHOJASTEH SALKUYEH AND FAEZEH TOUTOUNIAN Abstract. ILUS factorization has many desirable

More information

Total least squares. Gérard MEURANT. October, 2008

Total least squares. Gérard MEURANT. October, 2008 Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares

More information

A communication-avoiding thick-restart Lanczos method on a distributed-memory system

A communication-avoiding thick-restart Lanczos method on a distributed-memory system A communication-avoiding thick-restart Lanczos method on a distributed-memory system Ichitaro Yamazaki and Kesheng Wu Lawrence Berkeley National Laboratory, Berkeley, CA, USA Abstract. The Thick-Restart

More information

Sparse BLAS-3 Reduction

Sparse BLAS-3 Reduction Sparse BLAS-3 Reduction to Banded Upper Triangular (Spar3Bnd) Gary Howell, HPC/OIT NC State University gary howell@ncsu.edu Sparse BLAS-3 Reduction p.1/27 Acknowledgements James Demmel, Gene Golub, Franc

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix BIT 39(1), pp. 143 151, 1999 ILL-CONDITIONEDNESS NEEDS NOT BE COMPONENTWISE NEAR TO ILL-POSEDNESS FOR LEAST SQUARES PROBLEMS SIEGFRIED M. RUMP Abstract. The condition number of a problem measures the sensitivity

More information

Dedicated to Michael Saunders s 70th birthday

Dedicated to Michael Saunders s 70th birthday MINIMAL RESIDUAL METHODS FOR COMPLEX SYMMETRIC, SKEW SYMMETRIC, AND SKEW HERMITIAN SYSTEMS SOU-CHENG T. CHOI Dedicated to Michael Saunders s 70th birthday Abstract. While there is no lac of efficient Krylov

More information

S-Step and Communication-Avoiding Iterative Methods

S-Step and Communication-Avoiding Iterative Methods S-Step and Communication-Avoiding Iterative Methods Maxim Naumov NVIDIA, 270 San Tomas Expressway, Santa Clara, CA 95050 Abstract In this paper we make an overview of s-step Conjugate Gradient and develop

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Suares Problem Hongguo Xu Dedicated to Professor Erxiong Jiang on the occasion of his 7th birthday. Abstract We present

More information

Error Bounds for Iterative Refinement in Three Precisions

Error Bounds for Iterative Refinement in Three Precisions Error Bounds for Iterative Refinement in Three Precisions Erin C. Carson, New York University Nicholas J. Higham, University of Manchester SIAM Annual Meeting Portland, Oregon July 13, 018 Hardware Support

More information

Gram-Schmidt Orthogonalization: 100 Years and More

Gram-Schmidt Orthogonalization: 100 Years and More Gram-Schmidt Orthogonalization: 100 Years and More September 12, 2008 Outline of Talk Early History (1795 1907) Middle History 1. The work of Åke Björck Least squares, Stability, Loss of orthogonality

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 20 1 / 20 Overview

More information

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

GMRES ON (NEARLY) SINGULAR SYSTEMS

GMRES ON (NEARLY) SINGULAR SYSTEMS SIAM J. MATRIX ANAL. APPL. c 1997 Society for Industrial and Applied Mathematics Vol. 18, No. 1, pp. 37 51, January 1997 004 GMRES ON (NEARLY) SINGULAR SYSTEMS PETER N. BROWN AND HOMER F. WALKER Abstract.

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices Chapter 7 Iterative methods for large sparse linear systems In this chapter we revisit the problem of solving linear systems of equations, but now in the context of large sparse systems. The price to pay

More information

ANONSINGULAR tridiagonal linear system of the form

ANONSINGULAR tridiagonal linear system of the form Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal

More information

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve

More information

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES 1 PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES MARK EMBREE, THOMAS H. GIBSON, KEVIN MENDOZA, AND RONALD B. MORGAN Abstract. fill in abstract Key words. eigenvalues, multiple eigenvalues, Arnoldi,

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices.

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. A.M. Matsekh E.P. Shurina 1 Introduction We present a hybrid scheme for computing singular vectors

More information

Chapter 8 Cholesky-based Methods for Sparse Least Squares: The Benefits of Regularization

Chapter 8 Cholesky-based Methods for Sparse Least Squares: The Benefits of Regularization In L. Adams and J. L. Nazareth eds., Linear and Nonlinear Conjugate Gradient-Related Methods, SIAM, Philadelphia, 92 100 1996. Chapter 8 Cholesky-based Methods for Sparse Least Squares: The Benefits of

More information

WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS

WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS IMA Journal of Numerical Analysis (2002) 22, 1-8 WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS L. Giraud and J. Langou Cerfacs, 42 Avenue Gaspard Coriolis, 31057 Toulouse Cedex

More information

Lecture 9 Least Square Problems

Lecture 9 Least Square Problems March 26, 2018 Lecture 9 Least Square Problems Consider the least square problem Ax b (β 1,,β n ) T, where A is an n m matrix The situation where b R(A) is of particular interest: often there is a vectors

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM Electronic Transactions on Numerical Analysis. Volume 45, pp. 133 145, 2016. Copyright c 2016,. ISSN 1068 9613. ETNA ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

More information

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Numerical Methods. Elena loli Piccolomini. Civil Engeneering.  piccolom. Metodi Numerici M p. 1/?? Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement

More information

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems Part I: Review of basic theory of eigenvalue problems 1. Let A C n n. (a) A scalar λ is an eigenvalue of an n n A

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006

More information

Notes on Some Methods for Solving Linear Systems

Notes on Some Methods for Solving Linear Systems Notes on Some Methods for Solving Linear Systems Dianne P. O Leary, 1983 and 1999 and 2007 September 25, 2007 When the matrix A is symmetric and positive definite, we have a whole new class of algorithms

More information

Inexactness and flexibility in linear Krylov solvers

Inexactness and flexibility in linear Krylov solvers Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7) - IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy - October 15-19, 2007 in honor of Gérard Meurant for his 60 th

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

Augmented GMRES-type methods

Augmented GMRES-type methods Augmented GMRES-type methods James Baglama 1 and Lothar Reichel 2, 1 Department of Mathematics, University of Rhode Island, Kingston, RI 02881. E-mail: jbaglama@math.uri.edu. Home page: http://hypatia.math.uri.edu/

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues

More information