On the influence of eigenvalues on Bi-CG residual norms

Similar documents
On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

The Lanczos and conjugate gradient algorithms

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Iterative Methods for Linear Systems of Equations

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Math 504 (Fall 2011) 1. (*) Consider the matrices

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Iterative methods for Linear System

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Algorithms that use the Arnoldi Basis

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

M.A. Botchev. September 5, 2014

Iterative Methods for Sparse Linear Systems

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Course Notes: Week 1

Inexactness and flexibility in linear Krylov solvers

6.4 Krylov Subspaces and Conjugate Gradients

PRECONDITIONED ITERATIVE METHODS FOR LINEAR SYSTEMS, EIGENVALUE AND SINGULAR VALUE PROBLEMS. Eugene Vecharynski. M.S., Belarus State University, 2006

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

JURJEN DUINTJER TEBBENS

Block Bidiagonal Decomposition and Least Squares Problems

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

On the Preconditioning of the Block Tridiagonal Linear System of Equations

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Block Krylov Space Solvers: a Survey

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

Principles and Analysis of Krylov Subspace Methods

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Fast iterative solvers

Index. for generalized eigenvalue problem, butterfly form, 211

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Algebra C Numerical Linear Algebra Sample Exam Problems

Hessenberg eigenvalue eigenvector relations and their application to the error analysis of finite precision Krylov subspace methods

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

FEM and sparse linear system solving

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Krylov subspace projection methods

ETNA Kent State University

On the accuracy of saddle point solvers

HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE

1 Extrapolation: A Hint of Things to Come

Numerical Methods in Matrix Computations

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Arnoldi Methods in SLEPc

PROJECTED GMRES AND ITS VARIANTS

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

9.1 Preconditioned Krylov Subspace Methods

Lecture 9 Least Square Problems

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Augmented GMRES-type methods

GMRES ON (NEARLY) SINGULAR SYSTEMS

arxiv: v2 [math.na] 1 Feb 2013

c 2005 Society for Industrial and Applied Mathematics

Numerical Methods - Numerical Linear Algebra

Scientific Computing: An Introductory Survey

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Contents. Preface... xi. Introduction...

Recycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

Key words. cyclic subspaces, Krylov subspaces, orthogonal bases, orthogonalization, short recurrences, normal matrices.

Eigenvalue and Eigenvector Problems

Matrices, Moments and Quadrature, cont d

IDR(s) as a projection method

A stable variant of Simpler GMRES and GCR

Preconditioned inverse iteration and shift-invert Arnoldi method

On the properties of Krylov subspaces in finite precision CG computations

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Key words. GMRES method, convergence bounds, worst-case GMRES, ideal GMRES, field of values

DELFT UNIVERSITY OF TECHNOLOGY

Numerical linear algebra

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

IDR(s) A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations

Numerical behavior of inexact linear solvers

1 Conjugate gradients

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

Residual iterative schemes for largescale linear systems

Preconditioning for Nonsymmetry and Time-dependence

c 2008 Society for Industrial and Applied Mathematics

Transcription:

On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue du sergent Bauchat 75012 Paris, France. gerard.meurant@gmail.com 4th IMA Conference on Numerical Linear Algebra and Optimisation University of Birmingham, 4.9.2014 1 / 30

Outline 1 The problem and previous work 2 Prescribed behavior of GMRES and FOM 3 Prescribed behavior in Bi-CG 4 Bi-CG breakdowns 5 Conclusions 2 / 30

Introduction: The Problem We consider the solution of linear systems Ax = b where A C n n is non-normal and nonsingular, by a Krylov subspace method. 3 / 30

Introduction: The Problem We consider the solution of linear systems Ax = b where A C n n is non-normal and nonsingular, by a Krylov subspace method. On the one hand there are methods with relatively nice theoretical properties (orthogonal bases) using long recurrences like the GMRES method [Saad & Schultz 1986] or the FOM method [Arnoldi 1951]. In practice they need to be restarted. 3 / 30

Introduction: The Problem We consider the solution of linear systems Ax = b where A C n n is non-normal and nonsingular, by a Krylov subspace method. On the one hand there are methods with relatively nice theoretical properties (orthogonal bases) using long recurrences like the GMRES method [Saad & Schultz 1986] or the FOM method [Arnoldi 1951]. In practice they need to be restarted. On the other hand, there are methods like Bi-CG [Lanczos 1952, Fletcher 1974], QMR [Freund & Nachtigal 1991] and Bi-CGStab [van der Vorst 1992] with constant costs per iteration based on less natural projection processes. They can break down without having found the solution. 3 / 30

Introduction: Eigenvalues govern convergence? The convergence behavior of both types of methods is in practice often governed by the eigenvalue distribution. For instance, eigenvalues clustered around the origin seem to guarantee fast convergence; this is what is generally used to analyse the quality of preconditioners. 4 / 30

Introduction: Eigenvalues govern convergence? The convergence behavior of both types of methods is in practice often governed by the eigenvalue distribution. For instance, eigenvalues clustered around the origin seem to guarantee fast convergence; this is what is generally used to analyse the quality of preconditioners. There is, however, no sound theoretical explanation for the relation between eigenvalue distribution and convergence speed. On the contrary, there are theoretical results which imply there is, in general, no relation at all. 4 / 30

Introduction: Eigenvalues govern convergence? The convergence behavior of both types of methods is in practice often governed by the eigenvalue distribution. For instance, eigenvalues clustered around the origin seem to guarantee fast convergence; this is what is generally used to analyse the quality of preconditioners. There is, however, no sound theoretical explanation for the relation between eigenvalue distribution and convergence speed. On the contrary, there are theoretical results which imply there is, in general, no relation at all. The most convincing results showing that the GMRES method need not be governed by eigenvalues alone appeared in a series of papers by Arioli, Greenbaum, Pták and Strakoš [Greenbaum & Strakoš 1994, Greenbaum & Pták & Strakoš 1996, Arioli & Pták & Strakoš 1998]: 4 / 30

Introduction: Eigenvalues govern convergence? Theorem 1 [Greenbaum & Pták & Strakoš 1996] Let b = f 0 f 1 f 2 f n 1 > 0 be any non-increasing sequence of real positive values and let λ 1,..., λ n be any set of nonzero complex numbers. Then there exists a class of matrices A C n n and right-hand sides b C n such that the residual vectors r k generated by the GMRES method applied to A and b satisfy r k = f k, 0 k n, and spectrum(a) = {λ 1,..., λ n }. 5 / 30

Introduction: Eigenvalues govern convergence? For assessing the quality of a preconditioner P when GMRES is applied to PAx = Pb, PA non-symmetric, this means that analysis of the spectrum of PA alone is not enough. 6 / 30

Introduction: Eigenvalues govern convergence? For assessing the quality of a preconditioner P when GMRES is applied to PAx = Pb, PA non-symmetric, this means that analysis of the spectrum of PA alone is not enough. For instance in constraint preconditioning, the fact that spec(pa) is, say, spec (PA) = {1, 1 2 ± (5) 2 } does not suffice to guarantee fast convergence of GMRES when PA is non-symmetric. What is needed additionally, is the fact that the eigenvalues have maximal geometric multiplicity (i.e. that PA is non-derogatory). 6 / 30

Introduction: Eigenvalues govern convergence? For assessing the quality of a preconditioner P when GMRES is applied to PAx = Pb, PA non-symmetric, this means that analysis of the spectrum of PA alone is not enough. For instance in constraint preconditioning, the fact that spec(pa) is, say, spec (PA) = {1, 1 2 ± (5) 2 } does not suffice to guarantee fast convergence of GMRES when PA is non-symmetric. What is needed additionally, is the fact that the eigenvalues have maximal geometric multiplicity (i.e. that PA is non-derogatory). The main problem with possibly non-normal input matrices is that besides eigenvalues, the eigenspaces can strongly influence residual norms because they are not orthogonal to eachother. 6 / 30

Introduction: Objects governing convergence? Tools other than eigenvalues used to explain GMRES convergence include: the pseudo-spectrum (e.g. [Trefethen & Embree 2005]) the field of values (e.g. [Eiermann 1993]) the polynomial numerical hull (e.g. [Greenbaum 2002]) decomposition in normal plus small rank (e.g. [Huhtanen & Nevanlinna 2000]) 7 / 30

Introduction: Objects governing convergence? Tools other than eigenvalues used to explain GMRES convergence include: the pseudo-spectrum (e.g. [Trefethen & Embree 2005]) the field of values (e.g. [Eiermann 1993]) the polynomial numerical hull (e.g. [Greenbaum 2002]) decomposition in normal plus small rank (e.g. [Huhtanen & Nevanlinna 2000]) One may wonder whether the strong potential independence between convergence behavior and eigenvalues is just an artefact of the GMRES method? 7 / 30

Introduction: Objects governing convergence? Tools other than eigenvalues used to explain GMRES convergence include: the pseudo-spectrum (e.g. [Trefethen & Embree 2005]) the field of values (e.g. [Eiermann 1993]) the polynomial numerical hull (e.g. [Greenbaum 2002]) decomposition in normal plus small rank (e.g. [Huhtanen & Nevanlinna 2000]) One may wonder whether the strong potential independence between convergence behavior and eigenvalues is just an artefact of the GMRES method? For restarted GMRES and FOM, the same independence has been proved [Vecharinsky & Langou 2011, Schweitzer 2014?]. 7 / 30

The aim of the talk As all Krylov subspace methods for non-symmetric matrices (e.g. GMRES, FOM, Bi-CG, QMR, TFQMR, CGS, Bi-CGStab, IDR) project in different ways onto essentially the same Krylov subspaces, one may expect that similar results are possible for short recurrence Krylov subspace methods. In this talk, we concentrate on the theoretically simplest method with short recurrences, the Bi-CG method. we will try to show whether arbitrary convergence curves can be combined with arbitrary eigenvalues in the Bi-CG method. We know the answer in nearly all cases. if possible, we will try to show how linear systems can be constructed generating prescribed Bi-CG residual norms with input matrices having prescribed spectrum. 8 / 30

Outline 1 The problem and previous work 2 Prescribed behavior of GMRES and FOM 3 Prescribed behavior in Bi-CG 4 Bi-CG breakdowns 5 Conclusions 9 / 30

The GMRES and FOM methods In GMRES and FOM an orthogonal basis for K k (A, b) is constructed with the Arnoldi process. In the kth iteration it computes (when there is no breakdown) the decomposition AV k = V k H k + h k+1,k v k+1 e T 1 = V k+1 Hk, where the columns of V k = [v 1,..., v k ] (the Arnoldi vectors) contain an orthogonal basis for the kth Krylov subspace, K k (A, b) span{b, Ab,..., A k 1 b}. 10 / 30

The GMRES and FOM methods In GMRES and FOM an orthogonal basis for K k (A, b) is constructed with the Arnoldi process. In the kth iteration it computes (when there is no breakdown) the decomposition AV k = V k H k + h k+1,k v k+1 e T 1 = V k+1 Hk, where the columns of V k = [v 1,..., v k ] (the Arnoldi vectors) contain an orthogonal basis for the kth Krylov subspace, K k (A, b) span{b, Ab,..., A k 1 b}. H k (square), resp. Hk (rectangular) are upper Hessenberg matrices containing the coefficients of the long recurrences; [ ] H H k = k C (k+1) k. 0 h k+1,k 10 / 30

The GMRES method With initial guess x 0 = 0, GMRES iterates are given by x G k = V ky k, b e1 y k = min H k y y C k FOM iterates are given by x F k = V k y k, y k = H 1 k b e 1 if H k is nonsingular; otherwise they are not defined. 11 / 30

The GMRES method With initial guess x 0 = 0, GMRES iterates are given by x G k = V ky k, b e1 y k = min H k y y C k FOM iterates are given by x F k = V k y k, y k = H 1 k b e 1 if H k is nonsingular; otherwise they are not defined. Thus iterates and residual norms generated by GMRES and FOM are fully determined by the Hessenberg matrices H k, H k and b. Prescribing GMRES and FOM residual norms amounts to prescribing the entries of these Hessenberg matrices in the right way. 11 / 30

The GMRES method The kth GMRES residual vector can be characterized through r k = min b Ax, equivalently, x K k (A,b) rg k AK k (A, b), whereas the kth FOM residual vector is characterized through r F k K k(a, b). 12 / 30

The GMRES method The kth GMRES residual vector can be characterized through r k = min b Ax, equivalently, x K k (A,b) rg k AK k (A, b), whereas the kth FOM residual vector is characterized through r F k K k(a, b). The corresponding residual norms are related through to formula 1 1 rk F = rk G 1 2 rk 1 G. 2 Note that FOM residual norms need not be non-increasing and are not defined if the corresponding GMRES iterate stagnates. 12 / 30

The parametrization for FOM and GMRES Here is how one can prescribe residual norms, eigenvalues and Ritz values in FOM and GMRES [DT & Meurant 2013]: 13 / 30

The parametrization for FOM and GMRES Here is how one can prescribe residual norms, eigenvalues and Ritz values in FOM and GMRES [DT & Meurant 2013]: Choose a unitary matrix V and put b = V e 1 and A = V HV, H upper Hessenberg. 13 / 30

The parametrization for FOM and GMRES Here is how one can prescribe residual norms, eigenvalues and Ritz values in FOM and GMRES [DT & Meurant 2013]: Choose a unitary matrix V and put b = V e 1 and A = V HV, H upper Hessenberg. To force the desired eigenvalues, H is of the form H = U 1 CU, U nonsingular upper triangular, where C is the companion matrix for the prescribed spectrum. 13 / 30

The parametrization for FOM and GMRES Here is how one can prescribe residual norms, eigenvalues and Ritz values in FOM and GMRES [DT & Meurant 2013]: Choose a unitary matrix V and put b = V e 1 and A = V HV, H upper Hessenberg. To force the desired eigenvalues, H is of the form H = U 1 CU, U nonsingular upper triangular, where C is the companion matrix for the prescribed spectrum. To force FOM residual norms f(0),..., f(n 1), f(i) > 0, the first row g T of U can be chosen as 1 g k = f(k 1), k = 1,..., n. 13 / 30

The parametrization for FOM and GMRES Let A = V (U 1 CU)V, b = V e 1. 14 / 30

The parametrization for FOM and GMRES Let A = V (U 1 CU)V, b = V e 1. To force GMRES residual norms f(0) f(n 1) > 0, the first row g T of U can be chosen as g 1 = 1 f(0), g 1 k = f(k 1) 2 1, k = 2,..., n. f(k 2) 2 14 / 30

The parametrization for FOM and GMRES Let A = V (U 1 CU)V, b = V e 1. To force GMRES residual norms f(0) f(n 1) > 0, the first row g T of U can be chosen as g 1 = 1 f(0), g 1 k = f(k 1) 2 1, k = 2,..., n. f(k 2) 2 To force Ritz values, i.e. eigenvalues of H k, the remaining submatrix T of [ ] g U = T 0 T can be chosen to have entries satisfying k (λ ρ (k) i ) = 1 ( g k+1 + i=1 t k,k ) k t i,k λ i. i=1 14 / 30

Outline 1 The problem and previous work 2 Prescribed behavior of GMRES and FOM 3 Prescribed behavior in Bi-CG 4 Bi-CG breakdowns 5 Conclusions 15 / 30

Bi-Lanczos The FOM/GMRES pair of methods with long recurrences has an analogue among methods with short recurrences: The Bi-CG/QMR pair. 16 / 30

Bi-Lanczos The FOM/GMRES pair of methods with long recurrences has an analogue among methods with short recurrences: The Bi-CG/QMR pair. Bi-CG and QMR use bi-orthogonal bases of Krylov subspaces constructed with the Bi-Lanczos algorithm. In the kth iteration it computes (when there is no breakdown) the decomposition AV k = V k T k + h k+1,k v k+1 e T 1 = V k+1 T k, where the columns of V k = [v 1,..., v k ] span K k (A, b) and satisfy W k V k = diag(ω 1,..., ω k ), ω i 0 for a matrix W k whose columns span K k (A, b). The matrix T k (rectangular), resp. T k (square) is tridiagonal, thus allowing for short recurrences to generate the bi-orthogonal bases. 16 / 30

Bi-Lanczos In analogy with the FOM/GMRES pair, the kth QMR residual norm is r QMR k = V k+1 ( b e 1 T b e1 k y k ), y k = min T k y. y C k 17 / 30

Bi-Lanczos In analogy with the FOM/GMRES pair, the kth QMR residual norm is r QMR k = V k+1 ( b e 1 T b e1 k y k ), y k = min T k y. y C k The kth Bi-CG residual norm satisfies r BiCG k and does not exist for singular T k. = b t k+1,k e T k T 1 e 1 v k+1. k 17 / 30

Bi-Lanczos In analogy with the FOM/GMRES pair, the kth QMR residual norm is r QMR k = V k+1 ( b e 1 T b e1 k y k ), y k = min T k y. y C k The kth Bi-CG residual norm satisfies r BiCG k and does not exist for singular T k. = b t k+1,k e T k T 1 e 1 v k+1. With a correct scaling of the columns of V k+1, Bi-CG residual norms are fully determined by the entries of T k. (QMR residual norms also depend upon V k+1 ). We can therefore try to extend the construction to prescribe FOM convergence behavior. k 17 / 30

The Bi-CG parametrization To force desired eigenvalues and Bi-CG residual norms we can Choose a nonsingular matrix V with normalized columns and put b = V e 1 and A = V T V 1, T tridiagonal. 18 / 30

The Bi-CG parametrization To force desired eigenvalues and Bi-CG residual norms we can Choose a nonsingular matrix V with normalized columns and put b = V e 1 and A = V T V 1, T tridiagonal. Try to find a tridiagonal T allowing the decomposition T = U 1 CU, U nonsingular upper triangular, where C is the companion matrix for the prescribed spectrum and 18 / 30

The Bi-CG parametrization To force desired eigenvalues and Bi-CG residual norms we can Choose a nonsingular matrix V with normalized columns and put b = V e 1 and A = V T V 1, T tridiagonal. Try to find a tridiagonal T allowing the decomposition T = U 1 CU, U nonsingular upper triangular, where C is the companion matrix for the prescribed spectrum and where the first row g T of U has entries g k = 1, k = 1,..., n, f(k 1) if f(0),..., f(n 1), f(i) > 0 are the prescribed Bi-CG residual norms. 18 / 30

The Bi-CG parametrization Let us assume for the moment, that we wish to prescribe only convergence curves where all iterates are defined. That means, we assume that the entries 1 g k = r BiCG k 1, k = 1,..., n. of the first row of U are all nonzero. Equivalently, all leading principal submatrices T k of T are nonsingular. 19 / 30

The Bi-CG parametrization Let us assume for the moment, that we wish to prescribe only convergence curves where all iterates are defined. That means, we assume that the entries 1 g k = r BiCG k 1, k = 1,..., n. of the first row of U are all nonzero. Equivalently, all leading principal submatrices T k of T are nonsingular. Clearly T = U 1 CU, U nonsingular upper triangular, where C is a given companion matrix and the first row g T of U is a given vector is always upper Hessenberg. Note that there holds [Parlett 1967] U 1 = [e 1, T e 1,..., T n 1 e 1 ]. To obtain a tridiagonal T one can apply the Bi-Lanczos algorithm to C with some starting vector z. But the Bi-Lanczos process can break down... 19 / 30

The Bi-CG parametrization Theorem [Joubert 1992]: The Bi-Lanczos algorithm applied to a nonderogatory matrix runs till completion and no leading principal submatrix of the generated tridiagonal matrix will be singular for almost every starting vector, i.e. except for a measure zero set of vectors. 20 / 30

The Bi-CG parametrization Theorem [Joubert 1992]: The Bi-Lanczos algorithm applied to a nonderogatory matrix runs till completion and no leading principal submatrix of the generated tridiagonal matrix will be singular for almost every starting vector, i.e. except for a measure zero set of vectors. Thus let for an appropriate z, CZ = ZT, Ze 1 = z. Then T = Z 1 CZ will have the desired spectrum. 20 / 30

The Bi-CG parametrization Theorem [Joubert 1992]: The Bi-Lanczos algorithm applied to a nonderogatory matrix runs till completion and no leading principal submatrix of the generated tridiagonal matrix will be singular for almost every starting vector, i.e. except for a measure zero set of vectors. Thus let for an appropriate z, CZ = ZT, Ze 1 = z. Then T = Z 1 CZ will have the desired spectrum. If we define U = [e 1, T e 1,..., T n 1 e 1 ] 1, T can be decomposed as T = U 1 CU, U nonsingular upper triangular 20 / 30

The Bi-CG parametrization Theorem [Joubert 1992]: The Bi-Lanczos algorithm applied to a nonderogatory matrix runs till completion and no leading principal submatrix of the generated tridiagonal matrix will be singular for almost every starting vector, i.e. except for a measure zero set of vectors. Thus let for an appropriate z, CZ = ZT, Ze 1 = z. Then T = Z 1 CZ will have the desired spectrum. If we define U = [e 1, T e 1,..., T n 1 e 1 ] 1, T can be decomposed as T = U 1 CU, U nonsingular upper triangular To have the first row of U equal to our given g T, it suffices to scale T with a diagonal matrix D as ˆT = D 1 T D = (UD) 1 C(UD). 20 / 30

Outline 1 The problem and previous work 2 Prescribed behavior of GMRES and FOM 3 Prescribed behavior in Bi-CG 4 Bi-CG breakdowns 5 Conclusions 21 / 30

Bi-CG breakdown types So far, we assumed that we prescribe only convergence curves where all iterates are defined. Let us list all situations where the Bi-CG method can break down: Bi-CG can break down if the underlying Bi-Lanczos process breaks down: 22 / 30

Bi-CG breakdown types So far, we assumed that we prescribe only convergence curves where all iterates are defined. Let us list all situations where the Bi-CG method can break down: Bi-CG can break down if the underlying Bi-Lanczos process breaks down: In case of a lucky breakdown, a subdiagonal entry of T is zero and we have found an A- or A -invariant subspace. This gives, in exact arithmetics, the solution of the linear system or of the system with A. 22 / 30

Bi-CG breakdown types So far, we assumed that we prescribe only convergence curves where all iterates are defined. Let us list all situations where the Bi-CG method can break down: Bi-CG can break down if the underlying Bi-Lanczos process breaks down: In case of a lucky breakdown, a subdiagonal entry of T is zero and we have found an A- or A -invariant subspace. This gives, in exact arithmetics, the solution of the linear system or of the system with A. In case of a serious breakdown, two basis vectors v k and w k are orthogonal to eachother and the bi-orthogonality condition W k V k = diag(ω 1,..., ω k ), ω i 0 cannot be satisfied. One way to continue the Bi-Lanczos process is to use a look-ahead technique, i.e. to perform further iterations until for some i, w k+i v k+i 0. 22 / 30

Bi-CG breakdown types So far, we assumed that we prescribe only convergence curves where all iterates are defined. Let us list all situations where the Bi-CG method can break down: Bi-CG can break down if the underlying Bi-Lanczos process breaks down: In case of a lucky breakdown, a subdiagonal entry of T is zero and we have found an A- or A -invariant subspace. This gives, in exact arithmetics, the solution of the linear system or of the system with A. In case of a serious breakdown, two basis vectors v k and w k are orthogonal to eachother and the bi-orthogonality condition W k V k = diag(ω 1,..., ω k ), ω i 0 cannot be satisfied. One way to continue the Bi-Lanczos process is to use a look-ahead technique, i.e. to perform further iterations until for some i, w k+i v k+i 0. In case of an incurable breakdown, no look-ahead strategy will help to generate a pair of bi-orthognal bases (w k+i v k+i = 0 for all i). 22 / 30

Bi-CG breakdown types Bi-CG can also break down if the underlying Bi-Lanczos process runs to completion. 23 / 30

Bi-CG breakdown types Bi-CG can also break down if the underlying Bi-Lanczos process runs to completion. In this case, the breakdown is due to the fact that the current tridiagonal matrix T k is singular. 23 / 30

Bi-CG breakdown types Bi-CG can also break down if the underlying Bi-Lanczos process runs to completion. In this case, the breakdown is due to the fact that the current tridiagonal matrix T k is singular. This corresponds precisely to a zero entry on position k of the first row of U in the decomposition T = U 1 CU, U nonsingular upper triangular, C a companion matrix. 23 / 30

Bi-CG breakdown types Bi-CG can also break down if the underlying Bi-Lanczos process runs to completion. In this case, the breakdown is due to the fact that the current tridiagonal matrix T k is singular. This corresponds precisely to a zero entry on position k of the first row of U in the decomposition T = U 1 CU, U nonsingular upper triangular, C a companion matrix. If we wish to prescribe any Bi-CG convergence curve, we must also be able to prescribe curves with this type of breakdown, i.e. with possibly non-defined iterates. 23 / 30

Prescribing Bi-CG breakdowns Thus the question is: Can we find tridiagonal matrices T of the form T = U 1 CU, U nonsingular upper triangular, where C is a given companion matrix and the first row g T of U is a given vector with zeros on prescribed positions. 24 / 30

Prescribing Bi-CG breakdowns Thus the question is: Can we find tridiagonal matrices T of the form T = U 1 CU, U nonsingular upper triangular, where C is a given companion matrix and the first row g T of U is a given vector with zeros on prescribed positions. The answer is no, or not yet? But first, we have this proposition: 24 / 30

Prescribing Bi-CG breakdowns Thus the question is: Can we find tridiagonal matrices T of the form T = U 1 CU, U nonsingular upper triangular, where C is a given companion matrix and the first row g T of U is a given vector with zeros on prescribed positions. The answer is no, or not yet? But first, we have this proposition: Theorem [DT & Meurant 2014?]: Assume Bi-Lanczos with nonsingular input matrix A with initial vector b runs to completion. Let x i if and only of the ith leading principal submatrix of T is singular. Then whenever x k 1 =, we have x k. 24 / 30

Prescribing Bi-CG breakdowns Thus it is impossible in Bi-CG to have two subsequent iterations where iterates are not defined. 25 / 30

Prescribing Bi-CG breakdowns Thus it is impossible in Bi-CG to have two subsequent iterations where iterates are not defined. We were unable to find a way to generate a tridiagonal matrix with given spectrum and singular leading principal submatrices of particular (not subsequent) sizes. This seems to be an open or unconsidered problem in the literature. 25 / 30

Prescribing Bi-CG breakdowns Thus it is impossible in Bi-CG to have two subsequent iterations where iterates are not defined. We were unable to find a way to generate a tridiagonal matrix with given spectrum and singular leading principal submatrices of particular (not subsequent) sizes. This seems to be an open or unconsidered problem in the literature. j k T n 0 0. 25 / 30

Prescribing Bi-CG breakdowns The problem amounts to finding a starting vector for Bi-Lanczos applied to a companion matrix C such that the generated tridiagonal matrix has some singular leading principal submatrices. 26 / 30

Prescribing Bi-CG breakdowns The problem amounts to finding a starting vector for Bi-Lanczos applied to a companion matrix C such that the generated tridiagonal matrix has some singular leading principal submatrices. To force a leading principal submatrices to be singular, one can modify two entries of a starting vector. In this way we are able to influence the singularity of only the first half of the leading principal submatrices of T. 26 / 30

Prescribing Bi-CG breakdowns The problem amounts to finding a starting vector for Bi-Lanczos applied to a companion matrix C such that the generated tridiagonal matrix has some singular leading principal submatrices. To force a leading principal submatrices to be singular, one can modify two entries of a starting vector. In this way we are able to influence the singularity of only the first half of the leading principal submatrices of T. Similarly, one can use a backward eigenvalue-forcing strategy described in Parlett and Strang [Parlett & Strang 2008] to influence the singularity of only the second half of the leading principal submatrices of T. 26 / 30

Outline 1 The problem and previous work 2 Prescribed behavior of GMRES and FOM 3 Prescribed behavior in Bi-CG 4 Bi-CG breakdowns 5 Conclusions 27 / 30

Conclusions Any history of positive, finite Bi-CG residual norm is possible with any eigenvalues. 28 / 30

Conclusions Any history of positive, finite Bi-CG residual norm is possible with any eigenvalues. Additionally, infinite Bi-CG residual norms can be prescribed either in the first or in the second half of the iterations. 28 / 30

Conclusions Any history of positive, finite Bi-CG residual norm is possible with any eigenvalues. Additionally, infinite Bi-CG residual norms can be prescribed either in the first or in the second half of the iterations. Future work: Consequences for the QMR method, prescribing Ritz values of tridiagonal matrices. 28 / 30

Conclusions Any history of positive, finite Bi-CG residual norm is possible with any eigenvalues. Additionally, infinite Bi-CG residual norms can be prescribed either in the first or in the second half of the iterations. Future work: Consequences for the QMR method, prescribing Ritz values of tridiagonal matrices. For more details see: J. Duintjer Tebbens, G. Meurant: On the Convergence of QOR and QMR Krylov Methods for Solving Linear Systems, to be submitted. 28 / 30

Last but not least Thank you for your attention! 29 / 30

Related papers A. Greenbaum and Z. Strakoš, Matrices that generate the same Krylov residual spaces, IMA Vol. Math. Appl., 60 (1994), pp. 95 118. A. Greenbaum, V. Pták and Z. Strakoš, Any nonincreasing convergence curve is possible for GMRES, SIMAX, 17 (1996), pp. 465 469. M. Arioli, V. Pták and Z. Strakoš, Krylov sequences of maximal length and convergence of GMRES, BIT, 38 (1998), pp. 636 643. E. Vecharinsky and J. Langou, Any admissible cycle-convergence behavior is possible for restarted GMRES at its initial cycles, 18 (2011), Num. Lin. Alg. Appl., pp. 499 511. J. Duintjer Tebbens and G. Meurant, Any Ritz value behavior is possible for Arnoldi and for GMRES, SIMAX, 33 (2012), pp. 958 978. J. Duintjer Tebbens and G. Meurant, Prescribing the behavior of early terminating GMRES and Arnoldi iterations, Numer. Algorithms, 65 (2014), pp. 69 90. M. Schweitzer, Any cycle-convergence curve is possible for restarted FOM, technical report Univ. Wuppertal, Preprint BUW-IMACM 14/19, 2014. 30 / 30