The parallel rational Arnoldi algorithm

Similar documents
Solving large sparse eigenvalue problems

Parallelization of the rational Arnoldi algorithm. Berljafa, Mario and Güttel, Stefan. MIMS EPrint:

Fitting an artificial frequency response

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Rational Krylov Decompositions: Theory and Applications. Berljafa, Mario. MIMS EPrint:

Introduction to Arnoldi method

Exponential of a nonnormal matrix

International Space Station model

Square root of a symmetric matrix

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Barycentric interpolation via the AAA algorithm

The Lanczos and conjugate gradient algorithms

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

M.A. Botchev. September 5, 2014

6.4 Krylov Subspaces and Conjugate Gradients

Algorithms that use the Arnoldi Basis

Numerical Methods in Matrix Computations

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

PROJECTED GMRES AND ITS VARIANTS

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Generalized rational Krylov decompositions with an application to rational approximation. Berljafa, Mario and Güttel, Stefan. MIMS EPrint: 2014.

(Linear equations) Applied Linear Algebra in Geoscience Using MATLAB

Lecture 8: Fast Linear Solvers (Part 7)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Applied Linear Algebra in Geoscience Using MATLAB

Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

1 Extrapolation: A Hint of Things to Come

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Block Krylov Space Solvers: a Survey

lecture 2 and 3: algorithms for linear algebra

Arnoldi Methods in SLEPc

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Matrix functions and their approximation. Krylov subspaces

lecture 3 and 4: algorithms for linear algebra

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

Simple iteration procedure

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Preconditioned inverse iteration and shift-invert Arnoldi method

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Course Notes: Week 1

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

Research Matters. February 25, The Nonlinear Eigenvalue Problem. Nick Higham. Part III. Director of Research School of Mathematics

Iterative Methods for Linear Systems of Equations

Reduction of Finite Element Models of Complex Mechanical Components

EECS 275 Matrix Computation

FEM and sparse linear system solving

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

Electronic filter design using RKFUN arithmetic

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

Lab 1: Iterative Methods for Solving Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems

A communication-avoiding thick-restart Lanczos method on a distributed-memory system

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis

9.1 Preconditioned Krylov Subspace Methods

LSMR: An iterative algorithm for least-squares problems

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Preface to the Second Edition. Preface to the First Edition

Iterative Methods for Sparse Linear Systems

Scientific Computing: Solving Linear Systems

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Exponentials of Symmetric Matrices through Tridiagonal Reductions

IDR(s) as a projection method

MAA507, Power method, QR-method and sparse matrix representation.

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

APPLIED NUMERICAL LINEAR ALGEBRA

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Numerical linear algebra

Rational Krylov methods for linear and nonlinear eigenvalue problems

Index. for generalized eigenvalue problem, butterfly form, 211

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Contents. Preface... xi. Introduction...

The Conjugate Gradient Method

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

Augmented GMRES-type methods

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

On Solving Large Algebraic. Riccati Matrix Equations

An Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems

Stabilization and Acceleration of Algebraic Multigrid Method

Lecture 3: Inexact inverse iteration with preconditioning

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Compressing exterior Helmholtz problems

NUMERICS OF THE GRAM-SCHMIDT ORTHOGONALIZATION PROCESS

Lecture 18 Classical Iterative Methods

The rational Krylov subspace for parameter dependent systems. V. Simoncini

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Large-scale eigenvalue problems

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Transcription:

The parallel rational Arnoldi algorithm Mario Berljafa Stefan Güttel May 016 Contents 1 Introduction 1 Parallelization strategies 1 3 Basic usage 1 Near-optimal continuation strategy 3 5 Links to numerical examples 6 References 5 1 Introduction Since version. of the RKToolbox, the rat krylov function can simulate the parallel construction of a rational Krylov basis. This is done by imposing various nonzero patterns in a so-called continuation matrix T. Simply speaking, the j-th column of this matrix contains the coefficients of the linear combination of j rational Krylov basis vectors which have been used to compute the next (j + 1)-th basis vector. Therefore T is an upper triangular matrix. Parallelization strategies By imposing certain nonzero patters on the continuation matrix T, we can simulate a parallel rational Arnoldi algorithm where several linear system solves are performed synchronously. Note that the typically most expensive part in the rational Arnoldi algorithm is the solution of large linear systems of equations, hence parallelizing this component can yield significant savings in computation time. There are various continuation strategies (i.e., ways to form the matrix T ), each with a different numerical behaviour, and we refer to [1] for an overview. By default, the strategy ruhe is used, which determines the next admissible continuation vector (column of T ) by computing a QR factorization of a small matrix []. Other available parallelization strategies are last, almost-last, and near-optimal. The strategy almost-last is the one used by Skoogh [3]. The near-optimal strategy is the one developed and advocated in [1]. It uses a (cheap) 1

predictor method to anticipate the direction of the next Krylov basis vector(s) in order to determine a good continuation matrix T. By good we mean that the rational Krylov basis that is orthogonalized is well-conditioned, so that numerical instabilities in the Gram Schmidt process are limited. 3 Basic usage We now demonstrate at a simple example how to use the parallel option in rat krylov. Links to the practically relevant examples are given below. First, let us define a matrix A, a starting vector b, and a sequence of pole parameters, where p = consecutive poles are pairwise distinct. Having distinct poles is a necessary requirement for parallelization. load west079 A = west079 ; b = ones (79,1) ; xi = repmat ([ -50, -5, 5, 50], 1, ); Now we can run rat krylov to compute a rational Arnoldi decomposition AV m+1 K m = V m+1 H m, with the basis vectors in V m+1 being generated in parallel. This is done by specifying the continuation strategy and the p = parameter: param = struct ( ' continuation ', 'almost - last ',... 'p ',,... 'orth ', 'CGS ',... ' reorth ', 1,... ' waitbar ', 1 ); [V, K, H, out ] = rat_krylov (A, b, xi, param ); The other struct fields specify that classical Gram-Schmidt orthogonalization has been used with reorthogonalization, the columns of the matrix pencil (H m, K m ) have been rescaled in an attempt to improve the numerical conditioning, and a waitbar has been displayed during the computation. The output parameter out has several fields of interest, in particular, it returns the continuation matrix being used during the computation. Here is a spy plot of T : spy ( out.t)

0 6 8 10 1 1 16 18 0 5 10 15 nz = 16 We can clearly see how Skoogh s almost-last strategy works: in every parallel cycle, p = rational Krylov basis vectors are computed, each from the last vector obtained on the respective parallel process in the previous parallel cycle. The output parameter out.r corresponds to the R factor in the implicitly computed QR factorization of the rational Krylov basis. As is explained in much more detail in [1], its condition number is an indicator for the stability of the computation. We observe that with the almost-last strategy this condition number is rather high: R = out.r; D = fminsearch (@(x) cond (R* diag (x)), ones ( size (R, ), 1),... struct ( ' Display ','off ')); disp ( cond (R* diag (D))) 3.177 e +06 Near-optimal continuation strategy To overcome, or at least reduce, the numerical problems observed with canonical parallelization strategies (such as almost-last and last), rat krylov implements the near-optimal strategy suggested in [1]. With this strategy, a (cheap) predictor method is used to anticipate the direction of rational Krylov basis vectors before their actual computation, to infer good continuation pairs. param. continuation = 'near - optimal '; [V, K, H, out ] = rat_krylov (A, b, xi, param ); 3

The continuation matrix T has a more dense pattern compared to the almost-last strategy, but still allows for p = basis vectors to be computed in parallel: spy ( out.t) 0 6 8 10 1 1 16 18 0 5 10 15 nz = 11 The condition number of the basis being orthogonalized is significantly smaller: R = out.r; D = fminsearch (@(x) cond (R* diag (x)), ones ( size (R, ), 1),... struct ( ' Display ','off ')); disp ( cond (R* diag (D))) 5.506 e +00 In fact, this basis is so close to being orthogonal, that running classical Gram-Schmidt without orthogonalization would be fine. The near-optimal method implemented in this simple manner is, however, computationally expensive: in fact every linear system in the rational Arnoldi algorithm has been solved twice using backslash. In a highperformance implementation one would reuse existing LU factors, or switch to an inexact predictor method like, e.g., FOM or GMRES. The predictor method is specified by a function handle in the continuation solve field and the number of iterations is specified in continuation m. Demonstrations of this are given in the numerical examples linked below.

5 Links to numerical examples The numerical examples in [1] can be reproduced with the following scripts: Numerical illustration from [1, Sec. 3.] TEM example from [1, Sec. 5.1] Inlet example from [1, Sec. 5.] Waveguide example from [1, Sec. 5.3] 6 References [1] M. Berljafa and S. Güttel. Parallelization of the rational Arnoldi algorithm, SIAM J. Sci. Comput., 39(5):S197 S1, 017. [] A. Ruhe. Rational Krylov: A practical algorithm for large sparse nonsymmetric matrix pencils, SIAM J. Sci. Comput., 19(5):1535 1551, 1998. [3] D. Skoogh. A parallel rational Krylov algorithm for eigenvalue computations, in G. Goos et al., editors, Applied Parallel Computing, volume 151 of Lecture Notes in Computer Science, pages 51 56. Springer-Verlag, Berlin, 1998. 5