QD-MnHOD WITH NEWTON SHIFT F. L. BAUER TECHNICAL REPORT NO. 56 MARCH 1, 1967

Similar documents
11.3 Eigenvalues and Eigenvectors of a Tridiagonal Matrix

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Index. for generalized eigenvalue problem, butterfly form, 211

Eigenvalues and Eigenvectors

Introduction to Applied Linear Algebra with MATLAB

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Chapter 6. Algebraic eigenvalue problems Introduction Introduction 113. Das also war des Pudels Kern!

Eigenvalues and Eigenvectors

International Journal of Scientific Research and Reviews

Begin accumulation of transformation matrices. This block skipped when i=1. Use u and u/h stored in a to form P Q.

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

1 Number Systems and Errors 1

I-v k e k. (I-e k h kt ) = Stability of Gauss-Huard Elimination for Solving Linear Systems. 1 x 1 x x x x

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Bounds for Eigenvalues of Tridiagonal Symmetric Matrices Computed. by the LR Method. By Gene H. Golub

Solving large scale eigenvalue problems

Numerical Methods in Matrix Computations

Math 411 Preliminaries

In this section again we shall assume that the matrix A is m m, real and symmetric.

Singular and Invariant Matrices Under the QR Transformation*

Linear Least-Squares Data Fitting

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

arxiv: v1 [math.na] 5 May 2011

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Foundations of Matrix Analysis

Numerical Methods - Numerical Linear Algebra

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

EIGENVALUE PROBLEMS (EVP)

Iassume you share the view that the rapid

Linear algebra & Numerical Analysis

Introduction to Numerical Analysis

Block-tridiagonal matrices

TABLE OF CONTENTS INTRODUCTION, APPROXIMATION & ERRORS 1. Chapter Introduction to numerical methods 1 Multiple-choice test 7 Problem set 9

On the reduction of matrix polynomials to Hessenberg form

M 340L CS Homework Set 12 Solutions. Note: Scale all eigenvectors so the largest component is +1.

LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,

Lecture 12 Eigenvalue Problem. Review of Eigenvalues Some properties Power method Shift method Inverse power method Deflation QR Method

The Eigenvalue Problem: Perturbation Theory

Numerical methods for eigenvalue problems

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Eigenvalue Problems and Singular Value Decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

EECS 275 Matrix Computation

6.4 Krylov Subspaces and Conjugate Gradients

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING

FINITE-DIMENSIONAL LINEAR ALGEBRA

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Review problems for MA 54, Fall 2004.

Computational Linear Algebra

Lecture 10 - Eigenvalues problem

lecture 2 and 3: algorithms for linear algebra

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Econ Slides from Lecture 7

Numerical Analysis Lecture Notes

11.5 Reduction of a General Matrix to Hessenberg Form

Eigenvalues and Eigenvectors

Lecture 1 INF-MAT3350/ : Some Tridiagonal Matrix Problems

Math Linear Algebra Final Exam Review Sheet

SQUARE ROOTS OF 2x2 MATRICES 1. Sam Northshield SUNY-Plattsburgh

MTH 464: Computational Linear Algebra

Orthogonal iteration to QR

The QR Algorithm. Marco Latini. February 26, 2004

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Eigenvalues and eigenvectors

Eigenvalue and Eigenvector Problems

3 Matrix Algebra. 3.1 Operations on matrices

Scientific Computing: An Introductory Survey

Chasing the Bulge. Sebastian Gant 5/19/ The Reduction to Hessenberg Form 3

The Cayley-Hamilton Theorem and the Jordan Decomposition

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that

Chapter 4 No. 4.0 Answer True or False to the following. Give reasons for your answers.

Linear Algebra Review

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

Systems of Second Order Differential Equations Cayley-Hamilton-Ziebur

APPLIED NUMERICAL LINEAR ALGEBRA

Solving large scale eigenvalue problems

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

Numerical Methods I: Eigenvalues and eigenvectors

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

11.0 Introduction. An N N matrix A is said to have an eigenvector x and corresponding eigenvalue λ if. A x = λx (11.0.1)

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

CS227-Scientific Computing. Lecture 4: A Crash Course in Linear Algebra

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15.

Introduction to Matrices

Eigenvalues and Eigenvectors

Linear Algebra: Matrix Eigenvalue Problems

BLAS Specification Revisited. Linda Kaufman William Paterson University

Matrix Computations and Semiseparable Matrices

THE QR METHOD A = Q 1 R 1

On the influence of eigenvalues on Bi-CG residual norms

Transcription:

CS 56 QD-MnHOD WITH NEWTON SHIFT BY F.. BAUER TECHNICA REPORT NO. 56 MARCH, 967 Supported in part by a grant from the National Science Foundation. COMPUTER SC IENCE DEPARTMENT School of Humanities and Sciences STANFORD UNIVERS ITY

&D-METHOD WITH NEWTON SHIFT BY F.. Bauer* I- / : t *Mathematisches Institut der Technischen Hochschule, M&hen and Computer Science Department, Stanford University. This work was in part supported by NSF.

. For determining the eigenvalues of a symmetric, tridiagonal matrix, various techniques have proven to be satisfactory; in particular the bisection method and &R-transformation with shifts determined by the last diagonal element or by the last 2X2 principal minor. Bisection, based upon a Sturm sequence, allows one to concentrate on the determination of any prescribed set of roots, prescribed by intervals or by ordering numbers. The &R-transformation is faster,; however, it gives the eigenvalues in a non-predictable ordering and is therefore mainly advocated for the determination of all roots. Theoretically, for symmetric matrices, a &R-step is equivalent to two successive R-steps, and the R-transformation for a tridiagonal matrix is, apart from organizational details, identical with the qd-method. For non-positive definite matrices, however, the R-transformation cannot be guaranteed to be numerically stable unless pivotal interchanges are made. This has led to preference for the &R-transformation, which is always numerically stable. If, however, some of the smallest or some of the largest eigenvalues are wanted, then the &R-transformation will not necessarily give only f -these, and bisection might seem too slow with its fixed convergence rate of l/2. In this situation, Newton's method would be fine if the Newton correction can be computed sufficiently simply, since it will always tend, &I monotonically to the nearest root starting from a point outside the spectrum. Consequently, if one always worked with positive (or negative) definite matrices, there would be no objection to using the now stable qd-algorithm. In particular, for the determination of some of the I- /

smallest roots of a matrix known to be positive definite -- this problem arises frequently in connection with finite difference or Ritz approximations to analytical eigenvalue problems -- the starting value zero would be usually a quite good initial approximation. We shall show that for a qd-algorithm, the Newton correction can very easily be calculated, and accordingly a shift which avoids undershooting, or a lower bound. Since the last diagonal element gives an upper bound, the situation is quite satisfactory with respect to bounds. 2. et q(h) = (A - AI)-' be the resolvent of a matrix A. Then -$ g(h) = (A - hi)02. Assume that A is of Hessenberg form of order n 2 2 J viz, i A= with +l 's in the lower off-diagonal. Then since A - h is again of Hessenberg form, the co-factor of the (,n) element is +l 9 and therefore the (n,l) -element of (A - TI)-~ is -l/f(h), where f(h) is the characteristic polynomial of A. From the result above we conclude that the (n,l) element of (A - XI)-' is f'(h)/f2(h), and therefore 6(X) = -f.(h)/f'(h) = ee(a - ht)-'el/ez(a - XI)-"el is the Newton correction, X + 6(h) being the next approximation. f I i

3* The calculation of ee(a - li)-'el and ee(a -II)-~~~ can be based upon the solution of (A - xi)x = ale and cy2yt(a - XI) = ez by backward substitution starting with (x) n =l and (y),=l. Then cwl=cu2 (= a), l/a = ez(a - hi)-'el, and ytx/a2 = et(a - hi)m2e n ; or f(h) = cy, f'(h) = ytx, 6(h) = cy/y'x. The final result holds even when h is an eigenvalue, a then being zero and ytx k 0 unless A is defective. While for the approximation of eigenvectors this backsubstitution, known as Hyman's technique, cannot be generally advocated, it offers a simple way to the calculation of f' (h) = y'x and of the Newton correction. It can be used when Newton's method can safely be used; e.g., when the roots of the Hessenberg matrix A are known to be all real, in connection with some deflation technique. 4. For tridiagonal matrices, however, the R-transformation or the qd-algorithm gives the Newton correction as a by-product. In the qd version of the R-transformation we perform first for a certain value of h the triangular decomposition of A - XI, - _ ql - - e G- i A - h = 92 0 q 3 I- e20.. e. n-l

multiply the factors conversely and decompose again. 0.. I en-l 92 q 3 0 = A - h = % 0 9; cl. 3 :l e*l. 0 e n-l et2 0 The transformed matrix A' - h has the same characteristic polynomial and may serve as well to calculate the New-ton correction. ever, the solution of (A' - h > X is immediately given by = Ye How- G~r+lq&l x. x q xq, 3 n 9 a =l-i i = 'i ' ( - > n+l and likewise for the solution of yl(a' - AI) = cy2ee by YT & ❼ ❻ ❾ ❼ n+l i=l i'- 4

Note that QI = cy 2 -det(hi -A) is the determinantal invariant of the qd-algorithm. Hence the relation or rather 6(X) = ' %I(.q+ % ) (f q3 ). 7 T-l q2 %l - +). The quotients qi+l/qi in the nested product, however, are calculated as a matter of course in the R-step with the quotient rule e! = (qi+l/q~)ei The extra work amounts therefore to n-l multiplications, n-l additions of, and one division. The shift by' s(h) is now preferably made after the next inter+ mediate matrix A: - h is formed and is done, as usual, implicitly I I in the difference rule. Thus, a shift is made every second qd-step. As mentioned in the introduction, numerical stability requires A in the beginning to be essentially symmetric and positive definite; i.e., e >O and % >o. P This property will then be preserved. I- The quantities gl~l and q can be calculated also by a continued m C fraction recurrence directly from ai-h bi I-! A' - AI = 0 a; -A b;.. - 0 b' n-l an-a 5

i.e., 92 = a' -, n q i = a! - A:--b!/q ii4 ( i = n-l,...,l) qi = a; - X, q; = ai - h - biol qfw2 I (i = 2,3,...,n). For an essentially symmetric matrix; i.e., bci > 0, the components of x H and y together with -Q! form a Sturm sequence. Correspondingly, if all the % 0 and ' 4 k 0, the number of positive elements in the q-sequence counts the number of positive eigenbalues. This use of the continued fraction recurrence has some merits for the bisection method. The qd-transformation would not allow one to calculate the Sturm sequence in a stable way,- apart from the trivial case where all 9 >o. iterature: For the methods, concepts, and results used here, see J.H. Wilkinson, The Algebraic Eigenvalue Problem. Oxford, 965. A. J. Householder, New York, 964. The Theory of Matrices in Numerical Analysis,