Math 411 Preliminaries

Similar documents
Math 411 Preliminaries

1 Number Systems and Errors 1

Introduction to Applied Linear Algebra with MATLAB

Numerical solutions of nonlinear systems of equations

Numerical Methods in Matrix Computations

Applied Linear Algebra

APPLIED NUMERICAL LINEAR ALGEBRA

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Introduction to Numerical Analysis

Notes for Chapter 1 of. Scientific Computing with Case Studies

Numerical Methods - Numerical Linear Algebra

Arithmetic and Error. How does error arise? How does error arise? Notes for Part 1 of CMSC 460

Numerical Methods I Non-Square and Sparse Linear Systems

NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING

17 Solution of Nonlinear Systems

6.4 Krylov Subspaces and Conjugate Gradients

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Chapter 1 Computer Arithmetic

Course Notes: Week 1

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Computational Methods. Least Squares Approximation/Optimization

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Index. for generalized eigenvalue problem, butterfly form, 211

MATCOM and Selected MATCOM Functions

1 Backward and Forward Error

ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013

Introduction. Chapter One

S.F. Xu (Department of Mathematics, Peking University, Beijing)

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

TEACHING NUMERICAL LINEAR ALGEBRA AT THE UNDERGRADUATE LEVEL by Biswa Nath Datta Department of Mathematical Sciences Northern Illinois University

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Chapter 1 Mathematical Preliminaries and Error Analysis

Lecture 7. Floating point arithmetic and stability

Next topics: Solving systems of linear equations

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Iterative Methods. Splitting Methods

Numerical Analysis Lecture Notes

Linear Algebra. PHY 604: Computational Methods in Physics and Astrophysics II

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

Numerical Algorithms. IE 496 Lecture 20

Iterative methods for Linear System

Numerical Methods I Solving Nonlinear Equations

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Linear algebra & Numerical Analysis

Conjugate Gradient (CG) Method

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Preface. 2 Linear Equations and Eigenvalue Problem 22

Numerical Methods I Eigenvalue Problems

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

AMS526: Numerical Analysis I (Numerical Linear Algebra)

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

EECS 275 Matrix Computation

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Iterative Methods for Solving A x = b

Some minimization problems

LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Numerical Mathematics

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

TABLE OF CONTENTS INTRODUCTION, APPROXIMATION & ERRORS 1. Chapter Introduction to numerical methods 1 Multiple-choice test 7 Problem set 9

Applied Linear Algebra in Geoscience Using MATLAB

ECS 231 Computer Arithmetic 1 / 27

Numerical Methods I: Eigenvalues and eigenvectors

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Chapter 7 Iterative Techniques in Matrix Algebra

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS

Index. Copyright (c)2007 The Society for Industrial and Applied Mathematics From: Matrix Methods in Data Mining and Pattern Recgonition By: Lars Elden

Preliminary Examination, Numerical Analysis, August 2016

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

AIMS Exercise Set # 1

Mathematical preliminaries and error analysis

Orthogonal iteration to QR

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Elements of Floating-point Arithmetic

Convex Optimization CMU-10725

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

632 CHAP. 11 EIGENVALUES AND EIGENVECTORS. QR Method

Numerical optimization

Numerical methods part 2

Statistical Geometry Processing Winter Semester 2011/2012

Notes for CS542G (Iterative Solvers for Linear Systems)

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Introduction to Numerical Analysis

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

Introduction to Numerical Analysis

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

ACM 106a: Lecture 1 Agenda

Lecture Notes to be used in conjunction with. 233 Computational Techniques

MAT 610: Numerical Linear Algebra. James V. Lambers

NUMERICAL MATHEMATICS AND COMPUTING

Transcription:

Math 411 Preliminaries Provide a list of preliminary vocabulary and concepts Preliminary Basic Netwon s method, Taylor series expansion (for single and multiple variables), Eigenvalue, Eigenvector, Vector norms, 2-norm, infinite norm, Matrix norms, Condition number, Gauss Elimination Preliminary Use of Absolute error and Relative error and their relation to number of significant digits Solution of linear system of equations via Gauss elimination Stopping criteria in iterations Rounding error and truncation error Floating point system 1

Floating Point System Understand and handle error from floating point calculation Role of stability in numerical calculation Base, Sign bit, Exponent, Mantissa, underflow, overflow, rounding, chopping, roundoff error, absolute error, relative error, significant digits, truncation error, machine epsilon. Distribution of floating point numbers Addition of floating point numbers Loss of associativity and distributivity Catastrophic cancellation Prevention of catastrophic cancellation Forward error analysis Backward error analysis IEEE standard Some External Links: Basic Issues in Floating Point Arithmetic and Error Analysis, Supplementary Lecture Notes of J. Demmel on Floating Point System, University of California at Berkeley, September 1995. (local copy) Miscalculating Area and Angles of a Needle-like Triangle, Notes of W. Kahan on common misconceptions of floating point calculations, University of California at Berkeley, September 1997. (local copy) IEEE Floating Point Arithmetic, More advanced Lecture Notes of W. Kahan on the Status of IEEE Standard 754 for Binary Floating-Point Arithmetic, University of California at Berkeley, September 1995. (local copy) Goldberg, David, What every computer scientist should know about floating-point arithmetic, ACM Computing Surveys, Vol.23, No. 1 (March 1991), pp. 5-48 (Electronic version not available.) 2

Newton s method for several variables vector functions Newton s method derivations strength and weaknesses of the method interpretation of quadratic convergence Local convergence, Quadratic rate of convergence, Function Iteration, Jacobian matrix Newton s method derivations via Generalization of Newton s method for single variable functions Taylor series Quadratic rate of convergence and number of significant digits "Derivative" appears as a matrix Requires solution of a "totally new" linear system at each iteration step Non-convergence and slow convergence of Newton s method: Jacobian matrix may become singular Oscillation similar to one dimensional case Flat surface Multiple root Method properties: Strength Derivation of method may be obtained from Taylor series Local convergence when approximation is close to root Quadratic rate of convergence when approximation is close to root Reasonably easy to implement if Jacobian matrix can be symbolically calculated May be used to find complex roots Weaknesses Determination of starting guess may not be trivial Convergence or rate of convergence is not guaranteed when not close to the root Method may not converge or may converge very slowly Method requires evaluation of the Jacobian matrix as well as solution of a linear system at each iteration Stopping criteria choice not obvious Requires initial guess to be complex in order to find a complex root Can still view Newton s method as a fixed point iteration Quadratic convergence follows from general theory of fixed point iteration Theory for system of non-linear equations is still incomplete In implementation, some approximates Jacobian matrix via finite differencing (naïve) or "automatic differencing" (sophisticated) 3

Fixed point iteration in higher dimension Contraction mapping and Function iteration method Seidel s method to speed convergence Functional iteration, Seidel method Brouwer fixed point theorem (Review) Uniqueness of fixed point and derivative bounds Functional iteration and contraction mapping Use of Seidel s method to accelerate convergence Theoretical and Computable (though often pessimistic) error bounds Choice of initial guess not obvious Often used as a brute force approach or method of last resort even if derivative bounds are not satisfied 4

Quasi-Newton method- Broyden s method Broyden s method as an efficient way to apply concept from Newton s method Efficient computation of inverse of a matrix with a rank-1 update Rank 1 update, Sherman-Morrison formula Entries in Jacobian matrix may be approximated by finite differences Deteriotion of rate of convergence from quadratic to superlinear Initial computation of inverse requires O(n 3 ) operations using Gauss Elimination Sherman-Morrison implies subsequent calculation requires O(n 2 ) operations Hessian Broyden s method is implemented by explicitly calculating the inverse using Sherman Morrison formula. Gauss elimination is used only for constructing the inverse of the initial (exact or approximate) Jacobian matrix. 5

Steepest Descent method General solution of nonlinear equations Special case of solution of linear equations with a symmetric matrix Steepest descent direction, Quadratic form, Line search, Quadratic Interpolation Conversion of nonlinear system of equations to minimization problem Use of gradient direction as steepest descent direction Slow rate of convergence near minimum point Use of steepest descent method to generate initial guess for Newton or quasi-newton method For quadratic forms, performance deteriorates as condition number increases Line search has exact value for quadratic form Line search for general nonlinear problems is approximated by a quadratic minimization problem Conjugate direction methods Even if the original set of nonlinear equations do not admit a solution, a solution to the minimization problem can always be found using the steepest descent method. A theoretically optimal method (such as steepest descent method) is not necessarily the best numerical method 6

Conjugate Gradient method Conjugate direction, A-conjugate, (Theoretical) Finite termination Some External Links: Jonathan R. Shewchuk, An Introduction to the Conjugate Gradient Method Without the Agonizing Pain: http://www.cs.cmu.edu/~jrs/jrspapers.html#cg 7

Eigenvalue Problems Application giving rise to eigenvalue problems Properties of eigenvalues/ eigenvectors Eigenvalue Problem, Shift, Similarity transform Eigenvalues of diagonal and triangular matrices Eigenvalues of inverse matrix Gauss elimination does not preserve eigenvalues Similarity transformation preserves eigenvalues Change to eigenvalues through shifting, scalar product, matrix product Determinant as product of eigenvalues Gerschgorin theorem- location of eigenvalues Eigenvalues of real symmetric matrices are real Positive definite matrices 8

Power method Power method, Rayleigh quotient and inverse iteration Wielandt deflation Infinite norm, 2-norm and inner product, deflation Convergence of power method relies on existence of a dominant eigenvalue Normalization is needed within iteration to keep the entries of the eigenvector meaningful Eigenvector is obtained along with the eigenvalue Rayleigh quotient (for symmteric matrices) converges at twice the rate as regular power method LU factorization on A may be performed prior to applying the inverse power method or inverse iteration in order to speed up the calculation If the matrix is tridiagonal, each inverse iteration step is only O(n). Existence of multiple dominant eigenvalues of the same magnitude lead to either slow convergence or oscillation in the numerical result A matrix with the same set of eigenvalues as A, except the dominant eigenvalue is zeroed out, may be constructed using Wielandt deflation Power method x(new) = c A x(old) Largest (in mag.) eigenvalue Inverse power method A x(new) = c x(old) Smallest (in mag.) eigenvalue Shifted power method x(new) = c (A-s I) x(old) Eigenvalue farthest from shift s Inverse iteration (A-s I) x(new) = c x(old) Eigenvalue closest to shift s Here c denote a normalizing factor G. Peters and J.H. Wilkinson, Inverse Iteration, Ill-Conditioned Equations and Newton s Method, SIAM Review 21. No. 1 (1979), 339-360. (accessible through JSTOR) Some External Links: Even if the initial random vector does not have a component in the dominant direction, the presence of rounding error will eventually reintroduce a component in the dominant direction, and thus ensuring the success of the power method. If the shift s is close to an eigenvalue, the matrix A-sI is near singular and some care must be taken in the inverse iteration. However, even with poor conditioning, the inverse iteration performs well. 9

Householder Transformation Householder transformation to upper Hessenberg form QR factorization Orthogonal matrix, Upper Hessenberg form, QR factorization, catastrophic cancellation Orthogonal matrices have nice stability property: Qx = x in 2-norm, thus condition number is 1 Product of orthogonal matrices is orthogonal Each Householder matrix is constructed from the column vector that it needs to zero out Householder transformation is symmetric and orthogonal Pre- and post multiplication by a Householder transformation is a similarity transformation Construction of Householder transformations for a matrix The first component of the vector in the Householder transformation should be constructed with a sign chosen to avoid catastrophic cancellation QR factorization is closely related to Gram-Schmidt orthogonalization A symmetric matrix in upper Hessengberg form is tridiagonal In QR factorization, the matrix Q is the transpose of product of Householder transformations It is usually not necessary to compute the Householder transformation matrix explicitly, instead, use the normalized vector u or the unnormalized vector x associated with the Householder transformation. In QR factorization, the upper triangular matrix R obtained by premultiplication of Householder transformations does not possess the same set of eigenvalues as the original matrix. The QR factorization is not unique: by appropiately changing the signs of the entries in the matrices we can get many different factorizations 10

QR algorithm Rotation matrix, Real symmetric matrices have real eigenvalues QR algorithm produces similarity transforms even though QR factorization does not preserve eigenvalues QR algorithm applied to a matrix with real eigenvalues gives a diagonal matrix in the limit. QR algorithm is usually performed after a matrix is converted to an upper Hessenberg form Jacobi rotation matrix is orthogonal but usually non-symmetric The QR factorization process may be performed using Jacobi rotation (or fast Given s method) Construction of rotation matrices QR algorithm is closely related to the power method and hence shift may be employed to enhance performance- often convergence rate is cubic Choice of shifts Some External Links: A discussion of the QR algorithm implementation in matlab: http://www.mathworks.com/publications/newsletter/pdf/sum95cleve.pdf 11