Krylov methods for the computation of matrix functions

Size: px
Start display at page:

Download "Krylov methods for the computation of matrix functions"

Transcription

1 Krylov methods for the computation of matrix functions Jitse Niesen (University of Leeds) in collaboration with Will Wright (Melbourne University) Heriot-Watt University, March 2010

2 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions

3 Matrix functions via series Given a scalar function f : C C with Taylor series f (x) = a 0 + a 1 x a 2x a 3x 3 +, the matrix function f : C n n C n n is defined by f (X ) = a 0 + a 1 X a 2X a 3X 3 +. This talk concentrates on the matrix exponential exp(x ) = n=0 1 n! X n. Matrix functions inherit many properties of the corresponding scalar functions, e.g., solution of X = AX is X (t) = exp(at) X (0). Definition needs to be adapted if scalar function f has singularities.

4 Matrix functions via diagonalization Scalar function: f (x) = a 0 + a 1 x a 2x a 3x 3 + Matrix function: f (X ) = a 0 + a 1 X a 2X a 3X 3 + Matrix function on diagonal matrix reduces to scalar function d f (d 1 ) f 0 d =. 0 f (d 2 ) d n f (d n ) If X can be diagonalized, X = VDV 1, then f (X ) = a n (VDV 1 ) n = n=0 a n VD n V 1 = Vf (D)V 1. n=0 This yields another definition of matrix functions. Can be extended to non-diagonalizable matrices by continuity.

5 Matrix functions via contour integration The third definition of matrix functions is f (X ) = 1 f (z) (X zi ) 1 dz, 2πi where integral is along contour encircling the eigenvalues of X. Equivalent to previous definition by residue theorem; the integrand has poles at the eigenvalues. This definition is theoretically convenient because it always works. References 1. Gantmacher, The Theory of Matrices, Horn & Johnson, Topics in Matrix Analysis, N. Higham, Functions of Matrices, 2008.

6 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions

7 Centrality measures of a network A graph is a collection of nodes, some of which are connected by edges. We want to know which nodes are central. One centrality measure is the degree: the number of edges of a node. This is a local measure; wish to extend this to incorporate global information. Let X be the adjacency matrix of a graph: x ij is 1 if there is an edge between nodes i and j and 0 otherwise. The (i, j) entry of X 2 is k x ikx kj. This counts the number of paths of length 2 from i to j. The (i, i) entry of X n counts the number of n-cycles that i is on. In particular, the (i, i) entry of X 2 is the degree. Estrada & Rodríguez-Velázquez (2005) propose to use the diagonal entries of exp(x ) = n 1 n! X n as a centrality measure.

8 Matrix functions solve differential equations The solution of x = ax, x(0) = x 0 is x(t) = exp(at)x 0. The solution of x = Ax, x(0) = x 0 is x(t) = exp(ta)x 0. The solution of x = ax + b, x(0) = 0 is x(t) = t where ϕ 1 (z) = ( exp(z) 1 ) /z. 0 exp(aτ)b dτ = exp(aτ) 1 b = tϕ 1 (at)b. a The solution of x = Ax + b, x(t) = 0 is x(t) = tϕ 1 (ta)b. The solution of x = Ax + ct, x(0) = 0 is x(t) = t 2 ϕ 2 (ta)c, where ϕ 2 (z) = ( exp(z) 1 z ) /z 2. These results can be combined by superposition.

9 Exponential Euler method Solution of x (t) = Lx(t) + v 0 + tv t2 v 2 +, x(0) = x 0, where x(t), x 0, v 0, v 1,... are vectors and L is a matrix, is x(t) = exp(tl)x 0 + hϕ 1 (tl)v 0 + h 2 ϕ 2 (tl)v 1 + h 3 ϕ 3 (tl)v 2 + Consider x = Lx + N(x) (L = linear, N = nonlinear) Replace the nonlinear term with the constant N(x(0)) N(x(t)) (for small t) and use the results from the previous slide: x(t) = exp(tl)x(0) + ϕ 1 (tl) N(x(0)). This leads to the exponential Euler method x n+1 = exp(tl)x n + ϕ 1 (hl)n(x n ). This method is not affected by stiffness in L. (Certaine 1960)

10 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions

11 Computation using series The series exp(x ) = I + X X X 3 + is one method to compute the matrix exponential. Series converges slowly away from the origin, so combine with scaling and squaring. Set Y = 2 k X, compute exponential, square exp(y ) k times. This uses the identity exp(x ) = (exp( 1 2 X ))2. Further improvement is to use Padé approximation instead of polynomial approximation: exp(x) c 0 + c 1 X + c 2 X 2 + c 3 X 3 d 0 + d 1 X + d 2 X 2 + d 3 X 3 where the coefficients c i and d i are chosen so that the error is as small as possible for small x (in practice, we use more coefficients). Padé with scaling and squaring is the most popular general-purpose method for computing the matrix exponential. (Lawson 1967; Higham 2005; Al-Mohy & Higham 2009)

12 Computing matrix functions by diagonalization If X = VDV 1 then f (X ) = Vf (D)V 1. So compute matrix function by first diagonalizing the matrix. This works well for some matrices, in particular symmetric matrices. However, it fails if X is (close to) non-diagonalizable. Computing matrix functions by integration Use f (X ) = 1 2πi f (z) (X zi ) 1 dz and evaluate contour integral with trapezium rule. (Kassam & Trefethen 2005; Schmelzer & Trefethen 2007) This has its appeal, but is expensive, especially if eigenvalues are far apart or matrix is very non-normal. (Ashi, Cummings & Matthews 2009)

13 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions

14 The idea behind Krylov methods In exponential integrators, and other applications, X is a large matrix. On the other hand, we need not f (X ) but f (X )b. A matrix-free method uses only matrix-vector products. This leads to the Krylov subspace K m (X, b) = span{b, Xb, X 2 b,..., X m 1 b}. This basis is very ill-conditioned, so use Gram Schmidt (here called Arnoldi or Lanczos) to get an orthogonal basis: K m (X, b) = span{v 1, v 2, v 3,..., v m }. Put the vectors v j in an n-by-m matrix V m (where n is size of X ). Basis transformation is encoded in m-by-m matrix H m.

15 The idea behind Krylov methods II Considering the matrices as linear mappings: X is map from big space C n to itself. V m is projection from C n onto small K m (X, b). V T m is extension from K m (X, b) to C n ; V m V T m = Id. H m is projection of action of X on K m (X, b); H m = V m XV T m. Krylov methods use the approximation X Vm T H m V m. In the context of matrix functions, use f (X ) = a k X k a k (Vm T H m V m ) k n=k = k=0 a k Vm T (H m ) k V m = Vm T f (H m )V m. k=0 Krylov approximation replaces big matrix X by small matrix H m.

16 Steps towards a practical method Estimate error in Krylov approximation. Use error estimate to adaptively choose dimension m of Krylov subspace (and to correct approximation). Use recursion relation to combine several terms in one: exp(x )b 0 + ϕ 1 (X )b 1 + ϕ 2 (X )b 2 = w 0 + ϕ 2 (X )w 2. Compute ϕ function using trick ([ ]) X b exp = 0 0 [ exp(x ) ϕ1 (X )b 0 1 Problem: A priori estimate suggests that optimal m is proportional to spectral radius ρ(x ). Thus, introduce time stepping similar to scaling-and-squaring. This is implemented in the matlab code phipm. ].

17 Leja point interpolation Krylov methods project matrix function on Krylov subspace K m (X, b) = span{b, Xb, X 2 b,..., X m 1 b}. They approximate f (X )b by element in K m (X, b). Thus, approximation is of form p(x )b, with p a polynomial. In fact, Krylov method performs polynomial interpolation in Ritz values (eigenvalues of H m ). Good because Ritz values approximate eigenvalues of A. (Saad 1992) But Ritz values may not be the best interpolation points. If bound on eigenvalues of A is known, Leja points may be better and show good performance in experiments. (Caliari & Ostermann 2009; Tambue, Lord & Geiger 2010)

18 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions

19 Prizing derivatives The Black Scholes model assumes that the value S t of the underlying asset follows a geometric Brownian motion: ds t = µs t dt + σs t dw t. The no-arbitrage principle (no strategy guarantees a profit) implies that price u of derivative given s (price of asset) satisfies u t σ2 s 2 u u + rs s2 s where r is risk-free interest rate. ru = 0. The Heston model assumes that volatility σ is not constant: ds t = µs t dt + ν t S t dw S t, dν t = κ(η ν t ) dt + λ ν t dw ν t. Parameters are such that ν t > 0.

20 Heston PDE Heston model + no-arbitrage principle + magic yields u t = 1 2 νs2 2 u s 2 +ρλνs 2 u ν s λ2 ν 2 u u +rs ν2 s +κ(η ν) u ν ru. Add boundary conditions for modelling European option. Use standard second-order finite differences and incorporate boundary conditions to get ODE of the form u = Au + v 1, u(0) = v 0. The matrix A has size 5100 and non-zero elements. The ODE can be solved by evaluating ϕ-functions. In t Hout (2007) advocates the use of ADI. We compare his ADI schemes with standard code ode15s and our phipm (in matlab).

21 Error vs time maximum error Phipm Crank Nicolson Douglas Craig Hundsdorfer Ode15s cpu time

22 Krylov dimension m and step size τ m x τ t

23 Error estimate (blue) and actual error (red) 4 x error t

24 Conclusion Our matlab code looks good. But it needs more testing. Current work: Rewrite code in compiled languaged (C++). Investigate instability; compare exp l trick with direct method. Extend test to exotic options (American, Asian, barrier). Other work shows our code can be used in exponential integrators to solve semi-linear PDEs. Excellent with spectral discretiation, promising for FD yielding mildly stiff problems, disappointing for very stiff. Compare with other methods for evaluating ϕ functions, especially Leja point interpolation and also RD-rational approximations (Moret & Novati 2004). For more details, see: Niesen & Wright, A Krylov subspace algorithm for evaluating the ϕ-functions appearing in exponential integrators, arxiv: matlab code is available at my home page.

Exponential integration of large systems of ODEs

Exponential integration of large systems of ODEs Exponential integration of large systems of ODEs Jitse Niesen (University of Leeds) in collaboration with Will Wright (Melbourne University) 23rd Biennial Conference on Numerical Analysis, June 2009 Plan

More information

Exponential integrators and functions of the matrix exponential

Exponential integrators and functions of the matrix exponential Exponential integrators and functions of the matrix exponential Paul Matthews, Stephen Cox, Hala Ashi and Linda Cummings School of Mathematical Sciences, University of Nottingham, UK Introduction to exponential

More information

Linear algebra for exponential integrators

Linear algebra for exponential integrators Linear algebra for exponential integrators Antti Koskela KTH Royal Institute of Technology Beräkningsmatematikcirkus, 28 May 2014 The problem Considered: time integration of stiff semilinear initial value

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

Computing the Action of the Matrix Exponential

Computing the Action of the Matrix Exponential Computing the Action of the Matrix Exponential Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/ Joint work with Awad H. Al-Mohy 16th ILAS

More information

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms 2. First Results and Algorithms Michael Eiermann Institut für Numerische Mathematik und Optimierung Technische

More information

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Antti Koskela KTH Royal Institute of Technology, Lindstedtvägen 25, 10044 Stockholm,

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Multi-Factor Finite Differences

Multi-Factor Finite Differences February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit

More information

Comparison of various methods for computing the action of the matrix exponential

Comparison of various methods for computing the action of the matrix exponential BIT manuscript No. (will be inserted by the editor) Comparison of various methods for computing the action of the matrix exponential Marco Caliari Peter Kandolf Alexander Ostermann Stefan Rainer Received:

More information

Exponential integrators for semilinear parabolic problems

Exponential integrators for semilinear parabolic problems Exponential integrators for semilinear parabolic problems Marlis Hochbruck Heinrich-Heine University Düsseldorf Germany Innsbruck, October 2004 p. Outline Exponential integrators general class of methods

More information

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Co-sponsored School/Workshop on Integrable Systems and Scientific Computing June Exponential integrators for stiff systems

Co-sponsored School/Workshop on Integrable Systems and Scientific Computing June Exponential integrators for stiff systems 2042-2 Co-sponsored School/Workshop on Integrable Systems and Scientific Computing 15-20 June 2009 Exponential integrators for stiff systems Paul Matthews University of Nottingham U.K. e-mail: Paul.Matthews@nottingham.ac.uk

More information

1 Extrapolation: A Hint of Things to Come

1 Extrapolation: A Hint of Things to Come Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.

More information

Lecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically

Lecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically Finite difference and finite element methods Lecture 1 Scope of the course Analysis and implementation of numerical methods for pricing options. Models: Black-Scholes, stochastic volatility, exponential

More information

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14

More information

Kasetsart University Workshop. Multigrid methods: An introduction

Kasetsart University Workshop. Multigrid methods: An introduction Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Exponential Almost Runge-Kutta Methods for Semilinear Problems

Exponential Almost Runge-Kutta Methods for Semilinear Problems Exponential Almost Runge-Kutta Methods for Semilinear Problems John Carroll Eóin O Callaghan 9 January 3 Abstract We present a new class of one-step, multi-value Exponential Integrator (EI) methods referred

More information

The ReLPM Exponential Integrator for FE Discretizations of Advection-Diffusion Equations

The ReLPM Exponential Integrator for FE Discretizations of Advection-Diffusion Equations The ReLPM Exponential Integrator for FE Discretizations of Advection-Diffusion Equations Luca Bergamaschi 1, Marco Caliari 2, and Marco Vianello 3 1 Dip.to di Met. e Mod. Mat. per le Scienze Appl., Università

More information

Department of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Spring Semester 2015

Department of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Spring Semester 2015 Department of Mathematics, University of California, Berkeley YOUR OR 2 DIGIT EXAM NUMBER GRADUATE PRELIMINARY EXAMINATION, Part A Spring Semester 205. Please write your - or 2-digit exam number on this

More information

Some minimization problems

Some minimization problems Week 13: Wednesday, Nov 14 Some minimization problems Last time, we sketched the following two-step strategy for approximating the solution to linear systems via Krylov subspaces: 1. Build a sequence of

More information

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u

More information

Reduction to the associated homogeneous system via a particular solution

Reduction to the associated homogeneous system via a particular solution June PURDUE UNIVERSITY Study Guide for the Credit Exam in (MA 5) Linear Algebra This study guide describes briefly the course materials to be covered in MA 5. In order to be qualified for the credit, one

More information

MATH 5524 MATRIX THEORY Problem Set 4

MATH 5524 MATRIX THEORY Problem Set 4 MATH 5524 MATRIX THEORY Problem Set 4 Posted Tuesday 28 March 217. Due Tuesday 4 April 217. [Corrected 3 April 217.] [Late work is due on Wednesday 5 April.] Complete any four problems, 25 points each.

More information

Exponential multistep methods of Adams-type

Exponential multistep methods of Adams-type Exponential multistep methods of Adams-type Marlis Hochbruck and Alexander Ostermann KARLSRUHE INSTITUTE OF TECHNOLOGY (KIT) 0 KIT University of the State of Baden-Wuerttemberg and National Laboratory

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Jordan normal form notes (version date: 11/21/07)

Jordan normal form notes (version date: 11/21/07) Jordan normal form notes (version date: /2/7) If A has an eigenbasis {u,, u n }, ie a basis made up of eigenvectors, so that Au j = λ j u j, then A is diagonal with respect to that basis To see this, let

More information

Implementation of exponential Rosenbrock-type integrators

Implementation of exponential Rosenbrock-type integrators Implementation of exponential Rosenbrock-type integrators Marco Caliari a,b, Alexander Ostermann b, a Department of Pure and Applied Mathematics, University of Padua, Via Trieste 63, I-35121 Padova, Italy

More information

Numerical Methods for Differential Equations Mathematical and Computational Tools

Numerical Methods for Differential Equations Mathematical and Computational Tools Numerical Methods for Differential Equations Mathematical and Computational Tools Gustaf Söderlind Numerical Analysis, Lund University Contents V4.16 Part 1. Vector norms, matrix norms and logarithmic

More information

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b) Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1

More information

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable) Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :

More information

Solutions Preliminary Examination in Numerical Analysis January, 2017

Solutions Preliminary Examination in Numerical Analysis January, 2017 Solutions Preliminary Examination in Numerical Analysis January, 07 Root Finding The roots are -,0, a) First consider x 0 > Let x n+ = + ε and x n = + δ with δ > 0 The iteration gives 0 < ε δ < 3, which

More information

Rational Krylov methods for linear and nonlinear eigenvalue problems

Rational Krylov methods for linear and nonlinear eigenvalue problems Rational Krylov methods for linear and nonlinear eigenvalue problems Mele Giampaolo mele@mail.dm.unipi.it University of Pisa 7 March 2014 Outline Arnoldi (and its variants) for linear eigenproblems Rational

More information

1 Assignment 1: Nonlinear dynamics (due September

1 Assignment 1: Nonlinear dynamics (due September Assignment : Nonlinear dynamics (due September 4, 28). Consider the ordinary differential equation du/dt = cos(u). Sketch the equilibria and indicate by arrows the increase or decrease of the solutions.

More information

Math 21b Final Exam Thursday, May 15, 2003 Solutions

Math 21b Final Exam Thursday, May 15, 2003 Solutions Math 2b Final Exam Thursday, May 5, 2003 Solutions. (20 points) True or False. No justification is necessary, simply circle T or F for each statement. T F (a) If W is a subspace of R n and x is not in

More information

Accurate evaluation of divided differences for polynomial interpolation of exponential propagators

Accurate evaluation of divided differences for polynomial interpolation of exponential propagators Computing 80, 189 201 (2007) DOI 10.1007/s00607-007-0227-1 Printed in The Netherlands Accurate evaluation of divided differences for polynomial interpolation of exponential propagators M. Caliari, Padua

More information

Preliminary Examination, Numerical Analysis, August 2016

Preliminary Examination, Numerical Analysis, August 2016 Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any

More information

Lecture 4: Numerical solution of ordinary differential equations

Lecture 4: Numerical solution of ordinary differential equations Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor

More information

APPROXIMATION OF THE LINEAR THE BLOCK SHIFT-AND-INVERT KRYLOV SUBSPACE METHOD

APPROXIMATION OF THE LINEAR THE BLOCK SHIFT-AND-INVERT KRYLOV SUBSPACE METHOD Journal of Applied Analysis and Computation Volume 7, Number 4, November 2017, 1402 1416 Website:http://jaac-online.com/ DOI:10.11948/2017085 APPROXIMATION OF THE LINEAR COMBINATION OF ϕ-functions USING

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v

More information

An Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence

An Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence An Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence John Pearson MSc Special Topic Abstract Numerical approximations of

More information

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

Numerical methods for matrix functions

Numerical methods for matrix functions Numerical methods for matrix functions SF2524 - Matrix Computations for Large-scale Systems Lecture 13 Numerical methods for matrix functions 1 / 26 Reading material Lecture notes online Numerical methods

More information

Matrix Functions and their Approximation by. Polynomial methods

Matrix Functions and their Approximation by. Polynomial methods [ 1 / 48 ] University of Cyprus Matrix Functions and their Approximation by Polynomial Methods Stefan Güttel stefan@guettel.com Nicosia, 7th April 2006 Matrix functions Polynomial methods [ 2 / 48 ] University

More information

ANALYSIS OF SOME KRYLOV SUBSPACE APPROXIMATIONS TO THE MATRIX EXPONENTIAL OPERATOR

ANALYSIS OF SOME KRYLOV SUBSPACE APPROXIMATIONS TO THE MATRIX EXPONENTIAL OPERATOR ANALYSIS OF SOME KRYLOV SUBSPACE APPROXIMATIONS TO THE MATRIX EXPONENTIAL OPERATOR Y. SAAD Abstract. In this note we present a theoretical analysis of some Krylov subspace approximations to the matrix

More information

LARGE SPARSE EIGENVALUE PROBLEMS

LARGE SPARSE EIGENVALUE PROBLEMS LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems

More information

Stability of Krylov Subspace Spectral Methods

Stability of Krylov Subspace Spectral Methods Stability of Krylov Subspace Spectral Methods James V. Lambers Department of Energy Resources Engineering Stanford University includes joint work with Patrick Guidotti and Knut Sølna, UC Irvine Margot

More information

Affine Processes. Econometric specifications. Eduardo Rossi. University of Pavia. March 17, 2009

Affine Processes. Econometric specifications. Eduardo Rossi. University of Pavia. March 17, 2009 Affine Processes Econometric specifications Eduardo Rossi University of Pavia March 17, 2009 Eduardo Rossi (University of Pavia) Affine Processes March 17, 2009 1 / 40 Outline 1 Affine Processes 2 Affine

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Chapter 7: Symmetric Matrices and Quadratic Forms

Chapter 7: Symmetric Matrices and Quadratic Forms Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

Span & Linear Independence (Pop Quiz)

Span & Linear Independence (Pop Quiz) Span & Linear Independence (Pop Quiz). Consider the following vectors: v = 2, v 2 = 4 5, v 3 = 3 2, v 4 = Is the set of vectors S = {v, v 2, v 3, v 4 } linearly independent? Solution: Notice that the number

More information

HW2 - Due 01/30. Each answer must be mathematically justified. Don t forget your name.

HW2 - Due 01/30. Each answer must be mathematically justified. Don t forget your name. HW2 - Due 0/30 Each answer must be mathematically justified. Don t forget your name. Problem. Use the row reduction algorithm to find the inverse of the matrix 0 0, 2 3 5 if it exists. Double check your

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

Chapter 6: Orthogonality

Chapter 6: Orthogonality Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS November 8, 203 ANALYTIC FUNCTIONAL CALCULUS RODICA D. COSTIN Contents. The spectral projection theorem. Functional calculus 2.. The spectral projection theorem for self-adjoint matrices 2.2. The spectral

More information

Implicit-explicit exponential integrators

Implicit-explicit exponential integrators Implicit-explicit exponential integrators Bawfeh Kingsley Kometa joint work with Elena Celledoni MaGIC 2011 Finse, March 1-4 1 Introduction Motivation 2 semi-lagrangian Runge-Kutta exponential integrators

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Research CharlieMatters

Research CharlieMatters Research CharlieMatters Van Loan and the Matrix Exponential February 25, 2009 Nick Nick Higham Director Schoolof ofresearch Mathematics The School University of Mathematics of Manchester http://www.maths.manchester.ac.uk/~higham

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called

More information

MATHEMATICS (MATH) Calendar

MATHEMATICS (MATH) Calendar MATHEMATICS (MATH) This is a list of the Mathematics (MATH) courses available at KPU. For information about transfer of credit amongst institutions in B.C. and to see how individual courses transfer, go

More information

Classical iterative methods for linear systems

Classical iterative methods for linear systems Classical iterative methods for linear systems Ed Bueler MATH 615 Numerical Analysis of Differential Equations 27 February 1 March, 2017 Ed Bueler (MATH 615 NADEs) Classical iterative methods for linear

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Bounds and Error Estimates for Nonlinear Eigenvalue Problems

Bounds and Error Estimates for Nonlinear Eigenvalue Problems Bounds and Error Estimates for Nonlinear Eigenvalue Problems D. Bindel Courant Institute for Mathematical Sciences New York Univerity 8 Oct 28 Outline Resonances via nonlinear eigenproblems Sensitivity

More information

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM 33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM (UPDATED MARCH 17, 2018) The final exam will be cumulative, with a bit more weight on more recent material. This outline covers the what we ve done since the

More information

Introduction to numerical simulations for Stochastic ODEs

Introduction to numerical simulations for Stochastic ODEs Introduction to numerical simulations for Stochastic ODEs Xingye Kan Illinois Institute of Technology Department of Applied Mathematics Chicago, IL 60616 August 9, 2010 Outline 1 Preliminaries 2 Numerical

More information

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. Math 5327 Fall 2018 Homework 7 1. For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. 3 1 0 (a) A = 1 2 0 1 1 0 x 3 1 0 Solution: 1 x 2 0

More information

Quizzes for Math 304

Quizzes for Math 304 Quizzes for Math 304 QUIZ. A system of linear equations has augmented matrix 2 4 4 A = 2 0 2 4 3 5 2 a) Write down this system of equations; b) Find the reduced row-echelon form of A; c) What are the pivot

More information

Section 6.4. The Gram Schmidt Process

Section 6.4. The Gram Schmidt Process Section 6.4 The Gram Schmidt Process Motivation The procedures in 6 start with an orthogonal basis {u, u,..., u m}. Find the B-coordinates of a vector x using dot products: x = m i= x u i u i u i u i Find

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C. Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures

More information

MATH 350: Introduction to Computational Mathematics

MATH 350: Introduction to Computational Mathematics MATH 350: Introduction to Computational Mathematics Chapter V: Least Squares Problems Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2011 fasshauer@iit.edu MATH

More information

Malliavin Calculus in Finance

Malliavin Calculus in Finance Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x

More information

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0 LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =

More information

Approximation of functions of large matrices. Part I. Computational aspects. V. Simoncini

Approximation of functions of large matrices. Part I. Computational aspects. V. Simoncini Approximation of functions of large matrices. Part I. Computational aspects V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The problem Given

More information

Lecture 6, Sci. Comp. for DPhil Students

Lecture 6, Sci. Comp. for DPhil Students Lecture 6, Sci. Comp. for DPhil Students Nick Trefethen, Thursday 1.11.18 Today II.3 QR factorization II.4 Computation of the QR factorization II.5 Linear least-squares Handouts Quiz 4 Householder s 4-page

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Lecture 2: Computing functions of dense matrices

Lecture 2: Computing functions of dense matrices Lecture 2: Computing functions of dense matrices Paola Boito and Federico Poloni Università di Pisa Pisa - Hokkaido - Roma2 Summer School Pisa, August 27 - September 8, 2018 Introduction In this lecture

More information

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Accurate approximation of stochastic differential equations

Accurate approximation of stochastic differential equations Accurate approximation of stochastic differential equations Simon J.A. Malham and Anke Wiese (Heriot Watt University, Edinburgh) Birmingham: 6th February 29 Stochastic differential equations dy t = V (y

More information

NUMERICAL MATHEMATICS AND COMPUTING

NUMERICAL MATHEMATICS AND COMPUTING NUMERICAL MATHEMATICS AND COMPUTING Fourth Edition Ward Cheney David Kincaid The University of Texas at Austin 9 Brooks/Cole Publishing Company I(T)P An International Thomson Publishing Company Pacific

More information

Math 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name:

Math 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name: Math 34/84 - Exam Blue Exam Solutions December 4, 8 Instructor: Dr. S. Cooper Name: Read each question carefully. Be sure to show all of your work and not just your final conclusion. You may not use your

More information