Krylov methods for the computation of matrix functions
|
|
- Ashley Lester
- 5 years ago
- Views:
Transcription
1 Krylov methods for the computation of matrix functions Jitse Niesen (University of Leeds) in collaboration with Will Wright (Melbourne University) Heriot-Watt University, March 2010
2 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions
3 Matrix functions via series Given a scalar function f : C C with Taylor series f (x) = a 0 + a 1 x a 2x a 3x 3 +, the matrix function f : C n n C n n is defined by f (X ) = a 0 + a 1 X a 2X a 3X 3 +. This talk concentrates on the matrix exponential exp(x ) = n=0 1 n! X n. Matrix functions inherit many properties of the corresponding scalar functions, e.g., solution of X = AX is X (t) = exp(at) X (0). Definition needs to be adapted if scalar function f has singularities.
4 Matrix functions via diagonalization Scalar function: f (x) = a 0 + a 1 x a 2x a 3x 3 + Matrix function: f (X ) = a 0 + a 1 X a 2X a 3X 3 + Matrix function on diagonal matrix reduces to scalar function d f (d 1 ) f 0 d =. 0 f (d 2 ) d n f (d n ) If X can be diagonalized, X = VDV 1, then f (X ) = a n (VDV 1 ) n = n=0 a n VD n V 1 = Vf (D)V 1. n=0 This yields another definition of matrix functions. Can be extended to non-diagonalizable matrices by continuity.
5 Matrix functions via contour integration The third definition of matrix functions is f (X ) = 1 f (z) (X zi ) 1 dz, 2πi where integral is along contour encircling the eigenvalues of X. Equivalent to previous definition by residue theorem; the integrand has poles at the eigenvalues. This definition is theoretically convenient because it always works. References 1. Gantmacher, The Theory of Matrices, Horn & Johnson, Topics in Matrix Analysis, N. Higham, Functions of Matrices, 2008.
6 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions
7 Centrality measures of a network A graph is a collection of nodes, some of which are connected by edges. We want to know which nodes are central. One centrality measure is the degree: the number of edges of a node. This is a local measure; wish to extend this to incorporate global information. Let X be the adjacency matrix of a graph: x ij is 1 if there is an edge between nodes i and j and 0 otherwise. The (i, j) entry of X 2 is k x ikx kj. This counts the number of paths of length 2 from i to j. The (i, i) entry of X n counts the number of n-cycles that i is on. In particular, the (i, i) entry of X 2 is the degree. Estrada & Rodríguez-Velázquez (2005) propose to use the diagonal entries of exp(x ) = n 1 n! X n as a centrality measure.
8 Matrix functions solve differential equations The solution of x = ax, x(0) = x 0 is x(t) = exp(at)x 0. The solution of x = Ax, x(0) = x 0 is x(t) = exp(ta)x 0. The solution of x = ax + b, x(0) = 0 is x(t) = t where ϕ 1 (z) = ( exp(z) 1 ) /z. 0 exp(aτ)b dτ = exp(aτ) 1 b = tϕ 1 (at)b. a The solution of x = Ax + b, x(t) = 0 is x(t) = tϕ 1 (ta)b. The solution of x = Ax + ct, x(0) = 0 is x(t) = t 2 ϕ 2 (ta)c, where ϕ 2 (z) = ( exp(z) 1 z ) /z 2. These results can be combined by superposition.
9 Exponential Euler method Solution of x (t) = Lx(t) + v 0 + tv t2 v 2 +, x(0) = x 0, where x(t), x 0, v 0, v 1,... are vectors and L is a matrix, is x(t) = exp(tl)x 0 + hϕ 1 (tl)v 0 + h 2 ϕ 2 (tl)v 1 + h 3 ϕ 3 (tl)v 2 + Consider x = Lx + N(x) (L = linear, N = nonlinear) Replace the nonlinear term with the constant N(x(0)) N(x(t)) (for small t) and use the results from the previous slide: x(t) = exp(tl)x(0) + ϕ 1 (tl) N(x(0)). This leads to the exponential Euler method x n+1 = exp(tl)x n + ϕ 1 (hl)n(x n ). This method is not affected by stiffness in L. (Certaine 1960)
10 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions
11 Computation using series The series exp(x ) = I + X X X 3 + is one method to compute the matrix exponential. Series converges slowly away from the origin, so combine with scaling and squaring. Set Y = 2 k X, compute exponential, square exp(y ) k times. This uses the identity exp(x ) = (exp( 1 2 X ))2. Further improvement is to use Padé approximation instead of polynomial approximation: exp(x) c 0 + c 1 X + c 2 X 2 + c 3 X 3 d 0 + d 1 X + d 2 X 2 + d 3 X 3 where the coefficients c i and d i are chosen so that the error is as small as possible for small x (in practice, we use more coefficients). Padé with scaling and squaring is the most popular general-purpose method for computing the matrix exponential. (Lawson 1967; Higham 2005; Al-Mohy & Higham 2009)
12 Computing matrix functions by diagonalization If X = VDV 1 then f (X ) = Vf (D)V 1. So compute matrix function by first diagonalizing the matrix. This works well for some matrices, in particular symmetric matrices. However, it fails if X is (close to) non-diagonalizable. Computing matrix functions by integration Use f (X ) = 1 2πi f (z) (X zi ) 1 dz and evaluate contour integral with trapezium rule. (Kassam & Trefethen 2005; Schmelzer & Trefethen 2007) This has its appeal, but is expensive, especially if eigenvalues are far apart or matrix is very non-normal. (Ashi, Cummings & Matthews 2009)
13 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions
14 The idea behind Krylov methods In exponential integrators, and other applications, X is a large matrix. On the other hand, we need not f (X ) but f (X )b. A matrix-free method uses only matrix-vector products. This leads to the Krylov subspace K m (X, b) = span{b, Xb, X 2 b,..., X m 1 b}. This basis is very ill-conditioned, so use Gram Schmidt (here called Arnoldi or Lanczos) to get an orthogonal basis: K m (X, b) = span{v 1, v 2, v 3,..., v m }. Put the vectors v j in an n-by-m matrix V m (where n is size of X ). Basis transformation is encoded in m-by-m matrix H m.
15 The idea behind Krylov methods II Considering the matrices as linear mappings: X is map from big space C n to itself. V m is projection from C n onto small K m (X, b). V T m is extension from K m (X, b) to C n ; V m V T m = Id. H m is projection of action of X on K m (X, b); H m = V m XV T m. Krylov methods use the approximation X Vm T H m V m. In the context of matrix functions, use f (X ) = a k X k a k (Vm T H m V m ) k n=k = k=0 a k Vm T (H m ) k V m = Vm T f (H m )V m. k=0 Krylov approximation replaces big matrix X by small matrix H m.
16 Steps towards a practical method Estimate error in Krylov approximation. Use error estimate to adaptively choose dimension m of Krylov subspace (and to correct approximation). Use recursion relation to combine several terms in one: exp(x )b 0 + ϕ 1 (X )b 1 + ϕ 2 (X )b 2 = w 0 + ϕ 2 (X )w 2. Compute ϕ function using trick ([ ]) X b exp = 0 0 [ exp(x ) ϕ1 (X )b 0 1 Problem: A priori estimate suggests that optimal m is proportional to spectral radius ρ(x ). Thus, introduce time stepping similar to scaling-and-squaring. This is implemented in the matlab code phipm. ].
17 Leja point interpolation Krylov methods project matrix function on Krylov subspace K m (X, b) = span{b, Xb, X 2 b,..., X m 1 b}. They approximate f (X )b by element in K m (X, b). Thus, approximation is of form p(x )b, with p a polynomial. In fact, Krylov method performs polynomial interpolation in Ritz values (eigenvalues of H m ). Good because Ritz values approximate eigenvalues of A. (Saad 1992) But Ritz values may not be the best interpolation points. If bound on eigenvalues of A is known, Leja points may be better and show good performance in experiments. (Caliari & Ostermann 2009; Tambue, Lord & Geiger 2010)
18 Outline Definition of matrix functions via series via diagonalization via contour integration Motivation Centrality measure Exponential integrators Direct methods for computation of matrix exponential Padé approximation with scaling and squaring Krylov method for computation of matrix functions The basic idea Implementation (sketch) Leja point interpolation Experiment: Heston equation for prizing derivatives Conclusions
19 Prizing derivatives The Black Scholes model assumes that the value S t of the underlying asset follows a geometric Brownian motion: ds t = µs t dt + σs t dw t. The no-arbitrage principle (no strategy guarantees a profit) implies that price u of derivative given s (price of asset) satisfies u t σ2 s 2 u u + rs s2 s where r is risk-free interest rate. ru = 0. The Heston model assumes that volatility σ is not constant: ds t = µs t dt + ν t S t dw S t, dν t = κ(η ν t ) dt + λ ν t dw ν t. Parameters are such that ν t > 0.
20 Heston PDE Heston model + no-arbitrage principle + magic yields u t = 1 2 νs2 2 u s 2 +ρλνs 2 u ν s λ2 ν 2 u u +rs ν2 s +κ(η ν) u ν ru. Add boundary conditions for modelling European option. Use standard second-order finite differences and incorporate boundary conditions to get ODE of the form u = Au + v 1, u(0) = v 0. The matrix A has size 5100 and non-zero elements. The ODE can be solved by evaluating ϕ-functions. In t Hout (2007) advocates the use of ADI. We compare his ADI schemes with standard code ode15s and our phipm (in matlab).
21 Error vs time maximum error Phipm Crank Nicolson Douglas Craig Hundsdorfer Ode15s cpu time
22 Krylov dimension m and step size τ m x τ t
23 Error estimate (blue) and actual error (red) 4 x error t
24 Conclusion Our matlab code looks good. But it needs more testing. Current work: Rewrite code in compiled languaged (C++). Investigate instability; compare exp l trick with direct method. Extend test to exotic options (American, Asian, barrier). Other work shows our code can be used in exponential integrators to solve semi-linear PDEs. Excellent with spectral discretiation, promising for FD yielding mildly stiff problems, disappointing for very stiff. Compare with other methods for evaluating ϕ functions, especially Leja point interpolation and also RD-rational approximations (Moret & Novati 2004). For more details, see: Niesen & Wright, A Krylov subspace algorithm for evaluating the ϕ-functions appearing in exponential integrators, arxiv: matlab code is available at my home page.
Exponential integration of large systems of ODEs
Exponential integration of large systems of ODEs Jitse Niesen (University of Leeds) in collaboration with Will Wright (Melbourne University) 23rd Biennial Conference on Numerical Analysis, June 2009 Plan
More informationExponential integrators and functions of the matrix exponential
Exponential integrators and functions of the matrix exponential Paul Matthews, Stephen Cox, Hala Ashi and Linda Cummings School of Mathematical Sciences, University of Nottingham, UK Introduction to exponential
More informationLinear algebra for exponential integrators
Linear algebra for exponential integrators Antti Koskela KTH Royal Institute of Technology Beräkningsmatematikcirkus, 28 May 2014 The problem Considered: time integration of stiff semilinear initial value
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationComputing the Action of the Matrix Exponential
Computing the Action of the Matrix Exponential Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/ Joint work with Awad H. Al-Mohy 16th ILAS
More informationKrylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms
Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms 2. First Results and Algorithms Michael Eiermann Institut für Numerische Mathematik und Optimierung Technische
More informationApproximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method
Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method Antti Koskela KTH Royal Institute of Technology, Lindstedtvägen 25, 10044 Stockholm,
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationMulti-Factor Finite Differences
February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit
More informationComparison of various methods for computing the action of the matrix exponential
BIT manuscript No. (will be inserted by the editor) Comparison of various methods for computing the action of the matrix exponential Marco Caliari Peter Kandolf Alexander Ostermann Stefan Rainer Received:
More informationExponential integrators for semilinear parabolic problems
Exponential integrators for semilinear parabolic problems Marlis Hochbruck Heinrich-Heine University Düsseldorf Germany Innsbruck, October 2004 p. Outline Exponential integrators general class of methods
More informationMATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL
MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationCo-sponsored School/Workshop on Integrable Systems and Scientific Computing June Exponential integrators for stiff systems
2042-2 Co-sponsored School/Workshop on Integrable Systems and Scientific Computing 15-20 June 2009 Exponential integrators for stiff systems Paul Matthews University of Nottingham U.K. e-mail: Paul.Matthews@nottingham.ac.uk
More information1 Extrapolation: A Hint of Things to Come
Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.
More informationLecture 1. Finite difference and finite element methods. Partial differential equations (PDEs) Solving the heat equation numerically
Finite difference and finite element methods Lecture 1 Scope of the course Analysis and implementation of numerical methods for pricing options. Models: Black-Scholes, stochastic volatility, exponential
More informationEigenvalue Problems CHAPTER 1 : PRELIMINARIES
Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14
More informationKasetsart University Workshop. Multigrid methods: An introduction
Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationExponential Almost Runge-Kutta Methods for Semilinear Problems
Exponential Almost Runge-Kutta Methods for Semilinear Problems John Carroll Eóin O Callaghan 9 January 3 Abstract We present a new class of one-step, multi-value Exponential Integrator (EI) methods referred
More informationThe ReLPM Exponential Integrator for FE Discretizations of Advection-Diffusion Equations
The ReLPM Exponential Integrator for FE Discretizations of Advection-Diffusion Equations Luca Bergamaschi 1, Marco Caliari 2, and Marco Vianello 3 1 Dip.to di Met. e Mod. Mat. per le Scienze Appl., Università
More informationDepartment of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Spring Semester 2015
Department of Mathematics, University of California, Berkeley YOUR OR 2 DIGIT EXAM NUMBER GRADUATE PRELIMINARY EXAMINATION, Part A Spring Semester 205. Please write your - or 2-digit exam number on this
More informationSome minimization problems
Week 13: Wednesday, Nov 14 Some minimization problems Last time, we sketched the following two-step strategy for approximating the solution to linear systems via Krylov subspaces: 1. Build a sequence of
More informationChapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs
Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u
More informationReduction to the associated homogeneous system via a particular solution
June PURDUE UNIVERSITY Study Guide for the Credit Exam in (MA 5) Linear Algebra This study guide describes briefly the course materials to be covered in MA 5. In order to be qualified for the credit, one
More informationMATH 5524 MATRIX THEORY Problem Set 4
MATH 5524 MATRIX THEORY Problem Set 4 Posted Tuesday 28 March 217. Due Tuesday 4 April 217. [Corrected 3 April 217.] [Late work is due on Wednesday 5 April.] Complete any four problems, 25 points each.
More informationExponential multistep methods of Adams-type
Exponential multistep methods of Adams-type Marlis Hochbruck and Alexander Ostermann KARLSRUHE INSTITUTE OF TECHNOLOGY (KIT) 0 KIT University of the State of Baden-Wuerttemberg and National Laboratory
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationJordan normal form notes (version date: 11/21/07)
Jordan normal form notes (version date: /2/7) If A has an eigenbasis {u,, u n }, ie a basis made up of eigenvectors, so that Au j = λ j u j, then A is diagonal with respect to that basis To see this, let
More informationImplementation of exponential Rosenbrock-type integrators
Implementation of exponential Rosenbrock-type integrators Marco Caliari a,b, Alexander Ostermann b, a Department of Pure and Applied Mathematics, University of Padua, Via Trieste 63, I-35121 Padova, Italy
More informationNumerical Methods for Differential Equations Mathematical and Computational Tools
Numerical Methods for Differential Equations Mathematical and Computational Tools Gustaf Söderlind Numerical Analysis, Lund University Contents V4.16 Part 1. Vector norms, matrix norms and logarithmic
More informationx x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)
Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1
More informationA nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)
Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :
More informationSolutions Preliminary Examination in Numerical Analysis January, 2017
Solutions Preliminary Examination in Numerical Analysis January, 07 Root Finding The roots are -,0, a) First consider x 0 > Let x n+ = + ε and x n = + δ with δ > 0 The iteration gives 0 < ε δ < 3, which
More informationRational Krylov methods for linear and nonlinear eigenvalue problems
Rational Krylov methods for linear and nonlinear eigenvalue problems Mele Giampaolo mele@mail.dm.unipi.it University of Pisa 7 March 2014 Outline Arnoldi (and its variants) for linear eigenproblems Rational
More information1 Assignment 1: Nonlinear dynamics (due September
Assignment : Nonlinear dynamics (due September 4, 28). Consider the ordinary differential equation du/dt = cos(u). Sketch the equilibria and indicate by arrows the increase or decrease of the solutions.
More informationMath 21b Final Exam Thursday, May 15, 2003 Solutions
Math 2b Final Exam Thursday, May 5, 2003 Solutions. (20 points) True or False. No justification is necessary, simply circle T or F for each statement. T F (a) If W is a subspace of R n and x is not in
More informationAccurate evaluation of divided differences for polynomial interpolation of exponential propagators
Computing 80, 189 201 (2007) DOI 10.1007/s00607-007-0227-1 Printed in The Netherlands Accurate evaluation of divided differences for polynomial interpolation of exponential propagators M. Caliari, Padua
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationLecture 4: Numerical solution of ordinary differential equations
Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor
More informationAPPROXIMATION OF THE LINEAR THE BLOCK SHIFT-AND-INVERT KRYLOV SUBSPACE METHOD
Journal of Applied Analysis and Computation Volume 7, Number 4, November 2017, 1402 1416 Website:http://jaac-online.com/ DOI:10.11948/2017085 APPROXIMATION OF THE LINEAR COMBINATION OF ϕ-functions USING
More informationNotes on PCG for Sparse Linear Systems
Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More informationMATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.
MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v
More informationAn Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence
An Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence John Pearson MSc Special Topic Abstract Numerical approximations of
More informationLARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works
CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The
More information235 Final exam review questions
5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an
More informationNumerical methods for matrix functions
Numerical methods for matrix functions SF2524 - Matrix Computations for Large-scale Systems Lecture 13 Numerical methods for matrix functions 1 / 26 Reading material Lecture notes online Numerical methods
More informationMatrix Functions and their Approximation by. Polynomial methods
[ 1 / 48 ] University of Cyprus Matrix Functions and their Approximation by Polynomial Methods Stefan Güttel stefan@guettel.com Nicosia, 7th April 2006 Matrix functions Polynomial methods [ 2 / 48 ] University
More informationANALYSIS OF SOME KRYLOV SUBSPACE APPROXIMATIONS TO THE MATRIX EXPONENTIAL OPERATOR
ANALYSIS OF SOME KRYLOV SUBSPACE APPROXIMATIONS TO THE MATRIX EXPONENTIAL OPERATOR Y. SAAD Abstract. In this note we present a theoretical analysis of some Krylov subspace approximations to the matrix
More informationLARGE SPARSE EIGENVALUE PROBLEMS
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems
More informationStability of Krylov Subspace Spectral Methods
Stability of Krylov Subspace Spectral Methods James V. Lambers Department of Energy Resources Engineering Stanford University includes joint work with Patrick Guidotti and Knut Sølna, UC Irvine Margot
More informationAffine Processes. Econometric specifications. Eduardo Rossi. University of Pavia. March 17, 2009
Affine Processes Econometric specifications Eduardo Rossi University of Pavia March 17, 2009 Eduardo Rossi (University of Pavia) Affine Processes March 17, 2009 1 / 40 Outline 1 Affine Processes 2 Affine
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationMath 307 Learning Goals. March 23, 2010
Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent
More informationSpan & Linear Independence (Pop Quiz)
Span & Linear Independence (Pop Quiz). Consider the following vectors: v = 2, v 2 = 4 5, v 3 = 3 2, v 4 = Is the set of vectors S = {v, v 2, v 3, v 4 } linearly independent? Solution: Notice that the number
More informationHW2 - Due 01/30. Each answer must be mathematically justified. Don t forget your name.
HW2 - Due 0/30 Each answer must be mathematically justified. Don t forget your name. Problem. Use the row reduction algorithm to find the inverse of the matrix 0 0, 2 3 5 if it exists. Double check your
More informationLab 1: Iterative Methods for Solving Linear Systems
Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationAn Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84
An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationNovember 18, 2013 ANALYTIC FUNCTIONAL CALCULUS
November 8, 203 ANALYTIC FUNCTIONAL CALCULUS RODICA D. COSTIN Contents. The spectral projection theorem. Functional calculus 2.. The spectral projection theorem for self-adjoint matrices 2.2. The spectral
More informationImplicit-explicit exponential integrators
Implicit-explicit exponential integrators Bawfeh Kingsley Kometa joint work with Elena Celledoni MaGIC 2011 Finse, March 1-4 1 Introduction Motivation 2 semi-lagrangian Runge-Kutta exponential integrators
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationResearch CharlieMatters
Research CharlieMatters Van Loan and the Matrix Exponential February 25, 2009 Nick Nick Higham Director Schoolof ofresearch Mathematics The School University of Mathematics of Manchester http://www.maths.manchester.ac.uk/~higham
More informationScientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix
Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationRemark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called
More informationMATHEMATICS (MATH) Calendar
MATHEMATICS (MATH) This is a list of the Mathematics (MATH) courses available at KPU. For information about transfer of credit amongst institutions in B.C. and to see how individual courses transfer, go
More informationClassical iterative methods for linear systems
Classical iterative methods for linear systems Ed Bueler MATH 615 Numerical Analysis of Differential Equations 27 February 1 March, 2017 Ed Bueler (MATH 615 NADEs) Classical iterative methods for linear
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationBounds and Error Estimates for Nonlinear Eigenvalue Problems
Bounds and Error Estimates for Nonlinear Eigenvalue Problems D. Bindel Courant Institute for Mathematical Sciences New York Univerity 8 Oct 28 Outline Resonances via nonlinear eigenproblems Sensitivity
More information33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM
33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM (UPDATED MARCH 17, 2018) The final exam will be cumulative, with a bit more weight on more recent material. This outline covers the what we ve done since the
More informationIntroduction to numerical simulations for Stochastic ODEs
Introduction to numerical simulations for Stochastic ODEs Xingye Kan Illinois Institute of Technology Department of Applied Mathematics Chicago, IL 60616 August 9, 2010 Outline 1 Preliminaries 2 Numerical
More informationHomework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.
Math 5327 Fall 2018 Homework 7 1. For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable. 3 1 0 (a) A = 1 2 0 1 1 0 x 3 1 0 Solution: 1 x 2 0
More informationQuizzes for Math 304
Quizzes for Math 304 QUIZ. A system of linear equations has augmented matrix 2 4 4 A = 2 0 2 4 3 5 2 a) Write down this system of equations; b) Find the reduced row-echelon form of A; c) What are the pivot
More informationSection 6.4. The Gram Schmidt Process
Section 6.4 The Gram Schmidt Process Motivation The procedures in 6 start with an orthogonal basis {u, u,..., u m}. Find the B-coordinates of a vector x using dot products: x = m i= x u i u i u i u i Find
More informationChapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors
Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there
More informationLecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.
Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference
More informationPseudospectra and Nonnormal Dynamical Systems
Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures
More informationMATH 350: Introduction to Computational Mathematics
MATH 350: Introduction to Computational Mathematics Chapter V: Least Squares Problems Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2011 fasshauer@iit.edu MATH
More informationMalliavin Calculus in Finance
Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x
More informationc 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0
LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =
More informationApproximation of functions of large matrices. Part I. Computational aspects. V. Simoncini
Approximation of functions of large matrices. Part I. Computational aspects V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The problem Given
More informationLecture 6, Sci. Comp. for DPhil Students
Lecture 6, Sci. Comp. for DPhil Students Nick Trefethen, Thursday 1.11.18 Today II.3 QR factorization II.4 Computation of the QR factorization II.5 Linear least-squares Handouts Quiz 4 Householder s 4-page
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationLecture 2: Computing functions of dense matrices
Lecture 2: Computing functions of dense matrices Paola Boito and Federico Poloni Università di Pisa Pisa - Hokkaido - Roma2 Summer School Pisa, August 27 - September 8, 2018 Introduction In this lecture
More informationCS 450 Numerical Analysis. Chapter 5: Nonlinear Equations
Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80
More informationAccurate approximation of stochastic differential equations
Accurate approximation of stochastic differential equations Simon J.A. Malham and Anke Wiese (Heriot Watt University, Edinburgh) Birmingham: 6th February 29 Stochastic differential equations dy t = V (y
More informationNUMERICAL MATHEMATICS AND COMPUTING
NUMERICAL MATHEMATICS AND COMPUTING Fourth Edition Ward Cheney David Kincaid The University of Texas at Austin 9 Brooks/Cole Publishing Company I(T)P An International Thomson Publishing Company Pacific
More informationMath 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name:
Math 34/84 - Exam Blue Exam Solutions December 4, 8 Instructor: Dr. S. Cooper Name: Read each question carefully. Be sure to show all of your work and not just your final conclusion. You may not use your
More information