A fast and well-conditioned spectral method: The US method
|
|
- Bertha Foster
- 5 years ago
- Views:
Transcription
1 A fast and well-conditioned spectral method: The US method Alex Townsend University of Oxford Leslie Fox Prize, 24th of June 23 Partially based on: S. Olver & T., A fast and well-conditioned spectral method, to appear in SIAM Review, (23). Alex Townsend, University of Oxford 24th of June 23 /26
2 The problem Problem: Solve linear ordinary differential equations (ODEs) a N (x) d N u dx N + + a (x) du dx + a (x)u = f (x), x [, ] with K linear boundary conditions (Dirichlet, Neumann, u = c, etc.) Assumptions: a,..., a N, f are continuous with bounded variation. The leading term a N (x) is non-zero on the interval. The solution exists and is unique. Main philosophy: Replace a,..., a N, f by polynomial approximations, and solve for a polynomial approximation for u. The spectral method finds the coefficients in a Chebyshev series in O(m 2 n) operations: n u(x) α k T k (x), k= T k (x) = cos(k cos x). Alex Townsend, University of Oxford 24th of June 23 /26
3 Overview 6 4 degree(u) = 55 Conventional wisdom A comparison First-order ODEs Higher-order ODEs Stability and convergence Fast linear algebra PDEs Conclusion y u(x) x x Alex Townsend, University of Oxford 24th of June 23 2/26
4 Quotes about spectral methods It is known that matrices generated by spectral methods can be dense and ill-conditioned. [Chen 25] The idea behind spectral methods is to take this process to the limit, at least in principle, and work with a differentiation formula of infinite order and infinite bandwidth, i.e., a dense matrix. [Trefethen 2] The best advice is still this: use pseudospectral (collocation) methods instead of spectral (coefficient), and use Fourier series and Chebyshev polynomials in preference to more exotic functions. [Boyd 23] The Ultraspherical spectral method (US method) can result in almost banded, well-conditioned linear systems. Alex Townsend, University of Oxford 24th of June 23 3/26
5 Chebyshev spectral methods u (x) + cos(x)u(x) =, Collocation Coefficients Condition number u(±) =. US method Density Matrix structure Condition number Problem sizes Operation count Collocation Dense O(n2N ) n 5, O(n3 ) Alex Townsend, University of Oxford Coefficient Banded from below O(n2N ) n 5, O(mn2 ) 24th of June 23 US method Banded + rank-k O(n2(D ) ) n 7, O(m2 n) 4/26
6 First-order differential equations Differentiation Operator Multiplication Operator u (x) + a(x)u(x) = f(x) (D + SM[a]) u = Sf D : T C () M[a] : T T S : T C () β Hypergeometric functions Jacobi with w(x)=( x) α (+x) β T L Differentiation C () Conversion Jacobi Ultraspherical Hypergeometric functions α Alex Townsend, University of Oxford 24th of June 23 5/26
7 First-order operators Differentiation operator: D : T C (), [DLMF] { dt k dx = kc () k, k, D =, k =, Conversion operator: S : T C (), [DLMF] ( ) 2 C () k C () k 2, k 2, T k = 2 C (), k =, C (), k =, Multiplication operator: M[a] : T T, [DLMF] T j T k = 2 ( T j k + T j+k ), j, k S = Alex Townsend, University of Oxford 24th of June 23 6/26
8 First-order multiplication operator T j T k = 2 T j k + 2 T j+k 2a a a 2 a M[a] = a 2a a a a 2 a 2a a.. + a a 2 a 3 a a 2 a 3 a 4 a 5... a 3 a 2 a 2a.. a 3 a 4 a 5 a }{{}}{{} Toeplitz Hankel + rank- Multiplication is not a dense operator in finite precision. It is m-banded: a(x) = a k T k (x) = k= m ã k T k (x) + O(ɛ), k= where ã k are aliased Chebyshev coefficients. Alex Townsend, University of Oxford 24th of June 23 7/26
9 Imposing boundary conditions Dirichlet: u( ) = c: B = (,,,,..., ), T k ( ) = ( ) k. Neumann: u () = c: B = (,, 4, 9,..., ), T k () = k 2. Other conditions: u = c, u() = c, or u() + u (π/6) = c. Imposing boundary conditions by boundary bordering: u ( ) (x)+a(x)u(x) = f, B = Bu = c, D + SM[a] ( c f ). The finite section can be done exactly by projection P n = (I n, ): ( P n DP T n BP T n + P n SP T n+mp n+m M[a]P T n ) = ( c P n f ). Alex Townsend, University of Oxford 24th of June 23 8/26
10 First-order examples Example u (x) + x 3 u(x) = sin(2,x 2 ), u( ) =. The exact solution is u(x) = e x4 4 ( x e t4 4 ) sin(2,t 2 )dt. u(x) degree(u) = 239 time = 5.5s x N = ultraop(@(x,u) ); N.lbc = ; u = N \ f; Adaptively selects the discretisation size. Forms a chebfun object [Chebfun V4.2]. ũ u =.5 5. Alex Townsend, University of Oxford 24th of June 23 9/26
11 First-order examples Example 2 u (x) + u(x) =, u( ) =. + 5,x 2 The exact solution with a = 5, is ( u(x) = exp tan ( ax) + tan ( ) a). a.5 5 u(x) degree(u) = x Absolute error 5 Old chebop Preconditioned collocation US method x Alex Townsend, University of Oxford 24th of June 23 /26
12 Higher-order linear ODEs Higher-order diff operators: D λ : T C (λ), [DLMF] λ times dc (λ) { k 2λC (λ+) {}}{ = k, k, D λ = 2 λ (λ )! λ dx, k =, λ Higher-order conversion operators: S λ : C (λ) C (λ+), [DLMF] ( ) λ C (λ+) λ+k k C (λ+) k 2, k 2, λ λ+2 k = λ λ+ C (λ+) λ, k =, S λ = λ λ+ λ+3. C (λ+)., k =,..... C (λ) Multiplication operators: M λ [a] : C λ C λ, M [a] : C () C (), M [a] = Toeplitz + Hankel. β Jacobi with w(x)=( x) α (+x) β Differentiation Conversion Jacobi Ultraspherical α T L Alex Townsend, University of Oxford 24th of June 23 /26 C () C (2)
13 Construction of ultraspherical multiplication operators M λ [a] : C (λ) C (λ) represents multiplication with a(x) in C (λ). If a(x) = m j= a j C (λ) j (x), then it can be shown that M λ [a] is m-banded with (M λ [a]) j,k = k s=max(,k j) where (to prevent overflow issues) we have cs λ (j, k) = j + k + λ 2s s j + k + λ s s t= a 2s+j k c λ s (k, 2s + j k), j, k, t= j s λ + t + t j s 2λ + j + k 2s + t λ + j + k 2s + t t= t= λ + t + t k s + + t k s + λ + t. For efficiency, we use a recurence relation to generate c λ s (j, k). Alex Townsend, University of Oxford 24th of June 23 2/26
14 A high-order example a N (x) d N u dx N + an (x) d N u dx N + + a (x)u = f (x), Bu = c, ( B ) M N [a N ]D N + S N M N [a N ]D N + + S N S M [a ] Example 3: High-order ODE ( = u () (x) + cosh(x)u (8) (x) + cos(x)u (2) (x) + x 2 u(x) = u(±) =, u (±) =, u (k) (±) =, k = 2, 3, 4. c S N S f ).5 u(x) u n+ u n ( ) (ũ(x) + ũ( x)) 2 2 = x n Alex Townsend, University of Oxford 24th of June 23 3/26
15 Airy equation and regularity preserving Example 4 Consider Airy s equation for ɛ >, ) ( ) ɛu (x) xu(x) =, u( ) = Ai ( 3 /ɛ 3, u() = Ai /ɛ. Condition Number ɛ = 9 ɛ = 4 ɛ = n x 4 Relative error Relative error in derivative at x=/2 Ai () Ai (2) Ai () n The US method can accurately compute high derivatives of functions. For a different approach, see [Bornemann 2]. Alex Townsend, University of Oxford 24th of June 23 4/26
16 Stability when K = N Assumption: Number of boundary conditions = Differential order. Definition: Higher-order l 2 λ -norms For λ N, l 2 λ C is the Banach space with norm u 2 l = u 2 k 2 ( + k) 2λ <. λ k= Right diagonal preconditioner: R = 2 N (N )! I N N N+.... Alex Townsend, University of Oxford 24th of June 23 5/26
17 Stability when K = N Recall the notation: Solve Lu = f, with Bu = c, B : l 2 D C K, bounded, D. Theorem: Condition number is bounded with n ( B Let : l L) λ+ l2 λ be an invertible operator for some λ D. Then, as n, A n R n l 2 λ = O(), (A n R n ) l 2 λ = O(), where R n = P T n RP n, A n = P T n ( B L) P n. Proof uses integral reformulation ideas: R integrates the ODE to form a Fredholm operator, i.e., ( B R = I + K, K = compact. L) Alex Townsend, University of Oxford 24th of June 23 6/26
18 Convergence with K = N Theorem: Computed solution converges in high-order norms ( ) B Suppose f l 2 λ N+ for some λ D, and that : l L 2 λ+ l2 λ is an invertible operator. Define ( ) u n = A c n P n, S N S f then u P n u n l 2 λ+ C u P n P n u l 2 λ+. The constant, C, does not depend on n. You don t always need the complex plane to compute high derivatives. Relative error 5 5 Relative error in derivative at x=/2 Ai () Ai (2) Ai () n Alex Townsend, University of Oxford 24th of June 23 7/26
19 Fast linear algebra: The QTP factorisation Original matrix: After left Givens: After right Givens: Apply a partial factorisation on the left, followed by a partial factorisation on the right. Factorisation: A n = QTP in O(m 2 n) operations. No fill-in: The matrix T contains no more non-zeros than A n. For solving A n x = b: The matrix Q can be immediately applied to b. The matrix P can be stored using Demmel s trick [Demmel 997]. UMFPACK by Davis is faster for n 5, but has observed complexity of, roughly, O(n.25 ) with a very small constant. Alex Townsend, University of Oxford 24th of June 23 8/26
20 Constant coefficient PDEs Consider the constant coefficient PDE N N j i+j u a ij i x j = f (x, y), y a j= i= ij R defined on [a, b] [c, d] with linear boundary conditions satisfying continuity conditions. We discretise as a generalised Sylvester matrix equation k σ j A j XBj T = F, A j R n n, B j R n2 n2, j= where A j and B j are US discretisations, for example, A j = ( ) Alex Townsend, University of Oxford 24th of June 23 9/26
21 Determining the rank of a PDE operator Given a PDE with constant coefficients of the form, L = We can rewrite this as N N j j= i= a ij i+j i x j y, A = (a ij). L = D(y) T AD(x), D(x) = / x.. N / x N. Hence, if A = UΣV T is the truncated SVD of A with Σ R k k then L = D(y) T ( UΣV T ) D(x) = ( D(y) T U ) Σ ( V T D(x) ). This low rank representation for L tells us how to discretize and solve the PDE. Alex Townsend, University of Oxford 24th of June 23 2/26
22 Matrix equation solvers Rank-: A XB T = F. Solve A Y = F, then B X T = Y T. Rank-2: A XB T + A 2XB2 T = F. Generalised Sylvester solver [Jonsson & Kågström, 992]. Rank-k, k 3: Solve n n 2 n n 2 system with QTP factorization. O(n3 n 2) n & n 2 O(n 3 + n3 2 ) O(n n 2 ) Boundary conditions are imposed by carefully removing degrees of freedom in the matrix equation. Automatically forms and solves subproblems, if possible. Corner singularities can be partly resolved by domain subdivision. Alex Townsend, University of Oxford 24th of June 23 2/26
23 Examples Example 5: Helmholtz equation uxx + uyy + 2ω 2 u =, u(±, y ) = f (±, y ), u(x, ±) = f (x, ±), where f (x, y ) = cos(ωx) cos(ωy ). ω = 5 N = chebop2(@(u)... ); N.lbc=f(-,:); u=n\; y.5 Adaptively selects the discretisation size. Forms a chebfun2 object [T. & Trefethen 23]..5.5 x.5 For ω = 2, k u uk2 = For ω = 5 solve takes.27s for n = n2 = 26, for ω = 2 solve takes 32.s for n = n2 = 224. Alex Townsend, University of Oxford 24th of June 23 22/26
24 Conclusions The US method results in almost banded well-conditioned matrices. It requires O(m 2 n) operations to solve an ODE. It can be used as a preconditioner for the collocation method. The US method converges and is numerically stable. It can be extended to a spectral method for solving PDEs on rectangular domains. Alex Townsend, University of Oxford 24th of June 23 23/26
25 Historical example [Orszag 97] Orr-Sommerfeld equation (u 2u + u) 2iu i( x 2 )(u u) = λ(u u), u(±) = u (±) =. Eigenvalues of the Orr Sommerfeld operator imag real Alex Townsend, University of Oxford 24th of June 23 24/26
26 The end Thank you More information: S. Olver & T., A fast and well-conditioned spectral method, to appear in SIAM Review, (23). Alex Townsend, University of Oxford 24th of June 23 25/26
27 References F. Bornemann, Accuracy and stability of computing high-order derivatives of analytic functions by Cauchy integrals, Found. Comput. Math., (2), pp. 63. I. Jonsson & Kågström, Recursive blocked algorithms for solving triangular matrix equations Part II: Two-sided and generalized Sylvester and Lyapunov equations, ACM Trans. Math. Softw., 28 (22), pp S. Olver & T., A fast and well-conditioned spectral method, to appear in SIAM Review, (23). S. Orszag, Accurate solution of the Orr-Sommerfeld stability equation, J. Fluid Mech., 5 (97), pp T. & L. N. Trefethen, An extension of Chebfun to two dimensions, likely to appear in SISC, (23). L. N. Trefethen et al., The Chebfun software system, version 4.2, 2. Alex Townsend, University of Oxford 24th of June 23 26/26
28 Chebyshev polynomials An underappreciated fact: Every Chebyshev series can be rewritten as a palindromic Laurent series. The degree k Chebyshev polynomial (of the first kind) is defined by T k (x) = cos ( k cos x ) = 2 ( z k + z k), x [, ], where z = e ikθ and cos θ = x. T 25 (x).5.5 T 25 (θ) = cos(25θ).5.5 x θ Every Chebyshev series can be written as a palindromic Laurent series: N α k T k (x) = 2 k= N α k z k + α + 2 k= N α k z k, z = e ikθ. k= Alex Townsend, University of Oxford 24th of June 23 27/26
29 Demmel s trick for storing Givens rotations A Givens rotation zeros out one entry of a matrix and we can store the rotation information there. Givens {}}{ p Let s = sin θ and c = cos θ, where θ is rotation angle, then { s sign(c), s < c, p = sign(s) c, s c. To restore (up to a 8 degrees): If p <, then s = p and c = s 2, otherwise c = /p and s = c 2. Alex Townsend, University of Oxford 24th of June 23 28/26
The automatic solution of PDEs using a global spectral method
The automatic solution of PDEs using a global spectral method 2014 CBMS-NSF Conference, 29th June 2014 Alex Townsend PhD student University of Oxford (with Sheehan Olver) Supervised by Nick Trefethen,
More informationChebfun2: Exploring Constant Coefficient PDEs on Rectangles
Chebfun2: Exploring Constant Coefficient PDEs on Rectangles Alex Townsend Oxford University with Sheehan Olver, Sydney University FUN13 workshop, 11th of April 2013 Alex Townsend Constant coefficient PDEs
More informationarxiv: v1 [math.na] 4 Nov 2015
ON WELL-CONDITIONED SPECTRAL COLLOCATION AND SPECTRAL METHODS BY THE INTEGRAL REFORMULATION KUI DU arxiv:529v [mathna 4 Nov 25 Abstract Well-conditioned spectral collocation and spectral methods have recently
More informationN y. N x. y i x j, (1.1)
THE AUTOMATIC SOLUTION OF PARTIAL DIFFERENTIAL EQUATIONS USING A GLOBAL SPECTRAL METHOD ALEX TOWNSEND AND SHEEHAN OLVER Abstract. A spectral method for solving linear partial differential equations (PDEs)
More informationPART IV Spectral Methods
PART IV Spectral Methods Additional References: R. Peyret, Spectral methods for incompressible viscous flow, Springer (2002), B. Mercier, An introduction to the numerical analysis of spectral methods,
More information256 Summary. D n f(x j ) = f j+n f j n 2n x. j n=1. α m n = 2( 1) n (m!) 2 (m n)!(m + n)!. PPW = 2π k x 2 N + 1. i=0?d i,j. N/2} N + 1-dim.
56 Summary High order FD Finite-order finite differences: Points per Wavelength: Number of passes: D n f(x j ) = f j+n f j n n x df xj = m α m dx n D n f j j n= α m n = ( ) n (m!) (m n)!(m + n)!. PPW =
More informationFast Structured Spectral Methods
Spectral methods HSS structures Fast algorithms Conclusion Fast Structured Spectral Methods Yingwei Wang Department of Mathematics, Purdue University Joint work with Prof Jie Shen and Prof Jianlin Xia
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationOn solving linear systems arising from Shishkin mesh discretizations
On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October
More informationarxiv: v1 [math.na] 2 Jul 2015
A FAST AND WELL-CONDITIONED SPECTRAL METHOD FOR SINGULAR INTEGRAL EQUATIONS RICHARD MIKAEL SLEVINSKY AND SHEEHAN OLVER arxiv:507.00596v [math.na] Jul 05 Abstract. We develop a spectral method for solving
More informationMulti-Factor Finite Differences
February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit
More informationAn Accurate Fourier-Spectral Solver for Variable Coefficient Elliptic Equations
An Accurate Fourier-Spectral Solver for Variable Coefficient Elliptic Equations Moshe Israeli Computer Science Department, Technion-Israel Institute of Technology, Technion city, Haifa 32000, ISRAEL Alexander
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationLecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.
Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference
More informationGeorgia Tech PHYS 6124 Mathematical Methods of Physics I
Georgia Tech PHYS 612 Mathematical Methods of Physics I Instructor: Predrag Cvitanović Fall semester 2012 Homework Set #5 due October 2, 2012 == show all your work for maximum credit, == put labels, title,
More informationAlex Townsend Cornell University
Continuous analogues of Krylov methods for differential operators Alex Townsend Cornell University townsend@cornell.edu Joint work with: Marc Aurèle Gilles Based on Continuous analogues of Krylov methods
More informationA FAST SOLVER FOR ELLIPTIC EQUATIONS WITH HARMONIC COEFFICIENT APPROXIMATIONS
Proceedings of ALGORITMY 2005 pp. 222 229 A FAST SOLVER FOR ELLIPTIC EQUATIONS WITH HARMONIC COEFFICIENT APPROXIMATIONS ELENA BRAVERMAN, MOSHE ISRAELI, AND ALEXANDER SHERMAN Abstract. Based on a fast subtractional
More informationComputing the common zeros of two bivariate functions via Bézout resultants Colorado State University, 26th September 2013
Work supported by supported by EPSRC grant EP/P505666/1. Computing the common zeros of two bivariate functions via Bézout resultants Colorado State University, 26th September 2013 Alex Townsend PhD student
More information2.3. VECTOR SPACES 25
2.3. VECTOR SPACES 25 2.3 Vector Spaces MATH 294 FALL 982 PRELIM # 3a 2.3. Let C[, ] denote the space of continuous functions defined on the interval [,] (i.e. f(x) is a member of C[, ] if f(x) is continuous
More informationExplicit construction of rectangular differentiation matrices
IMA Journal of Numerical Analysis (206) 36, 68 632 doi:0.093/imanum/drv03 Advance Access publication on April 26, 205 Explicit construction of rectangular differentiation matrices Kuan Xu Mathematical
More informationLinear Solvers. Andrew Hazel
Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction
More informationn 1 f n 1 c 1 n+1 = c 1 n $ c 1 n 1. After taking logs, this becomes
Root finding: 1 a The points {x n+1, }, {x n, f n }, {x n 1, f n 1 } should be co-linear Say they lie on the line x + y = This gives the relations x n+1 + = x n +f n = x n 1 +f n 1 = Eliminating α and
More informationReview and problem list for Applied Math I
Review and problem list for Applied Math I (This is a first version of a serious review sheet; it may contain errors and it certainly omits a number of topic which were covered in the course. Let me know
More informationPoisson Solvers. William McLean. April 21, Return to Math3301/Math5315 Common Material.
Poisson Solvers William McLean April 21, 2004 Return to Math3301/Math5315 Common Material 1 Introduction Many problems in applied mathematics lead to a partial differential equation of the form a 2 u +
More informationMath 322. Spring 2015 Review Problems for Midterm 2
Linear Algebra: Topic: Linear Independence of vectors. Question. Math 3. Spring Review Problems for Midterm Explain why if A is not square, then either the row vectors or the column vectors of A are linearly
More informationA Numerical Solution of the Lax s 7 th -order KdV Equation by Pseudospectral Method and Darvishi s Preconditioning
Int. J. Contemp. Math. Sciences, Vol. 2, 2007, no. 22, 1097-1106 A Numerical Solution of the Lax s 7 th -order KdV Equation by Pseudospectral Method and Darvishi s Preconditioning M. T. Darvishi a,, S.
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationEfficient Solvers for Stochastic Finite Element Saddle Point Problems
Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite
More informationPseudospectra and Nonnormal Dynamical Systems
Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures
More informationSPECTRAL METHODS: ORTHOGONAL POLYNOMIALS
SPECTRAL METHODS: ORTHOGONAL POLYNOMIALS 31 October, 2007 1 INTRODUCTION 2 ORTHOGONAL POLYNOMIALS Properties of Orthogonal Polynomials 3 GAUSS INTEGRATION Gauss- Radau Integration Gauss -Lobatto Integration
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationIterative solvers for linear equations
Spectral Graph Theory Lecture 23 Iterative solvers for linear equations Daniel A. Spielman November 26, 2018 23.1 Overview In this and the next lecture, I will discuss iterative algorithms for solving
More informationCholesky factorisation of linear systems coming from finite difference approximations of singularly perturbed problems
Cholesky factorisation of linear systems coming from finite difference approximations of singularly perturbed problems Thái Anh Nhan and Niall Madden Abstract We consider the solution of large linear systems
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple
More informationNumerical tensor methods and their applications
Numerical tensor methods and their applications 8 May 2013 All lectures 4 lectures, 2 May, 08:00-10:00: Introduction: ideas, matrix results, history. 7 May, 08:00-10:00: Novel tensor formats (TT, HT, QTT).
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationNumerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018
Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationPartial Differential Equations
M3M3 Partial Differential Equations Solutions to problem sheet 3/4 1* (i) Show that the second order linear differential operators L and M, defined in some domain Ω R n, and given by Mφ = Lφ = j=1 j=1
More informationBlockMatrixComputations and the Singular Value Decomposition. ATaleofTwoIdeas
BlockMatrixComputations and the Singular Value Decomposition ATaleofTwoIdeas Charles F. Van Loan Department of Computer Science Cornell University Supported in part by the NSF contract CCR-9901988. Block
More informationProblem set 3: Solutions Math 207B, Winter Suppose that u(x) is a non-zero solution of the eigenvalue problem. (u ) 2 dx, u 2 dx.
Problem set 3: Solutions Math 27B, Winter 216 1. Suppose that u(x) is a non-zero solution of the eigenvalue problem u = λu < x < 1, u() =, u(1) =. Show that λ = (u ) 2 dx u2 dx. Deduce that every eigenvalue
More informationLecture 4 Orthonormal vectors and QR factorization
Orthonormal vectors and QR factorization 4 1 Lecture 4 Orthonormal vectors and QR factorization EE263 Autumn 2004 orthonormal vectors Gram-Schmidt procedure, QR factorization orthogonal decomposition induced
More information1 Sylvester equations
1 Sylvester equations Notes for 2016-11-02 The Sylvester equation (or the special case of the Lyapunov equation) is a matrix equation of the form AX + XB = C where A R m m, B R n n, B R m n, are known,
More informationFinal Exam May 4, 2016
1 Math 425 / AMCS 525 Dr. DeTurck Final Exam May 4, 2016 You may use your book and notes on this exam. Show your work in the exam book. Work only the problems that correspond to the section that you prepared.
More informationFinal: Solutions Math 118A, Fall 2013
Final: Solutions Math 118A, Fall 2013 1. [20 pts] For each of the following PDEs for u(x, y), give their order and say if they are nonlinear or linear. If they are linear, say if they are homogeneous or
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationMcGill University Department of Mathematics and Statistics. Ph.D. preliminary examination, PART A. PURE AND APPLIED MATHEMATICS Paper BETA
McGill University Department of Mathematics and Statistics Ph.D. preliminary examination, PART A PURE AND APPLIED MATHEMATICS Paper BETA 17 August, 2018 1:00 p.m. - 5:00 p.m. INSTRUCTIONS: (i) This paper
More informationRational Chebyshev pseudospectral method for long-short wave equations
Journal of Physics: Conference Series PAPER OPE ACCESS Rational Chebyshev pseudospectral method for long-short wave equations To cite this article: Zeting Liu and Shujuan Lv 07 J. Phys.: Conf. Ser. 84
More informationIntroduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim
Introduction - Motivation Many phenomena (physical, chemical, biological, etc.) are model by differential equations. Recall the definition of the derivative of f(x) f f(x + h) f(x) (x) = lim. h 0 h Its
More informationCompact symetric bilinear forms
Compact symetric bilinear forms Mihai Mathematics Department UC Santa Barbara IWOTA 2006 IWOTA 2006 Compact forms [1] joint work with: J. Danciger (Stanford) S. Garcia (Pomona
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationFast algorithms based on asymptotic expansions of special functions
Fast algorithms based on asymptotic expansions of special functions Alex Townsend MIT Cornell University, 23rd January 2015, CAM Colloquium Introduction Two trends in the computing era Tabulations Software
More informationMath 671: Tensor Train decomposition methods II
Math 671: Tensor Train decomposition methods II Eduardo Corona 1 1 University of Michigan at Ann Arbor December 13, 2016 Table of Contents 1 What we ve talked about so far: 2 The Tensor Train decomposition
More informationAn adaptive fast multipole boundary element method for the Helmholtz equation
An adaptive fast multipole boundary element method for the Helmholtz equation Vincenzo Mallardo 1, Claudio Alessandri 1, Ferri M.H. Aliabadi 2 1 Department of Architecture, University of Ferrara, Italy
More informationSolution of Linear Equations
Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass
More informationMath 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018
1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationSpectral Collocation Method for Parabolic Partial Differential Equations with Neumann Boundary Conditions
Applied Mathematical Sciences, Vol. 1, 2007, no. 5, 211-218 Spectral Collocation Method for Parabolic Partial Differential Equations with Neumann Boundary Conditions M. Javidi a and A. Golbabai b a Department
More informationMath 671: Tensor Train decomposition methods
Math 671: Eduardo Corona 1 1 University of Michigan at Ann Arbor December 8, 2016 Table of Contents 1 Preliminaries and goal 2 Unfolding matrices for tensorized arrays The Tensor Train decomposition 3
More informationComputing least squares condition numbers on hybrid multicore/gpu systems
Computing least squares condition numbers on hybrid multicore/gpu systems M. Baboulin and J. Dongarra and R. Lacroix Abstract This paper presents an efficient computation for least squares conditioning
More informationLecture 8: Boundary Integral Equations
CBMS Conference on Fast Direct Solvers Dartmouth College June 23 June 27, 2014 Lecture 8: Boundary Integral Equations Gunnar Martinsson The University of Colorado at Boulder Research support by: Consider
More informationClassical iterative methods for linear systems
Classical iterative methods for linear systems Ed Bueler MATH 615 Numerical Analysis of Differential Equations 27 February 1 March, 2017 Ed Bueler (MATH 615 NADEs) Classical iterative methods for linear
More informationIterative solvers for linear equations
Spectral Graph Theory Lecture 17 Iterative solvers for linear equations Daniel A. Spielman October 31, 2012 17.1 About these notes These notes are not necessarily an accurate representation of what happened
More informationCollocation based high dimensional model representation for stochastic partial differential equations
Collocation based high dimensional model representation for stochastic partial differential equations S Adhikari 1 1 Swansea University, UK ECCM 2010: IV European Conference on Computational Mechanics,
More informationContinuous analogues of matrix factorizations NASC seminar, 9th May 2014
Continuous analogues of matrix factorizations NSC seminar, 9th May 2014 lex Townsend DPhil student Mathematical Institute University of Oxford (joint work with Nick Trefethen) Many thanks to Gil Strang,
More informationNORMS ON SPACE OF MATRICES
NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationSecond Order Elliptic PDE
Second Order Elliptic PDE T. Muthukumar tmk@iitk.ac.in December 16, 2014 Contents 1 A Quick Introduction to PDE 1 2 Classification of Second Order PDE 3 3 Linear Second Order Elliptic Operators 4 4 Periodic
More informationS.F. Xu (Department of Mathematics, Peking University, Beijing)
Journal of Computational Mathematics, Vol.14, No.1, 1996, 23 31. A SMALLEST SINGULAR VALUE METHOD FOR SOLVING INVERSE EIGENVALUE PROBLEMS 1) S.F. Xu (Department of Mathematics, Peking University, Beijing)
More informationAn Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence
An Analysis of Five Numerical Methods for Approximating Certain Hypergeometric Functions in Domains within their Radii of Convergence John Pearson MSc Special Topic Abstract Numerical approximations of
More informationNumerical Methods for PDEs
Numerical Methods for PDEs Problems 1. Numerical Differentiation. Find the best approximation to the second drivative d 2 f(x)/dx 2 at x = x you can of a function f(x) using (a) the Taylor series approach
More informationLab 1: Iterative Methods for Solving Linear Systems
Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as
More informationAM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition
AM 205: lecture 8 Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition QR Factorization A matrix A R m n, m n, can be factorized
More informationMath 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam
Math 8, Linear Algebra, Lecture C, Spring 7 Review and Practice Problems for Final Exam. The augmentedmatrix of a linear system has been transformed by row operations into 5 4 8. Determine if the system
More informationWhat is A + B? What is A B? What is AB? What is BA? What is A 2? and B = QUESTION 2. What is the reduced row echelon matrix of A =
STUDENT S COMPANIONS IN BASIC MATH: THE ELEVENTH Matrix Reloaded by Block Buster Presumably you know the first part of matrix story, including its basic operations (addition and multiplication) and row
More informationScientific Computing
Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting
More informationKasetsart University Workshop. Multigrid methods: An introduction
Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available
More informationAN ALGORITHM FOR COMPUTING FUNDAMENTAL SOLUTIONS OF DIFFERENCE OPERATORS
AN ALGORITHM FOR COMPUTING FUNDAMENTAL SOLUTIONS OF DIFFERENCE OPERATORS HENRIK BRANDÉN, AND PER SUNDQVIST Abstract We propose an FFT-based algorithm for computing fundamental solutions of difference operators
More information2 Two-Point Boundary Value Problems
2 Two-Point Boundary Value Problems Another fundamental equation, in addition to the heat eq. and the wave eq., is Poisson s equation: n j=1 2 u x 2 j The unknown is the function u = u(x 1, x 2,..., x
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationCS 143 Linear Algebra Review
CS 143 Linear Algebra Review Stefan Roth September 29, 2003 Introductory Remarks This review does not aim at mathematical rigor very much, but instead at ease of understanding and conciseness. Please see
More informationFourier series on triangles and tetrahedra
Fourier series on triangles and tetrahedra Daan Huybrechs K.U.Leuven Cambridge, Sep 13, 2010 Outline Introduction The unit interval Triangles, tetrahedra and beyond Outline Introduction The topic of this
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationNumerical Computing without Discretization Woes
Numerical Computing without Discretization Woes Alex Townsend Cornell University townsend@cornell.edu give a visionary computational science talk Ben Zhang & Ricardo Baptista Discretization alleviating
More informationLinear Algebra. Session 12
Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)
More informationPH.D. PRELIMINARY EXAMINATION MATHEMATICS
UNIVERSITY OF CALIFORNIA, BERKELEY Dept. of Civil and Environmental Engineering FALL SEMESTER 2014 Structural Engineering, Mechanics and Materials NAME PH.D. PRELIMINARY EXAMINATION MATHEMATICS Problem
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 0
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 0 GENE H GOLUB 1 What is Numerical Analysis? In the 1973 edition of the Webster s New Collegiate Dictionary, numerical analysis is defined to be the
More informationIntroduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Bastian Steder
Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Bastian Steder Reference Book Thrun, Burgard, and Fox: Probabilistic Robotics Vectors Arrays of numbers Vectors represent
More informationMathematical Methods wk 2: Linear Operators
John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm
More informationIterative solvers for linear equations
Spectral Graph Theory Lecture 15 Iterative solvers for linear equations Daniel A. Spielman October 1, 009 15.1 Overview In this and the next lecture, I will discuss iterative algorithms for solving linear
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationElementary Linear Algebra
Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We
More informationA Brief Outline of Math 355
A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More information9. Numerical linear algebra background
Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization
More informationNumerical Experiments for Finding Roots of the Polynomials in Chebyshev Basis
Available at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 12, Issue 2 (December 2017), pp. 988 1001 Applications and Applied Mathematics: An International Journal (AAM) Numerical Experiments
More informationPositive Definite Kernels: Opportunities and Challenges
Positive Definite Kernels: Opportunities and Challenges Michael McCourt Department of Mathematical and Statistical Sciences University of Colorado, Denver CUNY Mathematics Seminar CUNY Graduate College
More informationSystems of Linear ODEs
P a g e 1 Systems of Linear ODEs Systems of ordinary differential equations can be solved in much the same way as discrete dynamical systems if the differential equations are linear. We will focus here
More informationNon-Intrusive Solution of Stochastic and Parametric Equations
Non-Intrusive Solution of Stochastic and Parametric Equations Hermann G. Matthies a Loïc Giraldi b, Alexander Litvinenko c, Dishi Liu d, and Anthony Nouy b a,, Brunswick, Germany b École Centrale de Nantes,
More informationMath 315: Linear Algebra Solutions to Assignment 7
Math 5: Linear Algebra s to Assignment 7 # Find the eigenvalues of the following matrices. (a.) 4 0 0 0 (b.) 0 0 9 5 4. (a.) The characteristic polynomial det(λi A) = (λ )(λ )(λ ), so the eigenvalues are
More information