Fast Eigenvalue Solutions
|
|
- Charity Bridges
- 6 years ago
- Views:
Transcription
1 Fast Eigenvalue Solutions Techniques! Steepest Descent/Conjugate Gradient! Davidson/Lanczos! Carr-Parrinello PDF Files will be available
2 Where HF/DFT calculations spend time Guess ρ Form H Diagonalize ρ O(N 4 ) Cutoffs O(N 2 ) O(N 3 ) Difficult to reduce: Lanczos or Conjugate Gradient Currently only important if N > 2000 Yes Did ρ Change? No Done
3 Rayleigh-Ritz Variational Principle Can always map a minimization problem into a diagonalization problem, and vice-versa λ 1 = min u, Au u u, u This isn t surprising to QMers, since we normally interchange the two
4 Iterative Eigenvector Optimization Ron already covered this Eigenvalue problem Hx =Ex (H-D)x = (EI-D)x x = (EI-D) -1 (H-D)x Define an iterative approach x (n+1) = (E (n) I-D) -1 (H-D)x (n) E (n) = x (n) Hx (n) /x (n) x (n) Factor out the correction x (n+1) = x (n) + d (n) d (n) = (E (n) I-D) -1 (H-E (n) I)x (n) D is the diagonal of H
5 The Residual Vector d The residual vector d (n) = (E (n) I-D) -1 (H-E (n) I)x (n)! Represents the direction of greatest change for the eigenvector x.! Known as the steepest descent vector (although this term is also used for (H-E (n) I)x (n) We have a direction, but we don t know how far to move.! Use Steepest Descent or Conjugate Gradient optimization
6 Tends to be slow, because you overshoot the minimum in each direction. Steepest descent optimization 1. Find d1 2. Minimize in 1-d along d1 3. Find d2 4. Minimize in 1-d along d2 5. Etc.
7 Conjugate Gradient optimization 1. Find d1 2. Minimize in 1-d along d1 CG is normally must faster than SD 3. Find conjugate direction d2: d2 d1 = 0 4. Minimize along d2 5. Etc.
8 One-dimensional optimization Given a direction d (SD or CG), how do we optimize?! x (n+1) = x (n) cos(θ) + d sin(θ)! Finding optimal theta normally only requires a few points.! Trig functions preserve normality
9 Orbital/band at a time optimization Trial eigenvector for band Calculate steepest descent vector Orthogonalize to all bands Iterate until converged Determine conjugate direction Orthogonalize to present band and normalize Calculate KS Energy for θ Compute optimum θ Compute new eigenvector
10 Davidson Diagonalization Optimize eigenvectors in a guess subspace Also uses residual vector d (n) = (E (n) I-D) -1 (H-E (n) I)x (n)! Rather than CG or SD, use d to augment subspace! Lanczos technique Derived for Configuration Interactions! N ~ 10,000! K = 1 Lanczos methods can be unstable! Davidson normally more stable, because of the subspace projection.
11 Subspace Optimization Rather than solve a problem in a large space (N ~ 10,000) project onto a smaller space (n ~ 50).! Small space has to contain dominant features of large space Example: Water with 1M basis functions! Only have 5 occupied states " 1s orbital on O " 2 O lone pairs " 2 OH bonds! Project onto a space that spans these five orbitals! Add onto this space until converged
12 Davidson #1: Subspace Transform 1. Choose an initial set of guess vectors {b i, i=1,ng} 2. Transform H into subspace spanned by {b i } G = B T HB 3. Compute eigenvectors v and values λ of G Dense technique (ng is small) If B fully spans the desired eigenvector, one of the eigenvalues λ k will be the exact one. Otherwise, we can iteratively optimize B to make it better. How do we do this? Using the idea of the residual vector d that we saw earlier.
13 Davidson #2: Optimizing the subspace 4. Compute the residual vector d k as before: d = (λ k I-D) -1 (G- λ k I)v k 5. Orthogonalize d to all of the {b i }; normalize d. 6. Set b ng+1 = d 7. Form the additional row and column of G G ng+1,i = B T Hb ng+1,i 8. Diagonalize G, as before. 9. If not converged, go to #4.
14 Davidson Timings
15 Timing Caveats These timings represent a best-case scenario!!! 1 eigenvalue, lots of functions! Your mileage may vary!!!
16 Higher roots with Davidson If we want more than one eigenvalue/vector, the guess vectors {b} converged after computing one are normally a good guess for another. Start with the lowest one, compute to highest. Not a good method to compute all eigenvalues/vectors! End up solving the full problem multiple times!
17 Compare/Contrast CG & Davidson Both use a subspace transform! We normally have the previous iteration s eigenvectors, and we may as well use them Both use the residual vector d! CG uses d to compute the conjugate direction! Dav uses d to augment the subspace Conventional wisdom! Davidson is faster! CG is more stable! Many more people use CG than Davidson
18 Caveats 1. Use a dense technique when n < 500 always. dsyevx from LAPACK is very hard to beat! Dense techniques are much more stable than sparse. 2. Keep an eye on nk/n nk is the number of desired roots, N is the size of the matrix. If nk/n > 0.25, probably best to use dense technique. 3. Precondition! Multiplying by the last iteration s eigenvectors is a good form of precondintioning.
19 Pure Lanczos Techniques (Zunger, 94) 1. Start with a random u 1 2. Solve β i+1 u i+1 = Hu i α i u i β i u i-1 α i = <u i H u i > β i determined by normalization condition of u i 3. Diagonalize tridiagonal matrix M M M α β β β β α β β α O
20 Lanczos #2 4. Check to see how many eigenvalues are converged; add vectors u i until the requisite number of states are converged. 5. The k th eigenvector is given by V k = Σ i b ki u k
21 Pure Lanczos notes Lanczos has a reputation of being unstable! Can also produce duplicate eigenvalues/vectors " Must project out! Can become non-orthogonal! Numerical errors often grow during iterations
22 Car-Parrinello Propagator Technique Different approach! Rather than minimizing the problem, propagate solutions Born-Oppenheimer approximation! Nuclei move much faster than electrons! Fix nuclei, minimize electrons Car-Parrinello technique does not solve for the Born- Oppenheimer surface! Propagated electrons can be shown to lie reasonably close to BO surface
23 Scheme for QM-MD 1. Fix the nuclei, solve for the wave function 2. Compute the forces on the nuclei due to the wave function 3. Propagate the nuclei using Verlet-type scheme 4. Go to #1
24 Car-Parrinello Start from observation that diagonalization and energy minimization are interchangeable Rather than minimizing the electron, propagate it in the same way one propagates the nuclear motions! Electrons must have fictitious masses! Treat as dynamic variables! Much faster than diagonalization Can also dynamically damp the electron to obtain Born-Oppenheimer solution
25 References! Iterative minimization techniques for ab initio total energy calculations. Payne, Teter, Allen, Arias, Joannopoulos. RMP 64, 1045 (1992).! Iterative calculation of a few of the lowest eigenvalues and corresponding eigenvectors of large real-symmetric matrices. Davidson. J. Comp. Phys. 17, 87 (1975).! Asymptotic convergence for iterative optimization in electronic structure. Lippert, Sears. PRB 61, (2000).! Large scale electronic structure calculations using the Lanczos method. Wang, Zunger. Comp. Mat. Sci 2, 326 (1994).! Unified approach for MD and DFT Car, Parrinello. PRL 55, 2471 (1985).
Preconditioned Eigenvalue Solvers for electronic structure calculations. Andrew V. Knyazev. Householder Symposium XVI May 26, 2005
1 Preconditioned Eigenvalue Solvers for electronic structure calculations Andrew V. Knyazev Department of Mathematics and Center for Computational Mathematics University of Colorado at Denver Householder
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1
More informationAlgorithms and Computational Aspects of DFT Calculations
Algorithms and Computational Aspects of DFT Calculations Part I Juan Meza and Chao Yang High Performance Computing Research Lawrence Berkeley National Laboratory IMA Tutorial Mathematical and Computational
More informationNotes for CS542G (Iterative Solvers for Linear Systems)
Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,
More informationNotes on Some Methods for Solving Linear Systems
Notes on Some Methods for Solving Linear Systems Dianne P. O Leary, 1983 and 1999 and 2007 September 25, 2007 When the matrix A is symmetric and positive definite, we have a whole new class of algorithms
More informationLinear Algebra Final Exam Study Guide Solutions Fall 2012
. Let A = Given that v = 7 7 67 5 75 78 Linear Algebra Final Exam Study Guide Solutions Fall 5 explain why it is not possible to diagonalize A. is an eigenvector for A and λ = is an eigenvalue for A diagonalize
More informationConjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures
Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures Stanimire Tomov 1, Julien Langou 1, Andrew Canning 2, Lin-Wang Wang 2, and Jack Dongarra 1 1 Innovative
More informationEigenvalue Problems CHAPTER 1 : PRELIMINARIES
Eigenvalue Problems CHAPTER 1 : PRELIMINARIES Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Mathematics TUHH Heinrich Voss Preliminaries Eigenvalue problems 2012 1 / 14
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More information, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are
Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues
More informationSparse matrix methods in quantum chemistry Post-doctorale cursus quantumchemie Han-sur-Lesse, Belgium
Sparse matrix methods in quantum chemistry Post-doctorale cursus quantumchemie Han-sur-Lesse, Belgium Dr. ir. Gerrit C. Groenenboom 14 june - 18 june, 1993 Last revised: May 11, 2005 Contents 1 Introduction
More informationCourse Notes: Week 4
Course Notes: Week 4 Math 270C: Applied Numerical Linear Algebra 1 Lecture 9: Steepest Descent (4/18/11) The connection with Lanczos iteration and the CG was not originally known. CG was originally derived
More informationAPPLIED NUMERICAL LINEAR ALGEBRA
APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation
More informationConjugate Gradients I: Setup
Conjugate Gradients I: Setup CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Conjugate Gradients I: Setup 1 / 22 Time for Gaussian Elimination
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationProject: Vibration of Diatomic Molecules
Project: Vibration of Diatomic Molecules Objective: Find the vibration energies and the corresponding vibrational wavefunctions of diatomic molecules H 2 and I 2 using the Morse potential. equired Numerical
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large
More informationMath Linear Algebra Final Exam Review Sheet
Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationThe amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.
AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationEnergy and Forces in DFT
Energy and Forces in DFT Total Energy as a function of nuclear positions {R} E tot ({R}) = E DF T ({R}) + E II ({R}) (1) where E DF T ({R}) = DFT energy calculated for the ground-state density charge-density
More informationLecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.
Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix
More informationConjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)
Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps
More informationTight-Binding Model of Electronic Structures
Tight-Binding Model of Electronic Structures Consider a collection of N atoms. The electronic structure of this system refers to its electronic wave function and the description of how it is related to
More informationJust solve this: macroscopic Maxwell s equations
ust solve this: macroscopic Maxwell s equations Faraday: E = B t constitutive equations (here, linear media): Ampere: H = D t + (nonzero frequency) Gauss: D = ρ B = 0 D = ε E B = µ H electric permittivity
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationIs there life after the Lanczos method? What is LOBPCG?
1 Is there life after the Lanczos method? What is LOBPCG? Andrew V. Knyazev Department of Mathematics and Center for Computational Mathematics University of Colorado at Denver SIAM ALA Meeting, July 17,
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More information3: Density Functional Theory
The Nuts and Bolts of First-Principles Simulation 3: Density Functional Theory CASTEP Developers Group with support from the ESF ψ k Network Density functional theory Mike Gillan, University College London
More informationAb initio Molecular Dynamics Born Oppenheimer and beyond
Ab initio Molecular Dynamics Born Oppenheimer and beyond Reminder, reliability of MD MD trajectories are chaotic (exponential divergence with respect to initial conditions), BUT... With a good integrator
More informationTopics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems
Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate
More informationConjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
More informationAb initio molecular dynamics. Simone Piccinin CNR-IOM DEMOCRITOS Trieste, Italy. Bangalore, 04 September 2014
Ab initio molecular dynamics Simone Piccinin CNR-IOM DEMOCRITOS Trieste, Italy Bangalore, 04 September 2014 What is MD? 1) Liquid 4) Dye/TiO2/electrolyte 2) Liquids 3) Solvated protein 5) Solid to liquid
More informationContents. Preface... xi. Introduction...
Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism
More informationCollege of William & Mary Department of Computer Science
Technical Report WM-CS-2009-06 College of William & Mary Department of Computer Science WM-CS-2009-06 Extending the eigcg algorithm to non-symmetric Lanczos for linear systems with multiple right-hand
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 20 1 / 20 Overview
More informationPart 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)
Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 21: Sensitivity of Eigenvalues and Eigenvectors; Conjugate Gradient Method Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationMath 1553, Introduction to Linear Algebra
Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level
More informationLecture 1: Systems of linear equations and their solutions
Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications
More informationIterative Diagonalization in Augmented Plane Wave Based Methods in Electronic Structure Calculations
* 0. Manuscript Click here to view linked References 1 1 1 1 0 1 0 1 0 1 0 1 0 1 Iterative Diagonalization in Augmented Plane Wave Based Methods in Electronic Structure Calculations P. Blaha a,, H. Hofstätter
More informationSection 4.5 Eigenvalues of Symmetric Tridiagonal Matrices
Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices Key Terms Symmetric matrix Tridiagonal matrix Orthogonal matrix QR-factorization Rotation matrices (plane rotations) Eigenvalues We will now complete
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More information9.1 Preconditioned Krylov Subspace Methods
Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method The minimization problem We are given a symmetric positive definite matrix R n n and a right hand side vector b R n We want to solve the linear system Find u R n such that
More informationMath 5630: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 2019
Math 563: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 219 hroughout, A R n n is symmetric and positive definite, and b R n. 1 Steepest Descent Method We present the steepest descent
More informationPractice Exam. 2x 1 + 4x 2 + 2x 3 = 4 x 1 + 2x 2 + 3x 3 = 1 2x 1 + 3x 2 + 4x 3 = 5
Practice Exam. Solve the linear system using an augmented matrix. State whether the solution is unique, there are no solutions or whether there are infinitely many solutions. If the solution is unique,
More informationproblem Au = u by constructing an orthonormal basis V k = [v 1 ; : : : ; v k ], at each k th iteration step, and then nding an approximation for the e
A Parallel Solver for Extreme Eigenpairs 1 Leonardo Borges and Suely Oliveira 2 Computer Science Department, Texas A&M University, College Station, TX 77843-3112, USA. Abstract. In this paper a parallel
More informationApplications of Matrix Functions Part III: Quantum Chemistry
Applications of Matrix Functions Part III: Quantum Chemistry Emory University Department of Mathematics and Computer Science Atlanta, GA 30322, USA Prologue The main purpose of this lecture is to present
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationNotes on PCG for Sparse Linear Systems
Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/
More informationLinear Solvers. Andrew Hazel
Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction
More informationLecture 22. r i+1 = b Ax i+1 = b A(x i + α i r i ) =(b Ax i ) α i Ar i = r i α i Ar i
8.409 An Algorithmist s oolkit December, 009 Lecturer: Jonathan Kelner Lecture Last time Last time, we reduced solving sparse systems of linear equations Ax = b where A is symmetric and positive definite
More informationEXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants
EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. Determinants Ex... Let A = 0 4 4 2 0 and B = 0 3 0. (a) Compute 0 0 0 0 A. (b) Compute det(2a 2 B), det(4a + B), det(2(a 3 B 2 )). 0 t Ex..2. For
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More informationPreface to the Second Edition. Preface to the First Edition
n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................
More informationDomain decomposition on different levels of the Jacobi-Davidson method
hapter 5 Domain decomposition on different levels of the Jacobi-Davidson method Abstract Most computational work of Jacobi-Davidson [46], an iterative method suitable for computing solutions of large dimensional
More informationSome definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts
Some definitions Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization M. M. Sussman sussmanm@math.pitt.edu Office Hours: MW 1:45PM-2:45PM, Thack 622 A matrix A is SPD (Symmetric
More informationRitz Value Bounds That Exploit Quasi-Sparsity
Banff p.1 Ritz Value Bounds That Exploit Quasi-Sparsity Ilse Ipsen North Carolina State University Banff p.2 Motivation Quantum Physics Small eigenvalues of large Hamiltonians Quasi-Sparse Eigenvector
More informationBlock Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems
Mitglied der Helmholtz-Gemeinschaft Block Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems Birkbeck University, London, June the 29th 2012 Edoardo Di Napoli Motivation and Goals
More informationMath 308 Practice Final Exam Page and vector y =
Math 308 Practice Final Exam Page Problem : Solving a linear equation 2 0 2 5 Given matrix A = 3 7 0 0 and vector y = 8. 4 0 0 9 (a) Solve Ax = y (if the equation is consistent) and write the general solution
More information1. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal
. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal 3 9 matrix D such that A = P DP, for A =. 3 4 3 (a) P = 4, D =. 3 (b) P = 4, D =. (c) P = 4 8 4, D =. 3 (d) P
More information1 Extrapolation: A Hint of Things to Come
Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.
More informationCOMP6237 Data Mining Covariance, EVD, PCA & SVD. Jonathon Hare
COMP6237 Data Mining Covariance, EVD, PCA & SVD Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and covariance) in terms of
More informationIterative methods for symmetric eigenvalue problems
s Iterative s for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 11, 2008 s 1 The power and its variants Inverse power Rayleigh quotient
More informationConjugate Gradient Method
Conjugate Gradient Method Hung M Phan UMass Lowell April 13, 2017 Throughout, A R n n is symmetric and positive definite, and b R n 1 Steepest Descent Method We present the steepest descent method for
More informationMathematical Physics II
Mathematical Physics II PHYS 50 Solution to Problem Set # In the first two problems you will learn (hopefully) how to use numerical computations to suggest analytic results to you. 1. Fixed BC: Tridiagonal
More informationLARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems
More informationIntroduction to first-principles modelling and CASTEP
to first-principles modelling and Phil Hasnip to + Atomistic Simulations If we know what the bonding in a material is beforehand, then we can often find good expressions for the forces between atoms, e.g.
More informationDesigning Information Devices and Systems I Discussion 4B
Last Updated: 29-2-2 9:56 EECS 6A Spring 29 Designing Information Devices and Systems I Discussion 4B Reference Definitions: Matrices and Linear (In)Dependence We ve seen that the following statements
More information7.2 Steepest Descent and Preconditioning
7.2 Steepest Descent and Preconditioning Descent methods are a broad class of iterative methods for finding solutions of the linear system Ax = b for symmetric positive definite matrix A R n n. Consider
More informationLarge-scale eigenvalue problems
ELE 538B: Mathematics of High-Dimensional Data Large-scale eigenvalue problems Yuxin Chen Princeton University, Fall 208 Outline Power method Lanczos algorithm Eigenvalue problems 4-2 Eigendecomposition
More informationLecture 7: Minimization or maximization of functions (Recipes Chapter 10)
Lecture 7: Minimization or maximization of functions (Recipes Chapter 10) Actively studied subject for several reasons: Commonly encountered problem: e.g. Hamilton s and Lagrange s principles, economics
More informationLARGE SPARSE EIGENVALUE PROBLEMS
LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems
More informationITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS
ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID
More informationIterative Methods for Linear Systems of Equations
Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method
More informationGradient Method Based on Roots of A
Journal of Scientific Computing, Vol. 15, No. 4, 2000 Solving Ax Using a Modified Conjugate Gradient Method Based on Roots of A Paul F. Fischer 1 and Sigal Gottlieb 2 Received January 23, 2001; accepted
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Lecture 5: Numerical Linear Algebra Cho-Jui Hsieh UC Davis April 20, 2017 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationSolutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015
Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See
More informationProblem 1: Solving a linear equation
Math 38 Practice Final Exam ANSWERS Page Problem : Solving a linear equation Given matrix A = 2 2 3 7 4 and vector y = 5 8 9. (a) Solve Ax = y (if the equation is consistent) and write the general solution
More informationElectronic energy optimisation in ONETEP
Electronic energy optimisation in ONETEP Chris-Kriton Skylaris cks@soton.ac.uk 1 Outline 1. Kohn-Sham calculations Direct energy minimisation versus density mixing 2. ONETEP scheme: optimise both the density
More informationAb initio molecular dynamics : BO
School on First Principles Simulations, JNCASR, 2010 Ab initio molecular dynamics : BO Vardha Srinivasan IISER Bhopal Why ab initio MD Free of parametrization and only fundamental constants required. Bond
More informationMath Final December 2006 C. Robinson
Math 285-1 Final December 2006 C. Robinson 2 5 8 5 1 2 0-1 0 1. (21 Points) The matrix A = 1 2 2 3 1 8 3 2 6 has the reduced echelon form U = 0 0 1 2 0 0 0 0 0 1. 2 6 1 0 0 0 0 0 a. Find a basis for the
More informationConjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures
Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures Stanimire Tomov 1, Julien Langou 1, Andrew Canning 2, Lin-Wang Wang 2, and Jack Dongarra 1 1 Innovative
More informationData Mining Lecture 4: Covariance, EVD, PCA & SVD
Data Mining Lecture 4: Covariance, EVD, PCA & SVD Jo Houghton ECS Southampton February 25, 2019 1 / 28 Variance and Covariance - Expectation A random variable takes on different values due to chance The
More informationFFTs in Graphics and Vision. The Laplace Operator
FFTs in Graphics and Vision The Laplace Operator 1 Outline Math Stuff Symmetric/Hermitian Matrices Lagrange Multipliers Diagonalizing Symmetric Matrices The Laplacian Operator 2 Linear Operators Definition:
More informationGEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS
Methods in Geochemistry and Geophysics, 36 GEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS Michael S. ZHDANOV University of Utah Salt Lake City UTAH, U.S.A. 2OO2 ELSEVIER Amsterdam - Boston - London
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationFrom Stationary Methods to Krylov Subspaces
Week 6: Wednesday, Mar 7 From Stationary Methods to Krylov Subspaces Last time, we discussed stationary methods for the iterative solution of linear systems of equations, which can generally be written
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationIndex. for generalized eigenvalue problem, butterfly form, 211
Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More information