Lab 2 Iterative methods and eigenvalue problems. Introduction. Iterative solution of the soap film problem. Beräkningsvetenskap II/NV2, HT (6)

Size: px
Start display at page:

Download "Lab 2 Iterative methods and eigenvalue problems. Introduction. Iterative solution of the soap film problem. Beräkningsvetenskap II/NV2, HT (6)"

Transcription

1 Beräkningsvetenskap II/NV2, HT (6) Institutionen för informationsteknologi Teknisk databehandling Besöksadress: MIC hus 2, Polacksbacken Lägerhyddsvägen 2 Postadress: Box Uppsala Telefon: (växel) Telefax: Hemsida: Department of Information Technology Scientific Computing Visiting address: MIC bldg 2, Polacksbacken Lägerhyddsvägen 2 Postal address: Box 337 SE Uppsala SWEDEN Telephone: (switch) Telefax: Web page: Lab 2 Iterative methods and eigenvalue problems Ideally, you should work together in groups, 2 3 students/group. At the end of the lab, submit your reflections and/or questions. Start by copying the files from bervet2/ht08/lab2 Introduction In this lab we study the dynamics of a trampoline and show how the linear system that arises in the soap film application can be solved efficiently using iterative methods. Iterative solution of the soap film problem In Lab 1, we solved the soap film problem using FEM. Here, we let the computational domain be Ω = [0 1] [0 1] (the unit square), and solve the problem using finite differences. The PDE is given by Ù(Ü Ý) = 0 (Ü Ý) ¾ Ω (1) Ù(Ü Ý) = Ö(Ü Ý) (Ü Ý) ¾ Ω Let Ü = 1 and Æ Ü+1 Ý = 1 Æ Ý+1. Then the discrete problem is Ù+1 2Ù +Ù 1 2 Ü + Ù +1 2Ù +Ù 1 2 Ý Ù = Ö(Ü Ý ) = 0 1 Æ Ü 1 Æ Ý Ü = 0 1 or Ý = 0 1 By ordering the unknown values Ù for example row by row, we can write down a linear system of equations Ù =. So far, we have solved all linear systems by computing the ÄÍ-factorization of and then computing Ù through forward and backward substitution. Now we will explore the effects of this in terms of memory requirements.

2 2 (6) Compute the matrix through A=soapfilm(Nx,Ny). Suitable values for Æ Ü and Æ Ý at this point are around Run the code mylu to study the sparsity patterns of, Ä, and Í. (In the figures, nz is the number of nonzero elements.) How many nonzero diagonals does have? How many nonzero diagonals do Ä and Í have together? Try to come up with a formula for the number of nonzero diagonals in Ä and Í involving Æ Ü and/or Æ Ý. Now we are going to solve the soap film problem with the boundary function given by Ö(0 Ý) = Ö(1 Ý) = cos(2ý) + 05 and Ö(Ü 0) = Ö(Ü 1) = cos(2ü) + 05 Because of the boundary function being non-smooth we will need a fine grid in order to get an accurate solution. Solve the problem using Æ Ü = Æ Ý = 100 through [A,b,x,y,u]=soapfilm(100,100); What is the size of the matrix and how many nonzero diagonals would Ä and Í have in this case? In many cases it is better to solve sparse linear systems using iterative methods. Typically, an iterative method uses less memory than ÄÍ-factorization, and it is often faster. The simplest of the iterative methods is Jacobi s method. Let be a diagonal matrix with the same values as on the diagonal. Ù = Ù + ( )Ù = Ù = 1 ( )Ù + 1

3 3 (6) From this, we can formulate the iterative method in the following way: Choose an arbitrary initial guess Ù 0 for the solution. Let Å = 1 ( ) and = 1. For = 0 1 repeat Ù+1 = ÅÙ + = Ù +1 Ù until tol A more sophisticated iterative method is GMRES, in which the residual Ö = Ù is minimized in spanö 0 Ö 0 Ö 0 in each iteration. Look at the algorithm for Jacobi s method. Why does the iterative method use less memory than ÄÍ-factorization? Solve the soap film problem with Æ Ü = Æ Ý = 100 using Jacobi s method ([u,it]=jacobi(a,b,tol)) and GMRES ([u,it]=gmres(a,b,tol)). Vary the tolerance and write down the numbers of iterations. Suitable tolerance values can be for example between 0.1 and Plot the number of iterations against the log of the tolerance (semilogx(tol,it)) and compare the results. The mathematics of a trampoline We are going to study eigenmodes of a trampoline. Any displacement state of a trampoline at a given time is a linear combination of the eigenmodes. We state the trampoline equation, Û ØØ (Ü Ý Ø) + 2 Û(Ü Ý Ø) = 0 (Ü Ý) ¾ Ω Ø ¾ [0 Ì ] Ù(Ü Ý Ø) = 0 (Ü Ý) ¾ Ω Ø ¾ [0 Ì ] Ù(Ü Ý Ø) = 0 (Ü Ý) ¾ Ω Ø ¾ [0 Ì ] Ù(Ü Ý 0) = 0 (Ü Ý) ¾ Ω Ù Ø (Ü Ý 0) = Ú 0 (Ü Ý) (Ü Ý) ¾ Ω where is the density and is the flexural rigidity of the trampoline. Further Ù(Ü Ý Ø) is the vertical displacement as a function of space and time, Ù(Ü Ý Ø) = 0 at the boundary Ω means that the vertical displacement is zero, i.e. the trampoline is attached to the frame. The other boundary condition Ù(Ü Ý Ø) = 0 models hinges.

4 4 (6) If we can find the eigenmodes of the differential operator 2 we can easily study the dynamics of the trampoline. We let Û denote eigenmode with corresponding eigenvalue i.e., 2 Û (Ü Ý) = Û (Ü Ý) (2) Any function can be written as a linear combination of eigenmodes (Hilbert- Schmidt Theorem). Here we get, Û(Ü Ý Ø) = ½ =1 sin ¼ ½ Ø Û (Ü Ý) (3) where the coefficients are determined from the second initial condition. It is clear that it is crucial to compute the eigenfunctions Û (Ü Ý) and the eigenvalues. If we let Ú and be the eigenvaules and eigenfunctions of the Laplace operator, Ú = Ú (4) we note that, 2 Ú = ( Ú ) = ( Ú ) = 2 Ú, i.e. Û = Ú and = 2. Now we will study a code that computes the eigenfunctions and eigenvalues to the discrete Laplace operator. Again we use finite differences for the discretization. Given these eigenfunctions and eigenvalues we automatically get the eigenfunctions and eigenvalues to the trampoline as Û = Ú and = 2. Data for the trampoline We now specify data for a trampoline, see Figure 1. We let Ω = [0 426] [0 213] (Ñ), = 940 (Ñ 3 ), and = À 3 (12(1 2 )), where À = 001 (half thickness in meters), = (Young s modulus in Pascal), = 05 (ideal elasticity). Note that the matrix we study here is just a scaled version of the one used in the iterative part of the Lab. The scaling is due to different sizes of the computational domain. Given this data it remains to compute the eigenfunctions and eigenvalues to equation (4).

5 5 (6) Figur 1: See Run the code trampoline.m with correct dimensions of the domain Ω. Study different eigenmodes of the trampoline. How does the frequency depend on the size of the eigenvalue? Can we see this in the solution (3)? If two kids are jumping at the same time, where should they stand and how should they jump in order to avoid canceling out each others jumps and thereby decrease the amplitude? The m-file trampoline.m uses an eigenvalue solver called eigs, which is a built in function in Matlab. Now you are going to write your own iterative eigenvalue solver using the power method. The power method and inverse iteration The power method is the simplest algorithm for computing the largest eigenvalue of a matrix. Assume that the Ò Ò matrix has the eigenvalues, = 1 Ò and that Ò. Also assume that the eigenvectors Ü, = 1 Ò are linearly independent. This means that any vector Þ can be written as Þ = È Ò =1 Ü, and Þ = È Ò =1 Ü = È Ò =1 Ü. Repeated multiplication with leads to Þ = Ò =1 Ü = 1 ¼ 1 Ü Ü Ò Ò 1 Ü Ò ½

6 6 (6) This means that Þ will approach the eigenvector associated with the largest eigenvalue 1 as ½. In practise we also need to normalize in each iteration to avoid very large or small entries in the eigenvector (remember an eigenvector is only defined up to a constant). Since we in this application are interested in the smallest eigenvalue rather than the largest, we will apply the power method on the inverse of. Remember that if we assume that is non-singular. Then Ü = Ü Ü = 1 Ü 1 Ü = 1 Ü That is, the eigenvectors stay the same, but the eigenvalues are inverted. Hence, the eigenvector corresponding to the largest eigenvalue of 1 is same as the one corresponding to the smallest eigenvalue of. Applying the power method to the inverse in order to compute the smallest eigenvalue is called inverse iteration. Choose an arbitrary initial guess Þ 0 for the eigenvector. For = 0 1 repeat Þ Ý = Þ 2 Solve Þ +1 = Ý 1 = Ý À Þ +1 until 1 tol The vectors Ý converge to Ü Ò and is an approximation of Ò. Note that a linear system of equations with the matrix must be solved in each iteration. Write a function [lambda,x]=myinvit(a) that applies inverse iteration to the matrix (that you get as output data from trampoline.m) and returns the smallest eigenvalue Ò and the corresponding eigenvector Ü Ò. Compare the results with trampoline.m. You can also check your results against the code invit.m.

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

Automatic Control III (Reglerteknik III) fall Nonlinear systems, Part 3

Automatic Control III (Reglerteknik III) fall Nonlinear systems, Part 3 Automatic Control III (Reglerteknik III) fall 20 4. Nonlinear systems, Part 3 (Chapter 4) Hans Norlander Systems and Control Department of Information Technology Uppsala University OSCILLATIONS AND DESCRIBING

More information

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,

More information

Finding small factors of integers. Speed of the number-field sieve. D. J. Bernstein University of Illinois at Chicago

Finding small factors of integers. Speed of the number-field sieve. D. J. Bernstein University of Illinois at Chicago The number-field sieve Finding small factors of integers Speed of the number-field sieve D. J. Bernstein University of Illinois at Chicago Prelude: finding denominators 87366 22322444 in R. Easily compute

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Domain decomposition on different levels of the Jacobi-Davidson method

Domain decomposition on different levels of the Jacobi-Davidson method hapter 5 Domain decomposition on different levels of the Jacobi-Davidson method Abstract Most computational work of Jacobi-Davidson [46], an iterative method suitable for computing solutions of large dimensional

More information

Kasetsart University Workshop. Multigrid methods: An introduction

Kasetsart University Workshop. Multigrid methods: An introduction Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 4: Iterative Methods PD

More information

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD Mathematical Question we are interested in answering numerically How to solve the following linear system for x Ax = b? where A is an n n invertible matrix and b is vector of length n. Notation: x denote

More information

F O R SOCI AL WORK RESE ARCH

F O R SOCI AL WORK RESE ARCH 7 TH EUROPE AN CONFERENCE F O R SOCI AL WORK RESE ARCH C h a l l e n g e s i n s o c i a l w o r k r e s e a r c h c o n f l i c t s, b a r r i e r s a n d p o s s i b i l i t i e s i n r e l a t i o n

More information

Reduction of Finite Element Models of Complex Mechanical Components

Reduction of Finite Element Models of Complex Mechanical Components Reduction of Finite Element Models of Complex Mechanical Components Håkan Jakobsson Research Assistant hakan.jakobsson@math.umu.se Mats G. Larson Professor Applied Mathematics mats.larson@math.umu.se Department

More information

Fast Fourier Transform Solvers and Preconditioners for Quadratic Spline Collocation

Fast Fourier Transform Solvers and Preconditioners for Quadratic Spline Collocation Fast Fourier Transform Solvers and Preconditioners for Quadratic Spline Collocation Christina C. Christara and Kit Sun Ng Department of Computer Science University of Toronto Toronto, Ontario M5S 3G4,

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Decompositions, numerical aspects Gerard Sleijpen and Martin van Gijzen September 27, 2017 1 Delft University of Technology Program Lecture 2 LU-decomposition Basic algorithm Cost

More information

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects Numerical Linear Algebra Decompositions, numerical aspects Program Lecture 2 LU-decomposition Basic algorithm Cost Stability Pivoting Cholesky decomposition Sparse matrices and reorderings Gerard Sleijpen

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

Final exam: Automatic Control II (Reglerteknik II, 1TT495)

Final exam: Automatic Control II (Reglerteknik II, 1TT495) Uppsala University Department of Information Technology Systems and Control Professor Torsten Söderström Final exam: Automatic Control II (Reglerteknik II, TT495) Date: October 6, Responsible examiner:

More information

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations Sparse Linear Systems Iterative Methods for Sparse Linear Systems Matrix Computations and Applications, Lecture C11 Fredrik Bengzon, Robert Söderlund We consider the problem of solving the linear system

More information

Solving Ax = b, an overview. Program

Solving Ax = b, an overview. Program Numerical Linear Algebra Improving iterative solvers: preconditioning, deflation, numerical software and parallelisation Gerard Sleijpen and Martin van Gijzen November 29, 27 Solving Ax = b, an overview

More information

Numerical Programming I (for CSE)

Numerical Programming I (for CSE) Technische Universität München WT 1/13 Fakultät für Mathematik Prof. Dr. M. Mehl B. Gatzhammer January 1, 13 Numerical Programming I (for CSE) Tutorial 1: Iterative Methods 1) Relaxation Methods a) Let

More information

LA PRISE DE CALAIS. çoys, çoys, har - dis. çoys, dis. tons, mantz, tons, Gas. c est. à ce. C est à ce. coup, c est à ce

LA PRISE DE CALAIS. çoys, çoys, har - dis. çoys, dis. tons, mantz, tons, Gas. c est. à ce. C est à ce. coup, c est à ce > ƒ? @ Z [ \ _ ' µ `. l 1 2 3 z Æ Ñ 6 = Ð l sl (~131 1606) rn % & +, l r s s, r 7 nr ss r r s s s, r s, r! " # $ s s ( ) r * s, / 0 s, r 4 r r 9;: < 10 r mnz, rz, r ns, 1 s ; j;k ns, q r s { } ~ l r mnz,

More information

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018 Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer 28. (Vector and Matrix Norms) Homework 3 Due: Tuesday, July 3, 28 Show that the l vector norm satisfies the three properties (a) x for x

More information

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University Part 3 Chapter 10 LU Factorization PowerPoints organized by Dr. Michael R. Gustafson II, Duke University All images copyright The McGraw-Hill Companies, Inc. Permission required for reproduction or display.

More information

ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley

ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley Submitteed to the International Conference on Independent Component Analysis and Blind Signal Separation (ICA2) ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA Mark Plumbley Audio & Music Lab Department

More information

Vectors. Teaching Learning Point. Ç, where OP. l m n

Vectors. Teaching Learning Point. Ç, where OP. l m n Vectors 9 Teaching Learning Point l A quantity that has magnitude as well as direction is called is called a vector. l A directed line segment represents a vector and is denoted y AB Å or a Æ. l Position

More information

Math 504 (Fall 2011) 1. (*) Consider the matrices

Math 504 (Fall 2011) 1. (*) Consider the matrices Math 504 (Fall 2011) Instructor: Emre Mengi Study Guide for Weeks 11-14 This homework concerns the following topics. Basic definitions and facts about eigenvalues and eigenvectors (Trefethen&Bau, Lecture

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

UNIQUE FJORDS AND THE ROYAL CAPITALS UNIQUE FJORDS & THE NORTH CAPE & UNIQUE NORTHERN CAPITALS

UNIQUE FJORDS AND THE ROYAL CAPITALS UNIQUE FJORDS & THE NORTH CAPE & UNIQUE NORTHERN CAPITALS Q J j,. Y j, q.. Q J & j,. & x x. Q x q. ø. 2019 :. q - j Q J & 11 Y j,.. j,, q j q. : 10 x. 3 x - 1..,,. 1-10 ( ). / 2-10. : 02-06.19-12.06.19 23.06.19-03.07.19 30.06.19-10.07.19 07.07.19-17.07.19 14.07.19-24.07.19

More information

Poisson Equation in 2D

Poisson Equation in 2D A Parallel Strategy Department of Mathematics and Statistics McMaster University March 31, 2010 Outline Introduction 1 Introduction Motivation Discretization Iterative Methods 2 Additive Schwarz Method

More information

Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods

Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods Solving Ill-Posed Cauchy Problems in Three Space Dimensions using Krylov Methods Lars Eldén Department of Mathematics Linköping University, Sweden Joint work with Valeria Simoncini February 21 Lars Eldén

More information

Optimal Control of PDEs

Optimal Control of PDEs Optimal Control of PDEs Suzanne Lenhart University of Tennessee, Knoville Department of Mathematics Lecture1 p.1/36 Outline 1. Idea of diffusion PDE 2. Motivating Eample 3. Big picture of optimal control

More information

Algebraic Multigrid Preconditioners for Computing Stationary Distributions of Markov Processes

Algebraic Multigrid Preconditioners for Computing Stationary Distributions of Markov Processes Algebraic Multigrid Preconditioners for Computing Stationary Distributions of Markov Processes Elena Virnik, TU BERLIN Algebraic Multigrid Preconditioners for Computing Stationary Distributions of Markov

More information

Differentiating Functions & Expressions - Edexcel Past Exam Questions

Differentiating Functions & Expressions - Edexcel Past Exam Questions - Edecel Past Eam Questions. (a) Differentiate with respect to (i) sin + sec, (ii) { + ln ()}. 5-0 + 9 Given that y =, ¹, ( -) 8 (b) show that = ( -). (6) June 05 Q. f() = e ln, > 0. (a) Differentiate

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 1: Introduction to Multigrid 1 12/02/09 MG02.prz Error Smoothing 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 Initial Solution=-Error 0 10 20 30 40 50 60 70 80 90 100 DCT:

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Domain decomposition for the Jacobi-Davidson method: practical strategies

Domain decomposition for the Jacobi-Davidson method: practical strategies Chapter 4 Domain decomposition for the Jacobi-Davidson method: practical strategies Abstract The Jacobi-Davidson method is an iterative method for the computation of solutions of large eigenvalue problems.

More information

JACOBI S ITERATION METHOD

JACOBI S ITERATION METHOD ITERATION METHODS These are methods which compute a sequence of progressively accurate iterates to approximate the solution of Ax = b. We need such methods for solving many large linear systems. Sometimes

More information

Solving linear equations with Gaussian Elimination (I)

Solving linear equations with Gaussian Elimination (I) Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

9. Iterative Methods for Large Linear Systems

9. Iterative Methods for Large Linear Systems EE507 - Computational Techniques for EE Jitkomut Songsiri 9. Iterative Methods for Large Linear Systems introduction splitting method Jacobi method Gauss-Seidel method successive overrelaxation (SOR) 9-1

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Solving PDEs with Multigrid Methods p.1

Solving PDEs with Multigrid Methods p.1 Solving PDEs with Multigrid Methods Scott MacLachlan maclachl@colorado.edu Department of Applied Mathematics, University of Colorado at Boulder Solving PDEs with Multigrid Methods p.1 Support and Collaboration

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Sparse Matrix Techniques for MCAO

Sparse Matrix Techniques for MCAO Sparse Matrix Techniques for MCAO Luc Gilles lgilles@mtu.edu Michigan Technological University, ECE Department Brent Ellerbroek bellerbroek@gemini.edu Gemini Observatory Curt Vogel vogel@math.montana.edu

More information

Lecture 3: Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method

More information

Final exam: Computer-controlled systems (Datorbaserad styrning, 1RT450, 1TS250)

Final exam: Computer-controlled systems (Datorbaserad styrning, 1RT450, 1TS250) Uppsala University Department of Information Technology Systems and Control Professor Torsten Söderström Final exam: Computer-controlled systems (Datorbaserad styrning, RT450, TS250) Date: December 9,

More information

4.3 Laplace Transform in Linear System Analysis

4.3 Laplace Transform in Linear System Analysis 4.3 Laplace Transform in Linear System Analysis The main goal in analysis of any dynamic system is to find its response to a given input. The system response in general has two components: zero-state response

More information

1. Fast Iterative Solvers of SLE

1. Fast Iterative Solvers of SLE 1. Fast Iterative Solvers of crucial drawback of solvers discussed so far: they become slower if we discretize more accurate! now: look for possible remedies relaxation: explicit application of the multigrid

More information

Section 4.1 The Power Method

Section 4.1 The Power Method Section 4.1 The Power Method Key terms Dominant eigenvalue Eigenpair Infinity norm and 2-norm Power Method Scaled Power Method Power Method for symmetric matrices The notation used varies a bit from that

More information

Solving Linear Systems

Solving Linear Systems Solving Linear Systems Iterative Solutions Methods Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Linear Systems Fall 2015 1 / 12 Introduction We continue looking how to solve linear systems of

More information

Block vs. Stream cipher

Block vs. Stream cipher Block vs. Stream cipher Idea of a block cipher: partition the text into relatively large (e.g. 128 bits) blocks and encode each block separately. The encoding of each block generally depends on at most

More information

Chapter 12: Iterative Methods

Chapter 12: Iterative Methods ES 40: Scientific and Engineering Computation. Uchechukwu Ofoegbu Temple University Chapter : Iterative Methods ES 40: Scientific and Engineering Computation. Gauss-Seidel Method The Gauss-Seidel method

More information

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Taylor s Theorem Can often approximate a function by a polynomial The error in the approximation

More information

Final exam: Automatic Control II (Reglerteknik II, 1TT495)

Final exam: Automatic Control II (Reglerteknik II, 1TT495) Uppsala University Department of Information Technology Systems and Control Professor Torsten Söderström Final exam: Automatic Control II (Reglerteknik II, TT495) Date: October 22, 2 Responsible examiner:

More information

Outline. 1 Partial Differential Equations. 2 Numerical Methods for PDEs. 3 Sparse Linear Systems

Outline. 1 Partial Differential Equations. 2 Numerical Methods for PDEs. 3 Sparse Linear Systems Outline Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems 1 Partial Differential Equations 2 Numerical Methods for PDEs 3 Sparse Linear Systems Michael T Heath Scientific

More information

T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a. A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r )

T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a. A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r ) v e r. E N G O u t l i n e T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r ) C o n t e n t s : T h i s w o

More information

Notes for CS542G (Iterative Solvers for Linear Systems)

Notes for CS542G (Iterative Solvers for Linear Systems) Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

Eigenvalue Problems Computation and Applications

Eigenvalue Problems Computation and Applications Eigenvalue ProblemsComputation and Applications p. 1/36 Eigenvalue Problems Computation and Applications Che-Rung Lee cherung@gmail.com National Tsing Hua University Eigenvalue ProblemsComputation and

More information

Inexact inverse iteration with preconditioning

Inexact inverse iteration with preconditioning Department of Mathematical Sciences Computational Methods with Applications Harrachov, Czech Republic 24th August 2007 (joint work with M. Robbé and M. Sadkane (Brest)) 1 Introduction 2 Preconditioned

More information

Derivation of the Kalman Filter

Derivation of the Kalman Filter Derivation of the Kalman Filter Kai Borre Danish GPS Center, Denmark Block Matrix Identities The key formulas give the inverse of a 2 by 2 block matrix, assuming T is invertible: T U 1 L M. (1) V W N P

More information

B œ c " " ã B œ c 8 8. such that substituting these values for the B 3 's will make all the equations true

B œ c   ã B œ c 8 8. such that substituting these values for the B 3 's will make all the equations true System of Linear Equations variables Ð unknowns Ñ B" ß B# ß ÞÞÞ ß B8 Æ Æ Æ + B + B ÞÞÞ + B œ, "" " "# # "8 8 " + B + B ÞÞÞ + B œ, #" " ## # #8 8 # ã + B + B ÞÞÞ + B œ, 3" " 3# # 38 8 3 ã + 7" B" + 7# B#

More information

Block-tridiagonal matrices

Block-tridiagonal matrices Block-tridiagonal matrices. p.1/31 Block-tridiagonal matrices - where do these arise? - as a result of a particular mesh-point ordering - as a part of a factorization procedure, for example when we compute

More information

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of Chapter 2 Linear Algebra In this chapter, we study the formal structure that provides the background for quantum mechanics. The basic ideas of the mathematical machinery, linear algebra, are rather simple

More information

1.6: 16, 20, 24, 27, 28

1.6: 16, 20, 24, 27, 28 .6: 6, 2, 24, 27, 28 6) If A is positive definite, then A is positive definite. The proof of the above statement can easily be shown for the following 2 2 matrix, a b A = b c If that matrix is positive

More information

Framework for functional tree simulation applied to 'golden delicious' apple trees

Framework for functional tree simulation applied to 'golden delicious' apple trees Purdue University Purdue e-pubs Open Access Theses Theses and Dissertations Spring 2015 Framework for functional tree simulation applied to 'golden delicious' apple trees Marek Fiser Purdue University

More information

Iterative Methods for Linear Systems

Iterative Methods for Linear Systems Iterative Methods for Linear Systems 1. Introduction: Direct solvers versus iterative solvers In many applications we have to solve a linear system Ax = b with A R n n and b R n given. If n is large the

More information

Name: INSERT YOUR NAME HERE. Due to dropbox by 6pm PDT, Wednesday, December 14, 2011

Name: INSERT YOUR NAME HERE. Due to dropbox by 6pm PDT, Wednesday, December 14, 2011 AMath 584 Name: INSERT YOUR NAME HERE Take-home Final UWNetID: INSERT YOUR NETID Due to dropbox by 6pm PDT, Wednesday, December 14, 2011 The main part of the assignment (Problems 1 3) is worth 80 points.

More information

CLASSICAL ITERATIVE METHODS

CLASSICAL ITERATIVE METHODS CLASSICAL ITERATIVE METHODS LONG CHEN In this notes we discuss classic iterative methods on solving the linear operator equation (1) Au = f, posed on a finite dimensional Hilbert space V = R N equipped

More information

Solution of Linear Equations

Solution of Linear Equations Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass

More information

ETIKA V PROFESII PSYCHOLÓGA

ETIKA V PROFESII PSYCHOLÓGA P r a ž s k á v y s o k á š k o l a p s y c h o s o c i á l n í c h s t u d i í ETIKA V PROFESII PSYCHOLÓGA N a t á l i a S l o b o d n í k o v á v e d ú c i p r á c e : P h D r. M a r t i n S t r o u

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Solution methods for the Cahn-Hilliard equation discretized by conforming and non-conforming finite elements

Solution methods for the Cahn-Hilliard equation discretized by conforming and non-conforming finite elements Solution methods for the Cahn-Hilliard equation discretized by conforming and non-conforming finite elements Petia Boyanova 2, Minh Do-Quang 3, Maya Neytcheva Uppsala University, Box 337, 75 5 Uppsala,

More information

Here are proofs for some of the results about diagonalization that were presented without proof in class.

Here are proofs for some of the results about diagonalization that were presented without proof in class. Suppose E is an 8 8 matrix. In what follows, terms like eigenvectors, eigenvalues, and eigenspaces all refer to the matrix E. Here are proofs for some of the results about diagonalization that were presented

More information

An Introduction to Optimal Control Applied to Disease Models

An Introduction to Optimal Control Applied to Disease Models An Introduction to Optimal Control Applied to Disease Models Suzanne Lenhart University of Tennessee, Knoxville Departments of Mathematics Lecture1 p.1/37 Example Number of cancer cells at time (exponential

More information

MIDTERM 1 - SOLUTIONS

MIDTERM 1 - SOLUTIONS MIDTERM - SOLUTIONS MATH 254 - SUMMER 2002 - KUNIYUKI CHAPTERS, 2, GRADED OUT OF 75 POINTS 2 50 POINTS TOTAL ) Use either Gaussian elimination with back-substitution or Gauss-Jordan elimination to solve

More information

Sierra/SD: A Massively Parallel Finite Element Code for Structural Dynamics and Acoustics Analysis

Sierra/SD: A Massively Parallel Finite Element Code for Structural Dynamics and Acoustics Analysis Sierra/SD: A Massively Parallel Finite Element Code for Structural Dynamics and Acoustics Analysis Jerry W. Rouse, Timothy F. Walsh, Greg D. Tipton, Garth M. Reese Sandia National Labs Albuquerque, NM

More information

Numerical Methods Process Systems Engineering ITERATIVE METHODS. Numerical methods in chemical engineering Edwin Zondervan

Numerical Methods Process Systems Engineering ITERATIVE METHODS. Numerical methods in chemical engineering Edwin Zondervan IERAIVE MEHODS Numerical methods in chemical engineering Edwin Zondervan 1 OVERVIEW Iterative methods for large systems of equations We will solve Laplace s equation for steady state heat conduction LAPLACE

More information

probability of k samples out of J fall in R.

probability of k samples out of J fall in R. Nonparametric Techniques for Density Estimation (DHS Ch. 4) n Introduction n Estimation Procedure n Parzen Window Estimation n Parzen Window Example n K n -Nearest Neighbor Estimation Introduction Suppose

More information

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in 86 Problem Set - Solutions Due Thursday, 29 November 27 at 4 pm in 2-6 Problem : (5=5+5+5) Take any matrix A of the form A = B H CB, where B has full column rank and C is Hermitian and positive-definite

More information

Multigrid absolute value preconditioning

Multigrid absolute value preconditioning Multigrid absolute value preconditioning Eugene Vecharynski 1 Andrew Knyazev 2 (speaker) 1 Department of Computer Science and Engineering University of Minnesota 2 Department of Mathematical and Statistical

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication.

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication. 7-6 Multiplying matrices by diagonal matrices is faster than usual matrix multiplication. The following equations generalize to matrices of any size. Multiplying a matrix from the left by a diagonal matrix

More information

Lab #2 - Two Degrees-of-Freedom Oscillator

Lab #2 - Two Degrees-of-Freedom Oscillator Lab #2 - Two Degrees-of-Freedom Oscillator Last Updated: March 0, 2007 INTRODUCTION The system illustrated in Figure has two degrees-of-freedom. This means that two is the minimum number of coordinates

More information

Solutions Problem Set 8 Math 240, Fall

Solutions Problem Set 8 Math 240, Fall Solutions Problem Set 8 Math 240, Fall 2012 5.6 T/F.2. True. If A is upper or lower diagonal, to make det(a λi) 0, we need product of the main diagonal elements of A λi to be 0, which means λ is one of

More information

Lecture 16 Methods for System of Linear Equations (Linear Systems) Songting Luo. Department of Mathematics Iowa State University

Lecture 16 Methods for System of Linear Equations (Linear Systems) Songting Luo. Department of Mathematics Iowa State University Lecture 16 Methods for System of Linear Equations (Linear Systems) Songting Luo Department of Mathematics Iowa State University MATH 481 Numerical Methods for Differential Equations Songting Luo ( Department

More information

Seminar 6: COUPLED HARMONIC OSCILLATORS

Seminar 6: COUPLED HARMONIC OSCILLATORS Seminar 6: COUPLED HARMONIC OSCILLATORS 1. Lagrangian Equations of Motion Let consider a system consisting of two harmonic oscillators that are coupled together. As a model, we will use two particles attached

More information

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C. Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference

More information

Cheat Sheet for MATH461

Cheat Sheet for MATH461 Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A

More information

Numerical algorithms for one and two target optimal controls

Numerical algorithms for one and two target optimal controls Numerical algorithms for one and two target optimal controls Sung-Sik Kwon Department of Mathematics and Computer Science, North Carolina Central University 80 Fayetteville St. Durham, NC 27707 Email:

More information

Edexcel GCE A Level Maths Further Maths 3 Matrices.

Edexcel GCE A Level Maths Further Maths 3 Matrices. Edexcel GCE A Level Maths Further Maths 3 Matrices. Edited by: K V Kumaran kumarmathsweebly.com kumarmathsweebly.com 2 kumarmathsweebly.com 3 kumarmathsweebly.com 4 kumarmathsweebly.com 5 kumarmathsweebly.com

More information

Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White

Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White Introduction to Simulation - Lecture 2 Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White Thanks to Deepak Ramaswamy, Michal Rewienski, and Karen Veroy Outline Reminder about

More information

CSE 554 Lecture 7: Alignment

CSE 554 Lecture 7: Alignment CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex

More information

Connection equations with stream variables are generated in a model when using the # $ % () operator or the & ' %

Connection equations with stream variables are generated in a model when using the # $ % () operator or the & ' % 7 9 9 7 The two basic variable types in a connector potential (or across) variable and flow (or through) variable are not sufficient to describe in a numerically sound way the bi-directional flow of matter

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Convergence Behavior of Left Preconditioning Techniques for GMRES ECS 231: Large Scale Scientific Computing University of California, Davis Winter

Convergence Behavior of Left Preconditioning Techniques for GMRES ECS 231: Large Scale Scientific Computing University of California, Davis Winter Convergence Behavior of Left Preconditioning Techniques for GMRES ECS 231: Large Scale Scientific Computing University of California, Davis Winter Quarter 2013 March 20, 2013 Joshua Zorn jezorn@ucdavis.edu

More information