Computational Methods. Systems of Linear Equations

Similar documents
Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Linear Algebraic Equations

Next topics: Solving systems of linear equations

The Solution of Linear Systems AX = B

Process Model Formulation and Solution, 3E4

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

JACOBI S ITERATION METHOD

Review Questions REVIEW QUESTIONS 71

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

30.3. LU Decomposition. Introduction. Prerequisites. Learning Outcomes

Algebra C Numerical Linear Algebra Sample Exam Problems

5 Solving Systems of Linear Equations

Gaussian Elimination and Back Substitution

MATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

Scientific Computing

Numerical Methods - Numerical Linear Algebra

Scientific Computing: An Introductory Survey

Direct Methods for Solving Linear Systems. Matrix Factorization

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 2. Systems of Linear Equations

Ax = b. Systems of Linear Equations. Lecture Notes to Accompany. Given m n matrix A and m-vector b, find unknown n-vector x satisfying

Chapter 7 Iterative Techniques in Matrix Algebra

1.Chapter Objectives

CS 323: Numerical Analysis and Computing

LINEAR SYSTEMS (11) Intensive Computation

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

Scientific Computing: Dense Linear Systems

Solving Linear Systems of Equations

2.1 Gaussian Elimination

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

CHAPTER 6. Direct Methods for Solving Linear Systems

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

MATH2210 Notebook 2 Spring 2018

CS412: Lecture #17. Mridul Aanjaneya. March 19, 2015

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

LU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b

Solving Linear Systems of Equations

MTH 464: Computational Linear Algebra

Computational Methods. Least Squares Approximation/Optimization

5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Numerical Linear Algebra

Solution of Linear Equations

MAT 343 Laboratory 3 The LU factorization

Linear Algebra and Matrix Inversion

Linear Systems of n equations for n unknowns

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Pivoting. Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3

Linear System of Equations

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Lecture 9. Errors in solving Linear Systems. J. Chaudhry (Zeb) Department of Mathematics and Statistics University of New Mexico

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

MATH 3511 Lecture 1. Solving Linear Systems 1

Basic Concepts in Linear Algebra

Iterative Methods. Splitting Methods

POLI270 - Linear Algebra

Review of Basic Concepts in Linear Algebra

A Review of Matrix Analysis

MH1200 Final 2014/2015

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

1 GSW Sets of Systems

Matrix decompositions

Introduction to PDEs and Numerical Methods Lecture 7. Solving linear systems

Chapter 2 - Linear Equations

Applied Linear Algebra

Lecture 18 Classical Iterative Methods

Solving Ax = b w/ different b s: LU-Factorization

AM 205: lecture 6. Last time: finished the data fitting topic Today s lecture: numerical linear algebra, LU factorization

4.2 Floating-Point Numbers

Lecture 7. Gaussian Elimination with Pivoting. David Semeraro. University of Illinois at Urbana-Champaign. February 11, 2014

AM 205: lecture 6. Last time: finished the data fitting topic Today s lecture: numerical linear algebra, LU factorization

Practical Linear Algebra: A Geometry Toolbox

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Matrices and systems of linear equations

Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White

Numerical Linear Algebra

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Numerical Analysis: Solving Systems of Linear Equations

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Computational Economics and Finance

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Gaussian Elimination without/with Pivoting and Cholesky Decomposition

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Linear Equations and Matrix

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Introduction to Numerical Analysis

AMS526: Numerical Analysis I (Numerical Linear Algebra)

5. Direct Methods for Solving Systems of Linear Equations. They are all over the place...

Matrix decompositions

Linear System of Equations

6 Linear Systems of Equations

CSE 160 Lecture 13. Numerical Linear Algebra

Solving Linear Systems

EXAMPLES OF CLASSICAL ITERATIVE METHODS

Scientific Computing: Solving Linear Systems

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,

Transcription:

Computational Methods Systems of Linear Equations Manfred Huber 2010 1

Systems of Equations Often a system model contains multiple variables (parameters) and contains multiple equations Multiple equations arise because problems have multiple outputs and multiple parameters Multiple equations can also arise from multiple measurements This might lead to equations that are not solvable Many iterative solutions to equation solving do not easily extend to solving systems of equations or equations in multiple variables Manfred Huber 2010 2

Systems of Linear Equations Linear equations are a special type which is easier to solve and has analytic solution methods Single linear equation with n variables corresponds to a hyperplane in n+1 dimensional space a 1 x 1 +...+ a n x n = b a T x = b Finding one analytic solution requires only one division Has usually an infinite number of solutions if n is larger than 1 Systems of linear equations consist of multiple LEs Solution to a sytem of linear equations corresponds to the intersection of multiple hyperplanes Manfred Huber 2010 3

Systems of Linear Equations A system of linear equations can be written as a matrix multiplication a 1,1 x 1 +...+ a 1,n x n = b 1 Ax = b Systems of linear equations do not always have a unique solution a m,1 x 1 +...+ a m,n x n = b m If there are too many equations there might be no solution If there are too few equations then the system might have multiple solutions Manfred Huber 2010 4

Solving Linear Systems To solve a linear system analytically it is typically transformed into a system for which a solution can be easily computed Diagonal system a 1,1 0 0 x = b 0 0 a n,n x 1 = b 1 a 1,1 x n = b n a n,n Triangular systems a 1,1 a 1,n x = b 0 0 a n,n x n = b n a n,n x n 1 = b n 1 a n 1,n x n a n 1,n 1 Manfred Huber 2010 5

Solving Linear Systems To transform a linear system into a different linear system a number of legal operations can be applied Transformations correspond to premultiplying both sides of the linear system by a nonsingular matrix Ax = b MAx = Mb if M is not singular Useful transformations: Permutation: Swaps 2 rows (equations) Row scaling: Scales each row by a scalar Row addition: Subtracts a row from another row Manfred Huber 2010 6

Solving Linear Systems The most important transformation matrix for transforming a system into triangular form is the elimination matrix which combines row scaling and row subtraction Elementary elimination matrix with pivot a k M k = I m k e T k = I 0 0 a k +1 a k a n ( 0 0 1 0 0) a k Manfred Huber 2010 7

Solving Linear Systems Elementary elimination matrices can be combined into one elimination matrix Elimination matrices are lower triangular and nonsingular The inverse of an elimination matrix simply swaps the sign for the off-diagonal terms M k M l = I m k e k T m l e l T Inverse is lower triangular L k = M k 1 = I + m k e k T Manfred Huber 2010 8

Naïve Gaussian Elimination For systems of n linear equations in n variables, a number of analytic solution methods exist Matrix Inversion x = A 1 b Naïve Gaussian elimination Reduce system of equations to upper diagonal form and back substitute to compute the values MAx = Mb, MA = U Transformation of the system occurs through elimination Add or subtract one equation from another Multiply equations with a non-zero constant Manfred Huber 2010 9

Naïve Gaussian Elimination Elimination step For every variable x i, starting with i=1 Subtract a j,i /a i,i times equation i from every equation j, j>i Back substitution step Once the elimination is complete back substitution computes the values For every variable x i, starting with i=n Compute x i by solving the i th equation using the previously computed values for x j, j>i Naïve Gaussian elimination fails if any a i,i is 0 Manfred Huber 2010 10

Gaussian Elimination Complexity of naïve Gaussian elimination (in terms of multiplications and additions) Elimination step: Back substitution step: Computation has to be repeated for every b To address the problem with a 0 on the diagonal we have to use additional operations n 1 i=1 (n i)(1+ (n i +1)) = n 3 3 + n 2 2 5 n 6 Swap equations (rows) and variables (columns) i= n ((n i) +1) = n 2 2 + n 2 Manfred Huber 2010 11 1

LU Factorization Gaussian elimination has to be recomputed every time A or b change Often in practical problems we have to solve the same linear system for different result values LU factorization resolves this by explicitly decomposing A into the upper traingular matrix and the inverse of the elimination matrix Only A is transformed through elimination Solving for b, forward and backward substitution are used Forward substitution with L Backward substitution with U Ly = b Ux = y A = LU Manfred Huber 2010 12

LU Factorization Complexity of LU factorization is approximately the same as for Gaussian Elimination Elimination step: n 3 Forward and back substitution step: 3 n 2 2 Only the forward and back substitution step has to be repeated for a new b Both methods (Gaussian Elimination and LU Factorization) are approximately 3 times faster than matrix inversion n 3 Manfred Huber 2010 13

Existence and Uniqueness Existence and uniqueness of solution depends on the equations and the target result value In systems with n equations and n variables: There exists a unique solution iff A is not singular There are infinitely many solutions iff A is singular and b is in the span of There is no solution iff A is singular and b is not in the span of A A is not singular if the following equivalent conditions apply: There are n equations that are not linearly dependent A is invertible, rank(a)=n, det(a) 0 Manfred Huber 2010 14

Error Measures and Norms To measure errors in multi-dimensional spaces, error vectors have to be reduced to scalars Vector and matrix norms allow to do this Properties of vector norms Positive x > 0 if x 0 Scalar multiplication Triangle inequality αx = α x for any scalar α x + y x + y x y x y Manfred Huber 2010 15

Vector Norms Common vector norms: p-norms 1-norm: 2-norm: x p = x 1 = x 2 = p n i=1 n x i=1 i x i n x 2 i=1 i p -norm: x = max i x i P-norms are related x 1 x 2 x for all x R n Manfred Huber 2010 16

Matrix Norms Matrix norms can be defined in terms of a number of properties similar to the vector norm A > 0 if A 0 Positive Scalar multiplication αa = α A for any scalar α Triangle inequality A + B A + B For each vector norm there is a special matrix norm that can be derived from it and has all the properties Ax Induced Matrix norm A = max x 0 x Manfred Huber 2010 17

Induced Matrix Norms Induced vector norms (operator norms) correspond to the maximum scaling the matrix applies to the vector in terms of the specific vector norm For vector p-norms: 1-norm: -norm: A 1 = max j A = max i n a i, j i=1 n a i, j j=1 Induced matrix norms have additional properties AB A B Ax A x Manfred Huber 2010 18

Sensitivity and Conditioning To estimate the sensitivity of solving a system of linear equations we have to calculate the forward and backward errors Forward error: Backward error (residual): (Relative) Condition number: Δx res x b = Δx x x x ˆ = Δx Aˆ x b = A(x + Δx) Ax b res = Ax A 1 res res x = AΔx A x A 1 res x res = A A 1 cond(a) = A A 1 Manfred Huber 2010 19

Sensitivity and Conditioning Properties of condition number cond(a) = for singular A Condition measures the ratio of maximum stretching to maximum shrinking cond(a) 1 cond(a) = cond(a) A A 1 = max x 0 Ax x To compute condition number the norm of the inverse is often approximated as the maximum ratio of a set of solutions A 1 x Ax min x 0 Ax x Manfred Huber 2010 20

Sensitivity and Conditioning Parameter sensitivity: Δx cond(a) ΔA x A Conditioning depends on the relation of the hyperplanes Manfred Huber 2010 21

Stability Naïve Gaussian Elimination and LU Factorization fail when a pivot is 0 While Gaussian elimination and LU Factorization have no truncation error, they introduce rounding error during the elimination and substitution steps Using small pivots (and thus large multiplication factors) can lead to swamping where the the equation being subtracted overwhelms the equation it is subtracted from Non-singular matrix becomes close to singular Algorithms are more stable if they use larger pivots Manfred Huber 2010 22

Partial Pivoting To improve stability, partial pivoting uses row swaps (permutation matrices) to ensure that always the largest remaining entry in a column is used as the pivot For Gaussian Elimination this is directly implemented by applying permutations between elimination steps For LU-Factorization this causes problems Ax = b P -1 is not lower triangular PAx = Pb Row permutations have to be handled separately PA = LU Manfred Huber 2010 23

PA=LU Factorization Permutations are tracked separately from eliminations leading to a factorization of PA PA = LU The solution step is extended to address row permutations Permute the elements in b using P z = Pb Forward substitution using L Ly = z Ux = y Back substitution using U PAx = LUx = Pb Manfred Huber 2010 24

Complete Pivoting Stability can be further improved by ordering the pivots from largest to smallest through column swaps Always use the largest element in the remaining submatrix below row k as the pivot Column swaps correspond to reordering the variables x For Gaussian Elimination this is directly implemented by applying permutations between elimination steps Ax = b APx = b For LU-Factorization row and column swaps have to be tracked separately PAQ = LU Manfred Huber 2010 25

Scaling and Iterative Refinement Scaling of rows and columns can be used to reduce rounding error and thus to improve stability Large differences in coefficients can decrease stability Sometimes scaling can improve the stability Iterative refinement allows to iteratively break down the residual to improve precision Ax = b x 0 r 0 = b Ax 0 Az = r 0 z 0,x 1 = x 0 + z 0 r 1 = r 0 Az 0 Can sometimes lead to improved precision However, residual calculation is sensitive to cancellation Manfred Huber 2010 26

Direct Solution Methods LU Factorization and Gaussian Elimination return a result without truncation error in O(n 3 ) multiplications. Fixed calculation complexity Guaranteed solution for nonsingular matrix A For very large n or for very sparse matrices the complexity of O(n 3 ) can be very high Iterative methods can be used Lower cost per iteration Convergence has to be analyzed Manfred Huber 2010 27

Iterative Solution Methods Fixed point methods allowed for iterative solutions in single equations Iteration in systems of equations is more difficult Convergence properties are more complex Complexity of iterative methods consists of evaluating the system of equations in each iteration, thus O(n 2 ) If approximate solution is known the number of iterations required can be low If A is sparse the complexity per iteration can drop further Manfred Huber 2010 28

Jacobi Method The Jacobi Method defines a fixed point system by rewriting each equation f i (x) to compute x i f i ( x n b i ) = a i, j x j = b i x i = Starting with an initial vector x (0), the next value is calculated by evaluating the equations using the previous iteration s value x (t-1) j=1 j [1..n], j i a i, j x j Is ensured to converge if A is strictly diagonally dominant (i.e. the coefficients on the diagonal are strictly larger than all other coefficients in the corresponding row a i,i Manfred Huber 2010 29

Jacobi Method Jacobi Method can be rewritten as strict fixed point iteration Ax = b (D + L + U) x = b Dx = ( b (L + U) x ) x = D 1 ( b (L + U) x ) Number of iterations required depends on the starting point For sparse matrices the function form is more efficient than the matrix form Manfred Huber 2010 30

Gauss-Seidel Method The Gauss-Seidel method is variation of the Jacobi method in which for each equation the most recent estimate (from the current iteration) is used rather than the result form the previous iteration Ax = b (D + L) x (t ) = b Ux t 1 x (t ) = D 1 ( b Ux (t 1) Lx (t ) ) Has the same convergence conditions as the Jacobi method Usually converges faster than Jacobi method Manfred Huber 2010 31

Successive Over-Relaxation Successive Over-Relaxation (SOR) further increases convergence speed by anticipating future changes and overshooting the iteration point of the Gauss- Seidel method by a relaxation parameter >1 x i t b t = (1 ω)x t 1 i a i, j x j j=1 i + ω For =1 SOR is equal to Gauss-Seidel i 1 a i,i n t 1 a i, j x j= i+1 j Usually converges faster than Gauss-Seidel for appropriate relaxation parameters Manfred Huber 2010 32

Systems of Linear Equations Systems of linear equations can be solved using either direct or iterative methods Direct methods have fixed computation costs (O(n 3 )) and incur no truncation error Iterative methods are variants of fixed point methods and have a smaller per iteration complexity (O(n 2 )) Less complex for large systems in which a good starting point is known Less complex for sparse matrices (i.e. systems of equations where each equation only depends on a subset of the variables Manfred Huber 2010 33