On the Stability of LU-Factorizations

Similar documents
. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in

CS412: Lecture #17. Mridul Aanjaneya. March 19, 2015

AMS526: Numerical Analysis I (Numerical Linear Algebra)

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Lecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II)

Linear Algebraic Equations

MODULE 7. where A is an m n real (or complex) matrix. 2) Let K(t, s) be a function of two variables which is continuous on the square [0, 1] [0, 1].

Review. Example 1. Elementary matrices in action: (a) a b c. d e f = g h i. d e f = a b c. a b c. (b) d e f. d e f.

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Section Gaussian Elimination

The Solution of Linear Systems AX = B

2.1 Gaussian Elimination

MAC1105-College Algebra. Chapter 5-Systems of Equations & Matrices

Linear Equations and Matrix

y b where U. matrix inverse A 1 ( L. 1 U 1. L 1 U 13 U 23 U 33 U 13 2 U 12 1

Matrix decompositions

Computational Linear Algebra

CHAPTER 6. Direct Methods for Solving Linear Systems

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4

MATH 3511 Lecture 1. Solving Linear Systems 1

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

Direct Methods for Solving Linear Systems. Matrix Factorization

I-v k e k. (I-e k h kt ) = Stability of Gauss-Huard Elimination for Solving Linear Systems. 1 x 1 x x x x

CS 323: Numerical Analysis and Computing

Draft. Lecture 12 Gaussian Elimination and LU Factorization. MATH 562 Numerical Analysis II. Songting Luo

Solving Linear Systems Using Gaussian Elimination

On the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

1.4 Gaussian Elimination Gaussian elimination: an algorithm for finding a (actually the ) reduced row echelon form of a matrix. A row echelon form

Matrix decompositions

Section 5.6. LU and LDU Factorizations

Elementary maths for GMT

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey

Linear Equations in Linear Algebra

Matrix Algebra. Matrix Algebra. Chapter 8 - S&B

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1

Numerical Linear Algebra

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form:

Hani Mehrpouyan, California State University, Bakersfield. Signals and Systems

Dylan Zwick. Fall Ax=b. EAx=Eb. UxrrrEb

Matrix & Linear Algebra

Math 4377/6308 Advanced Linear Algebra

POLI270 - Linear Algebra

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

MIDTERM 1 - SOLUTIONS

Topics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij

Formula for the inverse matrix. Cramer s rule. Review: 3 3 determinants can be computed expanding by any row or column

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

Lectures on Linear Algebra for IT

Lecture 12: Solving Systems of Linear Equations by Gaussian Elimination

Numerical Linear Algebra

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Lectures on Linear Algebra for IT

Gaussian Elimination and Back Substitution

Scientific Computing

Math 2331 Linear Algebra

7.6 The Inverse of a Square Matrix

Linear Algebra Math 221

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Linear Algebra and Matrix Inversion

Matrices and systems of linear equations

Methods for Solving Linear Systems Part 2

Linear Systems. Math A Bianca Santoro. September 23, 2016

Roundoff Analysis of Gaussian Elimination

SOLVING LINEAR SYSTEMS

Department of Aerospace Engineering AE602 Mathematics for Aerospace Engineers Assignment No. 4

Lecture Note 2: The Gaussian Elimination and LU Decomposition

MAT 2037 LINEAR ALGEBRA I web:

System of Linear Equations

Solving Systems of Linear Equations

Solving Linear Systems of Equations

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics

Example: 2x y + 3z = 1 5y 6z = 0 x + 4z = 7. Definition: Elementary Row Operations. Example: Type I swap rows 1 and 3

Pivoting. Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

CSE 160 Lecture 13. Numerical Linear Algebra

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

Next topics: Solving systems of linear equations

Determinants Chapter 3 of Lay

Matrices and Matrix Algebra.

Review Questions REVIEW QUESTIONS 71

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

4 Elementary matrices, continued

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Math 1314 Week #14 Notes

II. Determinant Functions

EA = I 3 = E = i=1, i k

Relationships Between Planes

This can be accomplished by left matrix multiplication as follows: I

Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations

MATH2210 Notebook 2 Spring 2018

MATRICES. a m,1 a m,n A =

CS513, Spring 2007 Prof. Amos Ron Assignment #5 Solutions Prepared by Houssain Kettani. a mj i,j [2,n] a 11

EBG # 3 Using Gaussian Elimination (Echelon Form) Gaussian Elimination: 0s below the main diagonal

Transcription:

Australian Journal of Basic and Applied Sciences 5(6): 497-503 2011 ISSN 1991-8178 On the Stability of LU-Factorizations R Saneifard and A Asgari Department of Mathematics Oroumieh Branch Islamic Azad University Oroumieh Iran Abstract: The purpose of this paper is to show that the incomplete LU factorizations of the -matrices by using the incomplete LU- factorizations of an H-matrix are at least as stable as the complete LUfactorizations of its comparison matrix We give also some new characterizations of the H-matrices in connection with their incomplete LU-factorizations Key words: Stability; Gaussian elimination; Incomplete LU factorization INTRODUCTION A well-known approach to solving the problem is the classical elimination scheme known as Gaussian elimination The basic idea is to reduce the system to an equivalent upper triangular system so that reduced upper triangular system can be solved easily using the back-substitution algorithm 11 Gaussian Elimination for Solving Ax=b: Consider the problem of solving the linear system of n equations in n unknowns: Or in matrix notation Ax=b where A=(a ij ) and b=(b 1 b 2 b n ) T The reduction process consists of n-1 steps In the following Step 1: At step 1 the unknown x 1 is eliminated from the second through the nth equations This is done by multiplying the first equation by The quantities becomes A (1) x = b (1) where the entries of follows: and adding it to the second through nth equations are called multipliers At the end of step 1 the system Ax=b and those of b (1) are related to the entries of A and b as (Note that are all zero) Corresponding Author: R Saneifard Department of Mathematics Oroumieh Branch Islamic Azad University Oroumieh Iran E-mail: srsaneeifard@yahoocom 497

Aust J Basic & Appl Sci 5(6): 497-503 2011 Step 2: At step 2 x 2 is eliminated from the third through the nth equations of A (1) x = b (1) second by multiplying the equations of A (1) x = b (1) by the multipliers and adding the result to the third through nth equations The system now becomes A (2) x = b (2) whose entries are given as follows: The other entries of A (2) and those of b 2 are the same as those of A (1) and b (1) respectively Step k: At step k the (n-k) multipliers are formed using them x k is eliminated from the (k+1) st through the th equations of A (k-1) x = b (k-1) The entries of A (k) and those of b (k) are given by The other entries of A (k) and b (k) are the same as those of A (k-1) and b (k-1) respectively Step n 1: At the end of the (n-1) st step the reduced matrix A (n-1) is upper triangular and the original vector b is transformed to b (n-1) 2 Definitions and Concepts of Stability: Definition 1: An algorithm Will be called forward stable if the computed solution sense is close to the exact x in some Definition 2: An algorithm is called backward stable if it produces an exact solution to a nearby problem Definition 3: An algorithm for solving Ax = b will be called stable if the computed solution is such that with E and δb small Definition 4: An algorithm is stable for a of matrices C if for every matrix A in C the computed solution by the algorithm is the exact solution of a nearby problem Thus for the Linear system problem Ax = b an algorithm is stable for a class of matrices C if every A0C and for each b it produces a computed solution that satisfies for some E and δb where (A+E) is close to A and b-δb is close to b 498

Aust J Basic & Appl Sci 5(6): 497-503 2011 21 Strong Weak and Mild Stability: Definition 5: An algorithm for solving the linear system problem Ax = b is strongly stable for a class of matrices C if for each A in C the computed solution is the exact solution of a nearby problem and the matrix (A+E) also belongs to C Definition 6: An algorithm is weakly stable for a class of matrices C if for each well conditioned matrix in C the algorithm produces an acceptable accurate solution Definition 7: An algorithm is mildly stable if it produces a solution that is close to the exact solution of a nearby problem 22 Gaussian Elimination without Pivoting: Set A = A (0) Step 1: Find an elementary matrix E 1 such that A (1) = E 1 A has zeros below the (11) entry in the first column That is A (1) has the from Note that it is sufficient only to find E 1 such that Then A (1) = E 1 A will have the preceding from and is the same as the matrix A (1) obtained at the end of step 1 of Algorithm Record the multipliers: Step 2: Find an elementary matrix such that A (2) = E 2 A (1) has zeros below the (22) entry in the second column The matrix E 2 can be constructed as follows First find an elementary matrix Ê 2 of order (n 1) such that Record the multipliers: m 32 m n2 ; i = 3n Then define A (2) = E 2 A (1) will then have zeros below the (22) entry in the second column 499

Aust J Basic & Appl Sci 5(6): 497-503 2011 Note that pre multiplication of A (1) by E 2 does not does not destroy zeros already created in A (1) This matrix A (2) is the same as the matrix A (2) of Algorithm Step k: In general at the kth step an elementary matrix E k is found such that A (k) = E k A (k-1) has zeros below the (kk) entry in the th column E k is computed in two successive steps First an elementary matrix Ê k of order n-k + 1 is constructed such that and then E k is defined as Here I k-1 is the matrix of the first (k-1) rows and columns of the n n identity matrix I The matrix A (k) = E k A (k-1) is the same as the matrix A (k) of Algorithm (*) Record the multipliers: m k+1k m ik ; i = k + 1n Step n 1: At the end of the (n-1) st step the matrix A (n-1) is upper triangular and the same matrix A (n-1) of Algorithm Obtain L and U: Set: U = A (n-1) Then from before we have U = L 1 A Because each E k is a unit lower triangular matrix lower triangular matrix having is along the diagonal so is the matrix L: therefore L 1-1 exists (Note that the product of two triangular matrices One type is a triangular matrix of the same type) Set L = L 1-1 Then the equation U = L 1 A becomes A = LU This factorization of A is known as LU factorization Definition 8: The entries are called pivots and the preceding process of obtaining LU factorization is known as Gaussian elimination without row interchanges It is commonly known as Gaussian elimination without pivoting The process is named after the German mathematician and astronomer Karl Friedrich Gauss (1777-1855) 500

23 Gaussian Elimination with partial pivoting: The process consists of n-1 steps Aust J Basic & Appl Sci 5(6): 497-503 2011 Step 1: Scan the first column of A to identify the largest element in magnitude in that column let is a r1 * Form a permutation matrix P 1 by interchanging the rows 1 and r 1 of the identity matrix and leaving the other rows unchanged * Form P 1 A by interchanging the rows r 1 and 1 of A * Find an elementary lower triangular matrix M 1 such that A (1) = M 1 P 1 A has zeros below the (11) entry on the first column It is sufficient to construct M 1 such that Note that Where and a il refers to the (ij) th entry of the permuted matrix P 1 A Save the multipliers m i1 i = 2n and record the row interchanges Step 2: Scan the second column of A (1) below the first row to identify the largest element in magnitude in that column Let the element be From the permutation matrix P 2 by interchanging the rows 2 and r 2 of the identity matrix and leaving the other rows unchanged From P 2 A (1) Next find an elementary lower triangular matrix M 2 such that A (2) = M 2 P 2 A (1) has zeros bellow the (22) entry on the second column M 2 is constructed as follows First construct an elementary matrix of order (n-1) such that Then define Note that (ij) refers to the th entry of the current matrix P 2 A (1) At the end of step 2 we will have 501

Aust J Basic & Appl Sci 5(6): 497-503 2011 Where Save the multipliers m i2 and record the row interchange Step k: In general at the th step scan the entry of the th column of the matrix A (k-1) below the row (k-1) to identity the pivot a rkk form the permutation matrix P k and find an elementary over triangular matrix M k such that has zeros below the (kk) entry on the th column M k is constructed first by construction of order (n-k+1) such that and then defining Where 0 is a matrix of zeros The elements to the (ik)th entries of the matrix Step n 1: At the end of the (n-1)st step the matrix A(n-1) will be an upper triangular matrix Form U: Set A (n-1) =U (1) Then Set: (2) Then we have from the preceding the following factorization of A: U = MA 24 Gaussian Elimination with complete Pivoting: In Gaussian elimination with complete pivoting at the th step the search for the pivots is made among all the entries of the submatrix below the first (k-1) rows Thus if the pivot is a rs to bring this pivot to the (kk) position the interchange of the rows r and k has to be followed by the interchange of the columns k and s This is equivalent to pre multiplying the matrix A (k-1) by a permutation matrix P k obtained by interchanging rows k and r and post multiplying P k A (k-1) by another permutation matrix Q k obtained by interchanging the columns k and s of the identity matrix I The ordinary Gaussian elimination is then applied to the matrix P k A (k-1) Q k ; that is an elementary lower triangular matrix M k is sought such that the matrix Has zeros on the kth column below the (kk) entry The matrix Mk can of course be computed in two smaller steps as before At the end of the (n-1) st step the matrix A (n-1) is an upper triangular matrix Set: A (n-1) = U (3) 25 Stability of Gaussian Elimination: Definition 9: The growth factor ρ is the ratio of the largest element (in magnitude) of A A (1) A (n-1) to the largest 502

Aust J Basic & Appl Sci 5(6): 497-503 2011 element (in magnitude) of A: where and Example 3: Let (a) Gaussian elimination without pivoting gives (b) Gaussian elimination with pivoting yields 26 Growth Factor of Gaussian Elimination for Complete Pivoting: For Gaussian elimination with complete pivoting Growth Factor of Gaussian Elimination for Partial Pivoting For Gaussian elimination with partial pivoting that is ρ can be as big as 2 n-1 Conclusion: The purpose of this paper is to show that the incomplete LU factorizations of the H-matrices by using the incomplete LU- factorizations of an H-matrix are at least as stable as the complete LU- factorizations of its comparison matrix We give also some new characterizations of the H-matrices in connection with their incomplete LU-factorizations REFERENCES Ahac AA JJ Buoni DD Olesky 1988 Stable LU-factorization of H-matrices Linear Algebra Appl 99: 97-110 Buoni JJ 1990 A stable method for the incomplete factorization of H-matrices Linear Algebra Appl 129: 143-154 Datta BN 1995 Numerical Linear algebra and applications Kolotilina L B Polman 1992 POLMAN on incomplete block factorization methods of generalized SSOR type for H-matrices Linear Algebra Appl 177: 111-136 Manteuffel TA 1980 An incomplete factorization technique for positive definite linear systems Math Comput 34: 473-497 Meijerink JA HA Van der Vorst 1977 An interative solution method for linear systems of which the coefficient matrix is a symmetric M-matrix Math Comput 31: 148-162 Messaoudi A 1995 On the stability of the incomplete LU-factorizations and character: Zations of H- matrices Num er Math 69: 321-331 Varga RS EB Saff V Mehrman 1980 incomplete factorization of matrices and connection with H- matrices SIAM J Number Anal 17: 787-793 503