Linear System of Equations

Size: px
Start display at page:

Download "Linear System of Equations"

Transcription

1 Linear System of Equations Linear systems are perhaps the most widely applied numerical procedures when real-world situation are to be simulated. Example: computing the forces in a TRUSS. F F 5. 77F F. 6585F6 F7. 77F F4. 755F6 F 4 8 F. 6585F F 7 F8. 755F F. 6585F F F6 F8 F. 77F4 F. 77F4 F F F. 755F F F. 77F F4 4 equations with 4 unknowns Methods for numerically solving ordinary and partial differential equations depend on Linear Systems of Equations. No of unknowns = number of equations i.e. nn system

2 Linear System of Equations a x a x a x b a x a x a x b a x a x a x b a a a a a a a a a x x x b b b A x = b n * n Matrix Column vector Column vector

3 Matrix Notation A matrix consists of a rectangular array of elements represented by a single symbol (example: [A]). An individual entry of a matrix is an element (example: a )

4 Matrix Notation A horizontal set of elements is called a row and a vertical set of elements is called a column. The first subscript of an element indicates the row while the second indicates the column. The size of a matrix is given as m rows by n columns, or simply m by n (or m x n). x n matrices are row vectors. m x matrices are column vectors. 4

5 Special Matrices Matrices where m=n are called square matrices. There are a number of special forms of square matrices: Symmetric A Diagonal A a a a Identity A Upper Triangular A a a a a a a Lower Triangular A a a a a a a Banded A a a a a a a a a 4 a 4 a 44 5

6 Matrix Operations Two matrices are considered equal if and only if every element in the first matrix is equal to every corresponding element in the second. This means the two matrices must be the same size. A=B if a ij = b ij for all i and j Matrix addition and subtraction are performed by adding or subtracting the corresponding elements. This requires that the two matrices be the same size. C=A+B c ij =a ij + b ij multiplication by a scalar is performed by multiplying each element by the same scalar. ga ga ga D=gA ga ga ga ga ga ga 6

7 Matrix Multiplication The elements in the matrix [C] that results from multiplying matrices [A] and [B] are calculated using: c ij n k a ik b kj 7

8 Matrix Inverse and Transpose The inverse of a square, nonsingular matrix [A] is that matrix which, when multiplied by [A], yields the identity matrix. [A][A] - =[A] - [A]=[I] The transpose of a matrix involves transforming its rows into columns and its columns into rows. (a ij ) T =a ji 8

9 Solving With MATLAB MATLAB provides two direct ways to solve systems of linear algebraic equations [A]{x}={b}: Left-division x = A\b Matrix inversion x = inv(a)*b Note: The matrix inverse is less efficient than left-division and also only works for square, non-singular systems. 9

10 Gauss Elimination method Forward elimination Starting with the first row, add or subtract multiples of that row to eliminate the first coefficient from the second row and beyond. Continue this process with the second row to remove the second coefficient from the third row and beyond. Stop when an upper triangular matrix remains. Back substitution Starting with the last row, solve for the unknown, then substitute that value into the next highest row. Because of the upper-triangular nature of the matrix, each row will contain only one more unknown.

11 Gauss Elimination method Number of flops for Gauss Elimination method: n Total flops Forward Elimination Back substitution n / % due to Elimination % % 6.68* * * % As the system gets larger, the cost of computing increases - Flops increases nearly orders of magnitude for every order of increase in the number of equations Most of the efforts is incurred in the forward elimination steps.

12 Gauss Elimination method clc; close all; format long e; % A= coefficient matrix % b=right hand side vector % x=solution vector A=[ -7 ; - 6; 5-5]; b=[7;4;6]; [m,n]=size(a); if m~=n error('matrix A must be square'); end Nb=n+; Aug=[A b]; % Augmented matrix, combines A and b %----forward elimilation for k=:n- % from row to n- if abs(a(k,k))<eps error('zero Pivot encountered') end for j=k+:n multiplier=aug(j,k)/aug(k,k); Aug(j,k:Nb)=Aug(j,k:Nb)-multiplier*Aug(k,k:Nb); end end

13 Gauss Elimination method %----Back substitution x=zeros(n,); x(n)=aug(n,nb)/aug(n,n); for k=n-:-: for j=k+:n Aug(k,Nb)=Aug(k,Nb)-Aug(k,j)*x(j); % computes b(k)-sum(akj*xj) end end x(k)=aug(k,nb)/aug(k,k);

14 Cost of computing Example : Estimate the time required to carry out back substitution on a system of 5 equations in 5 unknowns, on a computer where elimination takes second. For Elimination approx. flops, fe= n /=*5 / time needed, te= sec For Back substitution approx. flops, fb=n =5 time needed, tb=? Back substitution time=. sec Total computation time=. sec tb/te=fb/fe tb=te*fb/fe=/(*5)=. sec 4

15 Cost of computing Example : Assume that your computer can solve a * linear system Ax=b in second by Gauss elimination method. Estimate the time required to solve four systems of 5 equations in 5 unknowns with the same coefficient matrix, using LU factorization method. For Gauss Elimination approx. flops, fg= n /=* / time needed, tg= sec For LU approx. flops, fl=n /+kn =*5 /+*4*5 time needed, tl=tg*fl/fg=(*5^/+*4*5 )/(* /)=6 sec If we do the same 4 problems by Gauss Elimination: tg=4*5 / =6.5 sec Time save=46.5 sec %!!! 5

16 Forward and backward error Definition Let x c be an approximate solution of the linear system Ax=b. The residual is the vector r=b-ax c The backward error = b-ax c The forward error = x exact -x c Error magnification factor= relative forward error relative backward error x exact x x exact b Ax b c c 6

17 Sources of error There are two major sources of error in Gauss-elimination as we have discussed so far. ill-conditioning swamping Ill-conditioning -sensitivity of the solution to the input data..first we examine the effect of small change in the coefficients Let us consider a system: Ax=b..99x. Exact solution x exact.99.x. Now, we make a small change in the coefficient matrix A...98 New solution A x C Large change in soln 7

18 Sources of error:ill-conditioning. we examine the effect of small change in b for the same coefficient matrix A b b b b b solution solution Ax=b solver x C x C x c x exact x C Here we see, even though the inputs are close together we get very distinct outputs Therefore, for ill-conditioned system, small change in input, we get large change in output. How can we express this characteristic mathematically? 8

19 Sources of error:ill-conditioning Let us consider another example of ill-conditioned system:. x x. x exact x C. Backward error vector = r b Ax Forward error vector = x exact x C C b Ax C b..... Relative backward error =.5% Relative forward error = x exact x x exact C =. % r = b-axc =. x exact -x c =. For ill-conditioned system, error magnification factor is very high Error magnification factor =./(./.) = 44. 9

20 Condition number There is another quantity known as Condition number is also used to show that the system is ill-conditioned. Definition The condition number of a square matrix A, cond(a), is the maximum possible error magnification factor for Ax=b, over all right-hand side b. Theorem The condition number of a n*n matrix A is cond(a)= A * A - For ill-conditioned system, condition number is high If the coefficient matrix A have t-digit precision, and cond(a)= k The solution vector x is accurate upto (t-k) digits. Example: elements of a coefficient matrix A is accurate up to 5 digits, and cond(a)= 4 So, the solution is accurate (5-4) = digit only!!!

21 Sources of error:swamping A second significant source of error in Gauss elimination method is Swamping. Much easier to fix. The Gauss elimination method discussed so far, works well as long as the diagonal elements of coefficient matrix A (pivot) are non-zero. Note: The computation of the multiplier and the back substitution require divisions by the pivots. Consequently the algorithm can not be carried out if any of the pivots are zero or near to zero. Let us consider our original example again, x x x x x x x exact

22 Sources of error:swamping let us assume that the solution is to be computed on a hypothetical machine that does decimal floating point arithmetic with five significant digits. the first step of elimination produces, x x x R R R 5 R R R quite small compared with the other elements

23 Sources of error:swamping without interchanging the rows, the multiplier R R R x x x x C Finally, x exact But,

24 Sources of error:swamping So, where did things go wrong? There is no accumulation of rounding error caused by doing thousands of arithmetic operations The matrix is NOT close to singular Actually, The difficulty comes from choosing a small pivot at second step of the elimination. As a result the multiplier is.5 and the final equation involves coefficients that are times larger those in the original problem. That means the equation is overpowered/swamped. If the multipliers are all less than or equal to in magnitude, then computed solution can be proved to be satisfactory. Keeping the multipliers less than one in absolute value can be ensured by a process known as Pivoting. 4

25 5 Sources of error: Pivoting x x x If we allow pivoting (partial), i.e, by interchanging R and R, we have,..6.4 x C.5. R R R Solution is close to the exact!!!

26 Pivoting Problems arise with Gauss elimination if a coefficient along the diagonal (Pivot element) is (problem: division by ) or close to (problem: round-off error) One way to combat these issues is to determine the coefficient with the largest absolute value in the column below the pivot element. The rows can then be switched so that the largest element is the pivot element. This is called partial pivoting. If the rows to the right of the pivot element are also checked and columns switched, this is called complete pivoting. 6

27 Gauss Elimination with partial pivoting clc; close all; format long e; % A= coefficient matrix % b=right hand side vector % x=solution vector A=[ -7 ; - 6; 5-5]; b=[7;4;6]; [m,n]=size(a); if m~=n, error('matrix A must be square');end Nb=n+; Aug=[A b]; % Augmented matrix, combines A and b %----forward elimilation for k=:n- % from row to n- % partial pivoting [big, big_pos]=max(abs(aug(k:n,k))); pivot_pos=big_pos+k-; if pivot_pos~=k Aug([k pivot_pos],:)=aug([pivot_pos k],:); end % for j=k+:n multiplier=aug(j,k)/aug(k,k); Aug(j,k:Nb)=Aug(j,k:Nb)-multiplier*Aug(k,k:Nb); end end 7

28 Gauss Elimination method %----Back substitution x=zeros(n,); x(n)=aug(n,nb)/aug(n,n); for k=n-:-: for j=k+:n Aug(k,Nb)=Aug(k,Nb)-Aug(k,j)*x(j); % computes b(k)-sum(akj*xj) end end x(k)=aug(k,nb)/aug(k,k); 8

29 Scaling Scaling is the operation of adjusting the coefficients of a set of equations so that they are all of the same order of magnitude. Example x x x 5 x exact If we do pivoting without scaling than round-off error may occur. For the above system, computed solution is x C But if you scale the system (divide the maximum coefficient in each Row) & then do the pivoting, R R R R R R x x.5 x.5. x C... 9

30 Scaling So, whenever the coefficients in one column are widely different from those in another column, scaling is beneficial. When all values are about the same order of magnitude, scaling should be avoided, as additional round-off error incurred during the scaling operation itself may adversely affect the accuracy.

31 Tridiagonal systems function [xout]=tridiagonal(a,b,c,d) N=length(b); %---forward elimination for k=:n multiplier=a(k)/b(k-); b(k)=b(k)-multiplier*c(k-); d(k)=d(k)-multiplier*d(k-); end %-----back substitution x(n)=d(n)/b(n); for k=n-:-: x(k)=(d(k)-c(k)*x(k+))/b(k); end xout=x.'; % Main: format long e; clc; clear all; close all; a=[ - - -]; b=[ ]; c=[- - - ]; d=[ ]; [xout]=tridiagonal(a,b,c,d)

32 Cholesky factorization for symmetric coefficient matrix: a ij =a ji and A=A T Algorithm is based on the fact that symmetric matrix can be decomposed as A= U U T Ax=b U T Ux=b U T c=b Ux=c (i) Solve for c (ii) solve for x. MATLAB: U = chol(a) U is a upper triangular matrix, so that U T *U=A

33 Matrix inversion For square matrix If matrix A is square, there is another matrix A-, called the inverse of A, for which AA - =I=A - A Inverse can be computed in a column-by-column fashion by generating solutions with the unit vectors as the right-hand-side constants. Example: for a system b Solution of the system Ax =b will provide the column of A - Then use, b b Solution of the system Ax =b will provide the column of A - Solution of the system Ax =b will provide the column of A -

34 Matrix inversion finally, x x x A - =????????? our original system is Ax = b x=a - * b No of flops: multiplication: n n 6 6 Total flops n n 5 addition: -no of flops times Gauss elimination, more round-off error -computationally more expensive -Note: singular matrix (det A=) has no inverse -MATLAB: xc=inv(a)*b 5 6 n 4 n n 5 6 n 4

35 MATLAB \ operator It is useful to know that specific algorithm used by MATLAB when the \operator is invoked depends on the structure of the coefficient matrix A. To determine the structure of A and select the appropriate algorithm, MATLAB follows the following precedence:. If A is banded, then banded solvers are used (e.g. Thomas algorithm).. If A is a upper or lower triangular matrix, then the system is solved by a backward substitution algorithm for upper triangular matrix, or by a forward substitution algorithm for a lower triangular matrix.. If A is symmetric and has real positive diagonal elements, then a Cholesky factorization is applied first. 4. If none of the above criteria is fulfilled, then a general triangular factorization is computed by Gauss elimination with partial pivoting (lu). 5. If A is not square, i.e., rectangular, QR factorization is used. Note: points -5 are simplified in the context of our course. 5

36 Iterative methods Uses an initial guess and refine the guess at each step, converging to the solution vector. 6

37 Jacobi method function [xout iter_number]=jacobi(a,b,max_iter,tol) N=length(b); d=diag(a); % takes the main diagonal elements only x=zeros(n,); % initial guess vector for k=:max_iter x=(b-(a-diag(d))*x)./d; maxresidual=norm(b-a*x,inf); if maxresidual<tol break; end End iter_number=k; xout=x; % Main: format long e; clc; clear all; close all; A=[ -. -.;. 7 -.;. -. ]; b=[7.85;-9.; 7.4]; max_iter=;tol=e-5; [xc iter_number]=jacobi(a,b,max_iter,tol); 7

38 Gauss Seidel and Jacobi method: Graphical depiction 8

39 Gauss Seidel method function [xout iter_number]=gaussseidel(a,b,max_iter,tol) N=length(b); d=diag(diag(a)); % takes the main diagonal elements only x=zeros(n,); % initial guess vector U=triu(A,); % above the main diagonal L=tril(A,-); % below the main diagonal for k=:max_iter bl=b-u*x; for j=:n x(j)=(bl(j)-l(j,:)*x)./d(j,j); end maxresidual=norm(b-a*x,inf); if maxresidual<tol break; end End iter_number=k; xout=x; %

40 SOR method function [xout iter_number]=sor(a,b,relax,max_iter,tol) N=length(b); d=diag(diag(a)); x=zeros(n,); % initial guess vector U=triu(A,); % above the main diagonal L=tril(A,-); % below the main diagonal for k=:max_iter bl=relax*(b-u*x)+(-relax)*d*x; for j=:n x(j)=(bl(j)-relax*l(j,:)*x)./d(j,j); end maxresidual=norm(b-a*x,inf); if maxresidual<tol break; end end iter_number=k; xout=x; 4

41 4 Sparse matrix computations AS No row of A has more than 4 non-zero entries. Since fewer than 4n of the n potential entries are non zero. We may call this matrix Sprase.

42 Sparse matrix computations We want to solve a system of equations defined by AS for n= 5 What are the options? Treating the coefficient matrix A as a full matrix means storing n = elements each as a double precision floating point number. One double precision floating point number requires 8 bytes (64 bit) of storage. So, n = elements will require 8* bytes = 8 gigabytes (approx.) of RAM Not only the size is enemy, but so is time. The number of flops by Gauss Elimination will be order of n 5 If a PC runs on the order of a few GHz ( 9 cycle per second), an upper bound of floating point operations per second is around 8. Therefore, time required of Gauss Elimination= 5 / 8 = 7 seconds. Note: year =* 7 seconds. 4

43 Sparse matrix computations So, it is clear that Gauss Elimination method for this problem is not an overnight computation. On the other hand, One step of an iterative method will require *4n=8* 5 (approx.) operations. If the solution needs iterations in Jacobi method, total operations= 8 Time required in modern PC for Jacobi method will be = second (approx.) or less!!! 4

44 MATLAB functions for iterative methods Function pcg bicg cgs bicgstab gmres Method Preconditioned conjugate gradients method Biconjugate gradients method Conjugate gradients squared method Biconjugate gradients stabilized method Generalized minimum residual method Syntax: same for all of the functions. For example, xc = pcg(a,b,tol,maxit); 44

45 Rank Deficiency Kronecker-Capelli theorem: A system has a SOLUTION if and only if its coeffiecient matrix (denoted by A) and its augmented matrix (denoted by Ab) have the same Rank. Let, Rank of A matrix=r Number of equation=n If R=n, the system has unique solution (full Rank system) If R<n, the system has an infinite number of solution (Rank-deficient system) We can express the system in terms of (n-r) variables under-determined system (number of equation<number of unknowns) 45

46 Rank Deficiency Example : A=[ - ; 4 - ; - -4 ]; b=[ 5; 7; ]; Ab=[A b]; % Augmented matrix RA=rank(A); RAb=rank(Ab); xc=a\b Solution: RA= RAb= Warning: Matrix is singular to working precision. xc = NaN -Inf -Inf 46

47 Rank Deficiency Example : A=[ - ; 4 - ; - -4 ]; b=[ 5; 7; ]; Ab=[A b]; % Augmented matrix RA=rank(A); RAb=rank(Ab); xc=a\b xc=pinv(a)*b Solution: RA= RAb= Warning: Matrix is singular to working precision. xc = NaN NaN NaN xc = e e e+ 47

Linear System of Equations

Linear System of Equations Linear System of Equations Linear systems are perhaps the most widely applied numerical procedures when real-world situation are to be simulated. Example: computing the forces in a TRUSS. F F 5. 77F F.

More information

A Review of Matrix Analysis

A Review of Matrix Analysis Matrix Notation Part Matrix Operations Matrices are simply rectangular arrays of quantities Each quantity in the array is called an element of the matrix and an element can be either a numerical value

More information

The Solution of Linear Systems AX = B

The Solution of Linear Systems AX = B Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has

More information

Numerical Analysis Fall. Gauss Elimination

Numerical Analysis Fall. Gauss Elimination Numerical Analysis 2015 Fall Gauss Elimination Solving systems m g g m m g x x x k k k k k k k k k 3 2 1 3 2 1 3 3 3 2 3 2 2 2 1 0 0 Graphical Method For small sets of simultaneous equations, graphing

More information

Linear Algebraic Equations

Linear Algebraic Equations Linear Algebraic Equations 1 Fundamentals Consider the set of linear algebraic equations n a ij x i b i represented by Ax b j with [A b ] [A b] and (1a) r(a) rank of A (1b) Then Axb has a solution iff

More information

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

Computational Methods. Systems of Linear Equations

Computational Methods. Systems of Linear Equations Computational Methods Systems of Linear Equations Manfred Huber 2010 1 Systems of Equations Often a system model contains multiple variables (parameters) and contains multiple equations Multiple equations

More information

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University Part 3 Chapter 10 LU Factorization PowerPoints organized by Dr. Michael R. Gustafson II, Duke University All images copyright The McGraw-Hill Companies, Inc. Permission required for reproduction or display.

More information

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear

More information

Process Model Formulation and Solution, 3E4

Process Model Formulation and Solution, 3E4 Process Model Formulation and Solution, 3E4 Section B: Linear Algebraic Equations Instructor: Kevin Dunn dunnkg@mcmasterca Department of Chemical Engineering Course notes: Dr Benoît Chachuat 06 October

More information

Linear Algebraic Equations

Linear Algebraic Equations Linear Algebraic Equations Linear Equations: a + a + a + a +... + a = c 11 1 12 2 13 3 14 4 1n n 1 a + a + a + a +... + a = c 21 2 2 23 3 24 4 2n n 2 a + a + a + a +... + a = c 31 1 32 2 33 3 34 4 3n n

More information

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1 Matrix notation A nm : n m : size of the matrix m : no of columns, n: no of rows Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1 n = m square matrix Symmetric matrix Upper triangular matrix: matrix

More information

SOLVING LINEAR SYSTEMS

SOLVING LINEAR SYSTEMS SOLVING LINEAR SYSTEMS We want to solve the linear system a, x + + a,n x n = b a n, x + + a n,n x n = b n This will be done by the method used in beginning algebra, by successively eliminating unknowns

More information

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems. COURSE 9 4 Numerical methods for solving linear systems Practical solving of many problems eventually leads to solving linear systems Classification of the methods: - direct methods - with low number of

More information

Chapter 9: Gaussian Elimination

Chapter 9: Gaussian Elimination Uchechukwu Ofoegbu Temple University Chapter 9: Gaussian Elimination Graphical Method The solution of a small set of simultaneous equations, can be obtained by graphing them and determining the location

More information

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization

More information

2.1 Gaussian Elimination

2.1 Gaussian Elimination 2. Gaussian Elimination A common problem encountered in numerical models is the one in which there are n equations and n unknowns. The following is a description of the Gaussian elimination method for

More information

Matrix & Linear Algebra

Matrix & Linear Algebra Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

1.Chapter Objectives

1.Chapter Objectives LU Factorization INDEX 1.Chapter objectives 2.Overview of LU factorization 2.1GAUSS ELIMINATION AS LU FACTORIZATION 2.2LU Factorization with Pivoting 2.3 MATLAB Function: lu 3. CHOLESKY FACTORIZATION 3.1

More information

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Review of matrices. Let m, n IN. A rectangle of numbers written like A = Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an

More information

CS227-Scientific Computing. Lecture 4: A Crash Course in Linear Algebra

CS227-Scientific Computing. Lecture 4: A Crash Course in Linear Algebra CS227-Scientific Computing Lecture 4: A Crash Course in Linear Algebra Linear Transformation of Variables A common phenomenon: Two sets of quantities linearly related: y = 3x + x 2 4x 3 y 2 = 2.7x 2 x

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2 MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Direct Methods Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) Linear Systems: Direct Solution Methods Fall 2017 1 / 14 Introduction The solution of linear systems is one

More information

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i ) Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical

More information

CHAPTER 6. Direct Methods for Solving Linear Systems

CHAPTER 6. Direct Methods for Solving Linear Systems CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to

More information

MATH 3511 Lecture 1. Solving Linear Systems 1

MATH 3511 Lecture 1. Solving Linear Systems 1 MATH 3511 Lecture 1 Solving Linear Systems 1 Dmitriy Leykekhman Spring 2012 Goals Review of basic linear algebra Solution of simple linear systems Gaussian elimination D Leykekhman - MATH 3511 Introduction

More information

Lecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II)

Lecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II) Math 59 Lecture 2 (Tue Mar 5) Gaussian elimination and LU factorization (II) 2 Gaussian elimination - LU factorization For a general n n matrix A the Gaussian elimination produces an LU factorization if

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Decompositions, numerical aspects Gerard Sleijpen and Martin van Gijzen September 27, 2017 1 Delft University of Technology Program Lecture 2 LU-decomposition Basic algorithm Cost

More information

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects Numerical Linear Algebra Decompositions, numerical aspects Program Lecture 2 LU-decomposition Basic algorithm Cost Stability Pivoting Cholesky decomposition Sparse matrices and reorderings Gerard Sleijpen

More information

Next topics: Solving systems of linear equations

Next topics: Solving systems of linear equations Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:

More information

4.2 Floating-Point Numbers

4.2 Floating-Point Numbers 101 Approximation 4.2 Floating-Point Numbers 4.2 Floating-Point Numbers The number 3.1416 in scientific notation is 0.31416 10 1 or (as computer output) -0.31416E01..31416 10 1 exponent sign mantissa base

More information

CPE 310: Numerical Analysis for Engineers

CPE 310: Numerical Analysis for Engineers CPE 310: Numerical Analysis for Engineers Chapter 2: Solving Sets of Equations Ahmed Tamrawi Copyright notice: care has been taken to use only those web images deemed by the instructor to be in the public

More information

Section 9.2: Matrices.. a m1 a m2 a mn

Section 9.2: Matrices.. a m1 a m2 a mn Section 9.2: Matrices Definition: A matrix is a rectangular array of numbers: a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn In general, a ij denotes the (i, j) entry of A. That is, the entry in

More information

1 Number Systems and Errors 1

1 Number Systems and Errors 1 Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floating-point Representation...........

More information

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) D. ARAPURA Gaussian elimination is the go to method for all basic linear classes including this one. We go summarize the main ideas. 1.

More information

Gauss Elimination. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Gauss Elimination. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan Gauss Elimination Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan chanhl@mail.cgu.edu.tw Solving small numbers of equations by graphical method The location of the intercept provides

More information

Elementary Linear Algebra

Elementary Linear Algebra Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We

More information

Solving Linear Systems Using Gaussian Elimination. How can we solve

Solving Linear Systems Using Gaussian Elimination. How can we solve Solving Linear Systems Using Gaussian Elimination How can we solve? 1 Gaussian elimination Consider the general augmented system: Gaussian elimination Step 1: Eliminate first column below the main diagonal.

More information

MTH 464: Computational Linear Algebra

MTH 464: Computational Linear Algebra MTH 464: Computational Linear Algebra Lecture Outlines Exam 2 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University February 6, 2018 Linear Algebra (MTH

More information

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination Chapter 2 Solving Systems of Equations A large number of real life applications which are resolved through mathematical modeling will end up taking the form of the following very simple looking matrix

More information

The purpose of computing is insight, not numbers. Richard Wesley Hamming

The purpose of computing is insight, not numbers. Richard Wesley Hamming Systems of Linear Equations The purpose of computing is insight, not numbers. Richard Wesley Hamming Fall 2010 1 Topics to Be Discussed This is a long unit and will include the following important topics:

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.

More information

Chapter 2 Notes, Linear Algebra 5e Lay

Chapter 2 Notes, Linear Algebra 5e Lay Contents.1 Operations with Matrices..................................1.1 Addition and Subtraction.............................1. Multiplication by a scalar............................ 3.1.3 Multiplication

More information

Solving Linear Systems of Equations

Solving Linear Systems of Equations 1 Solving Linear Systems of Equations Many practical problems could be reduced to solving a linear system of equations formulated as Ax = b This chapter studies the computational issues about directly

More information

Matrix decompositions

Matrix decompositions Matrix decompositions How can we solve Ax = b? 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x = The variables x 1, x, and x only appear as linear terms (no powers

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Elementary maths for GMT

Elementary maths for GMT Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1

More information

Introduction to Applied Linear Algebra with MATLAB

Introduction to Applied Linear Algebra with MATLAB Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation

More information

5 Solving Systems of Linear Equations

5 Solving Systems of Linear Equations 106 Systems of LE 5.1 Systems of Linear Equations 5 Solving Systems of Linear Equations 5.1 Systems of Linear Equations System of linear equations: a 11 x 1 + a 12 x 2 +... + a 1n x n = b 1 a 21 x 1 +

More information

MATH2210 Notebook 2 Spring 2018

MATH2210 Notebook 2 Spring 2018 MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................

More information

Phys 201. Matrices and Determinants

Phys 201. Matrices and Determinants Phys 201 Matrices and Determinants 1 1.1 Matrices 1.2 Operations of matrices 1.3 Types of matrices 1.4 Properties of matrices 1.5 Determinants 1.6 Inverse of a 3 3 matrix 2 1.1 Matrices A 2 3 7 =! " 1

More information

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix EE507 - Computational Techniques for EE 7. LU factorization Jitkomut Songsiri factor-solve method LU factorization solving Ax = b with A nonsingular the inverse of a nonsingular matrix LU factorization

More information

Linear Algebra V = T = ( 4 3 ).

Linear Algebra V = T = ( 4 3 ). Linear Algebra Vectors A column vector is a list of numbers stored vertically The dimension of a column vector is the number of values in the vector W is a -dimensional column vector and V is a 5-dimensional

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

Lectures on Linear Algebra for IT

Lectures on Linear Algebra for IT Lectures on Linear Algebra for IT by Mgr. Tereza Kovářová, Ph.D. following content of lectures by Ing. Petr Beremlijski, Ph.D. Department of Applied Mathematics, VSB - TU Ostrava Czech Republic 2. Systems

More information

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4 Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix

More information

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD Mathematical Question we are interested in answering numerically How to solve the following linear system for x Ax = b? where A is an n n invertible matrix and b is vector of length n. Notation: x denote

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

POLI270 - Linear Algebra

POLI270 - Linear Algebra POLI7 - Linear Algebra Septemer 8th Basics a x + a x +... + a n x n b () is the linear form where a, b are parameters and x n are variables. For a given equation such as x +x you only need a variable and

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in Chapter 4 - MATRIX ALGEBRA 4.1. Matrix Operations A a 11 a 12... a 1j... a 1n a 21. a 22.... a 2j... a 2n. a i1 a i2... a ij... a in... a m1 a m2... a mj... a mn The entry in the ith row and the jth column

More information

9. Numerical linear algebra background

9. Numerical linear algebra background Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 2 Systems of Linear Equations Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Linear Equations in Linear Algebra

Linear Equations in Linear Algebra 1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION x 1,, x n A linear equation in the variables equation that can be written in the form a 1 x 1 + a 2 x 2 + + a n x n

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

Section 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.

Section 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns. Section 9.2: Matrices Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns. That is, a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn A

More information

Gaussian Elimination and Back Substitution

Gaussian Elimination and Back Substitution Jim Lambers MAT 610 Summer Session 2009-10 Lecture 4 Notes These notes correspond to Sections 31 and 32 in the text Gaussian Elimination and Back Substitution The basic idea behind methods for solving

More information

Gaussian Elimination -(3.1) b 1. b 2., b. b n

Gaussian Elimination -(3.1) b 1. b 2., b. b n Gaussian Elimination -() Consider solving a given system of n linear equations in n unknowns: (*) a x a x a n x n b where a ij and b i are constants and x i are unknowns Let a n x a n x a nn x n a a a

More information

Pivoting. Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3

Pivoting. Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3 Pivoting Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3 In the previous discussions we have assumed that the LU factorization of A existed and the various versions could compute it in a stable manner.

More information

Introduction to PDEs and Numerical Methods Lecture 7. Solving linear systems

Introduction to PDEs and Numerical Methods Lecture 7. Solving linear systems Platzhalter für Bild, Bild auf Titelfolie hinter das Logo einsetzen Introduction to PDEs and Numerical Methods Lecture 7. Solving linear systems Dr. Noemi Friedman, 09.2.205. Reminder: Instationary heat

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

CSE 160 Lecture 13. Numerical Linear Algebra

CSE 160 Lecture 13. Numerical Linear Algebra CSE 16 Lecture 13 Numerical Linear Algebra Announcements Section will be held on Friday as announced on Moodle Midterm Return 213 Scott B Baden / CSE 16 / Fall 213 2 Today s lecture Gaussian Elimination

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination Chapter 2 Solving Systems of Equations A large number of real life applications which are resolved through mathematical modeling will end up taking the form of the following very simple looking matrix

More information

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic A FIRST COURSE IN LINEAR ALGEBRA An Open Text by Ken Kuttler Matrix Arithmetic Lecture Notes by Karen Seyffarth Adapted by LYRYX SERVICE COURSE SOLUTION Attribution-NonCommercial-ShareAlike (CC BY-NC-SA)

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently

More information

MATRICES. a m,1 a m,n A =

MATRICES. a m,1 a m,n A = MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Numerical Analysis FMN011

Numerical Analysis FMN011 Numerical Analysis FMN011 Carmen Arévalo Lund University carmen@maths.lth.se Lecture 4 Linear Systems Ax = b A is n n matrix, b is given n-vector, x is unknown solution n-vector. A n n is non-singular

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,

More information

Consider the following example of a linear system:

Consider the following example of a linear system: LINEAR SYSTEMS Consider the following example of a linear system: Its unique solution is x + 2x 2 + 3x 3 = 5 x + x 3 = 3 3x + x 2 + 3x 3 = 3 x =, x 2 = 0, x 3 = 2 In general we want to solve n equations

More information

Numerical Linear Algebra

Numerical Linear Algebra Chapter 3 Numerical Linear Algebra We review some techniques used to solve Ax = b where A is an n n matrix, and x and b are n 1 vectors (column vectors). We then review eigenvalues and eigenvectors and

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Matrices. Chapter Definitions and Notations

Matrices. Chapter Definitions and Notations Chapter 3 Matrices 3. Definitions and Notations Matrices are yet another mathematical object. Learning about matrices means learning what they are, how they are represented, the types of operations which

More information

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey Copyright 2005, W.R. Winfrey Topics Preliminaries Echelon Form of a Matrix Elementary Matrices; Finding A -1 Equivalent Matrices LU-Factorization Topics Preliminaries Echelon Form of a Matrix Elementary

More information

Solving Linear Systems of Equations

Solving Linear Systems of Equations November 6, 2013 Introduction The type of problems that we have to solve are: Solve the system: A x = B, where a 11 a 1N a 12 a 2N A =.. a 1N a NN x = x 1 x 2. x N B = b 1 b 2. b N To find A 1 (inverse

More information

MTH501- Linear Algebra MCQS MIDTERM EXAMINATION ~ LIBRIANSMINE ~

MTH501- Linear Algebra MCQS MIDTERM EXAMINATION ~ LIBRIANSMINE ~ MTH501- Linear Algebra MCQS MIDTERM EXAMINATION ~ LIBRIANSMINE ~ Question No: 1 (Marks: 1) If for a linear transformation the equation T(x) =0 has only the trivial solution then T is One-to-one Onto Question

More information

Systems of Linear Equations and Matrices

Systems of Linear Equations and Matrices Chapter 1 Systems of Linear Equations and Matrices System of linear algebraic equations and their solution constitute one of the major topics studied in the course known as linear algebra. In the first

More information

CE 206: Engineering Computation Sessional. System of Linear Equations

CE 206: Engineering Computation Sessional. System of Linear Equations CE 6: Engineering Computation Sessional System of Linear Equations Gauss Elimination orward elimination Starting with the first row, add or subtract multiples of that row to eliminate the first coefficient

More information

Laboratory #3: Linear Algebra. Contents. Grace Yung

Laboratory #3: Linear Algebra. Contents. Grace Yung Laboratory #3: Linear Algebra Grace Yung Contents List of Problems. Introduction. Objectives.2 Prerequisites 2. Linear Systems 2. What is a Matrix 2.2 Quick Review 2.3 Gaussian Elimination 2.3. Decomposition

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 2. Systems of Linear Equations

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 2. Systems of Linear Equations Lecture Notes to Accompany Scientific Computing An Introductory Survey Second Edition by Michael T. Heath Chapter 2 Systems of Linear Equations Copyright c 2001. Reproduction permitted only for noncommercial,

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information