Lecture 1 INF-MAT : Chapter 2. Examples of Linear Systems
|
|
- Harvey Byrd
- 5 years ago
- Views:
Transcription
1 Lecture 1 INF-MAT : Chapter 2. Examples of Linear Systems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo August 26, 2010
2 Notation The set of natural numbers, integers, rational numbers, real numbers, and complex numbers are denoted by N, Z, Q, R, C, respectively. R n (C n ) is the set of n-tuples of real(complex) numbers which we will represent as column vectors. Thus x R n means x 1 x 2 x =., x n where x i R(C) for i = 1,..., n. Row vectors are normally identified using the transpose operation. Thus if x R n then x is a column vector and x T is a row vector.
3 Notation2 R m,n (C m,n ) is the set of m n matrices with real(complex) elements represented as a 11 a 12 a 1n a 21 a 22 a 2n A =.... a m1 a m2 a mn The element in the ith row and jth column of a matrix A will be denoted by a i,j, a ij, A(i, j) or (A) i,j.
4 Column Partition of a Matrix Product Matrix vector product. a1j x j Ax =. = a 1j. x j = amj x j a mj In particular Ae j = a :j. n x j a :j Product (AB)e j = A(Be j ) = Ab :j so AB = [Ab :1,..., Ab :n ] AB is partitioned by columns Partition by rows a T 1: B a T 2: AB = B. a T m:b j=1
5 Partitioned Matrices A rectangular matrix A can be partitioned into submatrices by drawing horizontal lines between selected rows and[ vertical ] lines between selected columns. For example, A = can be partitioned as [ ] A11 A (i) 12 = 4 5 6, (ii) [ ] a A 21 A :1, a :2, a :3 = a T 1: (iii) a T 2: = a T 3: , (iv) [ ] A 11, A 12 = The submatrices in a partition is often referred to as blocks and a partitioned matrix is sometimes called a block matrix..
6 Block Multiplication 1 B 1 B 2 p m A AB 1 AB 2 m AB = [Ab :1,..., Ab :r, Ab :r+1..., Ab :n ] = [AB 1, AB 2 ].
7 Block Multiplication 2 m A 1 A 2 B 1 s m A 1 B 1 A 2 B 2 B 2 p s p s p (AB) ij = a ik b kj = a ik b kj + a ik b kj j=1 j=1 j=s+1 = (A 1 B 1 ) ij + (A 2 B 2 ) ij = (A 1 B 1 + A 2 B 2 ) ij.
8 The General Case If A 11 A 1s B 11 B 1q A =.., B =.., A p1 A ps B s1 B sq and if all the matrix products in s C ij = A ik B kj, k=1 i = 1,..., p, j = 1,..., q are well defined then C 11 C 1q AB =... C p1 C pq
9 Products of Triangular Matrices a 11 a 12 a 1n b 11 b 12 b 1n c 11 c 12 c 1n 0 a 22 a 2n 0 b 22 b 2n.... = 0 c 22 c 2n a nn 0 0 b nn 0 0 c nn Lemma The product C = AB = (c ij ) of two upper(lower) triangular matrices A = (a ij ) and B = (b ij ) is upper(lower) triangular with diagonal elements c ii = a ii b ii for all i. Proof. Exercise.
10 Block Triangular Matrices Lemma Suppose [ ] A11 A A = 12 0 A 22 where A, A 11 and A 22 are square matrices. Then A is nonsingular if and only if both A 11 and A 22 are nonsingular. In that case [ A 1 A 1 = 11 A 1 11 A 12A 1 ] 22 0 A 1 (1) 22
11 Proof If A 11 and A 12 are nonsingular then [ A 1 11 A 1 11 A 12A 1 ] [ ] 22 A11 A 12 0 A 1 = 22 0 A 22 and A is nonsingular with the indicated inverse. [ ] I 0 = I 0 I
12 Proof Conversely, let B be the inverse of the nonsingular matrix A. We partition B conformally with A and have [ ] [ ] [ ] B11 B BA = 12 A11 A 12 I 0 = = I B 21 B 22 0 A 22 0 I Using block multiplication we find B 11 A 11 = I, B 21 A 11 = 0, B 21 A 12 + B 22 A 22 = I. The first equation implies that A 11 is invertible, this in turn implies that B 21 = 0 in the second equation, and then the third equation simplifies to B 22 A 22 = I. We conclude that also A 22 is invertible.
13 The Inverse Consider now a triangular matrix. Lemma An upper (lower) triangular matrix A = [a ij ] R n,n is nonsingular if and only if the diagonal elements a ii, i = 1,..., n are nonzero. In that case the inverse is upper (lower) triangular with diagonal elements a 1 ii, i = 1,..., n. Proof: We use induction on n. The result holds for n = 1: The 1-by-1 matrix A = (a 11 ) is invertible if and only if a 11 0 and in that case A 1 = (a 1 11 ). Suppose the result holds for n = k and let A = R k+1,k+1 be upper triangular.
14 Proof We partition A in the form [ ] Ak a A = k 0 a k+1,k+1 and note that A k R k,k is upper triangular. By Lemma 1.1 A is nonsingular if and only if A k and (a k+1,k+1 ) are nonsingular and in that case [ A A 1 1 k A 1 k = a ] ka 1 k+1,k+1 0 a 1. k+1,k+1 By the induction hypothesis A k is nonsingular if and only if the diagonal elements a 11,..., a kk of A k are nonzero and in that case A 1 k is upper triangular with diagonal elements a 1 ii, i = 1,..., k. The result for A follows.
15 Unit Triangular Matrices A matrix is unit triangular if it is triangular with 1 s on the diagonal. Lemma For a unit upper(lower) triangular matrix A R n,n : 1. A is invertible and the inverse is unit upper(lower) triangular. 2. The product of two unit upper(lower) triangular matrices is unit upper(lower) triangular. Proof. 1. follows from the inverse Lemma while the above Lemma implies 2.
16 Two Point Boundary Value Problem u (x) = f (x), x [0, 1] u(0) = 0, u(1) = 0, where f is a given continuous function on [0, 1]. Unique solution Numerical solution: finite difference method m N, h := 1/(m + 1), v j u(jh), j = 0,..., m + 1 Discrete system: Tv = v j 1 +2v j v j+1 = h 2 f (jh), j = 1,..., m, v 0 = v m+1 = v 1 v 2. v m 1 v m = h 2 f (h) f (2h). f ((m 1)h) f (mh) =: b.
17 LU Factorization d 1 c 1 a 2 d 2 c 2 1 r 1 c l = a n 1 d n 1 c n 1 l a n d n 1 n r n 1 c n 1 r n. d 1 = r 1, a k = l k r k 1, d k = l k c k 1 + r k, k = 2, 3,..., n. Algorithm (trifactor) 1. function [l,r]=trifactor(a,d,c) 2. r=d; l=d; 3. for k=2:length(d) 4. l(k)=a(k)/r(k-1); 5. r(k)=d(k)-l(k)*c(k-1); 6. end
18 Solving Tridiagonal Ax = b by LU Factorization Ax = L(Rx) = b y 1 = b 1, y k = b k l k y k 1, k = 2, 3,..., n, x n = y n /r n, x k = (y k c k x k+1 )/r k, k = n 1,..., 2, 1 Algorithm (trisolve) 1. function x=trisolve(l,r,c,b) 2. x=b; n=length(b); 3. for k=2:n 4. x(k)=b(k)-l(k)*x(k-1); 5. end 6. x(n)=x(n)/r(n); 7. for k=n-1:-1:1 8. x(k)=(x(k)-c(k)*x(k+1))/r(k); 9. end
19 O(n) Complexity l k = a k /r k 1, r k = d k l k c k 1, k = 2,..., n. n 1 divisions, n 1 multiplications, and n 1 subtractions. A total of 3n 3 flops (floating point arithmetic operations). y 1 = b 1, y k = b k l k y k 1, k = 2, 3,..., n, 2n 2 flops x n = y n /r n, x k = (y k c k x k+1 )/r k, k = n 1,..., 2, 1. 3n 2 flops A total of 3n 3 + 2n 2 + 3n 2 = 8n 7 flops. # flops is O(n). Gaussian elimination on a full matrix requires O(n 3 ) flops.
20 Diagonal Dominance The matrix A = [a ij ] C n,n is diagonally dominant if a ii j i a ij, i = 1,..., n. (2) It is strictly diagonally dominant if strict inequality holds for i = 1,..., n. T is diagonally dominant. Is it nonsingular? A square matrix A is singular if Ax = 0 for a nonzero vector x. A diagonally dominant matrix can be singular. [ ] [ ] A 1 = and A 2 = are both diagonally dominant and singular.
21 Diagonal Dominance and Nonsingularity d 1 c 1 a 2 d 2 c 2 A := a n 1 d n 1 c n 1 a n d n Theorem Suppose A = tridiag(a i, d i, c i ) C n,n is tridiagonal and diagonally dominant. If in addition d 1 > c 1 and a i 0 for i = 2,..., n 1, then A has a unique LU factorization. If in addition d n 0, then A is nonsingular. T is nonsingular. We can solve the system Tv = f in O(n) arithmetic operations using trifactor and trisolve.
22 Proof l k = a k /r k 1, r k = d k l k c k 1, k = 2,..., n. The matrix A has an LU factorization if the r k s are nonzero for k = 1,..., n 1. We show by induction on k that r k > c k for k = 1,..., n 1. Clearly r 1 = d 1 > c 1 Suppose c k 1 / r k 1 < 1 for some 2 k n 1 r k = d k l k c k 1 = d k a kc k 1 r k 1 d k a k c k 1 r k 1. a k 0 implies r k > d k a k c k. Thus r k > c k 0, for k = 1,..., n 1 and an LU factorization exists. It is unique since any LU factorization must satisfy the deriving equations. If d n 0 then r n > 0 regarless of wether a n is zero or nonzero. L and R have nonzero diagonal elements so the product A = LR is nonsingular.
23 The inverse of T We could solve the system Tv = b by using the inverse T 1 of T and simply compute the matrix vector product v = T 1 b. n=6: T 1 = All elements in T 1 are nonzero and the calculation of T 1 b requires O(n 2 ) flops. Using trifactor and trisolve only requires O(n) flops. So using the inverse is not a good idea.
24 An Interpolation Problem Consider a different problem. Given an integer n 2, x values x = [x 1,..., x n ] T with a := x 1 < < x n =: b. y values y = [y 1,..., y n ] T, derivative values σ a, σ b. Find a function g : [a, b] R such that g(x i ) = y i, for i = 1,..., n, g (a) = σ a, g (b) = σ b. The conditions g (a) = σ a, g (b) = σ b are called 1. derivative (clamped, Hermite) boundary conditions.
25 Special case n = 2: Cubic Hermite Interpolation f (x) = x 4, g(x) = 4x 3 4x 2, g is a cubic polynomial, g(0) = f (0), g(2) = f (2), g (0) = f (0), g (2) = f (2).
26 Special case: polynomial interpolation to Arctan f (x) = Arctan(10x) + π/2, 1 x 1. n + 2 interpolation conditions. g is a polynomial interpolant to f of degree n + 1 = 15 x i = 1 + 2(i 1)/(n 1), i = 1,..., n = 14. Large oscillations near the ends of the range.
27 Cubic Spline Definition Let n 3, and a = x 1 < x 2 <... < x n = b. A function g : [a, b] R of the form p 1 (x), if a x < x 2, p 2 (x), if x 2 x < x 3, g(x) :=. p n 2 (x), if x n 2 x < x n 1, p n 1 (x), if x n 1 x b, (3) is called a cubic spline with knots x = [x 1,..., x n ] T provided (i) Each p i is a polynomial of degree 3. (ii) p i 1 (x i ) = p i (x i ), p i 1 (x i) = p i (x i), p i 1 (x i) = p i (x i), i = 2,..., n 1.
28 Cubic Spline Interpolant Given x values x = [x 1,..., x n ] T with a := x 1 < < x n =: b. y values y = [y 1,..., y n ] T, Derivative values σ a, σ b. A cubic spline with knots x that satisfies g(x i ) = y i, for i = 1,..., n, g (a) = σ a, g (b) = σ b is called a cubic spline interpolant. Each of the n 1 cubic polynomials p i has 4 coefficients and there are 3 continuity conditions at each of the n 2 interior knots. Thus, a cubic spline appears to have 4(n 1) 3(n 2) = n + 2 degrees of freedom. This is the same as the number of interpolation conditions
29 Special case: cubic spline interpolation to Arctan f (x) = Arctan(10x) + π/2, 1 x 1. Compute cubic spline interpolant g for n = 14, x i = 1 + 2(i 1)/(n 1), i = 1,..., n. g is composed of 13 polynomial pieces.
30 PP Representation We will represent the polynomial pieces in shifted power form: p i (x) = c 1i + c 2i (x x i ) + c 3i (x x i ) 2 + c 4i (x x i ) 3. In general the pp form of a cubic spline g with knots x 1,..., x n is c 11 c 12 c 1,k x s := [x 1,..., x k ], C := c 21 c 22 c 2,k c 31 c 32 c 3,k R4,k, c 41 c 42 c 4,k shifts x s, coefficients: C. k = n 1.
31 Determining the interpolant Given a = x 1 <... < x n = b, y i R, i = 1,..., n, and any numbers s 1,..., s n. Suppose the pp form x s, C = [c ji ] of g is given by x T s = [x 1,..., x n 1 ] and c 1i := y i, c 2i := s i, c 3i := (3δ i 2s i s i+1 )/h i, c 4i := ( 2δ i + s i + s i+1 )/h 2 i, (4) where Then h i = x i+1 x i, g(x i ) = y i, g (x i ) = s i for i = 1,..., n. δ i = y i+1 y i h i. (5) p i 1 (x i ) = p i (x i ), p i 1 (x i) = p i (x i), i = 2,..., n 1.
32 The C 2 Conditions For a cubic spline we need in addition p i 1 (x i) = p i (x i), i = 2,..., n 1. p i (x) = c 1i + c 2i (x x i ) + c 3i (x x i ) 2 + c 4i (x x i ) 3. p i (x) = 2c 3,i + 6c 4,i (x x i ) p i (x i) = 2c 3,i, p i 1 (x i) = 2c 3,i 1 + 6c 4,i 1 h i 1. (1 0 = h i 1 h i 2 p i 1(x i ) 1 2 p i (x i ) ) ( ) = h i 1 h i c3,i 1 + 3c 4,i 1 h i 1 c i,3 = h i (3δ i 1 2s i 1 s i ) + 3h i ( 2δ i 1 + s i 1 + s i ) h i 1 (3δ i 2s i s i+1 ) = h i s i 1 + 2(h i 1 + h i )s i + h i 1 s i+1 3(h i δ i 1 + h i 1 δ i ). Rearranging we obtain h i s i 1 +2(h i 1 +h i )s i +h i 1 s i+1 = 3(h i δ i 1 +h i 1 δ i ), i = 2,..., n 1.
33 A Linear System h i s i 1 +2(h i 1 +h i )s i +h i 1 s i+1 = 3(h i δ i 1 +h i 1 δ i ), Multiplying the last expression by 2/(h i 1 + h i ) leads to the equations λ i = i = 2,..., n 1 λ i s i 1 + 4s i + µ i s i+1 = β i, i = 2,..., n 1, (6) 2h i h i 1 + h i, µ i = 2h i 1 h i 1 + h i, h i = x i+1 x i, β i = 3(λ i δ i 1 + µ i δ i ), δ i = y i+1 y i. h i Obtain the linear system 1 0 s 1 λ 2 4 µ 2 s 2 Ns = = λ n 1 4 µ n 1 s n s n σ a β 2. β n 1 σ b (7) =: b. (8)
34 The matrix N 1 0 λ 2 4 µ 2 N = λ n 1 4 µ n N is tridiagonal and strictly diagonally dominant. The algorithms trifactor and trisolve can be used to find s. N is nonsingular. The cubic spline interpolant exists and is unique.
35 Finding g Solve the linear systen Ns = b, Compute the pp representation, O(n) flops.
36 Uniform knots Assume h i = h > 0, all i. λ i = 2h i h i 1 +h i = 1, µ i = 2h i 1 h i 1 +h i = 1. β i = 3(λ i δ i 1 + µ i δ i ) = 3((y i y i 1 )/h + (y i+1 y i )/h = 3(y i+1 y i 1 )/h. 1 0 s 1 σ a s 2 3(y 3 y 1 )/h Ns = = s n 1 3(y n y n 2 )/h 0 1 s n σ b
37 Example, f (x) = x 4 x = [0, 1, 2] T, y = [0, 1, 16] T, σ a = 0, σ b = s s 2 = s 3 32 s 1 = 0, s 3 = 32, s 2 = 4. c 1,1 = y 1 = 0, c 2,1 = s 1 = 0, c 3,1 = (3δ 1 2s 1 s 2 )/h = = 1, c 4,1 = ( 2δ 1 + s 1 + s 2 )/h 2 = = 2, c 1,2 = y 2 = 1, c 2,2 = s 2 = 4, c 3,2 = (3δ 2 2s 2 s 3 )/h = = 5, c 4,2 = ( 2δ 2 + s 2 + s 3 )/h 2 = = 6, p 1 (x) = x 2 + 2x 3. p 2 (x) = 1 + 4(x 1) + 5(x 1) 2 + 6(x 1) 3 = x 13x 2 + 6x 3.
38 Smooth Piecewise Polynomial, f (x) = x 4 16 g(x) := { p 1 (x) = x 2 + 2x 3, if 0 x < 1, p 2 (x) = x 13x 2 + 6x 3, if 1 x 2,
39 Compute g(x), but where is x? Algorithm (findsubintervals) Given shifts x s = [x 1,..., x k ] and a real number r, an integer i is computed so that i = 1 if r < x(2) i = k if r x k and x i r < x i+1 otherwise. If r is a vector then a vector i is computed, such that the jth component of i gives the location of the jth component of r. 1. function i=findsubintervals(xs,r) 2. k=length(xs); m=length(r); 3. xs(1)=min(r)-1; 4. [sorted,j] = sort([xs(:) r(:) ]); 5. i = find(j>k)-(1:m);
40 Spline Evaluation Algorithm (cubppeval) Given a pp-representation (x s, C) of a cubic spline g together with x-values r R m. The vector q = g(r) is computed. 1. function q=cubppeval(xs,c,r) 2. i=findsubintervals(xs,r); q=r; 3. for j=1:length(r) 4. k=i(j); t=r(j)-xs(k); 5. q(j)=[1 t t 2 t 3 ]*C(:,k); 6. end
41 Plotting a Spline in MATLAB x=[0 1 2]; n=length(x); h=(x(n)-x(1))/199; r=x(1):h:x(n)+h/2; xs=[0 1]; C=[0 1; 0 4; -1 5; 2 6]; q=cubppeval(xs,c,r) plot(r,q);
42 Plot 16
43 The Physical Spline A cubic spline interpolant minimizes g (x) 2 dx over all C 2 interpolants. It is a mathematical model of a physical spline.
Lecture 2 INF-MAT : A boundary value problem and an eigenvalue problem; Block Multiplication; Tridiagonal Systems
Lecture 2 INF-MAT 4350 2008: A boundary value problem and an eigenvalue problem; Block Multiplication; Tridiagonal Systems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University
More informationLecture 1 INF-MAT3350/ : Some Tridiagonal Matrix Problems
Lecture 1 INF-MAT3350/4350 2007: Some Tridiagonal Matrix Problems Tom Lyche University of Oslo Norway Lecture 1 INF-MAT3350/4350 2007: Some Tridiagonal Matrix Problems p.1/33 Plan for the day 1. Notation
More informationLecture 2 INF-MAT : , LU, symmetric LU, Positve (semi)definite, Cholesky, Semi-Cholesky
Lecture 2 INF-MAT 4350 2009: 7.1-7.6, LU, symmetric LU, Positve (semi)definite, Cholesky, Semi-Cholesky Tom Lyche and Michael Floater Centre of Mathematics for Applications, Department of Informatics,
More information1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )
Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical
More informationChapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations
Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationLecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II)
Math 59 Lecture 2 (Tue Mar 5) Gaussian elimination and LU factorization (II) 2 Gaussian elimination - LU factorization For a general n n matrix A the Gaussian elimination produces an LU factorization if
More informationLecture Notes for Inf-Mat 3350/4350, Tom Lyche
Lecture Notes for Inf-Mat 3350/4350, 2007 Tom Lyche August 5, 2007 2 Contents Preface vii I A Review of Linear Algebra 1 1 Introduction 3 1.1 Notation............................... 3 2 Vectors 5 2.1 Vector
More informationDirect Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le
Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization
More informationThe Solution of Linear Systems AX = B
Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has
More informationGAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)
GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) D. ARAPURA Gaussian elimination is the go to method for all basic linear classes including this one. We go summarize the main ideas. 1.
More informationCHAPTER 10 Shape Preserving Properties of B-splines
CHAPTER 10 Shape Preserving Properties of B-splines In earlier chapters we have seen a number of examples of the close relationship between a spline function and its B-spline coefficients This is especially
More informationLecture4 INF-MAT : 5. Fast Direct Solution of Large Linear Systems
Lecture4 INF-MAT 4350 2010: 5. Fast Direct Solution of Large Linear Systems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 16, 2010 Test Matrix
More informationChapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in
Chapter 4 - MATRIX ALGEBRA 4.1. Matrix Operations A a 11 a 12... a 1j... a 1n a 21. a 22.... a 2j... a 2n. a i1 a i2... a ij... a in... a m1 a m2... a mj... a mn The entry in the ith row and the jth column
More informationMATRICES. a m,1 a m,n A =
MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of
More informationChapter 1 Matrices and Systems of Equations
Chapter 1 Matrices and Systems of Equations System of Linear Equations 1. A linear equation in n unknowns is an equation of the form n i=1 a i x i = b where a 1,..., a n, b R and x 1,..., x n are variables.
More informationSection 9.2: Matrices.. a m1 a m2 a mn
Section 9.2: Matrices Definition: A matrix is a rectangular array of numbers: a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn In general, a ij denotes the (i, j) entry of A. That is, the entry in
More information2.1 Gaussian Elimination
2. Gaussian Elimination A common problem encountered in numerical models is the one in which there are n equations and n unknowns. The following is a description of the Gaussian elimination method for
More informationInverses. Stephen Boyd. EE103 Stanford University. October 28, 2017
Inverses Stephen Boyd EE103 Stanford University October 28, 2017 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number
More informationLinear Systems and Matrices
Department of Mathematics The Chinese University of Hong Kong 1 System of m linear equations in n unknowns (linear system) a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.......
More informationMatrix & Linear Algebra
Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A
More informationSystems of Linear Equations and Matrices
Chapter 1 Systems of Linear Equations and Matrices System of linear algebraic equations and their solution constitute one of the major topics studied in the course known as linear algebra. In the first
More informationSection 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.
Section 9.2: Matrices Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns. That is, a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn A
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationSystems of Linear Equations and Matrices
Chapter 1 Systems of Linear Equations and Matrices System of linear algebraic equations and their solution constitute one of the major topics studied in the course known as linear algebra. In the first
More informationChapter 4 No. 4.0 Answer True or False to the following. Give reasons for your answers.
MATH 434/534 Theoretical Assignment 3 Solution Chapter 4 No 40 Answer True or False to the following Give reasons for your answers If a backward stable algorithm is applied to a computational problem,
More informationDirect Methods for Solving Linear Systems. Matrix Factorization
Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011
More informationLinear Algebra: Lecture notes from Kolman and Hill 9th edition.
Linear Algebra: Lecture notes from Kolman and Hill 9th edition Taylan Şengül March 20, 2019 Please let me know of any mistakes in these notes Contents Week 1 1 11 Systems of Linear Equations 1 12 Matrices
More informationCOURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.
COURSE 9 4 Numerical methods for solving linear systems Practical solving of many problems eventually leads to solving linear systems Classification of the methods: - direct methods - with low number of
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationComputing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices
Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 8, 2009 Today
More informationLINEAR ALGEBRA WITH APPLICATIONS
SEVENTH EDITION LINEAR ALGEBRA WITH APPLICATIONS Instructor s Solutions Manual Steven J. Leon PREFACE This solutions manual is designed to accompany the seventh edition of Linear Algebra with Applications
More informationGaussian Elimination and Back Substitution
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 4 Notes These notes correspond to Sections 31 and 32 in the text Gaussian Elimination and Back Substitution The basic idea behind methods for solving
More informationLinear Equations and Matrix
1/60 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Gaussian Elimination 2/60 Alpha Go Linear algebra begins with a system of linear
More information1 Piecewise Cubic Interpolation
Piecewise Cubic Interpolation Typically the problem with piecewise linear interpolation is the interpolant is not differentiable as the interpolation points (it has a kinks at every interpolation point)
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationLinear Algebra. Linear Equations and Matrices. Copyright 2005, W.R. Winfrey
Copyright 2005, W.R. Winfrey Topics Preliminaries Systems of Linear Equations Matrices Algebraic Properties of Matrix Operations Special Types of Matrices and Partitioned Matrices Matrix Transformations
More informationIntroduction. Vectors and Matrices. Vectors [1] Vectors [2]
Introduction Vectors and Matrices Dr. TGI Fernando 1 2 Data is frequently arranged in arrays, that is, sets whose elements are indexed by one or more subscripts. Vector - one dimensional array Matrix -
More informationPrepared by: M. S. KumarSwamy, TGT(Maths) Page
Prepared by: M. S. KumarSwamy, TGT(Maths) Page - 50 - CHAPTER 3: MATRICES QUICK REVISION (Important Concepts & Formulae) MARKS WEIGHTAGE 03 marks Matrix A matrix is an ordered rectangular array of numbers
More informationMATH 3511 Lecture 1. Solving Linear Systems 1
MATH 3511 Lecture 1 Solving Linear Systems 1 Dmitriy Leykekhman Spring 2012 Goals Review of basic linear algebra Solution of simple linear systems Gaussian elimination D Leykekhman - MATH 3511 Introduction
More informationMTH5112 Linear Algebra I MTH5212 Applied Linear Algebra (2017/2018)
MTH5112 Linear Algebra I MTH5212 Applied Linear Algebra (2017/2018) COURSEWORK 3 SOLUTIONS Exercise ( ) 1. (a) Write A = (a ij ) n n and B = (b ij ) n n. Since A and B are diagonal, we have a ij = 0 and
More informationLinear Algebra and Matrices
Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.
More informationPhys 201. Matrices and Determinants
Phys 201 Matrices and Determinants 1 1.1 Matrices 1.2 Operations of matrices 1.3 Types of matrices 1.4 Properties of matrices 1.5 Determinants 1.6 Inverse of a 3 3 matrix 2 1.1 Matrices A 2 3 7 =! " 1
More informationMath 502 Fall 2005 Solutions to Homework 3
Math 502 Fall 2005 Solutions to Homework 3 (1) As shown in class, the relative distance between adjacent binary floating points numbers is 2 1 t, where t is the number of digits in the mantissa. Since
More informationSubmatrices and Partitioned Matrices
2 Submatrices and Partitioned Matrices Two very important (and closely related) concepts are introduced in this chapter that of a submatrix and that of a partitioned matrix. These concepts arise very naturally
More informationA Review of Matrix Analysis
Matrix Notation Part Matrix Operations Matrices are simply rectangular arrays of quantities Each quantity in the array is called an element of the matrix and an element can be either a numerical value
More informationRadial Basis Functions I
Radial Basis Functions I Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo November 14, 2008 Today Reformulation of natural cubic spline interpolation Scattered
More informationReview of matrices. Let m, n IN. A rectangle of numbers written like A =
Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an
More informationI = i 0,
Special Types of Matrices Certain matrices, such as the identity matrix 0 0 0 0 0 0 I = 0 0 0, 0 0 0 have a special shape, which endows the matrix with helpful properties The identity matrix is an example
More informationMatrix Algebra Determinant, Inverse matrix. Matrices. A. Fabretti. Mathematics 2 A.Y. 2015/2016. A. Fabretti Matrices
Matrices A. Fabretti Mathematics 2 A.Y. 2015/2016 Table of contents Matrix Algebra Determinant Inverse Matrix Introduction A matrix is a rectangular array of numbers. The size of a matrix is indicated
More information9. Numerical linear algebra background
Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization
More informationMath 304 (Spring 2010) - Lecture 2
Math 304 (Spring 010) - Lecture Emre Mengi Department of Mathematics Koç University emengi@ku.edu.tr Lecture - Floating Point Operation Count p.1/10 Efficiency of an algorithm is determined by the total
More informationMidterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015
Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 The test lasts 1 hour and 15 minutes. No documents are allowed. The use of a calculator, cell phone or other equivalent electronic
More informationLecture Notes in Linear Algebra
Lecture Notes in Linear Algebra Dr. Abdullah Al-Azemi Mathematics Department Kuwait University February 4, 2017 Contents 1 Linear Equations and Matrices 1 1.2 Matrices............................................
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More information14.2 QR Factorization with Column Pivoting
page 531 Chapter 14 Special Topics Background Material Needed Vector and Matrix Norms (Section 25) Rounding Errors in Basic Floating Point Operations (Section 33 37) Forward Elimination and Back Substitution
More informationSteven J. Leon University of Massachusetts, Dartmouth
INSTRUCTOR S SOLUTIONS MANUAL LINEAR ALGEBRA WITH APPLICATIONS NINTH EDITION Steven J. Leon University of Massachusetts, Dartmouth Boston Columbus Indianapolis New York San Francisco Amsterdam Cape Town
More informationSPLINE INTERPOLATION
Spline Background SPLINE INTERPOLATION Problem: high degree interpolating polynomials often have extra oscillations. Example: Runge function f(x = 1 1+4x 2, x [ 1, 1]. 1 1/(1+4x 2 and P 8 (x and P 16 (x
More informationMatrix Arithmetic. j=1
An m n matrix is an array A = Matrix Arithmetic a 11 a 12 a 1n a 21 a 22 a 2n a m1 a m2 a mn of real numbers a ij An m n matrix has m rows and n columns a ij is the entry in the i-th row and j-th column
More informationLU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark
DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline
More informationMAT 2037 LINEAR ALGEBRA I web:
MAT 237 LINEAR ALGEBRA I 2625 Dokuz Eylül University, Faculty of Science, Department of Mathematics web: Instructor: Engin Mermut http://kisideuedutr/enginmermut/ HOMEWORK 2 MATRIX ALGEBRA Textbook: Linear
More informationCHAPTER 3 Further properties of splines and B-splines
CHAPTER 3 Further properties of splines and B-splines In Chapter 2 we established some of the most elementary properties of B-splines. In this chapter our focus is on the question What kind of functions
More informationReview of Vectors and Matrices
A P P E N D I X D Review of Vectors and Matrices D. VECTORS D.. Definition of a Vector Let p, p, Á, p n be any n real numbers and P an ordered set of these real numbers that is, P = p, p, Á, p n Then P
More informationStage-structured Populations
Department of Biology New Mexico State University Las Cruces, New Mexico 88003 brook@nmsu.edu Fall 2009 Age-Structured Populations All individuals are not equivalent to each other Rates of survivorship
More information5. Direct Methods for Solving Systems of Linear Equations. They are all over the place...
5 Direct Methods for Solving Systems of Linear Equations They are all over the place Miriam Mehl: 5 Direct Methods for Solving Systems of Linear Equations They are all over the place, December 13, 2012
More informationLeast Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. under-determined
More informationOrthonormal Transformations
Orthonormal Transformations Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 25, 2010 Applications of transformation Q : R m R m, with Q T Q = I 1.
More information10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )
c Dr. Igor Zelenko, Fall 2017 1 10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections 7.2-7.4) 1. When each of the functions F 1, F 2,..., F n in right-hand side
More informationLecture Summaries for Linear Algebra M51A
These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture
More informationSolving Linear Systems Using Gaussian Elimination. How can we solve
Solving Linear Systems Using Gaussian Elimination How can we solve? 1 Gaussian elimination Consider the general augmented system: Gaussian elimination Step 1: Eliminate first column below the main diagonal.
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 12: Gaussian Elimination and LU Factorization Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 10 Gaussian Elimination
More information1 Determinants. 1.1 Determinant
1 Determinants [SB], Chapter 9, p.188-196. [SB], Chapter 26, p.719-739. Bellow w ll study the central question: which additional conditions must satisfy a quadratic matrix A to be invertible, that is to
More informationFundamentals of Engineering Analysis (650163)
Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationSOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS We want to solve the linear system a, x + + a,n x n = b a n, x + + a n,n x n = b n This will be done by the method used in beginning algebra, by successively eliminating unknowns
More informationNext topics: Solving systems of linear equations
Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:
More informationThe matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.
) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to
More informationCS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:
CS100: DISCRETE STRUCTURES Lecture 3 Matrices Ch 3 Pages: 246-262 Matrices 2 Introduction DEFINITION 1: A matrix is a rectangular array of numbers. A matrix with m rows and n columns is called an m x n
More informationMatrices and Vectors
Matrices and Vectors James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 11, 2013 Outline 1 Matrices and Vectors 2 Vector Details 3 Matrix
More informationGraduate Mathematical Economics Lecture 1
Graduate Mathematical Economics Lecture 1 Yu Ren WISE, Xiamen University September 23, 2012 Outline 1 2 Course Outline ematical techniques used in graduate level economics courses Mathematics for Economists
More informationAMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems
AMS 209, Fall 205 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems. Overview We are interested in solving a well-defined linear system given
More informationCHAPTER 6. Direct Methods for Solving Linear Systems
CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to
More informationNecessary And Sufficient Conditions For Existence of the LU Factorization of an Arbitrary Matrix.
arxiv:math/0506382v1 [math.na] 19 Jun 2005 Necessary And Sufficient Conditions For Existence of the LU Factorization of an Arbitrary Matrix. Adviser: Charles R. Johnson Department of Mathematics College
More informationMatrix Algebra. Matrix Algebra. Chapter 8 - S&B
Chapter 8 - S&B Algebraic operations Matrix: The size of a matrix is indicated by the number of its rows and the number of its columns. A matrix with k rows and n columns is called a k n matrix. The number
More informationMatrices and Determinants
Chapter1 Matrices and Determinants 11 INTRODUCTION Matrix means an arrangement or array Matrices (plural of matrix) were introduced by Cayley in 1860 A matrix A is rectangular array of m n numbers (or
More informationCS412: Lecture #17. Mridul Aanjaneya. March 19, 2015
CS: Lecture #7 Mridul Aanjaneya March 9, 5 Solving linear systems of equations Consider a lower triangular matrix L: l l l L = l 3 l 3 l 33 l n l nn A procedure similar to that for upper triangular systems
More informationMA 138 Calculus 2 with Life Science Applications Matrices (Section 9.2)
MA 38 Calculus 2 with Life Science Applications Matrices (Section 92) Alberto Corso albertocorso@ukyedu Department of Mathematics University of Kentucky Friday, March 3, 207 Identity Matrix and Inverse
More informationJim Lambers MAT 610 Summer Session Lecture 1 Notes
Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra
More informationMATH2210 Notebook 2 Spring 2018
MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................
More informationGaussian Elimination -(3.1) b 1. b 2., b. b n
Gaussian Elimination -() Consider solving a given system of n linear equations in n unknowns: (*) a x a x a n x n b where a ij and b i are constants and x i are unknowns Let a n x a n x a nn x n a a a
More informationOHSx XM511 Linear Algebra: Solutions to Online True/False Exercises
This document gives the solutions to all of the online exercises for OHSx XM511. The section ( ) numbers refer to the textbook. TYPE I are True/False. Answers are in square brackets [. Lecture 02 ( 1.1)
More informationKevin James. MTHSC 3110 Section 2.1 Matrix Operations
MTHSC 3110 Section 2.1 Matrix Operations Notation Let A be an m n matrix, that is, m rows and n columns. We ll refer to the entries of A by their row and column indices. The entry in the i th row and j
More informationChapter 7. Tridiagonal linear systems. Solving tridiagonal systems of equations. and subdiagonal. E.g. a 21 a 22 a A =
Chapter 7 Tridiagonal linear systems The solution of linear systems of equations is one of the most important areas of computational mathematics. A complete treatment is impossible here but we will discuss
More informationMAT 610: Numerical Linear Algebra. James V. Lambers
MAT 610: Numerical Linear Algebra James V Lambers January 16, 2017 2 Contents 1 Matrix Multiplication Problems 7 11 Introduction 7 111 Systems of Linear Equations 7 112 The Eigenvalue Problem 8 12 Basic
More informationHani Mehrpouyan, California State University, Bakersfield. Signals and Systems
Hani Mehrpouyan, Department of Electrical and Computer Engineering, Lecture 26 (LU Factorization) May 30 th, 2013 The material in these lectures is partly taken from the books: Elementary Numerical Analysis,
More informationMultiplying matrices by diagonal matrices is faster than usual matrix multiplication.
7-6 Multiplying matrices by diagonal matrices is faster than usual matrix multiplication. The following equations generalize to matrices of any size. Multiplying a matrix from the left by a diagonal matrix
More informationMA2501 Numerical Methods Spring 2015
Norwegian University of Science and Technology Department of Mathematics MA2501 Numerical Methods Spring 2015 Solutions to exercise set 3 1 Attempt to verify experimentally the calculation from class that
More informationWilliam Stallings Copyright 2010
A PPENDIX E B ASIC C ONCEPTS FROM L INEAR A LGEBRA William Stallings Copyright 2010 E.1 OPERATIONS ON VECTORS AND MATRICES...2 Arithmetic...2 Determinants...4 Inverse of a Matrix...5 E.2 LINEAR ALGEBRA
More informationChapter 3. Linear and Nonlinear Systems
59 An expert is someone who knows some of the worst mistakes that can be made in his subject, and how to avoid them Werner Heisenberg (1901-1976) Chapter 3 Linear and Nonlinear Systems In this chapter
More informationOR MSc Maths Revision Course
OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision
More information