Lecture Notes for Inf-Mat 3350/4350, Tom Lyche

Size: px
Start display at page:

Download "Lecture Notes for Inf-Mat 3350/4350, Tom Lyche"

Transcription

1 Lecture Notes for Inf-Mat 3350/4350, 2007 Tom Lyche August 5, 2007

2 2

3 Contents Preface vii I A Review of Linear Algebra 1 1 Introduction Notation Vectors Vector Spaces Linear Independence and Bases Operations on Subspaces Sums and intersections of subspaces The quotient space Norms Convergence of Vectors Convergence of series of vectors Inner Products Orthogonality Projections and Orthogonal Complements Matrices Arithmetic Operations and Block Multiplication The Transpose Matrix Linear Systems The Inverse matrix Rank, Nullity, and the Fundamental Subspaces Linear Transformations and Matrices Orthonormal and Unitary Matrices Determinants Permutations Basic Properties of Determinants The Adjoint Matrix and Cofactor Expansion i

4 ii Contents 4.4 Computing Determinants Some Useful Determinant Formulas Gaussian Elimination Gaussian Elimination and LU-factorization Algoritms Operation count Pivoting Permutation matrices Gaussian elimination works mathematically Pivot strategies The PLU-Factorization An Algorithm for Finding the PLU-Factorization Eigenvalues and Eigenvectors The Characteristic Polynomial The characteristic equation Similarity Transformations Linear Independence of Eigenvectors Left Eigenvectors II Some Linear Systems with a Special Structure 73 7 Examples of Tridiagonal Systems Cubic Spline Interpolation Splines Cubic spline interpolation LU-factorization of a tridiagonal system A Two-Point Boundary Value Problem Numerical example An Eigenvalue Problem The buckling of a beam The eigenpairs of the matrix tridiag( 1, 2, 1) LU Factorization and Positive Definite Matrices Algebraic Properties of Triangular Matrices The LU and LDL T -Factorizations Positive Definite Matrices Definitions and examples When is a matrix positive definite? Algorithms Some Bivariate Problems and the Kronecker Product The Poisson Problem Block LU-factorization of a block-triangular matrix 106

5 Contents iii 9.2 The Kronecker Product Fast Direct Solution of a Large Linear System A Fast Poisson Solver based on Diagonalization A Fast Poisson Solver based on the Discrete Sine and Fourier Transforms The discrete sine transform (DST) The discrete Fourier transform (DFT) The fast Fourier transform (FFT) A Poisson solver based on the FFT Problems III Some Matrix Theory Orthonormal Eigenpairs and the Schur Form The Schur Form Hermitian and Normal Matrices The Spectral Theorem The Rayleigh Quotient and Minmax Theorems The Rayleigh Quotient Minmax Theorems The Hoffman-Wielandt Theorem Proof of the Real Schur Form The Singular Value Decomposition Singular Values and Singular Vectors Singular values The singular value decomposition Three forms of the SVD The SVD of A H Singular vectors The SVD of A H A and AA H Singular values of normal and positive definite matrices A geometric interpretation The Pseudo-inverse and Orthogonal Projections The pseudo-inverse Orthogonal projections The Minmax Theorem for Singular Values and the Hoffman- Wielandt Theorem Matrix Norms Matrix Norms Consistent and Subordinate Matrix Norms Operator norms The p-norms

6 iv Contents Unitary invariant matrix norms Absolute and monotone norms The Condition Number with Respect to Inversion Determining the Rank of a Matrix IV Iterative Methods for Linear Systems The Classical Iterative Methods Convergence and Spectral Radius Convergence in R m,n and C m,n The spectral radius Classical Iterative Methods; Component form Classical Iterative Methods; Matrix Form Convergence Convergence of SOR Numerical Example The Conjugate Gradient Method The Conjugate Gradient Algorithm Numerical Example Derivation and Basic Properties Convergence Minimization and Preconditioning Minimization Preconditioning Preconditioning Example A banded matrix Preconditioning V Orthonormal Transformations and Least Squares Orthonormal Transformations The QR decomposition and QR factorization QR and Gram-Schmidt The Householder Transformation Householder Triangulation Givens rotations Least Squares Examples Numerical Solution using the Normal Equations Numerical Solution using the QR-factorization QR and linear systems Numerical Solution using the Singular Value Factorization

7 Contents v 18.5 Perturbation Theory for Least Squares Perturbing the right hand side Perturbing the matrix Perturbation Theory for Singular Values VI Eigenvalues and Eigenvectors Numerical Eigenvalue Problems Perturbation of Eigenvalues Gerschgorin s theorem Unitary Similarity Transformation of a Matrix into Upper Hessenberg or Tridiagonal Form Computing a Selected Eigenvalue of a Symmetric Matrix The inertia theorem Approximating λ m Perturbation Proofs The Power and QR methods The Power method The inverse power method The QR-Algorithm The relation to the power method A convergence theorem The shifted QR-algorithms VII Appendix 259 A Computer Arithmetic 261 A.1 Absolute and Relative Errors A.2 Floating Point Numbers A.3 Rounding and Arithmetic Operations A.3.1 Rounding A.3.2 Arithmetic operations A.4 Backward Rounding-Error Analysis A.4.1 Computing a sum A.4.2 Computing an inner product A.4.3 Computing a matrix product B Differentiation of Vector Functions 271 C Some Inequalities 275 C.1 Convexity C.2 Inequalities D The Jordan Form 279

8 vi Contents D.1 The Jordan Form D.1.1 The minimal polynomial Bibliography 285 Index 287

9 Preface These lecture notes are the result of an ongoing project to write a text for a course in matrix analysis and numerical linear algebra given at the advanced undergraduate and beginning graduate level at the University of Oslo. In the first 6 chapters we give a quick review of basis linear algebra. Some of the results in this first part are not found in most introductory linear algebra courses. In particular we discuss algorithmic aspects of Gaussian elimination. The lectures can start with chapter 7 and then each of the remaining chapters 8-20 should correspond to one week of lectures. Oslo, 05 August, 2007 Tom Lyche vii

10 viii Preface

11 List of Algorithms 5.5 lufactor forwardsolve backsolve PLU-factorization Forward Substitution (column oriented) Backward Substitution (column oriented) locatesubinterval trifactor trisolve bandcholesky bandforwardsolve bandbacksolve Fast Poisson Solver Recursive FFT Stationary Iterative Methods Conjugate Gradient Iteration Testing Conjugate Gradient Preconditioned Conjugate Gradient Algorithm Generate a Householder transformation Upper Hessenberg linear system Solving least squares by Householder QR Householder reduction to Hessenberg form Assemble Householder transformations The Power Method Rayleigh quotient iteration ix

12 x List of Algorithms

13 List of Exercises xi

14 xii List of Exercises

15 List of Exercises xiii

16 xiv List of Exercises

17 List of Exercises xv A A A B D D D D D D D

18 xvi List of Exercises

19 Part I A Review of Linear Algebra 1

20

21 Chapter 1 Introduction 1.1 Notation The following sets will be used throughout these notes. 1. The set of natural numbers, integers, rational numbers, real numbers, and complex numbers are denoted by N, Z, Q, R, C, respectively. 2. R n is the set of n-tuples of real numbers which we will represent as column vectors. Thus x R n means x 1 x 2 x =.., where x i R for i = 1,..., n. Row vectors are normally identified using the transpose operation. Thus if x R n then x is a column vector and x T is a row vector. 3. R m,n is the set of m n matrices with real entries represented as a 11 a 12 a 1n a 21 a 22 a 2n A =.... a m1 a m2 a mn The entry in the ith row and jth column of a matrix A will be denoted by a i,j, a ij, A(i, j) or (A) i,j. We use the notations a 1j a 2j a.j =., a T at 2. i. = [a i1, a i2,..., a in ], A = [a.1, a.2,... a.n ] =.. a mj a T m. 3 x n a T 1.

22 4 Chapter 1. Introduction for the columns a.j and rows a T i. of A. We often drop the dots and write a j and a T i when no confusion can arise. If m = 1 then A is a row vector, if n = 1 then A is a column vector, while if m = n then A is a square matrix. In this text we will denote matrices by boldface capital letters A, B, C and vectors most often by boldface lower case letters x, y, z,. 4. The imaginary unit 1 is denoted by i. The complex conjugate and the modulus of a complex number z is denoted by z and z, respectively. Thus if z = x + iy = re iφ = r(cos φ + i sin φ) is a complex number then z = x iy = re iφ = cos φ i sin φ and z = x 2 + y 2 = r. Re(z) := x and Im(z) := y denote the real and imaginary part of the complex number z. 5. For matrices and vectors with complex entries we use the notation A C m,n and x C n. We identify complex row vectors using either the transpose T or the Hermitian (conjugate) transpose operation x H := x T = [x 1,..., x n ]. 6. The unit vectors in R n and C n are denoted by e 1 = 0, e 2 = 0, e 3 = 1,, e n = 0, while I n = I =: [δ ij ] n i,j=1, where δ ij :=. 0 { 1 if i = j, 0 otherwise, is the identity matrix of order n. Both the columns and the transpose of the rows of I are the unit vectors e 1, e 2,..., e n. 7. We use the following notations for diagonal- and tridiagonal n n matrices d d 2 0 diag(d i ) = diag(d 1,..., d n ) := = d d n d 1 c 1 a 2 d 2 c 2 B = tridiag(a i, d i, c i ) = tridiag(a, d, c) := d n, a n 1 d n 1 c n 1 a n d n Here b ii = d i for i = 1,..., n, b i+1,i = a i+1, b i,i+1 = c i for i = 1,..., n 1, and b ij = 0 otherwise. 8. We use the colon equal symbol v := e to indicate that the symbol v is defined by the expression e..

23 Chapter 2 Vectors This chapter contains a review of vector space concepts that will be useful in this text. we start by introducing a vector space. To define a vector space we need a field F, a set of vectors V, a way to combine vectors called vector addition, and a way to combine elements of F and V called scalar multiplication. In the first part of this section F will be an arbitrary field, but later the field will be the set of real or complex numbers with the usual arithmetic operations. 2.1 Vector Spaces Definition 2.1 A field is a set F together with two operations +, : F F F such that for all a, b, c F the following arithmetic rules hold (A0) there exists an element 0 F such that a + 0 = a. (Am) there exists an element ( a) F such that a + ( a) = 0. We define subtraction as a b := a + ( b). (Aa) a + (b + c) = (a + b) + c. (Ac) a + b = b + a. (M1) there exists an element 1 F such that a 1 = a. (Mi) if a 0 then there exists an element a 1 F such that a a 1 = 1. (Ma) a (b c) = (a b) c. (Mc) a b = b a. (D) a (b + c) = a b + a c. The requirements (A0), (Am), (Aa) are the axioms for a group. They state that (F, +) is a group, and since in addition (Ac) holds then (F, +) is by defintion an 5

24 6 Chapter 2. Vectors abelian group. The axioms (M1), (Mi), (Ma), (Mc) state that (F\{0}, ) is an abelian group. Often we drop the dot and write ab for the product a b. Examples of fields are R or C with ordinary addition and multiplication. Definition 2.2 A vector space over a field F is a set V together with two operations vector addition, + : V V V and scalar multiplication, : F V V such that for all a, b F and v, w V the following hold (V) (V, +) is an abelian group. (Va) (a b) v = a (b v). (Vd1) (a + b) v = a v + b v. (Vd2) a (v + w) = a v + a w. (M1) 1 v = v. We denote a vector space by (V, F) or by V if the underlying field is clear from the context. Definition 2.3 Let (V, F) be a vector space and S a nonempty subset of V. Then (S, F) is a subspace of (V, F) if (S, F) is itself a vector space. It follows that (S, F) is a subspace of (V, F) if S is closed under vector addition and scalar multiplication, i.e. as 1 + bs 2 S for all a, b F and all s 1, s 2 S. For any vector space (V, F) the two sets {0}, consisting only of the zero element in V, and V itself are subspaces. They are called the trivial subspaces. Here are some examples of vector spaces. Example 2.4 (The Vector Spaces R n and C n ) In the following chapters we will deal almost exclusively with the vector spaces R n = (R n, R), C n = (C n, C) and their subspaces. Addition and scalar multiplication are defined by v + w = v 1 + w 1.. v n + w n, av = av 1... av n Example 2.5 (Subspaces of R 2 and R 3 ) For a given vector x R n let S = {tx : t R}. Then S is a subspace of R n, in fact it represents a straight line passing through the origin. For n = 2 it can be shown that all nontrivial subspaces of R 2 are of this form. For n = 3 the nontrivial subspaces are all lines and all planes containing {0}.

25 2.1. Vector Spaces 7 Example 2.6 (The Vector Space C(I)) Let F = R and let C(I) be the set of all real valued functions f : I R which are defined and continuous on an interval I R. Here the vectors are functions in C(I). Vector addition and scalar multiplication are defined for all f, g C(I) and all a R by (f + g)(x) := f(x) + g(x), (af)(x) := af(x), for all x I. C(I) = ( C(I), R ) is a vector space since the sum of two continuous functions is continuous, a constant times a continuous function is continuous vector addition and scalar multiplication are defined point-wise, so the axioms for a vector space follows from properties of real numbers. Example 2.7 (The Vector Space Π n ) Let Π n (I) be the set of all polynomials of degree at most n defined on a subset I R or I C. We write simply Π n if I = R or I = C. With pointwise addition and scalar multiplication defined as in Example 2.6 the set (Π n (I), R) is a subspace of (C(I), R). Definition 2.8 (Linear Combinations) The sum c 1 v 1 + c 2 v c n v n with c i F and v i V for i = 1,..., n is called a linear combination of v 1,..., v n. We say that the linear combination is nontrivial if at least one of the c i s is nonzero. The set span{v 1,..., v n } := {c 1 v c n v n : c i F, i = 1,..., n} spanned by v 1,..., v n V is a subspace of (V, F). A vector space V is called finite dimensional if it has a finite spanning set; i.e. there exist n N and {v 1,..., v n } in V such that V = span{v 1,..., v n }. Exercise 2.9 Show that the 0 of vector addition is unique and that {0} is a subspace. Exercise 2.10 Show that 0 x = 0 for any x V. Exercise 2.11 Show that span{v 1,..., v n } is a subspace. Exercise 2.12 Show that span{v 1,..., v n } is the smallest subspace containing the vectors v 1,..., v n.

26 8 Chapter 2. Vectors 2.2 Linear Independence and Bases Definition 2.13 Let X := {v 1,..., v n } be a set of vectors in a vector space (V, F). We say that X is linearly dependent if we can find a nontrivial linear combination which is equal to zero. We say that X is linearly independent if it is not linearly dependent. In other words c 1 v c n v n = 0 for some c 1,..., c n F = c 1 = = c n = 0. The elements in a set of linearly independent vectors must all be nonzero and we have Lemma 2.14 Suppose v 1,..., v n span a vector space V and that w 1,..., w k are linearly independent vectors in V. Then k n. Proof. Suppose k > n. Write w 1 as a linear combination of elements from the set X 0 := {v 1,..., v n }, say w 1 = c 1 v 1 + +c n v n. Since w 1 0 not all the c s are equal to zero. Pick a nonzero c, say c i1. Then v i1 can be expressed as a linear combination of w 1 and the remaining v s. So the set X 1 := {w 1, v 1,..., v i1 1, v i1+1,..., v n } must also be a spanning set for V. We repeat this for w 2 and X 1. In the linear combination w 2 = d i1 w 1 + j i 1 d j v j, we must have d i2 0 for some i 2. Moreover i 2 i 1 for otherwise w 2 = d 1 w 1 contradicting the linear independence of the w s. So the set X 2 consisting of the v s with v i1 replaced by w 1 and v i2 replaced by w 2 is again a spanning set for V. Repeating this process n 2 more times we obtain a spanning set X n where all the v s have been replaced by w 1,..., w n. Since k > n we can then write w k as a linear combination of w 1,..., w n contradicting the linear independence of the w s. We conclude that k n. Definition 2.15 A finite set of vectors {v 1,..., v n } in a vector space (V, F) is a basis for (V, F) if 1. span{v 1,..., v n } = V. 2. {v 1,..., v n } is linearly independent. Theorem 2.16 Suppose (V, F) is a vector space and that S := {v 1,..., v n } is a spanning set for V. Then we can find a subset {v i1,..., v ik } of S that forms a basis for V. Proof. If {v 1,..., v n } is linearly dependent we can express one of the v s as a nontrivial linear combination of the remaining v s and drop that v from the spanning set. Continue this process until the remaining v s are linearly independent. They still span the vector space and therefore form a basis. Corollary 2.17 A vector space is finite dimensional if and only if it has a basis.

27 2.2. Linear Independence and Bases 9 Proof. Let V = span{v 1,..., v n } be a finite dimensional vector space. By Theorem 2.16 V has a basis. Conversely, if V = span{v 1,..., v n } and {v 1,..., v n } is a basis then it is by defintion a finite spanning set. Theorem 2.18 Every basis for a vector space V has the same number of elements. This number is called the dimension of the vector space and denoted dim V. Proof. Suppose X = {v 1,..., v n } and Y = {w 1,..., w k } are two bases for V. By Lemma 2.14 we have k n. Using the same Lemma with X and Y switched we obtain n k. We conclude that n = k. The set of unit vectors {e 1,..., e n } form a basis for both R n and C n. The dimension of the trivial subspace {0} is defined to be zero. Theorem 2.19 Every linearly independent set of vectors {v 1,..., v k } in a finite dimensional vector space V can be enlarged to a basis for V. Proof. If {v 1,..., v k } does not span V we can enlarge the set by one vector v k+1 which cannot be expressed as a linear combination of {v 1,..., v k }. The enlarged set is also linearly independent. Continue this process. Since the space is finite dimensional it must stop after a finite number of steps. It is convenient to introduce a matrix transforming a basis in a subspace into a basis for the space itself. Lemma 2.20 Suppose S is a subspace of a finite dimensional vector space (V, F) and let {s 1,..., s n } be a basis for S and {v 1,..., v m } a basis for V. Then each s j can be expressed as a linear combination of v 1,..., v m, say s j = m a ij v i for j = 1,..., n. (2.1) i=1 If x S then x = n j=1 c js j = m i=1 b iv i for some coefficients b := [b 1,..., b m ] T, c := [c 1,..., c n ] T. Moreover b = Ac, where A = [a ij ] C m,n. The matrix A has linearly independent columns. Proof. (2.1) holds since s j V and {v 1,..., v n } spans V. Since {s 1,..., s n } is a basis for S and {v 1,..., v m } a basis for V every x S can be written x = n j=1 c js j = m i=1 b iv i for some scalars (c j ) and (b i ). But then x = n j=1 (2.1) c j s j = n ( m ) m ( n ) m c j a ij v i = a ij c j vi = b i v i. j=1 i=1 Since {v 1,..., v m } is linearly independent it follows that b i = n j=1 a ijc j for i = 1,..., m or b = Ac. Finally, to show that A has linearly independent columns i=1 j=1 i=1

28 10 Chapter 2. Vectors suppose b := Ac = 0 for some c = [c 1,..., c n ] T. Define x := n j=1 c js j. Then x = m i=1 b iv i and since b = 0 we have x = 0. But since {s 1,..., s n } is linearly independent we have c = 0. The matrix A in Lemma 2.20 is called a change of basis matrix. Exercise 2.21 Show that the elements in a linearly independent set must be nonzero. Exercise 2.22 Show that the set of unit vectors {e 1,..., e n } form a basis both for R n and for C n. Why does this show that the dimension of R n and C n is n? 2.3 Operations on Subspaces Let R and S be two subsets of a vector space (V, F) and let a be a scalar. The sum, multiplication by scalar, union, and intersection of R and S are defined by R + S := {r + s : r R and s S}, (2.2) as := {as : s S}, (2.3) R S := {x : x R or x S}. (2.4) R S := {x : x R and x S}. (2.5) Exercise 2.23 Let R = {(x, y) : x 2 + y 2 1} be the unit disc in R 2 and set S = {(x, y) : (x 1 2 )2 + y 2 1}. Find R + S, 2S, R S,, and R S Sums and intersections of subspaces In many cases R and S will be subspaces. Then as = S and both the sum and intersection of two subspaces is a subspace of (V, F). Note however that the union R S of two subspaces is not necessarily a subspace. Exercise 2.24 Let R and S be two subspaces of a vector space (V, F). Show that as = S and that both R + S and R S are subspaces of (V, F). Example 2.25 For given vectors x, y R n with x and y linearly independent let R = span{x} and S = span{y}. Then R and S are subspaces of R n. For n = 2 we have R + S = R 2, while for n = 3 the sum represents a plane passing through the origin. We also see that R S = {0} and that R S is not a subspace. Exercise 2.26 Show the statements made in Example Theorem 2.27 Let R and S be two subspaces of a vector space (V, F). Then dim(r + S) = dim(r) + dim(s) dim(r S). (2.6)

29 2.3. Operations on Subspaces 11 Proof. Let {u 1,..., u p } be a basis for R S, where {u 1,..., u p } =, the empty set, in the case R S = {0}. We use Theorem 2.19 to extend {u 1,..., u p } to a basis {u 1,..., u p, r 1,..., r q } for R and a basis {u 1,..., u p, s 1,..., s t } for S. Every x R + S can be written as a linear combination of {u 1,..., u p, r 1,..., r q, s 1,..., s t } so these vectors span R + S. We show that they are linearly independent and hence a basis. Suppose u + r + s = 0, where u := p j=1 α ju j, r := q j=1 ρ jr j, and s := t j=1 σ js j. Now r = (u + s) belongs to both R and to S and hence r R S. Therefore r can be written as a linear combination of u 1,..., u p say r := p j=1 β ju j and at the same time as a linear combination of r 1,..., r q. But then 0 = p j=1 β ju j q j=1 ρ jr j and since {u 1,..., u p, r 1,..., r q } is linearly independent we must have β 1 = = β p = ρ 1 = = ρ q = 0 and hence r = 0. We now have u + s = 0 and by linear independence of {u 1,..., u p, s 1,..., s t } we obtain α 1 = = α p = σ 1 = = σ t = 0. We have shown that the vectors {u 1,..., u p, r 1,..., r q, s 1,..., s t } constitute a basis for R + S. The result now follows from a simple calculation dim(r + S) = p + q + t = (p + q) + (p + t) p = dim(r) + dim(s) dim(r S). From this theorem it follows that dim(r + S) = dim(r) + dim(s) provided R S = {0}. Definition 2.28 (Direct Sum) Let R and S be two subspaces of a vector space (V, F). If R S = {0} then the subspace R + S is called a direct sum and denoted R S. The subspaces R and S are called complementary in the subspace R S. Theorem 2.29 Let R and S be two subspaces of a vector space (V, F) and assume R S = {0}. Every x R S can be decomposed uniquely in the form x = r + s, where r R and s S. If {r 1,..., r k } is a basis for R and {s 1,..., s n } is a basis for S then {r 1,..., r k, s 1,..., s n } is a basis for R S. Proof. To show uniqueness, suppose we could write x = r 1 + s 1 = r 2 + s 2 for r 1, r 2 R and s 1, s 2 S. Then r 1 r 2 = s 2 s 1 and it follows that r 1 r 2 and s 2 s 1 belong to both R and S and hence to R S. But then r 1 r 2 = s 2 s 1 = 0 so r 1 = r 2 and s 2 = s 1. Thus uniqueness follows. Suppose {r 1,..., r k } is a basis for R and {s 1,..., s n } is a basis for S. Since dim(r + S) = dim(r) + dim(s) the vectors {r 1,..., r k, s 1,..., s n } span R + S. To show linear independence suppose k j=1 ρ jr j + n j=1 σ js j = 0. The first sum belongs to R and the second to S and the sum is a decomposition of 0. By uniqueness of the decomposition both sums must be zero. But then ρ 1 = = ρ k = σ 1 = = σ n = 0 and linear independence follows.

30 12 Chapter 2. Vectors The quotient space For the sum of two sets we write x + S := {x + s : s S} when one of the sets is a singleton set {x}. Suppose S is a subspace of a vector space (X, F). Since as = S we have The set a(x + S) + b(y + S) = (ax + by) + S, for all a, b F and all x, y S. is a vector space if we define X /S := {x + S : x X } (2.7) a(x + S) + b(y + S) := (ax + by) + S, for all a, b F and all x, y S. The space X /S is called the quotient space of X by S. The zero element in X /S is S itself. Moreover, if x + S = y + S then x y S. Exercise 2.30 Show that X /S is a vector space. Theorem 2.31 Suppose S is a subspace of a finite dimensional vector space (X, F). Then dim(s) + dim ( X /S ) = dim(x ). (2.8) Proof. Let n := dim(x ), k = dim(s), and let {s 1,..., s k } be a basis for S. By Theorem 2.19 we can extend it to a basis {s 1,..., s k, t k+1,..., t n } for X. The result will follow if we can show that {t k+1 + S,..., t n + S} is a basis for X /S. Recall that the zero element in X /S is S. To show linear independence suppose n j=k+1 a j(t j + S) = S for some a k+1,..., a n in F. Since n j=k+1 a js = S and the zero element in X /S is unique we must have n j=k+1 a jt j = 0 which implies that a k+1 = = a n = 0 by linear independence of the t s. It remains to show that span{t k+1 + S,..., t n + S} = X /S. Suppose x + S X /S. For some a 1,..., a n we have x = x 1 + x 2, where x 1 = k j=1 a js j and x 2 = n j=k+1 a jt j. Since x 1 + S = S we have x + S = x 2 + S = n j=k+1 a jt j + S = n j=k+1 a j(t j + S) X /S. 2.4 Norms To measure the size of a vector in a vector space (V, F) we use norms. Definition 2.32 (Norm) A norm in a vector space (V, F), where F = R or F = C, is a function : V R that satisfies for all x, y in V and all a in F 1. x 0 with equality if and only if x = 0. (positivity) 2. ax = a x. (homogeneity) 3. x + y x + y. (subadditivity)

31 2.4. Norms 13 The triple (V, F, ) is said to be a normed vector space and the inequality 3. is called the triangle inequality. Since x = x y + y x y + y we obtain x y x y. By symmetry x y = y x y x and we obtain the inverse triangle inequality x y x y, x, y V. (2.9) Consider now some specific vector norms. For the vector spaces R n and C n we define for p 1 the p-norms by The most important cases are: x p := ( n x j p) 1/p, (2.10) j=1 1. x 1 = n j=1 x j, (the one-norm or l 1 -norm) x := max 1 j n x j. (2.11) 2. x 2 = ( n j=1 x j 2) 1/2, the two-norm, l2 -norm, or Euclidian norm) 3. x = max 1 j n x j, (the infinity-norm, l -norm, or max norm.) The infinity norm is related to the other p-norms by lim x p = x for all x C n. (2.12) p This clearly holds for x = 0. For x 0 we write x p := x ( n j=1 ( x j x ) p ) 1/p. Now each term in the sum is not greater than one and at least one term is equal to one and we obtain x x p n 1/p x, p 1. (2.13) Since lim p n 1/p = 1 for any n N we see that (2.12) follows. It can be shown (cf. Appendix C) that the p-norm are norms in R n and in C n for any p with 1 p. The triangle inequality x + y p x p + y p is called Minkowski s inequality. To prove it one first establishes Hölder s inequality n x j y j x p y q, j=1 1 p + 1 = 1. (2.14) q The relation 1 p + 1 q = 1 means that if p = 1 then q = and if p = 2 then q = 2. (2.13) shows that the infinity norm and any other p-norm can be bounded in terms of each other. We define

32 14 Chapter 2. Vectors Definition 2.33 Two norms and in a finite dimensional vector space (V, F) are equivalent if there are positive constants m and M (depending only on the dimension of V) such that for all vectors x V we have The following result is proved in Appendix C. m x x M x. (2.15) Theorem 2.34 All vector norms in a finite dimensional vector space are equivalent. The inverse triangle inequality (2.9) shows that a norm is a continuous function V R. Exercise 2.35 Show that p is a vector norm in R n for p = 1, p =. Exercise 2.36 The set S p = {x R n : x p = 1} is called the unit sphere in R n with respect to p. Draw S p for p = 1, 2, for n = 2. Exercise 2.37 Let 1 p. Produce a vector x l such that x l = x l p and another vector x u such that x u p = n 1/p x u p. Thus the inequalities in (2.12) are sharp. Exercise 2.38 If 1 q p then x p x q n 1/q 1/p x p, x C n. Hint: For the rightmost inequality use Jensen s inequality Cf. Theorem C.2 with f(z) = z p/q and z i = x i q. For the left inequality consider first y i = x i / x, i = 1, 2,..., n. 2.5 Convergence of Vectors Consider an infinite sequence {x k } = x 0, x 1, x 2,... of vectors in R n. This sequence converges to zero if and only if each component sequence x k (j) converges to zero for j = 1,..., n. In terms of the natural basis we have x k = n j=1 x k(j)e k and another way of stating convergence to zero is that in terms of the basis {e 1,..., e n } for R n each coefficient x k (j) of x k converges to zero. Consider now a more general vector space. Definition 2.39 Let {v 1,..., v n } be a basis for a finite dimensional vector space (V, F), where F = R or F = C, and let {x k } be an infinite sequence of vectors in V with basis coefficients {c k }, i.e. x k = n j=1 c kjv j for each k. We say that {x k } converges to zero, or have the limit zero, if lim k c kj = 0 for j = 1,..., n. We say that {x k } converge to the limit x in V if x k x converges to zero. We write this as lim k x k = x or x k x (as k ).

33 2.5. Convergence of Vectors 15 This definition is actually independent of the basis chosen. If {w 1,..., w n } is another basis for V and x k = n j=1 b kjw j for each k then from Lemma 2.20 b k = Ac k for some nonsingular matrix A independent of k. Hence c k 0 if and only if b k 0. If {a k } and {b k } are sequences of scalars and {x k } and {y k } are sequences of vectors such that {a k } a, {b k } b, {x k } x, and {y k } y then {a k x k + b k y k } ax + by. This shows that scalar multiplication and vector addition are continuous functions with respect to this notion of limit. Corresponding to a basis {v 1,..., v n }, we define x c := max 1 j n c j where x = n c j v j. We leave as an exercise to show that this is a norm on V. Recall that any two norms on V are equivalent. This implies that for any other norm on V there are positive constants α, β such that any x = n j=1 c jv j satisfy x α max 1 j n c j and c j β x for j = 1,..., n. (2.16) Suppose now (V, F, ) is a normed vector space with F = R or F = C. The notion of limit can then be stated in terms of convergence in norm. Theorem 2.40 In a normed vector space we have x k x if and only if lim k x k x = 0. Proof. Suppose {v 1,..., v n } is a basis for the vector space and assume x k, x V. Then x k x = n j=1 c kjv j for some scalars c kj By (2.16) we see that j=1 1 β max k,j c kj x k x α max k,j c kj and hence x k x 0 lim k c kj 0 for each j x k x. Since all vector norms are equivalent we have convergence in any norm we can define on a finite dimensional vector space. Definition 2.41 Let (V, F, ) be a normed vector space and let {x k } in V be an infinite sequence. 1. {x k } is a Cauchy sequence if lim k,l (x k x l ) = 0 or equivalently lim k,l x k x l = 0. More precisely, for each ɛ > 0 there is an integer N N such that for each k, l N we have x k x l ɛ. 2. The normed vector space is said to be complete if every Cauchy sequence converges to a point in the space. 3. {x k } is called bounded if there is a positive number M such that x k M for all k.

34 16 Chapter 2. Vectors 4. {x nk } is said to be a subsequence of {x k } k 0 if 0 n 0 < n 1 < n 2. Theorem 2.42 In a finite dimensional vector space V the following hold: 1. A sequence in V is convergent if and only if it is a Cauchy sequence. 2. V is complete. 3. Every bounded sequence in V has a convergent subsequence. Proof. 1. Suppose x k x. By the triangle inequality x k x l x k x + x l x and hence x k x l 0. Conversely, let {v 1,..., v n } be a basis for V and {x k } a Cauchy sequence with x k = n j=1 c kjv j for each k. Then x k x l = n j=1 (c kj c lj )v j and since lim k,l (x k x l ) = 0 we have by definition of convergence lim k,l (c kj c lj ) = 0 for j = 1,..., n. Thus for each j we have a Cauchy-sequence {c kj } C and since C is complete {c kj } converges to some c j C. But then x k x := n j=1 c jv j V. 2. V is complete since we just showed that every Cauchy sequence converges to a point in the space. 3. Let {v 1,..., v n } be a basis for V and {x k } be a bounded sequence with x k = n j=1 c kjv j for each k. By (2.16) each coefficient sequence {c kj } k is a bounded sequence of complex numbers and therefore, by a well known property of complex numbers, has a convergent subsequence. In particular the sequence of v 1 coefficients {c k1 } has a convergent subsequence c ki,1. For the second component the sequence {c ki,2} has a convergent subsequence, say c li,2. Continuing with j = 3,..., n we obtain integers 0 m 0 < m 1 < such that {c mi,j} is a convergent subsequence of c kj for j = 1,..., n. But then {x mi } is a convergent subsequence of {x k } Convergence of series of vectors Consider now an infinite series m=0 y m of vectors in a vector space (V, F) with F = R or F = C. We say that the series converges if the sequence of partial sums {x k } given by x k = k m=0 y m converges. A sufficient condition for convergence is that m=0 y m converges for some vector norm. We say that the series converges absolutely if this is the case. Note that m=0 y m m=0 y m, and absolute convergence in one norm implies absolute convergence in any norm by Theorem In an absolute convergent series we may change the order of the terms without changing the value of the sum. Problems for Section 2.5 Exercise 2.43 Show that if {a k } a, {b k } b, {x k } x, and {y k } y then {a k x k + b k y k } ax + by. Exercise 2.44 Show that c is a norm.

35 2.6. Inner Products Inner Products An inner product or scalar product in a vector space (V, F), where F = R or F = C, is a function, mapping pairs of vectors into a scalar. We consider first the case where F = R. Definition 2.45 An inner product in a vector space (V, R) is a function V V R satisfying for all x, y, z V and all a, b R the following conditions: 1. x, x 0 with equality if and only if x = 0. (positivity) 2. x, y = y, x (symmetry) 3. ax + by, z = a x, z + b y, z. (linearity) The triple (V, R,,, ) is called a real inner product space The standard inner product in V = R n is given by x, y := x T y. It is clearly an inner product in R n. When the field of scalars is C the inner product is complex valued and properties 2. and 3. are altered as follows: Definition 2.46 An inner product in a vector space (V, C) is a function V V C satisfying for all x, y, z V and all a, b C the following conditions: 1. x, x 0 with equality if and only if x = 0. (positivity) 2. x, y = y, x (skew symmetry) 3. ax + by, z = a x, z + b y, z. (linearity) The triple (V, C,,, ) is called a complex inner product space Note the complex conjugate in 2. and that (Cf. Exercise 2.52) x, ay + bz = a x, y + b x, z. (2.17) The standard inner product in C n is given by x, y := x H y = n j=1 x jy j. It is clearly an inner product in C n. Suppose now (V, F,,, ) is an inner product space with F = R or F = C. We define the inner product norm by x := x, x, x V. For any vectors x, y V and scalar a F we have (Cf. Exercises 2.51 and 2.52) by linearity and symmetry the expansion x + ay 2 = x 2 + 2a x, y + a 2 y 2 (real case), (2.18) = x 2 + 2Re x, ay + a 2 y 2 (complex case), (2.19) where Re z and Im z denotes the real- and imaginary part of the complex number z.

36 18 Chapter 2. Vectors In the complex case we can write the inner product of two vectors as a sum of inner product norms. For any x, y V it follows from (2.19) that 4 x, y = x + y 2 x y 2 + i x iy 2 i x + iy 2, (2.20) where i = 1 and we used that Im(z) = Re( iz) for any z C. To show that the inner product norm is a norm in (V, R) we need the triangle inequality. To show it we start with a famous inequality. Theorem 2.47 (Cauchy-Schwarz inequality) For any x, y in a real or complex inner product space x, y x y with equality if and only if x and y are linearly dependent. Proof. The inequality is trivial if x, y = 0 so assume x, y = 0. Suppose first x, y R. We define the scalar a := x,y y, and use (2.18) to obtain 0 2 x + ay 2 = x 2 ( x, y ) 2 / y 2. Thus the inequality follows in the real case. Suppose next x, y is complex valued, say x, y = re iφ. We define b := e iφ and observe that b x, y = r is real valued and b = 1. Using the real case of the Cauchy-Schwarz inequality we find x, y = b x, y = bx, y bx y = x y which proves the inequality also in the complex case. We have equality if and only if x + ay = 0 which means that x and y are linearly dependent. Theorem 2.48 (Triangle Inequality) For any x, y in a real or complex inner product space x + y x + y. Proof. From the Cauchy-Schwarz inequality it follows that Re x, y x y. Using this on the inner product term in (2.19) with a = 1 we get x + y 2 x x y + y 2 = ( x + y ) 2. Taking square roots completes the proof. Theorem 2.49 (Parallelogram Identity) For all x, y in a real or complex inner product space x + y 2 + x y 2 = 2 x y 2. Proof. We set a = ±1 in the inner product expansion (2.19) and add the two equations.

37 2.7. Orthogonality 19 In the real case the Cauchy-Schwarz inequality implies that 1 x,y x y 1 for nonzero x and y so there is a unique angle θ in [0, π] such that cos θ = x, y x y. (2.21) This defines the angle between vectors in a real inner product space. Exercise 2.50 Suppose A R m,n has linearly independent columns. Show that x, y := x T A T Ay defines an inner product on R n. F Exercise 2.51 Show (2.18) Exercise 2.52 Show (2.17) and (2.19). Exercise 2.53 Show (2.20) Exercise 2.54 Show that in the complex case there is a unique angle θ in [0, π/2] such that x, y cos θ = x y. (2.22) 2.7 Orthogonality As in the previous section we assume that (V, F,,, ) is an inner product space with F = R or F = C. Also denotes the inner product norm. Definition 2.55 (Orthogonality) Two vectors x, y in a real or complex inner product space are called orthogonal or perpendicular, denoted as x y, if x, y = 0. The vectors are orthonormal if in addition x = y = 1. For orthogonal vectors it follows from (2.19) that the Pythagorean theorem holds x + y 2 = x 2 + y 2, if x y. Definition 2.56 (Orthogonal- and Orthonormal Bases) A set of vectors {v 1,..., v k } in a subspace S of a real or complex inner product space is called an orthogonal basis for S if it is a basis for S and v i, v j = 0 for i j. It is an orthonormal basis for S if it is a basis for S and v i, v j = δ ij for all i, j. A basis for an inner product space can be turned into an orthogonal- or orthonormal basis for the subspace by the following construction.

Lecture Notes for Mat-inf 4130, Tom Lyche

Lecture Notes for Mat-inf 4130, Tom Lyche Lecture Notes for Mat-inf 4130, 2016 Tom Lyche August 15, 2016 2 Contents Preface ix 0 A Short Review of Linear Algebra 1 0.1 Notation.............................. 1 0.2 Vector Spaces and Subspaces...................

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Lecture 2 INF-MAT : A boundary value problem and an eigenvalue problem; Block Multiplication; Tridiagonal Systems

Lecture 2 INF-MAT : A boundary value problem and an eigenvalue problem; Block Multiplication; Tridiagonal Systems Lecture 2 INF-MAT 4350 2008: A boundary value problem and an eigenvalue problem; Block Multiplication; Tridiagonal Systems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

G1110 & 852G1 Numerical Linear Algebra

G1110 & 852G1 Numerical Linear Algebra The University of Sussex Department of Mathematics G & 85G Numerical Linear Algebra Lecture Notes Autumn Term Kerstin Hesse (w aw S w a w w (w aw H(wa = (w aw + w Figure : Geometric explanation of the

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Lecture 1 INF-MAT3350/ : Some Tridiagonal Matrix Problems

Lecture 1 INF-MAT3350/ : Some Tridiagonal Matrix Problems Lecture 1 INF-MAT3350/4350 2007: Some Tridiagonal Matrix Problems Tom Lyche University of Oslo Norway Lecture 1 INF-MAT3350/4350 2007: Some Tridiagonal Matrix Problems p.1/33 Plan for the day 1. Notation

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

MAT 610: Numerical Linear Algebra. James V. Lambers

MAT 610: Numerical Linear Algebra. James V. Lambers MAT 610: Numerical Linear Algebra James V Lambers January 16, 2017 2 Contents 1 Matrix Multiplication Problems 7 11 Introduction 7 111 Systems of Linear Equations 7 112 The Eigenvalue Problem 8 12 Basic

More information

Mathematics Department Stanford University Math 61CM/DM Inner products

Mathematics Department Stanford University Math 61CM/DM Inner products Mathematics Department Stanford University Math 61CM/DM Inner products Recall the definition of an inner product space; see Appendix A.8 of the textbook. Definition 1 An inner product space V is a vector

More information

Lecture Summaries for Linear Algebra M51A

Lecture Summaries for Linear Algebra M51A These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture

More information

SUMMARY OF MATH 1600

SUMMARY OF MATH 1600 SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You

More information

Applied Linear Algebra

Applied Linear Algebra Applied Linear Algebra Peter J. Olver School of Mathematics University of Minnesota Minneapolis, MN 55455 olver@math.umn.edu http://www.math.umn.edu/ olver Chehrzad Shakiban Department of Mathematics University

More information

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Linear Algebra. Session 12

Linear Algebra. Session 12 Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

FINITE-DIMENSIONAL LINEAR ALGEBRA

FINITE-DIMENSIONAL LINEAR ALGEBRA DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

October 25, 2013 INNER PRODUCT SPACES

October 25, 2013 INNER PRODUCT SPACES October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal

More information

Linear Algebra. Min Yan

Linear Algebra. Min Yan Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information

Math113: Linear Algebra. Beifang Chen

Math113: Linear Algebra. Beifang Chen Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary

More information

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.) page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation

More information

Linear Algebra Highlights

Linear Algebra Highlights Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to

More information

Linear Algebra. Paul Yiu. Department of Mathematics Florida Atlantic University. Fall A: Inner products

Linear Algebra. Paul Yiu. Department of Mathematics Florida Atlantic University. Fall A: Inner products Linear Algebra Paul Yiu Department of Mathematics Florida Atlantic University Fall 2011 6A: Inner products In this chapter, the field F = R or C. We regard F equipped with a conjugation χ : F F. If F =

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Chapter 1 Eigenvalues and Eigenvectors Among problems in numerical linear algebra, the determination of the eigenvalues and eigenvectors of matrices is second in importance only to the solution of linear

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N. Math 410 Homework Problems In the following pages you will find all of the homework problems for the semester. Homework should be written out neatly and stapled and turned in at the beginning of class

More information

Numerical Linear Algebra

Numerical Linear Algebra University of Alabama at Birmingham Department of Mathematics Numerical Linear Algebra Lecture Notes for MA 660 (1997 2014) Dr Nikolai Chernov April 2014 Chapter 0 Review of Linear Algebra 0.1 Matrices

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

Lecture 1 INF-MAT : Chapter 2. Examples of Linear Systems

Lecture 1 INF-MAT : Chapter 2. Examples of Linear Systems Lecture 1 INF-MAT 4350 2010: Chapter 2. Examples of Linear Systems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo August 26, 2010 Notation The set of natural

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Vectors in Function Spaces

Vectors in Function Spaces Jim Lambers MAT 66 Spring Semester 15-16 Lecture 18 Notes These notes correspond to Section 6.3 in the text. Vectors in Function Spaces We begin with some necessary terminology. A vector space V, also

More information

The Singular Value Decomposition and Least Squares Problems

The Singular Value Decomposition and Least Squares Problems The Singular Value Decomposition and Least Squares Problems Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 27, 2009 Applications of SVD solving

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Hands-on Matrix Algebra Using R

Hands-on Matrix Algebra Using R Preface vii 1. R Preliminaries 1 1.1 Matrix Defined, Deeper Understanding Using Software.. 1 1.2 Introduction, Why R?.................... 2 1.3 Obtaining R.......................... 4 1.4 Reference Manuals

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45 address 12 adjoint matrix 118 alternating 112 alternating 203 angle 159 angle 33 angle 60 area 120 associative 180 augmented matrix 11 axes 5 Axiom of Choice 153 basis 178 basis 210 basis 74 basis test

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2 Contents Preface for the Instructor xi Preface for the Student xv Acknowledgments xvii 1 Vector Spaces 1 1.A R n and C n 2 Complex Numbers 2 Lists 5 F n 6 Digression on Fields 10 Exercises 1.A 11 1.B Definition

More information

Math 302 Outcome Statements Winter 2013

Math 302 Outcome Statements Winter 2013 Math 302 Outcome Statements Winter 2013 1 Rectangular Space Coordinates; Vectors in the Three-Dimensional Space (a) Cartesian coordinates of a point (b) sphere (c) symmetry about a point, a line, and a

More information

MATH 532: Linear Algebra

MATH 532: Linear Algebra MATH 532: Linear Algebra Chapter 5: Norms, Inner Products and Orthogonality Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2015 fasshauer@iit.edu MATH 532 1 Outline

More information

ELEMENTARY MATRIX ALGEBRA

ELEMENTARY MATRIX ALGEBRA ELEMENTARY MATRIX ALGEBRA Third Edition FRANZ E. HOHN DOVER PUBLICATIONS, INC. Mineola, New York CONTENTS CHAPTER \ Introduction to Matrix Algebra 1.1 Matrices 1 1.2 Equality of Matrices 2 13 Addition

More information

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same. Introduction Matrix Operations Matrix: An m n matrix A is an m-by-n array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read

More information

MATHEMATICS 217 NOTES

MATHEMATICS 217 NOTES MATHEMATICS 27 NOTES PART I THE JORDAN CANONICAL FORM The characteristic polynomial of an n n matrix A is the polynomial χ A (λ) = det(λi A), a monic polynomial of degree n; a monic polynomial in the variable

More information

Lecture 1: Review of linear algebra

Lecture 1: Review of linear algebra Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017 Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...

More information

Applied Linear Algebra

Applied Linear Algebra Applied Linear Algebra Gábor P. Nagy and Viktor Vígh University of Szeged Bolyai Institute Winter 2014 1 / 262 Table of contents I 1 Introduction, review Complex numbers Vectors and matrices Determinants

More information

Matrix Theory. A.Holst, V.Ufnarovski

Matrix Theory. A.Holst, V.Ufnarovski Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. under-determined

More information

Chapter 6 Inner product spaces

Chapter 6 Inner product spaces Chapter 6 Inner product spaces 6.1 Inner products and norms Definition 1 Let V be a vector space over F. An inner product on V is a function, : V V F such that the following conditions hold. x+z,y = x,y

More information

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space. Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

Math 307 Learning Goals

Math 307 Learning Goals Math 307 Learning Goals May 14, 2018 Chapter 1 Linear Equations 1.1 Solving Linear Equations Write a system of linear equations using matrix notation. Use Gaussian elimination to bring a system of linear

More information

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

Introduction to Applied Linear Algebra with MATLAB

Introduction to Applied Linear Algebra with MATLAB Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex

More information

1 Number Systems and Errors 1

1 Number Systems and Errors 1 Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floating-point Representation...........

More information

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v

More information

2. Every linear system with the same number of equations as unknowns has a unique solution.

2. Every linear system with the same number of equations as unknowns has a unique solution. 1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations

More information

Jim Lambers MAT 610 Summer Session Lecture 1 Notes

Jim Lambers MAT 610 Summer Session Lecture 1 Notes Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

2. Review of Linear Algebra

2. Review of Linear Algebra 2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Notes on Mathematics

Notes on Mathematics Notes on Mathematics - 12 1 Peeyush Chandra, A. K. Lal, V. Raghavendra, G. Santhanam 1 Supported by a grant from MHRD 2 Contents I Linear Algebra 7 1 Matrices 9 1.1 Definition of a Matrix......................................

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational

More information

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2 Norwegian University of Science and Technology Department of Mathematical Sciences TMA445 Linear Methods Fall 07 Exercise set Please justify your answers! The most important part is how you arrive at an

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Introduction to Numerical Analysis

Introduction to Numerical Analysis J. Stoer R. Bulirsch Introduction to Numerical Analysis Second Edition Translated by R. Bartels, W. Gautschi, and C. Witzgall With 35 Illustrations Springer Contents Preface to the Second Edition Preface

More information