5.6. PSEUDOINVERSES 101. A H w.


 Randall Edwards
 1 years ago
 Views:
Transcription
1 5.6. PSEUDOINVERSES 0 Corollary If A is a matrix such that A H A is invertible, then the leastsquares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and is an example of a Moore Penrose pseudoinverse. Theorem Suppose the vector ˆv V minimizes v over all v V satisfying T v = w. Then, ˆv [N T ] and, if R T is closed, ˆv = T u for some u W. Proof. Suppose ˆv / [N T ], then the orthogonal decomposition V = [N T ] + N T shows that the projection of ˆv onto [N T ] has smaller norm but still satisfies T ˆv = w. This gives a contradiction and shows that ˆv [N T ]. If R T is closed, then R T = [N T ] and ˆv = T u for some u W. Corollary If A is a matrix such that AA H is invertible, then the minimumnorm solution to Av = w is v = A H AA H) w. Proof. The theorem shows that v = A H u and Av = AA H u = w. Since AA H is invertible, this gives u = AA H ) w and computing v gives the desired result. The matrix A H AA H) is the right inverse of A and is another example of a MoorePenrose pseudoinverse. Definition Let T : V W be a bounded linear transformation, where V and W are Hilbert spaces, and R T is closed. For each w W, there is a unique vector ˆv of minimum norm in the set of vectors that minimize T v w. The pseudoinverse T is the transformation mapping each w W to its unique ˆv.
2 02 CHAPTER 5. LINEAR TRANSFORMATIONS AND OPERATORS
3 Chapter 6 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signalprocessing algorithms. Their applications include solving linear equations LU), decorrelating random variables LDLT,Cholesky), orthogonalizing sets of vectors QR), and finding lowrank matrix approximations SVD). Their usefulness is often twofold: they allow efficient computation of important quantities and they are often) designed to minimize roundoff error due to finiteprecision calculation. An algorithm is called numerically stable, for a particular set of inputs, if the error in the final solution is proportional to the roundoff error in the elementary field operations. 6. Triangular Systems A square matrix L F n n is called lower triangular or upper triangular) if all elements above or below) the main diagonal are zero. Likewise, a triangular matrix lower or upper) is a unit triangular if it has all ones on the main diagonal. A system of linear equations is called triangular if it can be represented by the matrix equation Ax = b where A is either upper or lower triangular. 6.. Solution by Substitution Let L F n n be a lower triangular matrix with entries l ij = [L] ij. The matrix equation Ly = b can be solved efficiently using forward substitution, which is 03
4 04 CHAPTER 6. MATRIX FACTORIZATION AND ANALYSIS defined by the recursion ) y j = j b j l ji y i, j =, 2,..., n. l jj i= Example 6... Consider the system y y 2 y 3 = 2 9. Applying the above recursion gives y = = y 2 = 2 ) = y 3 = 9 2 ) = 6. Let U F n n be an upper triangular matrix with entries u ij = [U] ij. The matrix equation U x = y can be solved efficiently using backward substitution, which is defined by the recursion ) x j = y j u ji x i, j = n, n,...,. u jj i=j+ Example Consider the system x x 2 x 3 = 6. Applying the above recursion gives x 3 = 6 2 = 3 x 2 = 3 3) = 8 x = 6 8)) = 3. The computational complexity of each substitution is roughly 2 n2 operations.
5 6.2. LU DECOMPOSITION 05 P Show that set of upper triangular matrices is a subalgebra of the set of all matrices. Since it is clearly a subspace, only two properties must be verified:. that the product of two upper triangular matrices is upper triangular 2. that the inverse of an upper triangular matrix is upper triangular 6..2 The Determinant The determinant deta) of a square matrix A F n n is a scalar which captures a number of important properties of that matrix. For example, A is invertible iff deta) 0 and the determinant satisfies detab) = deta) detb) for square matrices A, B. Mathematically, it is the unique function mapping matrices to scalars that is i) linear in each column, ii) negated by column transposition, and iii) satisfies deti) =. The determinant of a square matrix can be defined recursively using the fact that det [a]) = a. Let A F n n be an arbitrary square matrix with entries a ij = [A] ij. The i, j)minor of A is the determinant of the n ) n ) matrix formed by deleting the ith row and jth column of A. Fact 6..4 Laplace s Formula). The determinant of A is given by deta) = a ij ) i+j M ij = a ij ) i+j M ij, j= i= where M ij is the i, j)minor of A. Theorem The determinant of a triangular matrix is the product of its diagonal elements. Proof. For upper lower) triangular matrices, this can be shown by expanding the determinant along the first column row) to compute each minor. 6.2 LU Decomposition 6.2. Introduction LU decomposition is a generalization of Gaussian elimination which allows one to efficiently solve a system of linear equations Ax = b multiple times with different
6 06 CHAPTER 6. MATRIX FACTORIZATION AND ANALYSIS righthand sides. In its basic form, it is numerically stable only if the matrix is positive definite or diagonally dominant. A slight modification, known as partial pivoting, makes it stable for a very large class of matrices. Any square matrix A F n n can be factored as A = LU, where L is a unit lowertriangular matrix and U is an uppertriangular matrix. The following example uses elementary row operations to cancel, in each column, all elements below the main diagonal. These elementary row operations are represented using left multiplication by a unit lowertriangular matrix = = = 0 3 This allows one to write 2 4 = = = LU decomposition can also be used to efficiently compute the determinant of A. Since deta) = detlu) = detl) detu), the problem is reduced to computing the determinant of triangular matrices. Using Theorem 6..5, it is easy to see that detl) = and detu) = n i= u ii.
7 6.2. LU DECOMPOSITION Formal Approach To describe LU decomposition formally, we first need to describe the individual operations that are used to zero out matrix elements. Definition Let A F n n be an arbitrary matrix, α F be a scalar, and i, j {, 2,..., n}. Then, adding α times the jth row to the ith row an elementary rowaddition operation. Moreover, I + αe ij, where E ij e i e T j and e k is the k th standard basis vector, is the elementary rowaddition matrix which effects this operation via left multiplication. Example For example, elementary row operations are used to cancel the 2, ) matrix entry in 0 0 I E 2, )A = = Lemma The following identities capture the important properties of elementary rowoperation matrices:. i) E ij E kl = δ j,k E il ii) I + αe ij )I + βe kl ) = I + αe ij + βe kl iii) I + αe ij ) = I αe ij ) if i j. if j k Proof. This proof is left as an exercise. Now, consider the process for computing the LU decomposition of A. To initialize the process, we let A ) = A. In each round, we let ) L j = I aj) i,j E i,j i=j+ be the product of elementary row operation matrices which cancel the subdiagonal elements of the jth column. The process proceeds by defining A j+) = L j A j) so that A j+) has all zeros below the diagonal in the first j columns. After n rounds, the process terminates with a j) j,j U = A n) = L n L n 2 L A where L = L L 2 L n is unit lower triangular.
8 08 CHAPTER 6. MATRIX FACTORIZATION AND ANALYSIS Lemma From the structure of elementary row operation matrices, we see n Proof. First, we notice that i=j+ I + α ij E i,j ) = I + I + α ij E i,j ) = I + n i=j+ α ij E i,j α ij E i,j. for j =, 2,..., n. Expanding the product shows that any term with two E matrices must contain a product E i,j E l,j with l > i > j. By Lemma 6.2.3i, we see that this term must be zero because j l. Now, we can prove the main result via induction. First, we assume that k I + α ij E i,j ) = I + k Next, we find that if k n 2, then k+ k ) n I + α ij E i,j ) = I + α ij E i,j ) = = I + = I + = I + I + k+ k k+ k+ α ij E i,j ) I + α ij E i,j + α ij E i,j + α ij E i,j. k l=k+2 α ij E i,j. I + α l,k+ E l,k+ ) l=k+2 l=k+2 k l=k+2 α l,k+ E l,k+ ) ) α ij α l,k+ E i,j E l,k+ α ij α l,k+ E i,k+ δ j,l Finally, we point out that the base case k = is given by the initial observation. Theorem This process generates one column of L per round because a j) i,j if i < j a j) j,j [L] ij = if i = j 0 otherwise.
9 6.2. LU DECOMPOSITION 09 Proof. First, we note that L = L L 2 L n n n = j= n a) = i=j+ i= i=j+ n b) = i= i=j+ = I + n i= i=j+ I aj) i,j a j) j,j I aj) i,j a j) j,j I + aj) i,j a j) i,j a j) j,j a j) j,j E i,j )) E i,j ) E i,j ) E i,j, where a) follows from Lemma 6.2.3ii i.e., all matrices in the inside product commute) and b) follows from Lemma 6.2.3iii. Picking off the i, j) entry of L e.g., with e T i Le j ) gives the stated result. Finally, we note that the LU decomposition can be computed in roughly 2 3 n3 field operations Partial Pivoting Sometimes the pivot element a j) j,j can be very small or zero. In this case, the algorithm will either fail e.g., divide by zero) or return a very unreliable result. The algorithm can be easily modified to avoid this problem by swapping rows of A j) to increase the magnitude of the pivot element before each cancellation phase. This results in a decomposition of the form P A = LU, where P is a permutation matrix. In this section, we will describe LU decomposition with partial pivoting using the notation from the previous section. The main difference is that, in each round, we will define A j+) = M j P j A j) where P j is a permutation matrix. In particular, left multiplication by P j swaps row j with row p j, where p j = arg max i=j,j+,...,n aj) i,j. The matrix M j is now chosen to cancel the subdiagonal elements in jth column of P j A j). After n rounds, the resulting decomposition has the form A n) = M n P n M n 2P n 2 M P A = U.
10 0 CHAPTER 6. MATRIX FACTORIZATION AND ANALYSIS To show this can also be written in the desired form, we need to understand some properties of the permutations. First, we point that swapping two rows is a transposition and therefore Pj 2 be moved to the right. = I. Next, we will show that the permutations can Lemma Let M = I + k j= n i=j+ α ije ij and Q be a permutation matrix which swaps row l k + and row m > l. Then, QM = MQ where Therefore, we can write and P A = LU. M M = I + M k M α ij QE ij. A n) = n n 2 P n P 2 P A = U }{{}}{{} L P Proof. The proof is left as an exercise. 6.3 LDLT and Cholesky Decomposition If the matrix A C n n is Hermitian, then the LU decomposition allows the factorization A = LDL H, where L is unit lower triangular and D is diagonal. Since this factorization is typically applied to real matrices, it is referred to as LDLT decomposition. If A is also positive definite, then the diagonal elements of D are positive and we can write A = LD /2) LD /2) H. The form A = L LH, where L is lower triangular, is known as Cholesky factorization. To see this, we will describe the LDLT decomposition using the notation from LU decomposition starting from A ) = A. In the jth round, define L j to be the product of elementary rowoperation matrices which cancel the subdiagonal elements of the jth column A j). Then, define A j+) = L j A j) L H j and notice that A j+) is Hermitian because A j) is Hermitian. Next, notice that A j+) has zeros below the diagonal in the first j columns and zeros to the right of diagonal in the first j rows. This follows from the fact that the first j rows of A j) are not affected by applying L j on left. Therefore, applying L H j on the right also cancels
11 6.3. LDLT AND CHOLESKY DECOMPOSITION the elements to the right of the diagonal in the jth row. After n rounds, we find that D = A n) is a diagonal matrix. There are a number of redundancies in the computation described above. First off, the L matrix computed by LU decomposition is identical to the L matrix computed by LDLT decomposition. Therefore, one can save operations by defining A j+) = L j A j). Moreover, the elements to the right of the diagonal in A j) do not affect the computation at all. So, one can roughly half the number of additions and multiplies by only updating the lower triangular part of A j). The resulting computational complexity is roughly 3 n3 field operations Cholesky Decomposition For a positivedefinite matrix A, we can first apply the LDLT decomposition and then define L = LD /2. This gives the Cholesky decomposition L L H = LDL H = A. The Cholesky decomposition is typically used to compute whitening filters for random variables. For example, one can apply it to the correlation matrix R = E[XX H ] of a random vector X. Then, one can define Y = L X and see that E[Y Y H ] = E[ L XX H L H ] = L R L H = I. From this, one sees that Y is a vector of uncorrelated or white) random variables QR decomposition A complex matrix Q C n n is called unitary if Q H Q = QQ H = I. If all elements of the matrix are real, then it is called orthogonal and Q T Q = QQ T = I. Theorem Any matrix A C m n can be factored as A = QR, where Q is an m m unitary matrix, QQ H = I, and R is an m n uppertriangular matrix. Proof. To show this decomposition, we start by applying GramSchmidt Orthogonalization to the columns a,..., a n of A. This results in orthonormal vectors
12 2 CHAPTER 6. MATRIX FACTORIZATION AND ANALYSIS {q,..., q l }, where l = minm, n), such that a j = minj,l) i= r i,j q i for j =, 2,..., n. This gives an m l matrix Q = [q... q l ] and an l n uppertriangular matrix R, with entries [R] i,j = r i,j, such that A = QR. If m n, then l = m, Q is unitary, and the decomposition is complete. Otherwise, we must extend the orthonormal set {q,..., q l } to an orthonormal basis {q,..., q m } of C m. This gives an m m unitary matrix Q = [q... q m ]. Adding m n rows of zeros to the previous R matrix gives an m n matrix R such that A = Q R. 6.4 Hermitian Matrices and Complex Numbers Definition A square matrix Q R n n is orthogonal if Q T Q = QQ T = I. Definition A square matrix U C n n is unitary if U H U = UU H = I. It is worth noting that, for unitary resp. orthogonal) matrices, it suffices to check only that U H U = I resp. Q T Q = I) because U is invertible e.g., it has linearly independent columns) and U H U = I = I = UU = UU H U)U = UU H. A useful analogy between matrices and complex numbers is as follows. Hermitian matrices satisfying A H = A are analogous to real numbers, whose complex conjugates are equal to themselves. Unitary matrices satisfying U H U = I are analogous to complex numbers on the unit circle, satisfying zz =. Orthogonal matrices satisfying Q T Q = I are analogous to the real numbers z = ±, such that z 2 =. The transformation z = + jr jr maps real number r into the unit circle z =. Analogously, by Cayley s formula, U = I + jr)i jr), a Hermitian matrix R is mapped to a unitary matrix.
Matrix Factorization and Analysis
Chapter 7 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signalprocessing algorithms. Their
More informationThis can be accomplished by left matrix multiplication as follows: I
1 Numerical Linear Algebra 11 The LU Factorization Recall from linear algebra that Gaussian elimination is a method for solving linear systems of the form Ax = b, where A R m n and bran(a) In this method
More informationGaussian Elimination and Back Substitution
Jim Lambers MAT 610 Summer Session 200910 Lecture 4 Notes These notes correspond to Sections 31 and 32 in the text Gaussian Elimination and Back Substitution The basic idea behind methods for solving
More informationMath 407: Linear Optimization
Math 407: Linear Optimization Lecture 16: The Linear Least Squares Problem II Math Dept, University of Washington February 28, 2018 Lecture 16: The Linear Least Squares Problem II (Math Dept, University
More informationMath/CS 466/666: Homework Solutions for Chapter 3
Math/CS 466/666: Homework Solutions for Chapter 3 31 Can all matrices A R n n be factored A LU? Why or why not? Consider the matrix A ] 0 1 1 0 Claim that this matrix can not be factored A LU For contradiction,
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.
More informationMatrix decompositions
Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The
More informationLinear Systems and Matrices
Department of Mathematics The Chinese University of Hong Kong 1 System of m linear equations in n unknowns (linear system) a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.......
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a ToolBox Linear Equation Systems Discretization of differential equations: solving linear equations
More informationII. Determinant Functions
Supplemental Materials for EE203001 Students II Determinant Functions ChungChin Lu Department of Electrical Engineering National Tsing Hua University May 22, 2003 1 Three Axioms for a Determinant Function
More informationLinear Analysis Lecture 16
Linear Analysis Lecture 16 The QR Factorization Recall the GramSchmidt orthogonalization process. Let V be an inner product space, and suppose a 1,..., a n V are linearly independent. Define q 1,...,
More informationMATRICES ARE SIMILAR TO TRIANGULAR MATRICES
MATRICES ARE SIMILAR TO TRIANGULAR MATRICES 1 Complex matrices Recall that the complex numbers are given by a + ib where a and b are real and i is the imaginary unity, ie, i 2 = 1 In what we describe below,
More informationReview for Exam Find all a for which the following linear system has no solutions, one solution, and infinitely many solutions.
Review for Exam. Find all a for which the following linear system has no solutions, one solution, and infinitely many solutions. x + y z = 2 x + 2y + z = 3 x + y + (a 2 5)z = a 2 The augmented matrix for
More informationDeterminants Chapter 3 of Lay
Determinants Chapter of Lay Dr. Doreen De Leon Math 152, Fall 201 1 Introduction to Determinants Section.1 of Lay Given a square matrix A = [a ij, the determinant of A is denoted by det A or a 11 a 1j
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finitedimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)matrix Consider the following row operations on A (1) Swap the positions any
More informationMH1200 Final 2014/2015
MH200 Final 204/205 November 22, 204 QUESTION. (20 marks) Let where a R. A = 2 3 4, B = 2 3 4, 3 6 a 3 6 0. For what values of a is A singular? 2. What is the minimum value of the rank of A over all a
More informationFundamentals of Engineering Analysis (650163)
Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is
More information1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )
Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems PerOlof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationCHAPTER 6. Direct Methods for Solving Linear Systems
CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationEquality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.
Introduction Matrix Operations Matrix: An m n matrix A is an mbyn array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read
More informationG1110 & 852G1 Numerical Linear Algebra
The University of Sussex Department of Mathematics G & 85G Numerical Linear Algebra Lecture Notes Autumn Term Kerstin Hesse (w aw S w a w w (w aw H(wa = (w aw + w Figure : Geometric explanation of the
More informationNotes on Linear Algebra
1 Notes on Linear Algebra Jean Walrand August 2005 I INTRODUCTION Linear Algebra is the theory of linear transformations Applications abound in estimation control and Markov chains You should be familiar
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationLinear Algebra, part 3 QR and SVD
Linear Algebra, part 3 QR and SVD AnnaKarin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2012 Going back to least squares (Section 1.4 from Strang, now also see section 5.2). We
More informationlinearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice
3. Eigenvalues and Eigenvectors, Spectral Representation 3.. Eigenvalues and Eigenvectors A vector ' is eigenvector of a matrix K, if K' is parallel to ' and ' 6, i.e., K' k' k is the eigenvalue. If is
More informationCheat Sheet for MATH461
Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A
More informationDirect Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le
Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization
More informationOrthonormal Transformations and Least Squares
Orthonormal Transformations and Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 30, 2009 Applications of Qx with Q T Q = I 1. solving
More information1 Last time: leastsquares problems
MATH Linear algebra (Fall 07) Lecture Last time: leastsquares problems Definition. If A is an m n matrix and b R m, then a leastsquares solution to the linear system Ax = b is a vector x R n such that
More informationSection 6.4. The Gram Schmidt Process
Section 6.4 The Gram Schmidt Process Motivation The procedures in 6 start with an orthogonal basis {u, u,..., u m}. Find the Bcoordinates of a vector x using dot products: x = m i= x u i u i u i u i Find
More informationLinear Algebra Primer
Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................
More informationSUMMARY OF MATH 1600
SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tuberlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationApplied Linear Algebra
Applied Linear Algebra Gábor P. Nagy and Viktor Vígh University of Szeged Bolyai Institute Winter 2014 1 / 262 Table of contents I 1 Introduction, review Complex numbers Vectors and matrices Determinants
More informationVector and Matrix Norms. Vector and Matrix Norms
Vector and Matrix Norms Vector Space Algebra Matrix Algebra: We let x x and A A, where, if x is an element of an abstract vector space n, and A = A: n m, then x is a complex column vector of length n whose
More informationMATH 2050 Assignment 8 Fall [10] 1. Find the determinant by reducing to triangular form for the following matrices.
MATH 2050 Assignment 8 Fall 2016 [10] 1. Find the determinant by reducing to triangular form for the following matrices. 0 1 2 (a) A = 2 1 4. ANS: We perform the Gaussian Elimination on A by the following
More informationIndex. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)
page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation
More informationLinear Algebraic Equations
Linear Algebraic Equations 1 Fundamentals Consider the set of linear algebraic equations n a ij x i b i represented by Ax b j with [A b ] [A b] and (1a) r(a) rank of A (1b) Then Axb has a solution iff
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationAssignment 11 (C + C ) = (C + C ) = (C + C) i(c C ) ] = i(c C) (AB) = (AB) = B A = BA 0 = [A, B] = [A, B] = (AB BA) = (AB) AB
Arfken 3.4.6 Matrix C is not Hermition. But which is Hermitian. Likewise, Assignment 11 (C + C ) = (C + C ) = (C + C) [ i(c C ) ] = i(c C ) = i(c C) = i ( C C ) Arfken 3.4.9 The matrices A and B are both
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationNumerical Methods  Numerical Linear Algebra
Numerical Methods  Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods  Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationANSWERS. E k E 2 E 1 A = B
MATH 7 Final Exam Spring ANSWERS Essay Questions points Define an Elementary Matrix Display the fundamental matrix multiply equation which summarizes a sequence of swap, combination and multiply operations,
More informationThe QR Factorization
The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş BabeşBolyai University March 11, 2009 Radu Trîmbiţaş ( BabeşBolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector
More informationTopic 15 Notes Jeremy Orloff
Topic 5 Notes Jeremy Orloff 5 Transpose, Inverse, Determinant 5. Goals. Know the definition and be able to compute the inverse of any square matrix using row operations. 2. Know the properties of inverses.
More informationOrthogonal Transformations
Orthogonal Transformations Tom Lyche University of Oslo Norway Orthogonal Transformations p. 1/3 Applications of Qx with Q T Q = I 1. solving least squares problems (today) 2. solving linear equations
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More information1 Positive definiteness and semidefiniteness
Positive definiteness and semidefiniteness Zdeněk Dvořák May 9, 205 For integers a, b, and c, let D(a, b, c) be the diagonal matrix with + for i =,..., a, D i,i = for i = a +,..., a + b,. 0 for i = a +
More informationThere are six more problems on the next two pages
Math 435 bg & bu: Topics in linear algebra Summer 25 Final exam Wed., 8/3/5. Justify all your work to receive full credit. Name:. Let A 3 2 5 Find a permutation matrix P, a lower triangular matrix L with
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 78pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationChapter 5: Matrices. Daniel Chan. Semester UNSW. Daniel Chan (UNSW) Chapter 5: Matrices Semester / 33
Chapter 5: Matrices Daniel Chan UNSW Semester 1 2018 Daniel Chan (UNSW) Chapter 5: Matrices Semester 1 2018 1 / 33 In this chapter Matrices were first introduced in the Chinese Nine Chapters on the Mathematical
More informationMATH 240 Spring, Chapter 1: Linear Equations and Matrices
MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 20112012 A square matrix is symmetric if A T
More informationA Review of Linear Algebra
A Review of Linear Algebra Gerald Recktenwald Portland State University Mechanical Engineering Department gerry@me.pdx.edu These slides are a supplement to the book Numerical Methods with Matlab: Implementations
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationA = , A 32 = n ( 1) i +j a i j det(a i j). (1) j=1
Lecture Notes: Determinant of a Square Matrix Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk 1 Determinant Definition Let A [a ij ] be an
More informationHere are some additional properties of the determinant function.
List of properties Here are some additional properties of the determinant function. Prop Throughout let A, B M nn. 1 If A = (a ij ) is upper triangular then det(a) = a 11 a 22... a nn. 2 If a row or column
More informationNumerical Linear Algebra Homework Assignment  Week 2
Numerical Linear Algebra Homework Assignment  Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationSolution of Linear Equations
Solution of Linear Equations (Com S 477/577 Notes) YanBin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass
More information1 Inner Product and Orthogonality
CSCI 4/Fall 6/Vora/GWU/Orthogonality and Norms Inner Product and Orthogonality Definition : The inner product of two vectors x and y, x x x =.., y =. x n y y... y n is denoted x, y : Note that n x, y =
More informationNumerical Linear Algebra
Numerical Linear Algebra Direct Methods Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) Linear Systems: Direct Solution Methods Fall 2017 1 / 14 Introduction The solution of linear systems is one
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationThis MUST hold matrix multiplication satisfies the distributive property.
The columns of AB are combinations of the columns of A. The reason is that each column of AB equals A times the corresponding column of B. But that is a linear combination of the columns of A with coefficients
More informationPractical Linear Algebra: A Geometry Toolbox
Practical Linear Algebra: A Geometry Toolbox Third edition Chapter 12: Gauss for Linear Systems Gerald Farin & Dianne Hansford CRC Press, Taylor & Francis Group, An A K Peters Book www.farinhansford.com/books/pla
More informationSYLLABUS. 1 Linear maps and matrices
Dr. K. Bellová Mathematics 2 (10PHYBIPMA2) SYLLABUS 1 Linear maps and matrices Operations with linear maps. Prop 1.1.1: 1) sum, scalar multiple, composition of linear maps are linear maps; 2) L(U, V
More informationThe Solution of Linear Systems AX = B
Chapter 2 The Solution of Linear Systems AX = B 21 Uppertriangular Linear Systems We will now develop the backsubstitution algorithm, which is useful for solving a linear system of equations that has
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationDeterminants. Recall that the 2 2 matrix a b c d. is invertible if
Determinants Recall that the 2 2 matrix a b c d is invertible if and only if the quantity ad bc is nonzero. Since this quantity helps to determine the invertibility of the matrix, we call it the determinant.
More informationThe Determinant: a Means to Calculate Volume
The Determinant: a Means to Calculate Volume Bo Peng August 16, 2007 Abstract This paper gives a definition of the determinant and lists many of its wellknown properties Volumes of parallelepipeds are
More informationLecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation
Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite
More information18.06SC Final Exam Solutions
18.06SC Final Exam Solutions 1 (4+7=11 pts.) Suppose A is 3 by 4, and Ax = 0 has exactly 2 special solutions: 1 2 x 1 = 1 and x 2 = 1 1 0 0 1 (a) Remembering that A is 3 by 4, find its row reduced echelon
More informationRoundoff Analysis of Gaussian Elimination
Jim Lambers MAT 60 Summer Session 20090 Lecture 5 Notes These notes correspond to Sections 33 and 34 in the text Roundoff Analysis of Gaussian Elimination In this section, we will perform a detailed error
More information33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM
33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM (UPDATED MARCH 17, 2018) The final exam will be cumulative, with a bit more weight on more recent material. This outline covers the what we ve done since the
More informationMODEL ANSWERS TO THE THIRD HOMEWORK
MODEL ANSWERS TO THE THIRD HOMEWORK 1 (i) We apply Gaussian elimination to A First note that the second row is a multiple of the first row So we need to swap the second and third rows 1 3 2 1 2 6 5 7 3
More informationMATH2210 Notebook 2 Spring 2018
MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................
More informationMatrix Algebra for Engineers Jeffrey R. Chasnov
Matrix Algebra for Engineers Jeffrey R. Chasnov The Hong Kong University of Science and Technology The Hong Kong University of Science and Technology Department of Mathematics Clear Water Bay, Kowloon
More informationMath Lecture 26 : The Properties of Determinants
Math 2270  Lecture 26 : The Properties of Determinants Dylan Zwick Fall 202 The lecture covers section 5. from the textbook. The determinant of a square matrix is a number that tells you quite a bit about
More informationMathematical Methods wk 2: Linear Operators
John Magorrian, magog@thphysoxacuk These are workinprogress notes for the secondyear course on mathematical methods The most uptodate version is available from http://wwwthphysphysicsoxacuk/people/johnmagorrian/mm
More informationLU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b
AM 205: lecture 7 Last time: LU factorization Today s lecture: Cholesky factorization, timing, QR factorization Reminder: assignment 1 due at 5 PM on Friday September 22 LU Factorization LU factorization
More informationChapter 4. Determinants
4.2 The Determinant of a Square Matrix 1 Chapter 4. Determinants 4.2 The Determinant of a Square Matrix Note. In this section we define the determinant of an n n matrix. We will do so recursively by defining
More informationOrthonormal Transformations
Orthonormal Transformations Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 25, 2010 Applications of transformation Q : R m R m, with Q T Q = I 1.
More informationNo books, notes, any calculator, or electronic devices are allowed on this exam. Show all of your steps in each answer to receive a full credit.
MTH 309001 Fall 2016 Exam 1 10/05/16 Name (Print): PID: READ CAREFULLY THE FOLLOWING INSTRUCTION Do not open your exam until told to do so. This exam contains 7 pages (including this cover page) and 7
More informationLecture 3: QRFactorization
Lecture 3: QRFactorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QRfactorization of matrices It also outlines some applications of this factorization
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More information1 0 1, then use that decomposition to solve the least squares problem. 1 Ax = 2. q 1 = a 1 a 1 = 1. to find the intermediate result:
Exercise Find the QR decomposition of A =, then use that decomposition to solve the least squares problem Ax = 2 3 4 Solution Name the columns of A by A = [a a 2 a 3 ] and denote the columns of the results
More informationLecture 23: Trace and determinants! (1) (Final lecture)
Lecture 23: Trace and determinants! (1) (Final lecture) Travis Schedler Thurs, Dec 9, 2010 (version: Monday, Dec 13, 3:52 PM) Goals (2) Recall χ T (x) = (x λ 1 ) (x λ n ) = x n tr(t )x n 1 + +( 1) n det(t
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationImportant Matrix Factorizations
LU Factorization Choleski Factorization The QR Factorization LU Factorization: Gaussian Elimination Matrices Gaussian elimination transforms vectors of the form a α, b where a R k, 0 α R, and b R n k 1,
More informationDeterminants An Introduction
Determinants An Introduction Professor Je rey Stuart Department of Mathematics Paci c Lutheran University Tacoma, WA 9844 USA je rey.stuart@plu.edu The determinant is a useful function that takes a square
More informationMATH 235. Final ANSWERS May 5, 2015
MATH 235 Final ANSWERS May 5, 25. ( points) Fix positive integers m, n and consider the vector space V of all m n matrices with entries in the real numbers R. (a) Find the dimension of V and prove your
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13
STAT 309: MATHEMATICAL COMPUTATIONS I FALL 208 LECTURE 3 need for pivoting we saw that under proper circumstances, we can write A LU where 0 0 0 u u 2 u n l 2 0 0 0 u 22 u 2n L l 3 l 32, U 0 0 0 l n l
More informationMath 2912: Lecture Notes Northwestern University, Winter 2016
Math 2912: Lecture Notes Northwestern University, Winter 2016 Written by Santiago Cañez These are lecture notes for Math 2912, the second quarter of MENU: Intensive Linear Algebra and Multivariable Calculus,
More information