VII Selected Topics. 28 Matrix Operations
|
|
- Alan Anthony
- 5 years ago
- Views:
Transcription
1 VII Selected Topics Matrix Operations Linear Programming Number Theoretic Algorithms Polynomials and the FFT Approximation Algorithms 28 Matrix Operations We focus on how to multiply matrices and solve sets of simultaneous linear equations First we show how to solve a set of linear equations using LUP decompositions Then, we explore the close relationship between multiplying and inverting matrices Finally, we discus the class of symmetric positive-denite matrices and show how we can use them to nd a least-squares solution to an overdetermined set of linear equations MAT AA+DS, Fall Nov
2 Matrix inverses and ranks If a matrix has an inverse, it is called invertible, or nonsingular The vectors,, are linearly dependent if there exist coefcients,,, not all of which are zero, such that =0 E.g., the row vectors = (13), = (24), and = (4119) are linearly dependent since + 3 =0 The column rank of a nonzero matrix is the size of the largest set of linearly independent columns of MAT AA+DS, Fall Nov For any matrix its row rank = its column rank, so that we can simply refer to the rank of Square matrix has full rank if its rank is An matrix has full column rank if its rank is An matrix is positive-denite if >0for all -vectors 0 E.g., the identity matrix is positive-denite, since for any nonzero vector =, = >0 MAT AA+DS, Fall Nov
3 Matrices that arise in applications are often positive-denite due to the following theorem Theorem D.6 For any matrix with full column rank, the matrix is positive-denite. MAT AA+DS, Fall Nov Solving systems of linear equations Numerous applications need to solve sets of simultaneous linear equations We can formulate a linear system as a matrix equation in which each matrix or vector element belongs to a eld, typically the real numbers How to solve a system of linear equations using a method called LUP decomposition? We start with a set of linear equations in unknowns,, MAT AA+DS, Fall Nov
4 ,, A solution to these equations is a set of values for,, that satisfy all of the equations simultaneously We treat only the case in which there are exactly equations in unknowns MAT AA+DS, Fall Nov We can rewrite the equations as the matrixvector equation = Equivalently, letting = =,and = If is nonsingular (= invertible), it possesses an inverse, and is the solution vector MAT AA+DS, Fall Nov
5 is the unique solution to the equation Let there be two solutions, and Then and, letting denote an identity matrix, = ) = = MAT AA+DS, Fall Nov We are predominantly concerned with the case in which is nonsingular Equivalently, the rank of is equal to the number of unknowns If the number of equations is or, more generally, if the rank of is then the system is underdetermined Such a system typically has innitely many solutions, although it may have no solutions at all if the equations are inconsistent If the number of equations exceeds, the system is overdetermined, and there may not exist any solutions MAT AA+DS, Fall Nov
6 Later we address the problem of nding good approximate solutions to overdetermined systems of linear equations Returning to the problem of solving the system of equations in unknowns We could compute and then, multiply by, yielding This approach suffers in practice from numerical instability Fortunately, another approach LUP decomposition is numerically stable and has the further advantage of being faster in practice MAT AA+DS, Fall Nov Overview of LUP decomposition The idea behind LUP decomposition is to nd three matrices,, and such that is a unit lower-triangular matrix, is an upper-triangular matrix, and is a permutation matrix (has exactly one 1 in each row and column) Call matrices,, and satisfying the equation an LUP decomposition of the matrix Every nonsingular matrix has such a decomposition MAT AA+DS, Fall Nov
7 Computing an LUP decomposition for has the advantage that we can more easily solve linear systems when they are triangular, as is the case for both matrices and Once we have found an LUP decomposition for, we can solve equation by solving only triangular linear systems, as follows Multiplying both sides of by yields the equivalent equation, which, by Exercise D.1-4, amounts to permuting the original linear equations Using our decomposition, we obtain = MAT AA+DS, Fall Nov We can now solve this equation by solving two triangular linear systems Let us dene, where is the desired solution vector First, we solve the lower-triangular system for the unknown vector by a method called forward substitution Having solved for, we then solve the uppertriangular system for the unknown by back substitution MAT AA+DS, Fall Nov
8 Because the permutation matrix is invertible (Exercise D.2-3), multiplying both sides of by gives, so that Hence, the vector is our solution to : We still need to show how forward and back substitution work MAT AA+DS, Fall Nov Forward and back substitution Forward substitution can solve the lowertriangular system in time, given,, and For convenience, we represent the permutation compactly by an array [1.. ] The entry ] indicates that ] =1and =0for ] Thus, has in row and column, and has as its th element MAT AA+DS, Fall Nov
9 Since is unit lower-triangular, we can rewrite as ] ] ] ] The rst equation tells us that ] Knowing the value of, we can substitute it into the second equation, yielding ] And so forth MAT AA+DS, Fall Nov In general, we substitute,,, forward into the th equation to solve for : Having solved for, we solve for in using back substitution, which is similar to forward substitution Here, we solve the th equation rst and work backward to the rst equation Like forward substitution, this process runs in time MAT AA+DS, Fall Nov
10 Given,,, and, LUP-SOLVE solves for by combining forward and back substitution Assume that the dimension appears in the attribute and that the permutation matrix is represented by the array LUP-SOLVE ) let be a new vector of length 3. for =1to for downto 1 6. )/ 7. return MAT AA+DS, Fall Nov Consider the system of linear equations = = We wish to solve for the unknown The LUP decomposition is = = = MAT AA+DS, Fall Nov
11 Let us verify that = = = = MAT AA+DS, Fall Nov By forward substitution, we solve = for : = obtaining = by computing rst, then, and nally ] =8 ] = = 1.4 ] = = = 1.5 MAT AA+DS, Fall Nov
12 Using back substitution, we solve for : = thereby obtaining the desired answer 1.4 = by computing rst, then, and nally MAT AA+DS, Fall Nov Inverting matrices We do not generally use matrix inverses to solve systems of linear equations, rather use more stable techniques such as LUP decomposition We show how to use LUP decomposition to compute a matrix inverse Matrix multiplication and computing the inverse are equivalently hard problems (subject to technical conditions): we can use an algorithm for one to solve the other in the same asymptotic running time MAT AA+DS, Fall Nov
13 Computing a matrix inverse from an LUP decomposition Suppose that we have an LUP decomposition of a matrix :,, and such that Using LUP-SOLVE, we can solve an equation of the form in time ) Since the LUP decomposition depends on but not, we can run LUP-SOLVE on a second set of equations in additional time ) Once we have the LUP decomposition of, we can solve, in time, versions of the equation = that differ only in MAT AA+DS, Fall Nov Think of the equation which denes, the inverse of, as a set of distinct equations of the form Let denote the th column of, and recall that the unit vector is the th column of We can then solve the equation for by using the LUP decomposition for to solve each equation separately for Once we have the LUP decomposition of, we can compute from it in time Hence we can compute the inverse of a matrix in time ( ) MAT AA+DS, Fall Nov
14 Matrix multiplication and matrix inversion The speedups obtained for matrix multiplication translate to speedups for matrix inversion In fact, matrix inversion is equivalent to matrix multiplication, in the following sense If ) denotes the time to multiply two matrices, then we can invert a nonsingular matrix in time O( ) Moreover, if ) denotes the time to invert a nonsingular matrix, then we can multiply two matrices in time ( ) MAT AA+DS, Fall Nov Theorem 28.1 (Multiplication is no harder than inversion) If we can invert an matrix in time ), where and satises the regularity condition ), then we can multiply two matrices in time ).. Proof Let and be matrices whose matrix product we wish to compute. We dene the 3 matrix by 0 = 0 0 MAT AA+DS, Fall Nov
15 The inverse of is =, 0 0 and thus we can compute the product by taking the upper right submatrix of. We can construct matrix in ) time, which is )because we assume that, and we can invert in )time, by the regularity condition on ). We thus have = ( ). MAT AA+DS, Fall Nov () satises the regularity condition whenever ) for any constants >0and 0. The following proof relies on some properties of symmetric positive-denite matrices proved later. Theorem 28.2 (Inversion is no harder than multiplication) Suppose we can multiply two real matrices in time ), where and ) satises the two regularity conditions ) = ) for any in the range and 2 ) for some constant < 1. 2Then we can compute the inverse of any real nonsingular matrix in time ( ). MAT AA+DS, Fall Nov
16 28.3 Symmetric positive-denite matrices and least-squares approximation Lemma 28.3 Any positive-denite matrix is nonsingular. Proof Suppose that a matrix is singular. Then by Corollary D.3, there exists a nonzero vector such that =0. Hence, =0, and cannot be positive-denite. MAT AA+DS, Fall Nov Least-squares approximation One application of symmetric pos-def matrices arises in tting curves to given data points We are given a set of data points,,,, where we know that the are subject to measurement errors We would like to determine a function ) such that the approximation errors are small for = 1,2,, MAT AA+DS, Fall Nov
17 The form of depends on the problem at hand Here, we assume that it has the form of a linearly weighted sum, = The number of summands and the specic basis functions are chosen based on knowledge of the problem at hand A common choice is, meaning that is a polynomial of degree 1 in MAT AA+DS, Fall Nov Thus, given data points,,,,, we wish to calculate coefcients,, that minimize the approximation errors,, Choosing, we can calculate each exactly Such a high-degree ts the noise as well as the data, however, and gives poor results when used to predict for unseen values of We choose and hope that by choosing the coefcients well, we can obtain that nds the signicant patterns in the data points without paying undue attention to the noise MAT AA+DS, Fall Nov
18 Once we choose a value of, we end up with an overdetermined set of equations whose solution we wish to approximate Let = denote the matrix of values of the basis functions at the given points; that is, The desired -vector of coefcients is = MAT AA+DS, Fall Nov Then = = is the -vector of predicted values for Thus, ) ) ) is the -vector of approximation errors To minimize approximation errors, we choose to minimize the norm of the error vector MAT AA+DS, Fall Nov
19 This gives us a least-squares solution, since Because = = = we can minimize by differentiating with respect to each and then setting the result to 0: MAT AA+DS, Fall Nov = 2 =0 The equations for = 1,2,, are equivalent to the single matrix equation =0 or, equivalently (using Exercise D.1-2), to =0 which implies In statistics, this is called the normal equation MAT AA+DS, Fall Nov
20 The matrix is symmetric by Exercise D.1-2, and if has full column rank, then by Theorem D.6, is positive-denite as well Hence, exists, and the solution to the previous equation is = where the matrix = is the pseudoinverse of the matrix The pseudoinverse naturally generalizes the notion of a matrix inverse to the case in which is not square MAT AA+DS, Fall Nov Equation is the approximate solution to Whereas the solution is the exact solution to = MAT AA+DS, Fall Nov
21 As an example of producing a least-squares t, suppose that we have ve data points = (1,2) = (1,1) = (2,1) = (3,0) = (5,3) shown as black dots in the next figure We wish to t these points with a quadratic polynomial = + + MAT AA+DS, Fall Nov MAT AA+DS, Fall Nov
22 We start with the matrix of basis-function values = 1 = whose pseudoinverse is MAT AA+DS, Fall Nov Multiplying by, we obtain the coefcient vector = , which corresponds to the quadratic polynomial = as the closest-tting quadratic to the given data, in a least-squares sense MAT AA+DS, Fall Nov
Linear Equations in Linear Algebra
1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION x 1,, x n A linear equation in the variables equation that can be written in the form a 1 x 1 + a 2 x 2 + + a n x n
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More information1 Determinants. 1.1 Determinant
1 Determinants [SB], Chapter 9, p.188-196. [SB], Chapter 26, p.719-739. Bellow w ll study the central question: which additional conditions must satisfy a quadratic matrix A to be invertible, that is to
More informationLinear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)
Linear Algebra (part ) : Matrices and Systems of Linear Equations (by Evan Dummit, 206, v 202) Contents 2 Matrices and Systems of Linear Equations 2 Systems of Linear Equations 2 Elimination, Matrix Formulation
More informationSection 5.6. LU and LDU Factorizations
5.6. LU and LDU Factorizations Section 5.6. LU and LDU Factorizations Note. We largely follow Fraleigh and Beauregard s approach to this topic from Linear Algebra, 3rd Edition, Addison-Wesley (995). See
More information1 Matrices and Systems of Linear Equations
Linear Algebra (part ) : Matrices and Systems of Linear Equations (by Evan Dummit, 207, v 260) Contents Matrices and Systems of Linear Equations Systems of Linear Equations Elimination, Matrix Formulation
More information29 Linear Programming
29 Linear Programming Many problems take the form of optimizing an objective, given limited resources and competing constraints If we can specify the objective as a linear function of certain variables,
More informationThe matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.
) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to
More informationMATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year
1 MATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 2 Linear Systems and solutions Systems of linear
More informationLinear Algebra, 4th day, Thursday 7/1/04 REU Info:
Linear Algebra, 4th day, Thursday 7/1/04 REU 004. Info http//people.cs.uchicago.edu/laci/reu04. Instructor Laszlo Babai Scribe Nick Gurski 1 Linear maps We shall study the notion of maps between vector
More informationA TOUR OF LINEAR ALGEBRA FOR JDEP 384H
A TOUR OF LINEAR ALGEBRA FOR JDEP 384H Contents Solving Systems 1 Matrix Arithmetic 3 The Basic Rules of Matrix Arithmetic 4 Norms and Dot Products 5 Norms 5 Dot Products 6 Linear Programming 7 Eigenvectors
More informationMath 240 Calculus III
The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A
More informationChapter 3 Least Squares Solution of y = A x 3.1 Introduction We turn to a problem that is dual to the overconstrained estimation problems considered s
Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology 1 1 c Chapter
More informationLinear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space
Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................
More informationNew concepts: Span of a vector set, matrix column space (range) Linearly dependent set of vectors Matrix null space
Lesson 6: Linear independence, matrix column space and null space New concepts: Span of a vector set, matrix column space (range) Linearly dependent set of vectors Matrix null space Two linear systems:
More informationIntroduction to Determinants
Introduction to Determinants For any square matrix of order 2, we have found a necessary and sufficient condition for invertibility. Indeed, consider the matrix The matrix A is invertible if and only if.
More informationMatrix Operations: Determinant
Matrix Operations: Determinant Determinants Determinants are only applicable for square matrices. Determinant of the square matrix A is denoted as: det(a) or A Recall that the absolute value of the determinant
More informationMatrices and systems of linear equations
Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.
More informationNotes on Row Reduction
Notes on Row Reduction Francis J. Narcowich Department of Mathematics Texas A&M University September The Row-Reduction Algorithm The row-reduced form of a matrix contains a great deal of information, both
More informationVector Spaces, Orthogonality, and Linear Least Squares
Week Vector Spaces, Orthogonality, and Linear Least Squares. Opening Remarks.. Visualizing Planes, Lines, and Solutions Consider the following system of linear equations from the opener for Week 9: χ χ
More information12/1/2015 LINEAR ALGEBRA PRE-MID ASSIGNMENT ASSIGNED BY: PROF. SULEMAN SUBMITTED BY: M. REHAN ASGHAR BSSE 4 ROLL NO: 15126
12/1/2015 LINEAR ALGEBRA PRE-MID ASSIGNMENT ASSIGNED BY: PROF. SULEMAN SUBMITTED BY: M. REHAN ASGHAR Cramer s Rule Solving a physical system of linear equation by using Cramer s rule Cramer s rule is really
More informationContents. 6 Systems of First-Order Linear Dierential Equations. 6.1 General Theory of (First-Order) Linear Systems
Dierential Equations (part 3): Systems of First-Order Dierential Equations (by Evan Dummit, 26, v 2) Contents 6 Systems of First-Order Linear Dierential Equations 6 General Theory of (First-Order) Linear
More information5 and A,1 = B = is obtained by interchanging the rst two rows of A. Write down the inverse of B.
EE { QUESTION LIST EE KUMAR Spring (we will use the abbreviation QL to refer to problems on this list the list includes questions from prior midterm and nal exams) VECTORS AND MATRICES. Pages - of the
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationFraction-free Row Reduction of Matrices of Skew Polynomials
Fraction-free Row Reduction of Matrices of Skew Polynomials Bernhard Beckermann Laboratoire d Analyse Numérique et d Optimisation Université des Sciences et Technologies de Lille France bbecker@ano.univ-lille1.fr
More informationMAT 1302B Mathematical Methods II
MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More informationLU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark
DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline
More informationLinear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.
POLI 7 - Mathematical and Statistical Foundations Prof S Saiegh Fall Lecture Notes - Class 4 October 4, Linear Algebra The analysis of many models in the social sciences reduces to the study of systems
More informationAck: 1. LD Garcia, MTH 199, Sam Houston State University 2. Linear Algebra and Its Applications - Gilbert Strang
Gaussian Elimination CS6015 : Linear Algebra Ack: 1. LD Garcia, MTH 199, Sam Houston State University 2. Linear Algebra and Its Applications - Gilbert Strang The Gaussian Elimination Method The Gaussian
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationGetting Started with Communications Engineering
1 Linear algebra is the algebra of linear equations: the term linear being used in the same sense as in linear functions, such as: which is the equation of a straight line. y ax c (0.1) Of course, if we
More informationLinear Algebra: Characteristic Value Problem
Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number
More informationNotes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.
Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where
More informationChapters 5 & 6: Theory Review: Solutions Math 308 F Spring 2015
Chapters 5 & 6: Theory Review: Solutions Math 308 F Spring 205. If A is a 3 3 triangular matrix, explain why det(a) is equal to the product of entries on the diagonal. If A is a lower triangular or diagonal
More informationSolution Set 7, Fall '12
Solution Set 7, 18.06 Fall '12 1. Do Problem 26 from 5.1. (It might take a while but when you see it, it's easy) Solution. Let n 3, and let A be an n n matrix whose i, j entry is i + j. To show that det
More information4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial
Linear Algebra (part 4): Eigenvalues, Diagonalization, and the Jordan Form (by Evan Dummit, 27, v ) Contents 4 Eigenvalues, Diagonalization, and the Jordan Canonical Form 4 Eigenvalues, Eigenvectors, and
More information30.3. LU Decomposition. Introduction. Prerequisites. Learning Outcomes
LU Decomposition 30.3 Introduction In this Section we consider another direct method for obtaining the solution of systems of equations in the form AX B. Prerequisites Before starting this Section you
More informationReview Questions REVIEW QUESTIONS 71
REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of
More informationGlossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB
Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the
More informationIntroduction to Matrices
POLS 704 Introduction to Matrices Introduction to Matrices. The Cast of Characters A matrix is a rectangular array (i.e., a table) of numbers. For example, 2 3 X 4 5 6 (4 3) 7 8 9 0 0 0 Thismatrix,with4rowsand3columns,isoforder
More informationMATH 167: APPLIED LINEAR ALGEBRA Least-Squares
MATH 167: APPLIED LINEAR ALGEBRA Least-Squares October 30, 2014 Least Squares We do a series of experiments, collecting data. We wish to see patterns!! We expect the output b to be a linear function of
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationLecture 3: Linear Algebra Review, Part II
Lecture 3: Linear Algebra Review, Part II Brian Borchers January 4, Linear Independence Definition The vectors v, v,..., v n are linearly independent if the system of equations c v + c v +...+ c n v n
More informationLinear Algebra. and
Instructions Please answer the six problems on your own paper. These are essay questions: you should write in complete sentences. 1. Are the two matrices 1 2 2 1 3 5 2 7 and 1 1 1 4 4 2 5 5 2 row equivalent?
More informationMobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti
Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes
More information5 Eigenvalues and Diagonalization
Linear Algebra (part 5): Eigenvalues and Diagonalization (by Evan Dummit, 27, v 5) Contents 5 Eigenvalues and Diagonalization 5 Eigenvalues, Eigenvectors, and The Characteristic Polynomial 5 Eigenvalues
More informationx n -2.5 Definition A list is a list of objects, where multiplicity is allowed, and order matters. For example, as lists
Vectors, Linear Combinations, and Matrix-Vector Mulitiplication In this section, we introduce vectors, linear combinations, and matrix-vector multiplication The rest of the class will involve vectors,
More informationLecture 21: The decomposition theorem into generalized eigenspaces; multiplicity of eigenvalues and upper-triangular matrices (1)
Lecture 21: The decomposition theorem into generalized eigenspaces; multiplicity of eigenvalues and upper-triangular matrices (1) Travis Schedler Tue, Nov 29, 2011 (version: Tue, Nov 29, 1:00 PM) Goals
More informationMATH PRACTICE EXAM 1 SOLUTIONS
MATH 2359 PRACTICE EXAM SOLUTIONS SPRING 205 Throughout this exam, V and W will denote vector spaces over R Part I: True/False () For each of the following statements, determine whether the statement is
More informationLinear Algebra Handout
Linear Algebra Handout References Some material and suggested problems are taken from Fundamentals of Matrix Algebra by Gregory Hartman, which can be found here: http://www.vmi.edu/content.aspx?id=779979.
More informationPractical Linear Algebra: A Geometry Toolbox
Practical Linear Algebra: A Geometry Toolbox Third edition Chapter 12: Gauss for Linear Systems Gerald Farin & Dianne Hansford CRC Press, Taylor & Francis Group, An A K Peters Book www.farinhansford.com/books/pla
More informationSection 1.1: Systems of Linear Equations
Section 1.1: Systems of Linear Equations Two Linear Equations in Two Unknowns Recall that the equation of a line in 2D can be written in standard form: a 1 x 1 + a 2 x 2 = b. Definition. A 2 2 system of
More informationMultiplying matrices by diagonal matrices is faster than usual matrix multiplication.
7-6 Multiplying matrices by diagonal matrices is faster than usual matrix multiplication. The following equations generalize to matrices of any size. Multiplying a matrix from the left by a diagonal matrix
More informationPUTNAM PROBLEM SOLVING SEMINAR WEEK 7 This is the last meeting before the Putnam. The Rules. These are way too many problems to consider. Just pick a
PUTNAM PROBLEM SOLVING SEMINAR WEEK 7 This is the last meeting before the Putnam The Rules These are way too many problems to consider Just pick a few problems in one of the sections and play around with
More informationLinear Equations in Linear Algebra
1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION,, 1 n A linear equation in the variables equation that can be written in the form a a a b 1 1 2 2 n n a a is an where
More informationSolution Set 3, Fall '12
Solution Set 3, 86 Fall '2 Do Problem 5 from 32 [ 3 5 Solution (a) A = Only one elimination step is needed to produce the 2 6 echelon form The pivot is the in row, column, and the entry to eliminate is
More information7. Dimension and Structure.
7. Dimension and Structure 7.1. Basis and Dimension Bases for Subspaces Example 2 The standard unit vectors e 1, e 2,, e n are linearly independent, for if we write (2) in component form, then we obtain
More informationLinear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.
Homework # due Thursday, Oct. 0. Show that the diagonals of a square are orthogonal to one another. Hint: Place the vertices of the square along the axes and then introduce coordinates. 2. Find the equation
More informationGaussian Elimination and Back Substitution
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 4 Notes These notes correspond to Sections 31 and 32 in the text Gaussian Elimination and Back Substitution The basic idea behind methods for solving
More informationPOLI270 - Linear Algebra
POLI7 - Linear Algebra Septemer 8th Basics a x + a x +... + a n x n b () is the linear form where a, b are parameters and x n are variables. For a given equation such as x +x you only need a variable and
More informationLinear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions
Linear Systems Carlo Tomasi June, 08 Section characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More information1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)
Notes for 2017-01-30 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty
More informationLinearly-solvable Markov decision problems
Advances in Neural Information Processing Systems 2 Linearly-solvable Markov decision problems Emanuel Todorov Department of Cognitive Science University of California San Diego todorov@cogsci.ucsd.edu
More informationLecture Summaries for Linear Algebra M51A
These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture
More informationchapter 5 INTRODUCTION TO MATRIX ALGEBRA GOALS 5.1 Basic Definitions
chapter 5 INTRODUCTION TO MATRIX ALGEBRA GOALS The purpose of this chapter is to introduce you to matrix algebra, which has many applications. You are already familiar with several algebras: elementary
More informationσ 11 σ 22 σ pp 0 with p = min(n, m) The σ ii s are the singular values. Notation change σ ii A 1 σ 2
HE SINGULAR VALUE DECOMPOSIION he SVD existence - properties. Pseudo-inverses and the SVD Use of SVD for least-squares problems Applications of the SVD he Singular Value Decomposition (SVD) heorem For
More informationAMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems
AMS 209, Fall 205 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems. Overview We are interested in solving a well-defined linear system given
More informationSingular Value Decomposition
Singular Value Decomposition (Com S 477/577 Notes Yan-Bin Jia Sep, 7 Introduction Now comes a highlight of linear algebra. Any real m n matrix can be factored as A = UΣV T where U is an m m orthogonal
More informationMATH 22A: LINEAR ALGEBRA Chapter 4
MATH 22A: LINEAR ALGEBRA Chapter 4 Jesús De Loera, UC Davis November 30, 2012 Orthogonality and Least Squares Approximation QUESTION: Suppose Ax = b has no solution!! Then what to do? Can we find an Approximate
More informationDetermine whether the following system has a trivial solution or non-trivial solution:
Practice Questions Lecture # 7 and 8 Question # Determine whether the following system has a trivial solution or non-trivial solution: x x + x x x x x The coefficient matrix is / R, R R R+ R The corresponding
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationMAT 2037 LINEAR ALGEBRA I web:
MAT 237 LINEAR ALGEBRA I 2625 Dokuz Eylül University, Faculty of Science, Department of Mathematics web: Instructor: Engin Mermut http://kisideuedutr/enginmermut/ HOMEWORK 2 MATRIX ALGEBRA Textbook: Linear
More informationMath Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88
Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant
More informationWorksheet for Lecture 25 Section 6.4 Gram-Schmidt Process
Worksheet for Lecture Name: Section.4 Gram-Schmidt Process Goal For a subspace W = Span{v,..., v n }, we want to find an orthonormal basis of W. Example Let W = Span{x, x } with x = and x =. Give an orthogonal
More informationSolution of Linear Equations
Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationLinear Algebra March 16, 2019
Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented
More informationPolynomial functions over nite commutative rings
Polynomial functions over nite commutative rings Balázs Bulyovszky a, Gábor Horváth a, a Institute of Mathematics, University of Debrecen, Pf. 400, Debrecen, 4002, Hungary Abstract We prove a necessary
More information3 Matrix Algebra. 3.1 Operations on matrices
3 Matrix Algebra A matrix is a rectangular array of numbers; it is of size m n if it has m rows and n columns. A 1 n matrix is a row vector; an m 1 matrix is a column vector. For example: 1 5 3 5 3 5 8
More informationReview of Basic Concepts in Linear Algebra
Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra
More informationChapter 1. Matrix Algebra
ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface
More informationLesson U2.1 Study Guide
Lesson U2.1 Study Guide Sunday, June 3, 2018 2:05 PM Matrix operations, The Inverse of a Matrix and Matrix Factorization Reading: Lay, Sections 2.1, 2.2, 2.3 and 2.5 (about 24 pages). MyMathLab: Lesson
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationReview of Matrices and Block Structures
CHAPTER 2 Review of Matrices and Block Structures Numerical linear algebra lies at the heart of modern scientific computing and computational science. Today it is not uncommon to perform numerical computations
More informationA linear equation in two variables is generally written as follows equation in three variables can be written as
System of Equations A system of equations is a set of equations considered simultaneously. In this course, we will discuss systems of equation in two or three variables either linear or quadratic or a
More informationUNIT 3 MATRICES - II
Algebra - I UNIT 3 MATRICES - II Structure 3.0 Introduction 3.1 Objectives 3.2 Elementary Row Operations 3.3 Rank of a Matrix 3.4 Inverse of a Matrix using Elementary Row Operations 3.5 Answers to Check
More informationEquality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.
Introduction Matrix Operations Matrix: An m n matrix A is an m-by-n array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read
More informationCHAPTER 6. Direct Methods for Solving Linear Systems
CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to
More information[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]
Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the
More informationMATH2210 Notebook 2 Spring 2018
MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationSingular Value Decomposition
Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our
More informationContents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces
Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v 250) Contents 2 Vector Spaces 1 21 Vectors in R n 1 22 The Formal Denition of a Vector Space 4 23 Subspaces 6 24 Linear Combinations and
More informationA Field Extension as a Vector Space
Chapter 8 A Field Extension as a Vector Space In this chapter, we take a closer look at a finite extension from the point of view that is a vector space over. It is clear, for instance, that any is a linear
More informationIndex. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)
page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation
More informationOrthonormal Bases; Gram-Schmidt Process; QR-Decomposition
Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 205 Motivation When working with an inner product space, the most
More information5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns
5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns (1) possesses the solution and provided that.. The numerators and denominators are recognized
More information