Lecture 1: Review of linear algebra

Similar documents
ELE/MCE 503 Linear Algebra Facts Fall 2018

Definitions for Quizzes

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

MAT Linear Algebra Collection of sample exams

6 Inner Product Spaces

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Lecture Summaries for Linear Algebra M51A

Online Exercises for Linear Algebra XM511

Linear Algebra. Session 12

ORTHOGONALITY AND LEAST-SQUARES [CHAP. 6]

Typical Problem: Compute.

235 Final exam review questions

Chapter 3 Transformations

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Elementary linear algebra

Introduction to Numerical Linear Algebra II

Review of Some Concepts from Linear Algebra: Part 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Lecture 7: Positive Semidefinite Matrices

Chap 3. Linear Algebra

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear Algebra Massoud Malek

Foundations of Matrix Analysis

Linear System Theory

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Lecture 2: Linear operators

Control Systems. Linear Algebra topics. L. Lanari

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Linear Algebra Review

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Linear Algebra Lecture Notes-II

1. General Vector Spaces

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

Basic Elements of Linear Algebra

MA 265 FINAL EXAM Fall 2012

Applied Linear Algebra in Geoscience Using MATLAB

Algebra II. Paulius Drungilas and Jonas Jankauskas

Fall 2016 MATH*1160 Final Exam

Basic Concepts in Linear Algebra

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

2. Every linear system with the same number of equations as unknowns has a unique solution.

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Lecture 2: Linear Algebra

Review of Basic Concepts in Linear Algebra

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Review of Linear Algebra

Stat 159/259: Linear Algebra Notes

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Linear Algebra Highlights

b 1 b 2.. b = b m A = [a 1,a 2,...,a n ] where a 1,j a 2,j a j = a m,j Let A R m n and x 1 x 2 x = x n

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

(v, w) = arccos( < v, w >

EXAM. Exam 1. Math 5316, Fall December 2, 2012

The following definition is fundamental.

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

CS 143 Linear Algebra Review

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Exercise Sheet 1.

Linear Algebra in Actuarial Science: Slides to the lecture

NOTES on LINEAR ALGEBRA 1

Definition (T -invariant subspace) Example. Example

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Recall the convention that, for us, all vectors are column vectors.

Chapter 2 Subspaces of R n and Their Dimensions

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Lecture 2: Linear Algebra Review

Basic Calculus Review

MAT 610: Numerical Linear Algebra. James V. Lambers

2. Review of Linear Algebra

Generalized Eigenvectors and Jordan Form

SYLLABUS. 1 Linear maps and matrices

Vector Spaces and Linear Transformations

Functional Analysis Review

Linear Algebra Review. Vectors

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Matrix Algebra: Summary

Linear Algebra I for Science (NYC)

Elementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

MATH36001 Generalized Inverses and the SVD 2015

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Matrix Vector Products

MATH Spring 2011 Sample problems for Test 2: Solutions

Lecture Notes 1: Vector spaces

Linear Algebra Methods for Data Mining

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

MTH 2032 SemesterII

Spring 2014 Math 272 Final Exam Review Sheet

Linear Algebra Practice Problems

Transcription:

Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations Eigenvalues and eigenvectors Linear algebra with applications (L1) Review of linear algebra 1 / 32

Notation R real numbers, Z integers, N natural numbers R n n-dimensional real vector space R m n space of real m n matrices LHS := RHS the LHS is defined by the RHS A the transposed of A Linear algebra with applications (L1) Review of linear algebra 2 / 32

Linear functions f : R n R m function mapping vectors in R n to vectors in R m Interpretation replacements of y = f(x): x given input, y corresponding output x f y f is a linear function if and only if superposition holds: f(αx + βv) = αf(x)+βf(v), for all α,β R, x,v R n f is linear A R m n, such that f(x) = Ax, for all x R n A is a matrix representing the linear function f Q: How can you find a matrix representation of a linear function f, if you are allowed to evaluate f at points x R n of your choice? Linear algebra with applications (L1) Review of linear algebra 3 / 32

Examples of linear functions Scalar function of a scalar argument y = tan(α)x, where α [0,2π) y α x Identity function x = f(x), for all x R n is a linear function represented by the identity matrix 1 I n :=... 1 R n n Linear algebra with applications (L1) Review of linear algebra 4 / 32

Matrix vector multiplication Partition A R m n elementwise, column-wise, and row-wise a 11 a 1n r 1 A =.. = c 1 c n =. a m1 a mn r m The matrix vector product y = Ax can be written as y 1.. = y m n j=1 a 1jx j. = n j=1 a mjx j n j=1 r 1 x c j x j =. r m x Interpretation: a ij gain factor from the jth input x j to the ith output y i. (e.g., a ij = 0 means that jth input has no influence on ith output.) Linear algebra with applications (L1) Review of linear algebra 5 / 32

Linearlization Consider a differentiable function f : R n R m. Then for given x 0 R n [ [ ] fi y = f(x 0 + x) f(x 0 ) +A x where A = aij = }{{} x j y 0 ]. x0 When the input deviation x = x x 0 is small, the output deviation ỹ := y y 0 is approximately a linear function of x, y f ỹ = A x Linear approximation of f : x y at x = x 0 y 0 x 0 x Linear algebra with applications (L1) Review of linear algebra 6 / 32

Rank of a matrix and inversion the set of vectors {v 1,...,v n } is independent if α 1 v 1 + +α n v n = 0 only if α 1 = = α n = 0 rank of a matrix number of lin. indep. columns (or rows) A R m n is full row rank (f.r.r.) if rank(a) = m Interpretation: A not f.r.r. there are redundant outputs Inversion problem: given y R m, find x, such that y = Ax. Interpretation: design an input that achieves a desired output. When is the inversion problem solvable? Is the solution unique? Linear algebra with applications (L1) Review of linear algebra 7 / 32

Inversion problem Given y R m, find x, such that y = Ax. Solution may not exist, be unique, or there may be many solutions. (Why it is not possible to have a finite number of solutions?) Interpretations: Control: x is a control input, y is a desired outcome Estimation: x is a vector of parameters, y is a set of measurements Typically in control, the solution is nonunique and we aim to find the best one. in estimation, there is no solution and we aim to find the best approximation. Linear algebra with applications (L1) Review of linear algebra 8 / 32

Inverse of a matrix If m = n = rank(a), then there exists a matrix A 1, such that Then for all y R m AA 1 = A 1 A = I m. y = (AA 1 ) y = A(A 1 y) = Ax. }{{}}{{} I x The inversion problem is solvable and the solution is unique. Q: Can you find a matrix representation of a linear function f, from given values y 1,...,y n of f at given points x 1,...,x n? If so, how? Linear algebra with applications (L1) Review of linear algebra 9 / 32

Vector and matrix norms Mathematical formalisation of the geometric notion of size or distance. Norm is a function x : x R that satisfies the following properties: Nonnegativity: x 0 for all x Definiteness: x = 0 x = 0 Homogeneity: αx = α x for all x and α Triangle inequality: x + y x + y Examples: Vector 2-norm: x 2 := x1 2 + +x n 2 = x x, for all x R n Frobenius matrix norm: A F := m i=1 n j=1 a2 ij, for all A Rm n Unit ball: U = {x x 1} Linear algebra with applications (L1) Review of linear algebra 10 / 32

Least-squares solution Assumption m n = rank(a), i.e., A is full column rank. The inversion problem typically has no solution. The least-squares solution minimizes the approximation error The matrix x ls = (A A) 1 A y =: A + y y Ax }{{} 2 := e e 2 1 + +e2 m = e e. A + := (A A) 1 A (if m > n = rank(a)) is called pseudo-inverse of A. Linear algebra with applications (L1) Review of linear algebra 11 / 32

Notes x ls is a linear function of y (given by the pseudo inverse matrix A + ) If A is square x ls = A 1 y (in other words A + = A 1 ) x ls is an exact solution if Ax = y has an exact solution ŷ = Ax ls = A(A A) 1 A y is a least-squares approximation of y Statistical interpretation: assume that y = Ax 0 + e where e is zero mean Gaussian random vector with covariance σ 2 I Then x ls is the best linear unbiased estimator for x 0. Linear algebra with applications (L1) Review of linear algebra 12 / 32

Least-norm solution Assumption n m = rank(a), i.e., A is full row rank. The inversion problem has infinitely many solution. The least-norm solution x ln = A (AA ) 1 y =: A + y minimizes the 2-norm of the solution x, i.e., The matrix minimize x 2 subject to Ax = y A + := A (AA ) 1 (if n > m = rank(a)) is called pseudo-inverse of A. Linear algebra with applications (L1) Review of linear algebra 13 / 32

Set of all solutions {x Ax = y } = {x p + z Az = 0} where x p is a particular solution, i.e., Ax p = y. Note that x ln = A (AA ) 1 y is a particular solution Ax ln = (AA )(AA ) 1 y = y. Moreover, x ln is the minimum 2-norm solution. Linear algebra with applications (L1) Review of linear algebra 14 / 32

Inner product The inner product of two vectors a,b R n is defined as a,b := a b = n k=1 a k b k. Matrix matrix product H = GF, F : R p n, G : R n m gives pm inner products between the rows of G and the columns of F g 1 g 1,f 1 g 1,f p H = GF =. f 1 f p =.. g m g m,f 1 g m,f p The Gram matrix of the vectors f 1,...,f m is defined by f 1... f m [ ] f 1 f m Linear algebra with applications (L1) Review of linear algebra 15 / 32

Cauchy-Schwarz inequality x y x y equality holds if and only if x = αy, for some α R or x = 0. Application: optimization of a linear function over the unit ball Given y R n maximize x y subject to x 1 The solution follows from the Cauchy-Schwarz inequality x opt = y y Linear algebra with applications (L1) Review of linear algebra 16 / 32

Angle between vectors The angle between the vectors x,y R n is defined as (x,y) = cos 1 x y x y x 0 and y are aligned if y = αx, for some α 0 In this case, (x,y) = 0. x 0 and y are opposite if y = αx, for some α 0 In this case, (x,y) = π. x and y are orthogonal (denoted x y) if x y = 0 In this case, (x,y) = π/2. Q: Given y R n, which x minimize x y subject to x 1? Linear algebra with applications (L1) Review of linear algebra 17 / 32

Subspace, basis, and dimension V R n is a subspace of a vector space R n if V is a vector space v,w V = αv + βw V, for all α,β R The set {v 1,...,v n } is a basis of V if v 1,...,v n span V, i.e., V = span(v 1,...,v n ) := {α 1 v 1 + +α n v n α 1,...,α n R} {v 1,...,v n } is an independent set of vectors. dim(v ) number of basis vectors (does not depend on the basis) Linear algebra with applications (L1) Review of linear algebra 18 / 32

Null space of a matrix (kernel) kernel of A the set of vectors mapped to zero by f(x) := Ax ker(a) := {x R n Ax = 0} y = A(x + x), for all x ker(a) Interpretation: ker(a) is the uncertainty in finding x, given y. Interpretation: ker(a) is the freedom in the x s that achieve y. ker(a) = {0} f(x) := Ax is one-to-one ker(a) = {0} A is full column rank Linear algebra with applications (L1) Review of linear algebra 19 / 32

Range of a matrix (image) image of A the set of all vectors obtainable by f(x) := Ax image(a) := {Ax x R n } image(a) = span of the columns of A image(a) = set of vectors y for which Ax = y has a solution image(a) = R m f(x) := Ax is onto (image(f) = R m ) image(a) = R m A is full row rank Linear algebra with applications (L1) Review of linear algebra 20 / 32

Change of basis standard basis vectors in R n the columns e 1,...,e n of I n Elements of x R n are coordinates of x w.r.t. standard basis. A new bases is given by the columns t 1,...,t n of T R n n. The coordinates of x in the new basis are x 1,..., x n, such that x = x 1 t 1 + + x n t n = T x = x = T 1 x T 1 transforms standard basis coordinates x into T -coordinates Linear algebra with applications (L1) Review of linear algebra 21 / 32

Similarity transformation Consider linear operator f : R n R n, given by f(x) = Ax, A R n n. Change standard basis to basis defined by columns of T R n n. The matrix representation of f changes to T 1 AT : x = T x, y = T ỹ = ỹ = (T 1 AT) x A T 1 AT similarity transformation of A Linear algebra with applications (L1) Review of linear algebra 22 / 32

Eigenvalues and eigenvectors λ C is eigenvalue of A R n n : there is v 0, s.t. Av = λv : λi n A is singular Any nonzero v C n such that Av = λv is called an eigenvector of A associate with the eigenvalue λ. Meaning of λ and v: the action of A in the direction defined by v is equivalent to scalar multiplication by λ Characteristic polynomial of A: p A (λ) := det(λi n A), deg(p A ) = n λ is an eigenvalue of A if and only if λ is a root of p A Geometric multiplicity of λ: dim ( (λi n A) ) Algebraic multiplicity of λ: multiplicity of the root λ of p A Linear algebra with applications (L1) Review of linear algebra 23 / 32

Eigenvalue decomposition Suppose {v 1,...,v n } is a lin. indep. set of eigenvectors of A R n n Av i = λ i v i, for i = 1,...,n written in a matrix form is V is nonsingular, so that A [ ] v 1 v n = [ 1 ] v 1 v n λ... }{{}}{{} V V λ n } {{ } Λ AV = V Λ = V 1 AV = Λ Linear algebra with applications (L1) Review of linear algebra 24 / 32

Diagonalization by similarity transformation V is nonsingular since by assumption {v 1,...,v n } is lin. indep. similarity transformation with T = V 1 diagonalizes A Conversely if there is a nonsingular V C n n, such that V 1 AV = Λ then Av i = λ i v i and {v 1,...,v n } is a lin. indep. set of eigenvectors A is diagonalizable if there is nonsingular T, such that TAT 1 is diagonal there is a set of n lin. indep. eigenvectors of A if A is not diagonalizable, it is called defective Linear algebra with applications (L1) Review of linear algebra 25 / 32

Jordan canonical form Distinct eigenvalues = diagonalizable matrix (converse not true) Prototypical example of a defective matrix: λ 1 λ...... 1 λ Defective matrices have an eigenvalue which algebraic multiplicity is higher than the corresponding geometric multiplicity. Jordan form: generalization of TAT 1 = Λ for defective matrices Linear algebra with applications (L1) Review of linear algebra 26 / 32

Eigenvalues and eigenvectors of symmetric matrix Theorem: A symmetric matrix A has real eigenvalues and a full set of eigenvectors, that can be chosen to form an orthonormal set. Symmetric matrix = real eigenvalues orthonormal eigenvectors Linear algebra with applications (L1) Review of linear algebra 27 / 32

Summary f is linear if superposition holds f(αx + βv) = αf(x)+βf(v) f is linear there is matrix A, such that f(x) = Ax image (column span, range) of A image(a) := {Ax x R n } kernel (null space) of A ker(a) := {x R n Ax = 0} V R n is subspace of R n if αv + βw V for all v,w V basis of a subspace set of linearly indep. vectors that span V dimension of a subspace the number of basis vectors image(a) and ker(a) are subspaces Linear algebra with applications (L1) Review of linear algebra 28 / 32

rank of A number of linearly independent rows (or columns) dim ( image(a) ) = rank(a), coldim(a) dim(ker(a)) = rank(a) A is full row rank if rank(a) = rowdim(a) A is full column rank if rank(a) = coldim(a) A is full rank if either full row or column rank A is nonsingular if A is square and full rank inversion problem: given y = Ax, find x A + is left inverse of A if A + A = I solution of the inversion problem: x = A + y, A + A = I Linear algebra with applications (L1) Review of linear algebra 29 / 32

left inverse exists iff A is full column rank least-squares left inverse A ls = (A A) 1 A A + is right inverse of A if AA + = I right inverse exists iff A is full row rank least-norm right inverse A ln = A (AA ) 1 A 1 is inverse of A if A 1 A = A 1 A = I for A to have inverse, A should be square and full rank 2-norm of a vector x = x x, unit ball {x x 1} inner product of a,b R n a,b := a b Linear algebra with applications (L1) Review of linear algebra 30 / 32

Cauchy-Schwarz inequality: a b a b a,b R n are orthogonal if a,b = 0 similarity transformation A T 1 AT, T nonsingular eigenvalue decomposition A = T 1 ΛT, Λ diagonal characteristic polynomial of A p A (λ) := det(λi A) symmetric matrix A = A = real eigenvalues orthonormal eigenvectors Linear algebra with applications (L1) Review of linear algebra 31 / 32

References Introductory texts: G. Strang, Introduction to linear algebra G. Strang, Linear algebra and its applications, C. Meyer, Matrix analysis and applied linear algebra, SIAM, 2000 Advanced texts: R. Bellman, Introduction to matrix analysis, 1970 R. Horn & Johnson, Matrix analysis, Cambridge Univ. Press, 1985 R. Horn & Johnson, Topics in matrix analysis, CUP, 1991 Linear algebra with applications (L1) Review of linear algebra 32 / 32