Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Similar documents
APPENDIX A Some Linear Algebra

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Norms, Condition Numbers, Eigenvalues and Eigenvectors

Errors for Linear Systems

2.3 Nilpotent endomorphisms

= = = (a) Use the MATLAB command rref to solve the system. (b) Let A be the coefficient matrix and B be the right-hand side of the system.

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

MEM 255 Introduction to Control Systems Review: Basics of Linear Algebra

332600_08_1.qxp 4/17/08 11:29 AM Page 481

8.6 The Complex Number System

DIFFERENTIAL FORMS BRIAN OSSERMAN

Homework Notes Week 7

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

More metrics on cartesian products

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Quantum Mechanics for Scientists and Engineers. David Miller

Formulas for the Determinant

THE CHINESE REMAINDER THEOREM. We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens

Singular Value Decomposition: Theory and Applications

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.

Vector Norms. Chapter 7 Iterative Techniques in Matrix Algebra. Cauchy-Bunyakovsky-Schwarz Inequality for Sums. Distances. Convergence.

2 More examples with details

Bezier curves. Michael S. Floater. August 25, These notes provide an introduction to Bezier curves. i=0

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

Linear Approximation with Regularization and Moving Least Squares

MAT 578 Functional Analysis

From Biot-Savart Law to Divergence of B (1)

Week 2. This week, we covered operations on sets and cardinality.

p 1 c 2 + p 2 c 2 + p 3 c p m c 2

Quantum Mechanics I - Session 4

The Order Relation and Trace Inequalities for. Hermitian Operators

Section 8.3 Polar Form of Complex Numbers

Math 217 Fall 2013 Homework 2 Solutions

Unit 5: Quadratic Equations & Functions

MATH 241B FUNCTIONAL ANALYSIS - NOTES EXAMPLES OF C ALGEBRAS

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0

CHAPTER III Neural Networks as Associative Memory

Affine transformations and convexity

Foundations of Arithmetic

The Geometry of Logit and Probit

COMPLEX NUMBERS AND QUADRATIC EQUATIONS

Spectral Graph Theory and its Applications September 16, Lecture 5

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

PRIMES 2015 reading project: Problem set #3

Lecture 3: Probability Distributions

5 The Rational Canonical Form

Example: (13320, 22140) =? Solution #1: The divisors of are 1, 2, 3, 4, 5, 6, 9, 10, 12, 15, 18, 20, 27, 30, 36, 41,

Difference Equations

Poisson brackets and canonical transformations

Complex Numbers. x = B B 2 4AC 2A. or x = x = 2 ± 4 4 (1) (5) 2 (1)

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Linear Algebra and its Applications

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Lecture 10 Support Vector Machines II

SL n (F ) Equals its Own Derived Group

Additional Codes using Finite Difference Method. 1 HJB Equation for Consumption-Saving Problem Without Uncertainty

CHAPTER 4. Vector Spaces

Causal Diamonds. M. Aghili, L. Bombelli, B. Pilgrim

Affine and Riemannian Connections

MATH Sensitivity of Eigenvalue Problems

NOTES ON SIMPLIFICATION OF MATRICES

Lecture 10 Support Vector Machines. Oct

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Workshop: Approximating energies and wave functions Quantum aspects of physical chemistry

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Lecture 6/7 (February 10/12, 2014) DIRAC EQUATION. The non-relativistic Schrödinger equation was obtained by noting that the Hamiltonian 2

Asymptotics of the Solution of a Boundary Value. Problem for One-Characteristic Differential. Equation Degenerating into a Parabolic Equation

Lecture 12: Discrete Laplacian

a b a In case b 0, a being divisible by b is the same as to say that

12 MATH 101A: ALGEBRA I, PART C: MULTILINEAR ALGEBRA. 4. Tensor product

Ballot Paths Avoiding Depth Zero Patterns

P A = (P P + P )A = P (I P T (P P ))A = P (A P T (P P )A) Hence if we let E = P T (P P A), We have that

1 Vectors over the complex numbers

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

Random Walks on Digraphs

The Second Eigenvalue of Planar Graphs

Digital Signal Processing

Structure and Drive Paul A. Jensen Copyright July 20, 2003

Lecture Notes on Linear Regression

20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness.

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Appendix for Causal Interaction in Factorial Experiments: Application to Conjoint Analysis

9 Characteristic classes

w ). Then use the Cauchy-Schwartz inequality ( v w v w ).] = in R 4. Can you find a vector u 4 in R 4 such that the

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1.

Complex Numbers Alpha, Round 1 Test #123

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 16

DISCRIMINANTS AND RAMIFIED PRIMES. 1. Introduction A prime number p is said to be ramified in a number field K if the prime ideal factorization

FINITELY-GENERATED MODULES OVER A PRINCIPAL IDEAL DOMAIN

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

PHYS 705: Classical Mechanics. Calculus of Variations II

Appendix B. Criterion of Riemann-Stieltjes Integrability

Restricted Lie Algebras. Jared Warner

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

LECTURE V. 1. More on the Chinese Remainder Theorem We begin by recalling this theorem, proven in the preceeding lecture.

Transcription:

Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could choose our norms analogous to the way we dd for vector norms; e.g., we could assocate the number max j a j. However, ths s actually not very useful because remember our goal s to study lnear systems A x = b. The general defnton of a matrx norm s a map from all m n matrces to IR 1 whch satsfes certan propertes. However, the most useful matrx norms are those that are generated by a vector norm; agan the reason for ths s that we want to solve A x = b so f we take the norm of both sdes of the equaton t s a vector norm and on the left hand sde we have the norm of a matrx tmes a vector. We wll defne an nduced matrx norm as the largest amount any vector s magnfed when multpled by that matrx,.e., A = max A x Note that all norms on the rght hand sde are vector norms. We wll denote a vector and matrx norm usng the same notaton; the dfference should be clear from the argument. We say that the vector norm on the rght hand sde nduces the matrx norm on the left. Note that sometmes the defnton s wrtten n an equvalent way as A = sup A x What s I? Clearly t s just one. The problem wth the defnton s that t doesn t tell us how to compute a matrx norm for a general matrx A. The followng theorem gves us a way to calculate matrx norms nduced by the l and l 1 norms; the matrx norm nduced by l 2 norm wll be addressed later after we have ntroduced egenvalues. Theorem Let A be an m n matrx. Then A = max 1 m A 1 = max 1 j n [ m [ n =1 ] a j ] a j (max absolute row sum) (max absolute column sum) 1

Determne A and A 1 where A = 4 3 0 12 20 We have A 1 = max{(1 + 3 + 20), (2 + 1), (4 + 12 + 2)} = max{24, 3, 18} = 24 A = max{(1 + 2 + 4), (3 + 12), (20 + 1 + 2)} = max{7, 15, 23} = 23 Proof We wll prove that A s the maxmum row sum (n absolute value). We wll do ths by provng that A max 1 m [ n Frst recall that f A x = b then b = ] a j and then showng a j x j b = max For the frst nequalty we know that by defnton Now lets smplfy the numerator to get A x = max A = max a j x j max b = max A x A max 1 m a j x j a j x j max [ n a j ] a j Thus the rato reduces to A x n max a j = max a j and hence A = max A x Now for the second nequalty we know that max a j. A A y y 2

for any y IR n because equalty n the defnton holds here for the maxmum of ths rato. So now we wll choose a partcular y and we wll construct t so that t has y = 1. Frst let p be the row where A has ts maxmum row sum (or there are two rows, take the frst),.e., max a j = a pj Now we wll take the entres of y to be ±1 so ts nfnty norm s one. Specfcally we choose y = { 1 f apj 0 1 f a pj < 0 Defnng y n ths way means that a j y j = a j. Usng ths and the fact that y = 1 we have A A y y = max a j y j a pj y j = n a pj = a pj but the last quantty on the rght s must the maxmum row sum and the proof s complete. What are the propertes that any matrx norm (and thus an nduced norm) satsfy? You wll recognze most of them as beng analogous to the propertes of a vector norm. () A 0 and = 0 only f a j = 0 for all, j. () ka = k A for scalars k () AB A B (v) A + B A + B A very useful nequalty s A x A for any nduced norm Why s ths true? A = max A x A x A x A The Condton Number of a Matrx We sad one of our goals was to determne f small changes n our data of a lnear system produces small changes n the soluton. Now lets assume we want to solve A x = b, b 0 but nstead we solve A y = b + δb that s, we have perturbed the rght hand sde by a small about δb. We assume that A s nvertble,.e., A 1 exsts. For smplcty, we have not perturbed the coeffcent matrx 3

A. What we want to see s how much y dffers from x. Lets wrte y as x + δx and so our change n the soluton wll be δx. The two systems are A x = b A( x + δx) = b + δb What we would lke to get s an estmate for the relatve change n the soluton,.e., δx n terms of the relatve change n b where denotes any nduced vector norm. Subtractng these two equatons gves A δx = δb whch mples δx = A 1 δb Now we take the (vector) norm of both sdes of the equaton and then use our favorte nequalty above δx = A 1 δb A 1 δb Remember our goal s to get an estmate for the relatve change n the soluton so we have a bound for the change. What we need s a bound for the relatve change. Because A x = b we have A x = b b A 1 b 1 A Now we see that f we dvde our prevous result for δx by A > 0 we can use ths result to ntroduce b n the denomnator. We have δx A A 1 δb A Multplyng by A gves the desred result A 1 δb b δx A A 1 δb b If the quantty A A 1 s small, then ths means small relatve changes n b result n small relatve changes n the soluton but f t s large, we could have a large relatve change n the soluton. Defnton The condton number of a square matrx A s defned as K(A) A A 1 Note that the condton number depends on what norm you are usng. We say that a matrx s well-condtoned f K(A) s small and ll-condtoned otherwse. 4

Fnd the condton number of the dentty matrx usng the nfnty norm. Clearly t s one because the nverse of the dentty matrx s tself. Fnd the condton number for each of the followng matrces usng the nfnty norm.. Explan geometrcally why you thnk ths s happenng. A 1 = A 4 3 2 = 0.998 2 Frst we need to fnd the nverse of each matrx and then take the norms. Note the followng trck for takng the nverse of a 2 2 matrx A 1 1 = 1 3 2 A 1 500 500 5 4 1 2 = 249.5 250 Now K (A 1 ) = (7)(1) = 7 K (A 2 ) = (3)(1000) = 3000 Calculate the K (A) where A s a b A = c d and comment on when the condton number wll be large. A 1 1 d b = ad bc c a So 1 1 K (A) = max{ a + b, c + d } max{ d + b, c + a } ad bc ad bc Consequently f the determnant ad bc 0 then the condton number can be qute large;.e., when the matrx s almost not nvertble,.e., almost sngular. A classc example of an ll condtoned matrx s the Hlbert matrx whch we have already encountered Here are some of the condton numbers (usng the matrx norm nduced by the l 2 vector norm). n approxmate K 2 (A) 2 19 3 524 4 15,514 5 476,607 6 1.495 10 7 5

Vector (or Lnear) Spaces What we want to do now s take the concept of vectors and the propertes of Eucldean space IR n and generalze them to a collecton of objects wth two operatons defned, addton and scalar multplcaton. We want to defne these operatons so that the usual propertes hold,.e., x + y = y + x, k(x + y) = kx + ky for objects other than vectors, such as matrces, contnuous functons, etc. We want to be able to add two elements of our set and get another element of the set. Defnton A vector or lnear space V s a set of objects, whch we wll call vectors, for whch addton and scalar multplcaton are defned and satsfy the followng propertes. () x + y = y + x () x + (y + z) = (x + y) + z () there exsts a zero element 0 V such that x + 0 = 0 + x = x (v) for each x V there exsts x V such that x + ( x) = 0 (v) 1x = x (v) (α + β)x = αx + βx, for scalars α, β (v) α(x + y) = αx + αy (v) (αβ)x = α(βx) IR n s a vector space wth the usual operaton of addton and scalar multpl- caton. The set of all m n matrces wth the usual defnton of addton and scalar multplcaton forms a vector space. All polynomals of degree less than or equal to two on [ 1, 1] form a vector space wth usual defnton of addton and scalar multplcaton. Here p(x) = a 0 + a 1 x + a 2 x 2 All contnuous functons defned on [0, 1]. Note that f v, w V then v + w V ; we say that V s closed under addton. Also f v V and k s a scalar, kv V and t s closed under scalar multplcaton. Defnton A subspace S of a vector space V s a nonempty subset of V such that f we take any lnear combnaton of vectors n S t s also n S. A lne s a subspace of IR 2 because f we take add two ponts on the lne, the result s stll a pont on the lne and f we multply any pont on the lne by a constant then t s stll on the lne. matrces. All lower trangular matrces form a subspace of the vector space of all square 6

Is the set of all ponts on the gven lnes a subspace of IR 2? y = x y = 2x + 1 The frst forms a subspace but the second does not. It s not closed under addton or scalar multplcaton, e.g., 2(x, 2x + 1) = (2x, 4x + 2) whch s not on the lne y = 2x + 1 Lnear ndependence, spannng sets and bass We now want to nvestgate how to characterze a vector space, f possble. To do ths, we need the concepts of () lnear ndependence/dependence of vectors () a spannng set () the dmenson of a vector space (v) the bass of a fnte dmensonal vector space In IR 2 the vectors v 1 = (1, 0) T and v 2 = (0, 1) T are specal because () every other vector n IR 2 can be wrtten as a lnear combnaton of these two vectors, e.g., f x = (x 1, x 2 ) T then x = x 1 v 1 + x 2 v 2. () the vectors v 1, v 2 are dfferent n the sense that the only way we can combne them and get the zero vector s wth zero coeffcents,.e., C 1 v 1 + C 2 v 2 = 0 C 1 = C 2 = 0 Of course there are many other sets of two vectors n IR 2 that have the same propertes. Defnton Let V be a vector space. Then the set of vectors { v }, = 1,..., n () span V f any w V can be wrtten as a lnear combnaton of the v,.e., () are lnearly ndependent f w = C j v j C j v j = 0 C = 0, otherwse they are lnearly dependent. Note that ths says the only way we can combne the vectors and get the zero vector s f all coeffcents are zero. () form a bass for V f they are lnearly ndependent and span V. 7

Let V be the vector space of all polynomals of degree less than or equal to two on [0, 1]. Then the set of vectors (.e., polynomals) n V {1, x, x 2, 2 + 6x} span V but they are NOT lnearly ndependent because 2(1) + 6(x) + 0(x 2 ) (2 + 6x) = 0 and thus they do not form a bass for V. The set of vectors {1, x} are lnearly ndependent but they do not span V because, e.g., x 2 V and t can t be wrtten as a lnear combnaton of {1, x}. Defnton The number of elements n the bass for V s called the dmenson of V. It s the smallest number of vectors needed to completely characterze V,.e., that span V. Fnd a bass for () the vector space of all 2 2 matrces and () the vector space of all 2 2 symmetrc matrces. Gve the dmenson. A bass for all 2 2 matrces conssts of the four matrces 1 0, 0 1,, 1 0 0 1 and so ts dmenson s 4. A bass for all 2 2 symmetrc matrces conssts of the three matrces 1 1,, 0 1 1 0 and so ts dmenson s 2. Four Important Spaces 1. The column space (or equvalently the range) of A, where A s m n matrx s all lnear combnatons of the columns of A. We denote ths by R(A). By defnton t s a subspace of IR m. Consder the overdetermned system 3 4 x = b 1 b y 2 5 6 b 3 We know that ths s solvable f there exsts x, y such that x 1 3 5 + y 2 4 = 6 b 1 b 2 b 3 whch says that b has to be n the column space (or range) of A. 8

Ths says that an equvalent statement to A x = b beng solvable s that b s n the range or column space of A. 2. The null space of A, denoted N (A), where A s m n matrx s the set of all vectors z IR n such that A z = 0. Fnd the null space of each matrx A 1 = 3 4 A 2 = 2 4 A 3 = For A 1 we have that N (A 1 ) = 0 because the matrx s nvertble. To see ths we could take the determnant or perform GE and get the result. For A 2 we see that N (A 2 ) s α( 2x 2, x 2 ) T,.e., the all ponts on the lne through the orgn y =.5x. To see ths consder GE for the system ( 2 4 ) ( Ths says that x 2 s arbtrary and x 1 = 2x 2. For A 3 we see that N (A 2 ) s all of IR 2. ) 0 x 2 = 0, x 1 + 2x 2 = 0 What are the possble null spaces of a 3 3 matrx? For an nvertble matrx t s the () zero vector n IR 3, we could have () a lne through the orgn, () a plane through the orgn or all of (v) IR 3. 9