Math 52: Course Summary

Similar documents
Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Linear algebra and differential equations (Math 54): Lecture 10

2. FUNCTIONS AND ALGEBRA

Lecture 6: Finite Fields

Linear Algebra, Summer 2011, pt. 3

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Math 110, Spring 2015: Midterm Solutions

Linear Algebra, Summer 2011, pt. 2

CMU CS 462/662 (INTRO TO COMPUTER GRAPHICS) HOMEWORK 0.0 MATH REVIEW/PREVIEW LINEAR ALGEBRA

Writing proofs for MATH 51H Section 2: Set theory, proofs of existential statements, proofs of uniqueness statements, proof by cases

NOTES ON LINEAR ALGEBRA CLASS HANDOUT

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Modern Algebra Prof. Manindra Agrawal Department of Computer Science and Engineering Indian Institute of Technology, Kanpur

Complex Numbers. Rich Schwartz. September 25, 2014

2 Metric Spaces Definitions Exotic Examples... 3

2 Systems of Linear Equations

Projections and Least Square Solutions. Recall that given an inner product space V with subspace W and orthogonal basis for

Math 126: Course Summary

Notes on Linear Algebra I. # 1

Linear Algebra I. Ronald van Luijk, 2015

Lecture 10: Powers of Matrices, Difference Equations

Section 3.1: Definition and Examples (Vector Spaces)

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Metric spaces and metrizability

One-to-one functions and onto functions

MATH 583A REVIEW SESSION #1

0.2 Vector spaces. J.A.Beachy 1

Linear Algebra and Robot Modeling

Math 101: Course Summary

Chapter 4. Solving Systems of Equations. Chapter 4

1.1.1 Algebraic Operations

Eigenvectors and Hermitian Operators

Introduction: Pythagorean Triplets

Math 123, Week 2: Matrix Operations, Inverses

LINEAR ALGEBRA KNOWLEDGE SURVEY

Answers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3

MATH 115, SUMMER 2012 LECTURE 12

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

Row Reduction

Linear Algebra Basics

Econ Slides from Lecture 8

1.1 Basic Algebra. 1.2 Equations and Inequalities. 1.3 Systems of Equations

Math 114: Course Summary

Honors Advanced Mathematics Determinants page 1

Math (P)Review Part I:

Induction 1 = 1(1+1) = 2(2+1) = 3(3+1) 2

Questionnaire for CSET Mathematics subset 1

Abstract Vector Spaces and Concrete Examples

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)

Contents. 1 Vectors, Lines and Planes 1. 2 Gaussian Elimination Matrices Vector Spaces and Subspaces 124

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Calculus (Math 1A) Lecture 5

Final Exam Practice Problems Answers Math 24 Winter 2012

Vector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n.

MATH 12 CLASS 4 NOTES, SEP

Commutative Rings and Fields

Sometimes the domains X and Z will be the same, so this might be written:

ELLIPTIC CURVES BJORN POONEN

Page 52. Lecture 3: Inner Product Spaces Dual Spaces, Dirac Notation, and Adjoints Date Revised: 2008/10/03 Date Given: 2008/10/03

Definition (T -invariant subspace) Example. Example

Linear Independence Reading: Lay 1.7

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

Some Notes on Linear Algebra

Math 309 Notes and Homework for Days 4-6

Introduction to Linear Algebra

CH 59 SQUARE ROOTS. Every positive number has two square roots. Ch 59 Square Roots. Introduction

Chapter 1: Linear Equations

2. Prime and Maximal Ideals

a (b + c) = a b + a c

2 so Q[ 2] is closed under both additive and multiplicative inverses. a 2 2b 2 + b

Figure 1: Doing work on a block by pushing it across the floor.

Math From Scratch Lesson 37: Roots of Cubic Equations

Generating Function Notes , Fall 2005, Prof. Peter Shor

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

Linear Algebra March 16, 2019

The Growth of Functions. A Practical Introduction with as Little Theory as possible

Solving simple equations, a review/primer

Chapter 1: Linear Equations

MTH 2032 Semester II

Math Lecture 3 Notes

REVIEW OF DIFFERENTIAL CALCULUS

0. Introduction 1 0. INTRODUCTION

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16

Vector, Matrix, and Tensor Derivatives

Linear Algebra. Min Yan

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

MAT1302F Mathematical Methods II Lecture 19

Elementary Linear Algebra

Math 416, Spring 2010 More on Algebraic and Geometric Properties January 21, 2010 MORE ON ALGEBRAIC AND GEOMETRIC PROPERTIES

Usually, when we first formulate a problem in mathematics, we use the most familiar

Linear Algebra. Preliminary Lecture Notes

(, ) : R n R n R. 1. It is bilinear, meaning it s linear in each argument: that is

TheFourierTransformAndItsApplications-Lecture28

1 Reductions from Maximum Matchings to Perfect Matchings

Getting Started with Communications Engineering

ABSTRACT VECTOR SPACES AND THE CONCEPT OF ISOMORPHISM. Executive summary

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Spanning, linear dependence, dimension

Euclidean Space. This is a brief review of some basic concepts that I hope will already be familiar to you.

Transcription:

Math 52: Course Summary Rich Schwartz September 2, 2009 General Information: Math 52 is a first course in linear algebra. It is a transition between the lower level calculus courses and the upper level math courses. I would say that Math 52 has four general aims: To clarify and unify all the stuff you learned in high school about linear equations and matrices. To introduce you to higher dimensional Euclidean spaces and the easiest kinds of mappings between these spaces the linear maps. To introduce you to the abstract formalism of vector spaces and linear transformations. The upper level math classes are filled with abstract structures like this, and the ones in M52 are a good warm-up. To introduce you to the idea of proving theorems in mathematics. Proofs abound in the upper level courses, and M52 gives an easy context in which to get started. M52 is traditionally taken after finishing the calculus series, but there is some argument for taking M52 in advance of several variable calculus e.g. M18, M20, M35. Several variable calculus uses linear algebra implicitly, and the constructions are more awkward without the linear algebra formalisms. Math 54 is a somewhat more advanced version of Math 52. In math 54, less attention is paid to concrete calculations with matrices and such, and more attention is paid to the theoretical side of the subject. High School Algebra Revisited: Suppose you want to solve the linear system 3x + 4y = 2 7x 3y = 5 1

It doesn t hurt to multiply both sides of the first equation by 3 and both sides of the second equation by 4. This gives 9x + 12y = 6 28x 12y = 20 It doesn t hurt to add the equations together. This gives 37x = 26 = x = 26/37. A similar trick will get y. Continuing with the example above, the only thing really important is the 3 2 array of numbers 3 4 2 7 3 5. You can simplify this array by adding multiples of the first column to the second column or vice versa. After some effort, you get down to an array that looks like 26 1 0 0 1 This final array tells us that x = 26/37 and y = 1/37. The same idea works for systems of equations with more variables. You might have an equation like 37 1 37 a 11 x 1 + a 12 x 2 + a 13 x 4 + a 14 x 4 = b 1 a 21 x 1 + a 22 x 2 + a 23 x 4 + a 24 x 4 = b 2, a 31 x 1 + a 32 x 2 + a 33 x 4 + a 34 x 4 = b 3 a 41 x 1 + a 42 x 2 + a 43 x 4 + a 44 x 4 = b 4 where a 11 etc. and b 1 etc are numbers and the variables x 1, x 2, x 3, x 4 are the unknowns. The important thing in solving this system is the 5 4 array a 11 a 12 a 13 a 14 b 1 a 21 a 22 a 23 a 24 b 2 a 31 a 32 a 33 a 34 b 3 a 41 a 42 a 43 a 44 b 4 You can solve this system (or see that it has no solutions) by adding multiples of one row to another until you have simplified the thing as much as possible. The first main thing you learn in M52 is an algorithmic way to do this, called 2

Gaussian elimination. This is really nothing more than a systematization of what you already knew how to do. Higher dimensional space: Here s another view of our first example above. Each of the two equations specifies a line in the plane, and then we are finding the point where these lines intersect. What about the second example? If there were just 3 variables, then each of the equations would specify a plane in space, and we would be looking for the point where the 3 planes intersect. Incidentally, one can see geometrically why such systems might not have solutions. For instance, the lines and planes might be parallel. In the actual second example, we could say that each of the equations in 4 variables specifies a copy of 3-dimensional space sitting inside 4-dimensional space, and we are looking for the quadruple intersection point. In M52 you learn that n-dimensional Euclidean space is nothing more than the set of all n-tuples (x 1,..., x n ) where these variables can take on any real number. Some people find this to be quite exhilarating. Most people are mystified by four dimensional space, and here you can think about a billion dimensional space without really sweating. Otjer people find this to be a letdown, because the concrete description we ve just given completely demystifies higher dimensional space. Once you have the underlying space, you can talk about distances between points. The distance between (x 1,..., x n ) and (y 1,..., y n ) is (x 1 y 1 ) 2 +... + (x n y n ) 2. This is really just a generalization of the Pythagorean theorem. Once you have distance, you can talk about rotations of higher dimensional space, mirror reflections, translations, and so on. This is all done in M52. So, this part of M52 gives you a foundation for working with higher dimensional Euclidean geometry. Linear Mappings of Euclidean Space: A function of one variable takes a number as input and spits out another number. The simplest kinds of functions have the form f(x) = ax. These functions, called linear functions, are characterized by the following two properties. f(x + y) = f(x) + f(y) f(bx) = bf(x). (On the right side, I mean b times f(x).) 3

They are known as linear functions. n-dimensional Euclidean space is denoted by R n. To generalize the notion of a function, we could say that a map from R n to R m is a gadget that takes a point in R n as input and spits out a point in R m. For instance f(x 1, x 2, x 3 ) = (x 2, 5x 1 + 7x 2 x 3 ) is a map from R 3 to R 2. In general, we write f : R n R m. Since it makes sense to add and scale points in R n, we call f linear if it satisfies the two properties mentioned above, namely f(x + y) = f(x) + f(y) and f(bx) = bf(x). Note that we have suddenly made a jump in abstraction. In our last equations, x and y represent points in R n. That is, x = (x 1,..., x n ) and y = (y 1,..., y n ) and x + y = (x 1 + y 1,..., x n + y n ) and bx = (bx 1,..., bx n ). Just as the single number a characterizes the linear function f(x) = ax, the linear maps are characterized by m n matrices. These are just like the arrays of numbers we discussed above. To facilitate the study of linear maps, one defines a way for matrices to act on vectors, as follows: x a 11 a 12 a 13 a 1 14 x a 11 x 1 + a 12 x 2 + a 13 x 3 + a 14 x 4 a 21 a 22 a 23 a 2 24 x a 31 a 32 a 33 a 3 = a 21 x 1 + a 22 x 2 + a 23 x 3 + a 24 x 4. 34 a x 31 x 1 + a 32 x 2 + a 33 x 3 + a 34 x 4 4 We took m = 3 and n = 4 just to make the notation easier, but the general case is essentially the same. Inspecting this last equation, we see that our 4 3 matrix sets up a map from R 4 to R 3. The input is (x 1, x 2, x 3, x 4 ) and the output is the right hand side of the equation. This map turns out to be linear, and all linear maps have this form. In calculus, you learned about composing functions, and perhaps about composing maps. (Remember the chain rule.) When linear maps are represented as matrices, the composition of linear maps corresponds to an operation called matrix multiplication. The example we gave above illustrates how the products of 3 4 matrix and a 4 1 matrix is a 3 1 matrix. In general, the product of an m n matrix and an n p matrix is an m p matrix. The complete correspondence between linear maps and matrices is worked out in M54. 4

Vector Spaces: A real vector space is a set V of elements called vectors. This set has two operations on it, addition and scaling by real numbers. The addition is denoted by (+) and the scaling is written without an intermediate symbol: rx means scale x by the real number r.. These rules obey the following axioms. (v 1 + v 2 ) + v 3 = v 1 + (v 2 + v 3 ) for all v 1, v 2, v 3 V. v 1 + v 2 = v 2 + v 1 for all v 1, v 2 V. There exists an element, called 0, such that 0+v = v +0 for all v V. For any v V there exists w V such that v + w = 0. r(sx) = (rs)x. r(v 1 + v 2 ) = rv 1 + rv 2. Here are some examples of vector spaces. The space R n is a vector space. When considered as a vector space, the points in R n are called vectors. The addition and scaling of vectors are the obvious ones. Let P n denote the space of real polynomials of degree n. Polynomials in P n are added and scaled in the obvious way, e.g. 3(1 + 2x + 7x 2 ) = 3 + 6x + 21x 2. The space of functions f : R R. Given f and g, the function h = f + g has the property that h(x) + f(x) + g(x). One point of introducing vector spaces is that a lot of basic properties of R n carry over to other examples. Sometimes the added abstraction clarifies what is going on. For instance, Fourier series, a subject from analysis, can be nicely formulated in terms of vector spaces. Bases, Dimension, Isomorphism: Once vector spaces are defined in M52, their basic properties are studied in a systematic way. I ll give you a feel for the sort of results you ll see. A subset of vectors v 1,..., v k V is called dependent if there are constants a 1,..., a k, not all zero, such that a 1 v 1 +... + a k v k = 0. 5

Otherwise, the set is called independent. A set of independent vectors is called a basis if it is maximal: The inclusion of any new vector creates a dependent set. One foundational result is this: If a vector space has a finite basis, then all of the bases have the same number of elements. This common number is called the dimension of the vector space. Linear maps between vector spaces are defined just as for R n. A linear map T : V W between vector spaces is called an isomorphism if T is one-to-one and onto. In this case, V and W are said to be isomorphic. Intuitively, this means that V and W are really the same vector space, up to a different convention for naming the elements. One standard result in M52 is that an n-dimensional vector space is isomorphic to R n. For instance, a basis for the vector space of quadratic polynomials is given by {1, x, x 2 }, and so P 2 is isomorphic to R 3. An isomorphism is given by T(ax 2 + bc + c) = (a, b, c). Our third example of a vector space is infinite dimensional. It is not isomorphic to R n for any n. Quadratic Forms and Inner Products: A quadratic form on a real vector space V is an operation Q which maps pairs of vectors to the reals. Q satisfies the following axioms. Q(v 1, v 2 ) = Q(v 2, v 1 ) for all v 1, v 2 V. Q(v 1 + v 2, v 3 ) = Q(v 1, v 3 ) + Q(v 2, v 3 ) for all v 1, v 2, v 3 V. Q(rv) = rq(v) for all r R and all v V. The dot product on R n is a classic example of a quadratic form. The formula is Q(x, y) = x 1 y 1 +... + x n y n. The dot product is a special case of an inner product. The quadratic form Q is an inner product if it is such that Q(v, v) > 0 for all nonzero v V. Given a pair (V, Q), where V is a vector space and Q is a quadratic form, one can consider O(V, Q), the set of linear isomorphisms from V to itself such that Q(T(v), T(w)) = Q(v, w) for all v, w V. Then V = R n and Q is the dot product, we are just talking about the set of rotations of R n that fix the origin. Other examples of Q and V (when Q is not necessarily an inner product) lead to very interesting geometrical objects, like hyperbolic geometry and Minkowski spacetime. In particular, the linear algebra in M52 is a nice framework for special relativity. 6

Eigenvalues and Eigenvectors: Let T : V V be a linear transformation from V to itself. An eigenvector of T is a vector v such that T(v) = λv, for some real λ. The number λ is called the eigenvalue associated to the eigenvector. In M52 you learn the significance of eigenvalues and eigenvectors, as well as methods for computing. An example will illustrate some of the significance of these concepts. Suppose that you have two kinds of cells, Type A and Type B. Each day, a Type A cell mutates to a Type B cell. At the same time, each Type B cell splits into a Type A cell and a Type B cell. Starting with, say, a single cell of Type A, one might wonder roughly how many cells one has after N days. One can think of the growth of these cells as a linear transformation on R 2. The point (x, y) represents a population of x cells of Type A and y cells of Type B. Then T(x, y) = (y, x + y), represents the population the next day. Let φ = 1 + 5 = 1.618... 2 be the famous golden ratio. It turns out that T has a basis of eigenvectors, given by v 1 = (φ, 1); v 2 = ( φ, 1), with corresponding eigenvalues λ 1 = φ; λ 2 = 1/φ. Since {v 1, v 2 } is a basis, and other vector will be a combination of these. For such a combination, of the form a 1 v 1 + a 2 v 2, we compute In general, we have T(a 1 v 1 + a 2 v 2 ) = a 1 T(v 1 ) + a 2 T(v 2 ) = a 1 λ 1 v 1 + a 2 λ 2 v 2. T n (a 1 v 2 + a 2 v + 2) = a 1 λ n 1v 1 + a 2 λ n 2v 2. 7

Note that λ 1 > 1 and λ 2 < 1. So, for large n, the second term is negligible. That is T n (a 1 v 1 + a 2 v 2 ) φ n (a 1 v 1 ). This tells us that we can expect an exponential growth of cells, with the growth rate being φ. That is, after many days have passed, the population of cells just about grows by a factor of φ each day. In general, studying the eigenvalues and eigenvectors of a linear transformation gives you a lot of information about what happens when this linear transformation is applied over and over again. Our example is a prototypical application of linear algebra to fields such as tumor growth or population genetics. 8