Usually, when we first formulate a problem in mathematics, we use the most familiar

Similar documents
MATH 115, SUMMER 2012 LECTURE 12

22A-2 SUMMER 2014 LECTURE Agenda

Dot Products, Transposes, and Orthogonal Projections

MAT1302F Mathematical Methods II Lecture 19

MAT 1302B Mathematical Methods II

Linear Algebra V = T = ( 4 3 ).

Eigenvalues & Eigenvectors

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Lecture 10: Powers of Matrices, Difference Equations

Roberto s Notes on Linear Algebra Chapter 10: Eigenvalues and diagonalization Section 3. Diagonal matrices

Linear Programming and its Extensions Prof. Prabha Shrama Department of Mathematics and Statistics Indian Institute of Technology, Kanpur

(, ) : R n R n R. 1. It is bilinear, meaning it s linear in each argument: that is

Linear Algebra & Geometry why is linear algebra useful in computer vision?

This last statement about dimension is only one part of a more fundamental fact.

15. LECTURE 15. I can calculate the dot product of two vectors and interpret its meaning. I can find the projection of one vector onto another one.

Matrices and Linear transformations

Linear Algebra, Summer 2011, pt. 2

MATH 310, REVIEW SHEET 2

Matrix-Vector Products and the Matrix Equation Ax = b

Image Registration Lecture 2: Vectors and Matrices

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Math (P)Review Part I:

Linear Algebra, Vectors and Matrices

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15.

Numerical Methods Lecture 2 Simultaneous Equations

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

MATH 12 CLASS 2 NOTES, SEP Contents. 2. Dot product: determining the angle between two vectors 2

Repeated Eigenvalues and Symmetric Matrices

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

Homework 3 Solutions Math 309, Fall 2015

Topic 15 Notes Jeremy Orloff

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

Basic Linear Algebra in MATLAB

Modern Algebra Prof. Manindra Agrawal Department of Computer Science and Engineering Indian Institute of Technology, Kanpur

Lecture 9: Elementary Matrices

Chapter 1: Useful definitions

Math 3191 Applied Linear Algebra

Linear Models Review

MATH Topics in Applied Mathematics Lecture 2-6: Isomorphism. Linear independence (revisited).

Section 4.6 Negative Exponents

Introduction to Vectors

Vector, Matrix, and Tensor Derivatives

Getting Started with Communications Engineering

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

Section 13.4 The Cross Product

Announcements Monday, November 13

Notes on multivariable calculus

Math 217: Eigenspaces and Characteristic Polynomials Professor Karen Smith

Vectors and matrices: matrices (Version 2) This is a very brief summary of my lecture notes.

Lecture 22. r i+1 = b Ax i+1 = b A(x i + α i r i ) =(b Ax i ) α i Ar i = r i α i Ar i

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Linear Algebra for Beginners Open Doors to Great Careers. Richard Han

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

Math 220 F11 Lecture Notes

Lecture 3: Linear Algebra Review, Part II

Lecture 15, 16: Diagonalization

Chapter 8. Rigid transformations

Matrices related to linear transformations

Examples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.

3 Fields, Elementary Matrices and Calculating Inverses

1 Matrices and Systems of Linear Equations. a 1n a 2n

Introduction to Matrices

Mechanics, Heat, Oscillations and Waves Prof. V. Balakrishnan Department of Physics Indian Institute of Technology, Madras

Lesson 3-2: Solving Linear Systems Algebraically

Name Solutions Linear Algebra; Test 3. Throughout the test simplify all answers except where stated otherwise.

Euclidean Space. This is a brief review of some basic concepts that I hope will already be familiar to you.

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Solving Systems of Equations Row Reduction

Topics in linear algebra

1 Review of the dot product

Math Lecture 3 Notes

Lecture 7. Econ August 18

Matrices and Deformation

Linear Algebra Exercises

Differential Equations

Cosets and Lagrange s theorem

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

(II.B) Basis and dimension

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear algebra and differential equations (Math 54): Lecture 10

M. Matrices and Linear Algebra

An Introduction To Linear Algebra. Kuttler

Spectral Theorem for Self-adjoint Linear Operators

Tutorial 2 - Learning about the Discrete Fourier Transform

Announcements Monday, November 13

Eigenvalues, Eigenvectors, and an Intro to PCA

22.2. Applications of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer. Lecture 5, January 27, 2006

Chapter 4. Solving Systems of Equations. Chapter 4

L2-1. Of space and subspace 10 Oct 2015

Chapter Two Elements of Linear Algebra

Eigenvalues, Eigenvectors, and an Intro to PCA

Chapter 1: Linear Equations

Eigenvalues, Eigenvectors, and an Intro to PCA

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

Elementary Linear Algebra

21 Symmetric and skew-symmetric matrices

Vectors. 1 Basic Definitions. Liming Pang

Transcription:

Change of basis Usually, when we first formulate a problem in mathematics, we use the most familiar coordinates. In R, this means using the Cartesian coordinates x, y, and z. In vector terms, this is equivalent to using what we ve called the standard basis in R ; that is, we write x y z = x + y + z = xe + ye + ze, where {e, e, e } is the standard basis. But, as you know, for any particular problem, there is often another coordinate system that simplifies the problem. For example, to study the motion of a planet around the sun, we put the sun at the origin, and use polar or spherical coordinates. This happens in linear algebra as well. Example: Let s look at a simple system of two first order linear differential equations x = x + x () x = x + x. Here, we are seeking functions x (t), and x (t) such that the equations hold simultaneously. Now there s no problem solving a single differential equation like x = x. In fact, we can see by inspection that x(t) = ce t is a solution for any scalar c. The difficulty with the system of two equations is that x and x are "coupled", and the

two equations must be solved simulataneously. There are a number of straightforward ways to solve the system () which you ll learn when you take a course in differential equations, and we won t worry about that here. But there s also a sneaky way to solve () by changing coordinates. We ll do this at the end of the lecture. First, we need to see what happens in general when we change the basis. For simplicity, we re just going to work in R ; generalization to higher dimensions is (really!) straightforward. Suppose we have a basis {e, e } for R. It doesn t have to be the standard basis. Then, by the definition of basis, any vector v R can be written as a linear combination of e and e. That is, there exist scalars c, c such that v = c e +c e. The numbers c and c are called the coordinates of v in the basis {e, e }. And v e = is called the coordinate vector of v in the basis {e, e }. c c Proposition: The coordinates of the vector v are unique. Proof: Suppose there are two sets of coordinates for v. That is, suppose that v = c e + c e, and v = d e + d e. Subtracting the two expressions for v gives = (c d )e + (c d )e. But {e, e } is linearly independent, so the coefficients in this expression must vanish: c d = c d =. That is, c = d and c = d, and the coordinates are unique,

as claimed. Example: Let {e, e } = and suppose v = We can find the coordinate vector v e in this basis in the usual way, by solving a system of linear equations. We are looking for numbers c and c such that In matrix form, this reads c + c Av e = v, = where A = and v = We solve for v e by multiplying both sides by A : v e = A v = (/7) = (/7) 9 = 9/7 /7 Notation In this section, we ll develope a compact shorthand notation for the above computation that is easy to remember. Start with an arbitrary basis {e, e } and an arbitrary vector

v. We know that v = c e + c e, where is the coordinate vector. c c = v e We see that the expression for v is a linear combination of two column vectors. And we know that such a thing can be obtained by writing down a certain matrix product: If we define the matrix E = (e.e ) then the expression for v can be simply written as v = E v e. Now suppose that {f, f } is another basis for R. Then the same vector v can also be written uniquely as a linear combination of these vectors. Of course it will have different coordinates, and a different coordinate vector v f. In matrix form, we ll have v = F v f. Exercise: Let {f, f } be given by If v = (same vector as above) find v f and verify that v = F v f = E v e. 4

Remark: This works just the same in R n, where e is n n, and v e is n. Continuing along with our examples, since E is a basis, the vectors f and f can each be written as linear combinations of e and e. So there exist scalars a, b, c, d such that f = = a + b f = = c + d We won t worry about the precise values of a, b, c, d, since you can easily solve for them. Define the change of basis matrix P by P = a b c d Note that this is the transpose of what you might think it should be; this is because we re doing column operations, and it s the first column of P which takes linear combinations of the columns of E and replaces the first column of E with the first column of F, and so on. In matrix form, we have F = E P and, of course, E = F P. Exercise: Find a, b, c, d and the change of basis matrix from E to F. Given the change of basis matrix, we can figure out everything else we need to know.

Suppose v has the known coordinates v e in the basis E, and F = E P. Then v = E v e = F P v e = F v f. Remember that the coordinate vector is unique. This means that v f = P v e. (If P changes the basis from E to F, then P changes the coordinates.) Compare this with the example at the end of the first section. Of course this has to be the case: the vector v exists as a geometric object. The coordinates and bases are just different ways of talking about the same thing. For any nonsingular matrix P, the following holds: v = E v e = E P P v e = G v g, where P is the change of basis matrix from E to G: G = E P, and P v e = v g are the coordinates of the vector v in this basis. This notation is consistent with the standard basis as well. e = and e = Since we have E = I, and v = I v Remark: When we change from the standard basis to the basis {e, e }, the corresponding matrices are I (for the standard basis) and E. So according to what s just been shown, the change of basis matrix will be the matrix P which satisfies E = I P. In other words, the change of basis matrix in this case is just the matrix E. 6

First example, cont d We can write the system of differential equations in matrix form as v = v = Av, () where the dot indicates d/dt. We change from the standard basis to F via the matrix F = Then, according to what we ve just worked out, we ll have v f = F v, and taking derivatives, v f = F v. So using v = F v f and substituting into (), we find F v f = AF v f, or v f = F AF v f. Now an easy computation shows that F AF = 4 and in the new coordinates, we have the system v f = 4v f v f = v f In the new coordinates, the system is now decoupled and easily solved to give v f = c e 4t v f = c e t, where c, c are arbitrary constants of integration. We can now transform back to the original (standard) basis to get the solution in the original coordinates: v = F v f = v v = c e 4t c e t = c e 4t + c e t c e 4t c e t 7

A reasonable question at this point is "How does one come up with this new basis F? It clearly was not chosen at random. The answer has to do with the eigenvalues and eigenvectors of the coefficient matrix of the differential equation, namely the matrix A = All of which brings us to the subject of the next lecture. 8