Section.8 Matrices as Linear Transformations Up to this point in the course, we have thought of matrices as stand alone constructions, objects of interest in their own right. We have learned multiple matrix operations and their properties, and have seen several special types of matrices. In this section, we will readjust our focus: we re going to discuss a geometric interpretation of matrices, and begin to understand matrices as objects which can be used to alter the shapes of vectors in Euclidean space. For the rest of this section, and indeed throughout much of the course, we will need to keep in mind a dual interpretation of notation such as. Of course, we know that this object is a matrix; however, it has a geometric interpretation, as a vector in dimensional Euclidean space. We build this correspondence by thinking of the matrix ( ) as the point (,, ) in three-dimensional space; indeed, we can go farther, and identify ( ) with the vector whose tip is the point (,, ): It is an intuitive fact that every matrix corresponds in a similar way to another vector in dimensional space. Of course, this concept works for other column vectors as well; keep it in mind as we progress through this section: Key Point. The column matrix c c. c n
Section.8 can be identified uniquely with the vector in n dimensional space whose tip is at the point (c, c,..., c n ). In other words, choosing a particular n column matrix is the same as choosing a vector in n dimensional space, and we will often emphasize this fact by referring to a column matrix as a column vector or simply a vector. Throughout the remainder of this section, we will use notation such as R n to indicate n dimensional Euclidean space; so R refers to the plane, R to three dimensional space, etc. Linear Transformations You have spent much of our time in previous mathematics classes on the study of functions. In particular, the majority of your time in calculus was spent studying the properties of real-valued functions functions whose output is a real number. In this section we are going to begin to study vector-valued functions functions whose output is a vector. For example, consider the function below: f( x x x ) x x x + x + x This function f has a three dimensional vector as its input, and a three dimensional vector as its output. Let s evaluate f at a few different vectors to try to get a feel for what s going on here: f( ) + +. 8. So Next, let s calculate f( ) f( ). 8 + + 4 9.
Section.8 Clearly, f is just a function; it sends one three dimensional vector to another one. Vector-valued functions can be quite difficult to work with, but those that have nice properties (and are thus easier to work with) have a special name: Definition. Let u and v be vectors in R n, and let k be a real number. A linear transformation f from R n R m is any function from n dimensional Euclidean space to m dimensional Euclidean space satisfying the following properties: () f(u + v) f(u) + f(v) () f(ku) kf(u). In other words, a linear transformation is (first of all) a function that turns n dimensional vectors into m dimensional vectors, and second, behaves nicely (see () and () above). If f is a linear transformation from R n R m, we often write f : R n R m to specifically indicate the domain and codomain of f. A Geometric View of Matrices As promised at the beginning of this section, we are now going to switch to a discussion of a geometric interpretation of matrices. To understand this new interpretation, let s start with a simple example. Let A, and consider the column vectors v, v, and v. Since v, v, and v are, they can be interpreted as vectors in three dimensional space, and are accordingly graphed below in red, blue, and green respectively:
Section.8 Now since the matrix A is, we know that we can multiply each of our vectors by A. In fact, each of the matrices Av, Av, and Av are matrices column vectors, all of which can be interpreted again as vectors in three dimensional space! Let s calculate each of Av, Av, and Av, and compare the graphs of these new vectors to the graphs of the old ones v, v, and v. We know that Av + + + + + + 8. 4
Section.8 Similarly, Av + + + + + +, and Av 4 + + + 9 + + 9 + 4 9. Now that we have calculated the vectors Av, Av, and Av, we should compare their graphs with the graphs of the original v, v, and v. All six vectors are plotted below; the new vectors are plotted with dashed lines, in colors corresponding to the colors of their original counterpart: 5
Section.8 Upon close inspection, the graphic above reveals some interesting information: it appears that applying A to the original vectors stretched them up and shifted them back a bit. Indeed, we could apply A to any three dimensional vector, or to the entire space; in a sense, we can think of A as a function that can be used to transform the shape of three dimensional space. Let s start with the same set of vectors v, v, and v, graphed below:
Section.8 Now let s multiply each of the vectors v, v, and v by the matrix B ( Of course, each of Bv, Bv, and Bv is a matrix, and thus a two dimensional vector. You should check for yourself that ( ) ( ) ( ) 5 Bv, Bv, and Bv. The three new vectors are graphed below on the (two dimensional) xy plane: ). 7
Section.8 Of course, we can still include them in a graph of the original vectors, as vectors lying on the xy plane (recall that v and Bv are both graphed in red, etc.). In this case, multiplying the vectors by B, in a sense, transformed them from three dimensional vectors to two dimensional ones. Let s look at one more example, using the same two dimensional vectors from the previous example; we ll rename them ( ) ( ) ( ) 5 w, w, and w ; 8
Section.8 the vectors are graphed below in red, blue, and green respectively: In this example, let s multiply each of the vectors above by the matrix C. Since C is, the resulting vectors will be. You should verify that 7 9 Cw, Cw, and Cw. 5 The resulting three dimensional vectors are graphed below, along with the original w, w, and w. It is clear that the matrix R has transformed our two dimensional vectors into three dimensional ones: 9
Section.8 Matrices and Linear Transformations The next theorem shows how the idea of a linear transformation is related to matrices: Theorem.8.. () Every m n matrix is a linear transformation from R n to R m. () If f is a linear transformation from R n to R m, then f has a representation as an m n matrix. Essentially, the theorem above says that linear transformations and matrices are two ways of thinking about exactly the same thing. We have actually already seen an example of this phenomenon: at the beginning of this section, we discussed the linear transformation f( x ) x x x. x + x + x Later, we studied the action of the matrix A x on three dimensional vectors. If you were paying close attention, you may have noticed that f(v ) Av and f(v ) Av. It turns out that A is actually the matrix form of the linear transformation given by f (you should check this claim yourself). Remark. We have now seen two different interpretations of matrices as objects of study in their own right, and as functions on Euclidean space. There are occasions on which we wish to emphasize one interpretation over the other; accordingly, we will introduce a bit of new notation to help us distinguish between the interpretations. If we wish to emphasize that an m n matrix A is a linear transformation (or function) from R n to R m, we will write T A (x) to indicate the product Ax, that is T A (x) Ax. The notation T A (x) simply indicates that we are thinking of A as a function acting on Euclidean space. When we wish to think of our matrices simply as matrices, we will continue to use matrix notation such as Ax to indicate the product of a matrix A and vector x.
Section.8 Let s reexamine the examples we saw in the beginning of this section. In the first, we looked at the matrix A. We can now think of A as a linear transformation from R to R, and write Similarly, the matrix T A : R R. B ( is a linear transformation from R to R, and C is a linear transformation from R to R. ) Standard Bases To help us better understand the way in which matrices act as linear transformations, we need to introduce the concept of standard basis vectors. You are already familiar with some of the standard basis vectors; for example, in R (the plane), the standard basis vectors are ( ) ( ) e and e. Similarly, in R, the standard basis vectors are e, e, and e n. Key Point. Standard basis vectors are important because any vector in R n is just a linear combination of the standard basis vectors for R n ; in a sense, you can use standard basis vectors to build any vector you want, indeed to build the entire space. You may have guessed the general definition for standard basis vectors, given below: Definition. N dimensional Euclidean space R n has n standard basis vectors, denoted by e., e., and e n..
Section.8 We will see that standard basis vectors have many uses; the first we will learn is discussed below. Finding the Matrix of a Linear Transformation Earlier, we saw that matrices and linear transformations are exactly the same objects. On one hand, every matrix can be thought of as a function which has all of the lovely properties of a linear transformation. On the other hand, the action T (x) of a linear transformation T on a vector x can always be described by matrix multiplication, T (x) Ax for some matrix A. This observation leads to a natural question: given a linear transformation T, how do we find the associated matrix A? The answer to the question turn out to be quite simple: Theorem. If T is a linear transformation from R n R m, then its matrix representation is the m n matrix given by the formula A T (e ) T (e )... T (e n ), where e, e, etc. are the standard basis vectors of the domain of T, R n. The matrix notation above indicates that the first column of A is T (e ), the second is T (e ), etc. In other words, to understand the matrix of a linear transformation, we simply need to understand how the transformation treats the standard basis vectors. Example Find the matrix A of the linear transformation ( ) x x x T ( ) x x. x + x As indicated by the theorem, we simply need to determine where the transformation T sends the standard basis vectors; this will give us the columns of the matrix A. Before we make any calculations, however, let s make sure that we understand what we re looking for. T is a linear transformation from R R, so the correct matrix A should be. Since the domain of T is R, we need to determine where T sends each of the standard basis vectors e ( ) and e ( ).
Section.8 Let s make the calculations: since T ( ( x x ) x x ) x x + x, we know that T (e ) ( ) T ( ) +. Similarly, T (e ) ( ) T ( ) +. Now T (e ) and T (e ) form the columns of the matrix form A of T, so we know that A. You should check for yourself that T and A treat vectors in the same way that is, that for any two dimensional vector x. T (x) Ax