15B. Isometries and Functionals 1

Similar documents
12. Special Transformations 1

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

MATH Linear Algebra

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Kevin James. MTHSC 206 Section 12.5 Equations of Lines and Planes

Inner Product Spaces

12.5 Equations of Lines and Planes

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

Linear Algebra Practice Problems

4 Linear operators and linear functionals

6. Linear Transformations.

Outline. Linear maps. 1 Vector addition is commutative: summands can be swapped: 2 addition is associative: grouping of summands is irrelevant:

V (v i + W i ) (v i + W i ) is path-connected and hence is connected.

LINEAR ALGEBRA W W L CHEN

CHAPTER VIII HILBERT SPACES

Linear Algebra. Workbook

and The important theorem which connects these various spaces with each other is the following: (with the notation above)

SECTION v 2 x + v 2 y, (5.1)

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

4.3 Equations in 3-space

Pullbacks, Isometries & Conformal Maps

October 25, 2013 INNER PRODUCT SPACES

Exercise Sheet 1.

Math Exam 2, October 14, 2008

Review of linear algebra

Isometries, Local Isometries, Riemannian Coverings and Submersions, Killing Vector Fields

Linear Algebra. Session 12

Properties of Transformations

Introduction to Real Analysis Alternative Chapter 1

Math Real Analysis II

VECTORS AND THE GEOMETRY OF SPACE

Chapter 3. Vector spaces

Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space

Section 6.2: THE KERNEL AND RANGE OF A LINEAR TRANSFORMATIONS

Linear Algebra (Math-324) Lecture Notes

1. General Vector Spaces

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

There are two things that are particularly nice about the first basis

Chapter 6 Inner product spaces

2 Garrett: `A Good Spectral Theorem' 1. von Neumann algebras, density theorem The commutant of a subring S of a ring R is S 0 = fr 2 R : rs = sr; 8s 2

LINES IN P 3. Date: December 12,

Spring 2014 Math 272 Final Exam Review Sheet

Infinite-Dimensional Triangularization

FUNCTIONAL ANALYSIS LECTURE NOTES: COMPACT SETS AND FINITE-DIMENSIONAL SPACES. 1. Compact Sets

Elementary linear algebra

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra

B 1 = {B(x, r) x = (x 1, x 2 ) H, 0 < r < x 2 }. (a) Show that B = B 1 B 2 is a basis for a topology on X.

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

MAT 445/ INTRODUCTION TO REPRESENTATION THEORY

Categories and Quantum Informatics: Hilbert spaces

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Rings. EE 387, Notes 7, Handout #10

is an isomorphism, and V = U W. Proof. Let u 1,..., u m be a basis of U, and add linearly independent

Commutative Banach algebras 79

Recall that any inner product space V has an associated norm defined by

Linear Algebra I. Ronald van Luijk, 2015

Chapter 2: Linear Independence and Bases

Rings. Chapter 1. Definition 1.2. A commutative ring R is a ring in which multiplication is commutative. That is, ab = ba for all a, b R.

4.6 Bases and Dimension

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Linear Algebra. Min Yan

Math 3108: Linear Algebra

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Stereographic projection and inverse geometry

The Gram-Schmidt Process 1

Abstract Vector Spaces and Concrete Examples

Definitions for Quizzes

Math Linear Algebra II. 1. Inner Products and Norms

1. Subspaces A subset M of Hilbert space H is a subspace of it is closed under the operation of forming linear combinations;i.e.,

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

y 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.

M.6. Rational canonical form

Math 4310 Solutions to homework 7 Due 10/27/16

Exam questions with full solutions

Lecture 23: 6.1 Inner Products

REAL LINEAR ALGEBRA: PROBLEMS WITH SOLUTIONS

MATH PROBLEM SET 6

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

MATH 423 Linear Algebra II Lecture 12: Review for Test 1.

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Last week we presented the following expression for the angles between two vectors u and v in R n ( ) θ = cos 1 u v

LECTURE 5, FRIDAY

I teach myself... Hilbert spaces

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

CHAPTER 5. Birkhoff Orthogonality and a Revisit to the Characterization of Best Approximations. 5.1 Introduction

Math 702 Problem Set 10 Solutions

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Chapter 2 Linear Transformations

MATH 112 QUADRATIC AND BILINEAR FORMS NOVEMBER 24, Bilinear forms

Equivalence Relations

Math 110, Spring 2015: Midterm Solutions

Math 255 Honors: Gram-Schmidt Orthogonalization on the Space of Polynomials

FUNCTIONAL ANALYSIS LECTURE NOTES: ADJOINTS IN HILBERT SPACES

2.3. VECTOR SPACES 25

Recall the convention that, for us, all vectors are column vectors.

ADJOINTS, ABSOLUTE VALUES AND POLAR DECOMPOSITIONS

2.2 Annihilators, complemented subspaces

Transcription:

5B. Isometries and Functionals Isometries If the important functions between vector spaces V and W are those that preserve linearity (i.e., linear transformations), then the important functions between inner product spaces should be those that preserve the inner product. To this end, we make the following natural definition: a linear transformation T:V W between inner product spaces is isometric if it satisfies the condition T(X),T(Y) = X, Y for every X and Y in V. (Note that the inner product on the left is the one in W while the one on the right is the one in V. Also, the term isometric refers to having the same distance : an isometric map preserves lengths of vectors as it passes from domain to codomain space. Of course, it also preserves angles.) Proposition An isometric linear transformation has trivial kernel. Proof Suppose T(X) = 0. Then X 2 = X, X = T(X),T(X) = T(X) 2 = 0 2 = 0, so X = 0. //

5B. Isometries and Functionals 2 This last proposition shows that all isometric maps are one-to-one. (Smith calls such maps isometric embeddings.) So if the linear transformation T:V W is isometric and V and W have the same dimension, the map must be an isomorphism as well; we call an isometric isomorphism simply an isometry, and we say that the spaces V and W are isometric. Proposition Any two inner product spaces of the same dimension are isometric. Proof Suppose inner product spaces V and W are both of dimension n. Then we can find orthonormal bases for each, say {V,V 2,,V n } for V and {W, W 2,,W n } for W. Let T:V W be the linear transformation that is the extension of the assignments T(V i ) = W i. Then T is isometric, for any X and Y in V can be written the form so that X = a V + a 2 V 2 +L+ a n V n Y = b V +b 2 V 2 +L+b n V n

5B. Isometries and Functionals 3 T(X),T(Y) = T( a i V i ),T( b j V j ) i n j n = a i T(V i ), b j T(V j ) i n j n = a i b j T(V i ),T(V j ) i, j n = a i b j V i,v j i, j n = a i V i, b j V j i n j n = X,Y. // Corollary Every n-dimensional inner product space is isomorphic to R n (endowed with the standard dot product). //

5B. Isometries and Functionals 4 Functionals Recall from Exercise 9.9(3) that any finitedimensional vector space V is isomorphic to its dual space V* = L(V,R). We will investigate next how having a scalar product helps to make an important connection between the inner product space V and its dual V*. The vectors in V* are real-valued linear functions on V. They are often called (linear) functionals on V; we will denote them with letters like f, g, etc., to emphasize their nature as functions on V. One example of a functional makes use of the inner product on V. Let A be a fixed vector in V and define for arbitrary vectors X in V, f(x) = X, A ; then f is a functional (why is it linear?). We will see that every functional on V can be represented this way. For instance, consider the functional f:r 3 R given by f (x, y,z) = 3x 2y + z; we represent it in the above form by writing f (x, y,z) = (x, y,z) (3, 2,), i.e., f corresponds to the dot product with the fixed vector A = (3, 2,).

5B. Isometries and Functionals 5 Observe that we have a geometric interpretation for this as well: the kernel of the functional f corresponds to the plane 3x 2y + z = 0 in R 3, and A is a normal vector to this plane (every vector in the kernel of f is orthogonal to A). We now propose to work out the details of how this situation generalizes in abstract inner product spaces. If S is any set of vectors in an inner product space V, we define the orthogonal complement of S to be the set S of vectors in V orthogonal to all vectors in S: S = {X V X, Y = 0 for all Y S }. Proposition If S is any set of vectors in an inner product space V, then S is a subspace of V. Further, only 0 can lie in both S and S. Proof If X and X lie in S, then X + X,Y = X, Y + X, Y = 0 and rx,y = r X,Y = 0 for all Y S. So S is a subspace of V. Also, if X lies in both S and S, then X, X = 0, so X = 0, whence X = 0. //

5B. Isometries and Functionals 6 Corollary Let W be a subspace of the inner product space V. Then V = W W. Proof By the proposition, W is also subspace of V and their intersection is trivial. To prove the result, we need only show that V = W + W. So let {W, W 2,,W k } be an orthonormal basis for W and suppose that V is any vector in V. Let X = V, W W +L+ V,W k W k and set Y = V X. Then necessarily V = X + Y with X in W. But since any W vector in W has the form W = a W +L+ a k W k for suitable scalars a,,a k, we can compute Y, W = Y,a W +L+ a k W k and since = a Y, W +L+ a k Y, W k = a [ V, W X,W ] +L + a k [ V,W k X,W k ]

5B. Isometries and Functionals 7 X, W i = V, W W +L+ V, W k W k,w i = V,W W, W i +L+ V, W k W k,w i = V,W i W i, W i = V,W i it follows that Y, W = 0 for every W in W. That is, Y lies in W. So V = W + W, completing the proof. // Corollary Let W be a k-dimensional subspace of the n-dimensional inner product space V. Then W has dimension n k. // The results we have just worked out provide the technical details that allow us to prove the central theorem of this discussion:

5B. Isometries and Functionals 8 Theorem [The Riesz Representation Theorem] Let V be an inner product space. Then for every functional f V*, there exists a corresponding unique vector A V with the property that f(x) = X, A. That is, every functional can be represented as an inner product against some vector in V. Proof If f is the zero map, then we can take A = 0 and we re done. So from here on, we may assume that f is not always zero on V. Since Im(f) is not trivial and is also a subspace of R, it follows that Im(f) = R. Thus, dim(im(f)) =, so by the Dimension Theorem, dim(ker(f)) = n where dimv = n. Let {V,V 2,,V n } be an orthonormal basis for ker(f). Since the subspace ker(f) must satisfy V = ker(f) ker(f), it follows that ker(f) is onedimensional. If V 0 is a unit vector in ker(f), then {V 0, V, V 2,,V n } must be an orthonormal basis for all of V.

5B. Isometries and Functionals 9 So if X is any vector in V, we can write X = X,V 0 V 0 + X, V V +L+ X,V n V n so that f (X) = X,V 0 f (V 0 ) + X,V f (V ) +L+ X,V n f (V n ) = X,V 0 f (V 0 ) = X,f (V 0 )V 0 It follows that the vector A = f (V 0 )V 0 satisfies the condition desired: f(x) = X, A. The choice of A is unique, for if B is another vector with the property that f(x) = X,B, then X, A = X,B for all X in V, so in particular, A B, A = A B,B A B, A B = 0 A B 2 = 0 A B = 0 A = B //

5B. Isometries and Functionals 0 The Riesz Representation Theorem, applied to the inner product space R 3 (under the standard dot product), guarantees that for every linear map (functional) f:r 3 R there is an associated vector A = (a,b,c) so that f has the form f (x, y,z) = (x, y,z) (a,b,c) = ax +by +cz. In fact, as the proof of the theorem indicates, A determines a -dimensional subspace of R 3 orthogonal to the 2-dimensional kernel of f, namely the plane with equation (x, y,z) (a,b,c) = ax +by + cz = 0; that is, A is a vector normal to the plane. While none of this is new information, we can produce some new results by looking at how the theorem applies in other inner product spaces. Consider the inner product space P k (R) (with the integral metric p(x),q(x) = p(x)q(x) dx). In this setting, the Riesz Representation Theorem guarantees that to any functional, like the evaluation-at-a map f a (p(x)) = p(a), there corresponds a fixed polynomial q a (x) so that

5B. Isometries and Functionals (*) p(a) = f a (p(x)) = p(x),q a (x) = p(x)q a (x) dx. That is, the polynomial q a (x) has the remarkable property that computation of the integral p(x)q a (x) dx is identical to evaluating the polynomial p(x) at x = a, and this is true simultaneously for all polynomials p(x) of degree up to k. For instance, with k = 2 and a = 0, we want a polynomial q 0 (x) = a 0 + a x + a 2 x 2 satisfying the condition (*) above. Following the procedure outlined in the proof of the theorem, we first find a polynomial r(x) which is orthogonal to the kernel of the functional. The p(x) in this kernel satisfy f 0 (p(x)) = p(0) = 0, so p(x) must have the form b x +b 2 x 2. So r(x) satisfies 0 = b x +b 2 x 2,r(x) = b x r(x) dx +b 2 x 2 r(x) dx for every choice of b,b 2. Thus, both integrals above must vanish: if we put r(x) = c 0 + c x + c 2 x 2, then 0 = x r(x) dx = c 0 xdx +c x 2 dx + c 2 x 3 dx 0 = x 2 r(x) dx = c 0 x 2 dx +c x 3 dx + c 2 x 4 dx

5B. Isometries and Functionals 2 becomes 0 = c 0 0 + c 2 +c 3 2 0 0 = c 0 2 + c 3 0 +c 2 2 5 whence c = 0 and c 2 = 5 3 c 0. So r(x) = c( 5 3 x2 ) for any nonzero choice of c. The procedure requires that we normalize r(x), so we compute its length: r(x),r(x) = c 2 ( 5 3 x2 ) 2 dx = c 2 ( 0 = c 2 (2 20 9 + 0 9 ) = 8 9 c2 3 x2 + 25 9 x4 ) dx To make this length equal, we force c 2 = 9 8. Finally, according to the procedure, we put q 0 (x) = f 0 (r(x)) r(x) = r(0) r(x) = c c( 5 3 x2 ) = 9 8 5 8 x2 It follows that every polynomial p(x) in P 2 (R) satisfies the formula p(0) = p(x)( 9 5 8 8 x2 ) dx.