Inner product spaces. Layers of structure:

Similar documents
2. Review of Linear Algebra

Linear Algebra Massoud Malek

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Lecture Notes 1: Vector spaces

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms.

The following definition is fundamental.

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

2. Signal Space Concepts

Math Linear Algebra II. 1. Inner Products and Norms

Fourier and Wavelet Signal Processing

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

Projection Theorem 1

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Cambridge University Press The Mathematics of Signal Processing Steven B. Damelin and Willard Miller Excerpt More information

Geometric interpretation of signals: background

1. General Vector Spaces

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

INNER PRODUCT SPACE. Definition 1

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Vectors in Function Spaces

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

NOTES ON VECTORS, PLANES, AND LINES

Vector Geometry. Chapter 5

4.1 Distance and Length

Lecture 1: Systems of linear equations and their solutions

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear Algebra, Summer 2011, pt. 3

October 25, 2013 INNER PRODUCT SPACES

Mathematics 530. Practice Problems. n + 1 }

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Elements of Convex Optimization Theory

LINEAR ALGEBRA W W L CHEN

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

CHAPTER II HILBERT SPACES

Linear Algebra. Min Yan

Designing Information Devices and Systems I Fall 2018 Lecture Notes Note 21

Normed & Inner Product Vector Spaces

Linear Normed Spaces (cont.) Inner Product Spaces

ORTHOGONALITY AND LEAST-SQUARES [CHAP. 6]

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Vector Spaces. Commutativity of +: u + v = v + u, u, v, V ; Associativity of +: u + (v + w) = (u + v) + w, u, v, w V ;

Definitions and Properties of R N

Lecture 2: Linear Algebra Review

Part 1a: Inner product, Orthogonality, Vector/Matrix norm

The Hilbert Space of Random Variables

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Introduction to Geometry

Linear Algebra. Session 12

Elementary linear algebra

4.2. ORTHOGONALITY 161

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Applied Linear Algebra in Geoscience Using MATLAB

Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u u 2 2 = u 2. Geometric Meaning:

Grothendieck s Inequality

Class notes: Approximation

Vector Calculus. Lecture Notes

Inner Product Spaces

Hilbert spaces. 1. Cauchy-Schwarz-Bunyakowsky inequality

Numerical Linear Algebra

There are two things that are particularly nice about the first basis

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?

Lecture 23: 6.1 Inner Products

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Contents. 1 Vectors, Lines and Planes 1. 2 Gaussian Elimination Matrices Vector Spaces and Subspaces 124

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Linear Algebra. Preliminary Lecture Notes

Mathematics Department Stanford University Math 61CM/DM Inner products

SUMMARY OF MATH 1600

(, ) : R n R n R. 1. It is bilinear, meaning it s linear in each argument: that is

Introduction to Signal Spaces

. = V c = V [x]v (5.1) c 1. c k

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

Lecture 20: 6.1 Inner Products

Linear Algebra I for Science (NYC)

y 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Chapter 6: Orthogonality

Chapter Two Elements of Linear Algebra

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

MATH 423/ Note that the algebraic operations on the right hand side are vector subtraction and scalar multiplication.

Linear Algebra. Preliminary Lecture Notes

1 Mathematical preliminaries

Chapter 2. Vectors and Vector Spaces

Notions such as convergent sequence and Cauchy sequence make sense for any metric space. Convergent Sequences are Cauchy

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

,, rectilinear,, spherical,, cylindrical. (6.1)

Lecture # 3 Orthogonal Matrices and Matrix Norms. We repeat the definition an orthogonal set and orthornormal set.

Eigenvectors and Hermitian Operators

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

Functional Analysis Review

Spectral Theorem for Self-adjoint Linear Operators

The Transpose of a Vector

Transcription:

Inner product spaces Layers of structure: vector space normed linear space inner product space The abstract definition of an inner product, which we will see very shortly, is simple (and by itself is pretty boring). But it gives us just enough mathematical structure to make sense of many important and fundamental problems. Consider the following motivating example in the plane R 2. Let T be a one dimensional subspace (i.e. a line through the origin). Now suppose we are given another vector x. What is the closest point in T to x? 49

x x 0 x x 0 The salient feature of this point x is that x x v for all v T. So all we need to define this optimality property is the notion of orthogonality which follows immediately from defining an inner product. More on this later... Definition: An inner product on a (real- or complex-valued) vector space S is a mapping that obeys 1 1. x, y = y, x 2. For any a, b C, : S S C ax + by, z = a x, z + b y, z 1 We are using a to denote the complex conjugate of a scalar a, and x H to denote the conjugate transpose of a vector x. 50

3. x, x 0 and x, x = 0 x = 0 Standard Examples: 1. S = R N, x, y = N x n y n n=1 = y T x 2. S = C N, x, y = N x n y n n=1 = y H x 3. S = L 2 ([a, b]), x, y = b a x(t)y(t) dt Slightly less standard examples: 1. S = R M N (the set of M N matrices with real entries) X, Y = trace(y T X) = M m=1 N X m,n Y m,n n=1 (Recall that trace(x) is the sum of the entries on the diagonal of X.) This is called the trace inner product or Frobenius inner product or Hilbert-Schmidt inner product. 51

2. S = zero-mean Gaussian random variables with finite variance, X, Y = E[XY ] 3. S = differentiable real-valued continuous-time signals on R, x, y = x(t)y(t) dt + x (t)y (t) dt, where x (t) is the derivative of x(t). This is called a Sobolev inner product. 4. S = signals x(θ, φ) on the sphere in 3D Difference in average temperature, 2010 vs. 1971-2000 A natural (and valid) inner product is x, y = π θ=0 2π φ=0 x(θ, φ) y(θ, φ) sin θ dφ dθ If we think of θ as being latitude and φ as longitude, the sin θ can be interpreted as a weight for the size of the circle of equal latitude (these get smaller as you go towards the poles). A linear vector space equipped with an inner product is called an inner product space. 52

Induced norms A valid inner product induces a valid norm by x = x, x (Check this on your own as an exercise.) It is not hard to see that in R N /C N, the standard inner product induces the l 2 norm. Properties of induced norms In addition to the triangle inequality, x + y x + y, induced norms obey some very handy inequalities (note that these are not necessarily true for norms in general, only for norms induced by an inner product): 1. Cauchy-Schwarz Inequality x, y x y Equality is achieved above when (and only when) x and y are colinear: a C such that y = ax. 2. Pythagorean Theorem x, y = 0 x + y 2 = x 2 + y 2 53

The left-hand side above also implies that x y 2 = x 2 + y 2. 3. Parallelogram Law x + y 2 + x y 2 = 2 x 2 + 2 y 2. You can prove this by expanding x + y 2 = x + y, x + y and similarly for x y 2. 4. Polarization Identity Angles between vectors Re { x, y } = x + y 2 x y 2 In R 2 (and R 3 ), we are very familiar with the geometrical notion of an angle between two vectors. x 4 y We have x, y = x y cos θ Notice that this relationship depends only on norms and inner products. Therefore, we can extend the definition to any inner product space. 54

Definition: The angle between two vectors x and y in an inner product space is x, y cos θ = x y, where the norm is the one induced by the inner product. Definition: Vectors x and y in an inner product space are orthogonal to one another if x, y = 0. 55

Example: (Weighted inner product) S = R 2, x, y Q = y T Qx, where Q = [ ] 4 0 0 1 so x, y = 4x 1 y 1 + x 2 y 2. What is the norm induced by this inner product? Draw the unit ball B Q = {x R 2 : x Q = 1}. Find a vector which is orthogonal to x = [ 1 1] under, Q. 56

Linear approximation in a Hilbert space Consider the following problem: Let S be a Hilbert space, and let T be a subspace of S. Given an x S, what is the closest point ˆx T? x ˆx T In other words, find ˆx T that minimizes x ˆx ; given x, we want to solve the following optimization program minimize y T x y, (1) where the norm above is the one induced by the inner product. This problem has a unique solution which is characterized by the orthogonality principle. Theorem: Let S be a Hilbert space, and let T be a finite dimensional subspace 2. Given an arbitrary x S, 1. there is exactly one ˆx T such that x ˆx T, (2) meaning x ˆx, y = 0 for all y T, and 2 The same results hold when T is infinite dimensional and is closed. We do not prove the infinite dimensional case just because it requires some analysis of infinite sequences which, while not really that difficult, kind of distract from the overall geometrical picture we are trying to paint here. 57

2. this ˆx is the closest point in T to x; that is, ˆx is the unique minimizer to (1). Proof: We will show that the first part is true in the next section of the notes, where we show how to explicitly calculate such an ˆx. For the second part, let ˆx be the vector which obeys ê = x ˆx T. Let y be any other vector in T, and set We will show that e = x y. e > ê (i.e. that x y > x ˆx ). Note that e 2 = x y 2 = ê (y ˆx) 2 = ê (y ˆx), ê (y ˆx) = ê 2 + y ˆx 2 ê, y ˆx y ˆx, ê. Since y ˆx T and ê T, and so ê, y ˆx = 0, and y ˆx, ê = 0, e 2 = ê 2 + y ˆx 2. Since all three quantities in the expression above are positive and we see that y ˆx > 0 y ˆx, y ˆx e > ê. 58

ˆx ê y x e T Computing the best approximation The orthogonality principle gives us a concrete procedure for actually computing the optimal point ˆx. Let N be the dimension of T, and let v 1,..., v N be a basis for T. We want to find coefficients a 1,..., a N C such that ˆx = a 1 v 1 + a 2 v 2 + + a N v N. The orthogonality principle tells us that This means the a n must obey x x ˆx, v n = 0 for n = 1,..., N. N a k v k, v n = 0 for n = 1,..., N, k=1 or moving things around, x, v n = N a k v k, v n for n = 1,..., N. k=1 59

Since x and the {v n } are given, we know both the x, v n and the v k, v n. We are left with a set of N linear equations with N unknowns: v 1, v 1 v 2, v 1 v N, v 1 v 1, v 2 v 2, v 2 v N, v 2..... v 1, v N v N, v N a 1 a 2. a N = x, v 1 x, v 2. x, v N The matrix on the left hand side above is called the Gram matrix or Grammian of the basis {v n }. In more compact notation, we want to find a C N such that Ga = b, where b n = x, v n and G k,n = v n, v k. Two notes on the structure of G: G is guaranteed to be invertible because the {v n } are linearly independent. We can comfortably write a = G 1 b. G is conjugate symmetric ( Hermitian ): G = G H, where G H is the conjugate transpose of G (take the transpose, then take the complex conjugate of all the entries). This fact has algorithmic implications when it comes time to actually solve the system of equations. 60

Uniqueness It should be clear that if e, v k = 0 for all of the basis vectors v 1,..., v N, then e, y = 0 for all y T. The converse is also true: if e, y 0 for some y T not equal to 0, then e, v k 0 for at least one of the v k. With the work above, this means that a necessary and sufficient condition for x ˆx, y = 0 for all y T is to have ˆx = N a n v n, where a satisfies Ga = b. n=1 Since G is square and invertible, there is exactly one such a, and hence exactly one ˆx that obeys the condition x ˆx T. 61

Example: Let 1 1 T = Span 1, 1, x = 1 1 2 1 3 Find the solution to minimize y T x y 2. (Recall that 2 in R 3 is induced by the standard inner product x, y = 3 n=1 x n y n.) Here is a plot of T and x: 62

Solution: We have G = b = and so G 1 = This means that a = from which we synthesize the answer ˆx = We can also check that the error is orthogonal to the approximation x ˆx, ˆx = 63

Technical Details: Completeness, convergence, and Hilbert Spaces As we can see from the properties above, inner product spaces have almost all of the geometrical properties of the familiar Euclidean space R 3 (or more generally R N ). In fact, in the coming sections, we will see that any space with an inner product defined (which comes with its induced norm) is directly analogous to Euclidean space. To make all of this work nicely in infinite dimensions, we need a technical condition on S called completeness. Roughly, this means that there are no points missing from the space. More precisely, if we have an infinite sequence of vectors in S that get closer and closer to one another, then they converge to something in S. Even more precisely, we call an inner product space complete if every Cauchy sequence is a convergent sequence; that is, for every sequence x 1, x 2,... S for which lim x m x n = 0, will also have lim x n = x S, min(m,n) n where is the norm induced by the inner product. An example of a space which is not complete is continuous, bounded functions on [0, 1]. It is easy to come up with a sequence of functions which are all continuous but converge to a discontinuous function. All finite dimensional inner product spaces are complete, as is L 2 ([a, b]) and L 2 (R) (and in fact, every example of an inner product space we have given above). The former is a basic result in mathematical analysis, the latter is a result from the modern theory of Lebesgue integration. Determining whether or not a space is complete is far 64

outside the scope of this course; it is enough for us to know what this concept means. An inner product space which is also complete is called a Hilbert space. As all of the inner product spaces we will encounter in this course are complete, we will use this descriptor from now on. The Wikipedia pages on these topics are actually pretty good: http://en.wikipedia.org/wiki/hilbert_space http://en.wikipedia.org/wiki/complete_metric_space The point of asking that the space be complete is that it gives us confidence in writing expressions like x(t) = α n ψ n (t). n=1 What is on the left is a sum of an infinite number of terms; the equality above means that as we include more and more terms in this sum, it converges to something which we call x(t). There are different ways we might define convergence, depending on how much of a role we want the order of terms to play in the result. But we say that n=1 α n ψ n (t), where the ψ n (t) are in a Hilbert space S, is convergent if there is an x(t) such that N x(t) α n ψ n (t) 0 as N. n=1 If S is complete, we know that x(t) will also be in S. 65

Technical Details: Proofs of properties of induced norms 1. Cauchy-Schwarz Set z = x and notice that z, y = 0, since z, y = x, y x, y y 2 y, We can write x in terms of y and z as and since y z, Thus x = x, y y 2 y 2 = 0. x, y y 2 y + z, x 2 x, y 2 = y 2 + z 2 y 4 x, y 2 = + z 2. y 2 x, y 2 = x 2 y 2 z 2 y 2 x 2 y 2. We have equality above if and only if z = 0. If z = 0, then x is co-linear with y, as x = αy, with α = x, y y 2. Conversely, if x = αy for some α C, then z = αy α y, y y 2 y = 0. 66

2. Pythagorean Theorem x + y 2 = x + y, x + y = x, x + x, y + y, x + y, y = x 2 + x, y + x, y + y 2 = x 2 + y 2 (since x, y = 0) 3. Parallelogram Law. As above, we have and so x + y 2 = x 2 + x, y + x, y + y 2 = x 2 + 2 Re { x, y } + y 2, x y 2 = x 2 x, y x, y + y 2 = x 2 2 Re { x, y } + y 2, x + y 2 + x y 2 = 2 x 2 + 2 y 2. 4. Polarization Identity. Using the expansions above, we quickly see that x + y 2 x y 2 = 4 Re { x, y }. 67