which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

Similar documents
MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Section 7.5 Inner Product Spaces

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MTH 309Y 37. Inner product spaces. = a 1 b 1 + a 2 b a n b n

Linear Algebra Massoud Malek

There are two things that are particularly nice about the first basis

Recall that any inner product space V has an associated norm defined by

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Chapter 3 Transformations

complex dot product x, y =

(v, w) = arccos( < v, w >

(v, w) = arccos( < v, w >

Applied Linear Algebra in Geoscience Using MATLAB

(v, w) = arccos( < v, w >

Algebra II. Paulius Drungilas and Jonas Jankauskas

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Diagonalizing Matrices

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 6: Orthogonality

MTH 2032 SemesterII

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

,, rectilinear,, spherical,, cylindrical. (6.1)

Math 24 Spring 2012 Sample Homework Solutions Week 8

INNER PRODUCT SPACE. Definition 1

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Lecture 1: Systems of linear equations and their solutions

PART II : Least-Squares Approximation

Vectors. Vectors and the scalar multiplication and vector addition operations:

Upon successful completion of MATH 220, the student will be able to:

MATH Linear Algebra

LINEAR ALGEBRA QUESTION BANK

The Gram Schmidt Process

The Gram Schmidt Process

Exercises * on Linear Algebra

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Lecture 10: Vector Algebra: Orthogonal Basis

2. Review of Linear Algebra

Math Linear Algebra II. 1. Inner Products and Norms

October 25, 2013 INNER PRODUCT SPACES

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Vectors in Function Spaces

MTH 2310, FALL Introduction

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math Linear Algebra

Introduction to Signal Spaces

Vectors Coordinate frames 2D implicit curves 2D parametric curves. Graphics 2008/2009, period 1. Lecture 2: vectors, curves, and surfaces

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Solutions to Review Problems for Chapter 6 ( ), 7.1

Inner Product Spaces

Math Real Analysis II

Linear Algebra II. 7 Inner product spaces. Notes 7 16th December Inner products and orthonormal bases

Solutions: Problem Set 3 Math 201B, Winter 2007

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

Conceptual Questions for Review

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

LINEAR ALGEBRA W W L CHEN

Assignment #9: Orthogonal Projections, Gram-Schmidt, and Least Squares. Name:

Physics 6303 Lecture 11 September 24, LAST TIME: Cylindrical coordinates, spherical coordinates, and Legendre s equation

Chapter 4 Euclid Space

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Cambridge University Press The Mathematics of Signal Processing Steven B. Damelin and Willard Miller Excerpt More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0

Inner Product Spaces An inner product on a complex linear space X is a function x y from X X C such that. (1) (2) (3) x x > 0 for x 0.

Review problems for MA 54, Fall 2004.

The Shortest Vector Problem (Lattice Reduction Algorithms)

Further Mathematical Methods (Linear Algebra) 2002

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

The 'linear algebra way' of talking about "angle" and "similarity" between two vectors is called "inner product". We'll define this next.

Linear Models Review

Linear Algebra. Min Yan

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Chapter 2. Vectors and Vector Spaces

BASIC ALGORITHMS IN LINEAR ALGEBRA. Matrices and Applications of Gaussian Elimination. A 2 x. A T m x. A 1 x A T 1. A m x

Homework 11 Solutions. Math 110, Fall 2013.

Linear Algebra. Session 12

Linear Algebra. and

Exercises for Unit I (Topics from linear algebra)

1 angle.mcd Inner product and angle between two vectors. Note that a function is a special case of a vector. Instructor: Nam Sun Wang. x.

Exercises for Unit I (Topics from linear algebra)

1. General Vector Spaces

The geometry of least squares

ELEMENTARY LINEAR ALGEBRA

MATH 167: APPLIED LINEAR ALGEBRA Chapter 3

This pre-publication material is for review purposes only. Any typographical or technical errors will be corrected prior to publication.

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra

Inner product spaces. Layers of structure:

Chapter Two Elements of Linear Algebra

Practice Exam. 2x 1 + 4x 2 + 2x 3 = 4 x 1 + 2x 2 + 3x 3 = 1 2x 1 + 3x 2 + 4x 3 = 5

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS

7. Dimension and Structure.

Math 113 Final Exam: Solutions

Math 54. Selected Solutions for Week 5

Matrix Algebra: Vectors

Transcription:

MODULE 6 Topics: Gram-Schmidt orthogonalization process We begin by observing that if the vectors {x j } N are mutually orthogonal in an inner product space V then they are necessarily linearly independent. For suppose that x i, x j = 0 for i j and α j x j = 0 then taking the inner product of this equation with x k shows that α k x k, x k = 0 so that α k = 0. Hence only zero coefficients are possible. We shall now observe the simplification which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that M = span{x 1,..., x N } V and x i, x j = 0 for i j. Let y be a given vector in V. According to Module 4 the orthogonal projection P y in M is given as where with and P y = α j x j A α = b, A ij = x j, x i b i = y, x i, has to be solved to obtain the coefficients {α j }. But because the {x j } are all orthogonal the matrix A is diagonal so that α i = y, x i x i, x i Example. Let f L 2 ( π, π) and let M = span{1, cos t,..., cos Nt, sin t,..., sin Nt}. 30

It is straightforward to verify that and that cos mt, sin nt = π π cos mt sin nt dt = 0 for all m, n cos mt, cos nt = sin mt, sin nt = 0 for m n. Hence all the elements spanning M are mutually orthogonal and therefore linearly independent and thus a basis of M. The orthogonal projection of f onto M is then given by P f = α n cos nt + β n sin nt n=0 n=1 where the coefficients are found explicitly as α n = f, cos nt cos nt, cos nt n = 0, 1,..., N and β n = f, sin nt sin nt, sin nt n = 1,..., N. These coefficients are known as the Fourier coefficients of f in L 2 ( π, π) and P f is the Nth partial sum of the Fourier series of f. This partial sum is the best approximation, in the mean square sense, over the interval ( π, π) in terms of the given sine and cosine functions (i.e., in terms of a so-called trigonometric polynomial). If in an application the linearly independent vectors {x j } spanning M are not orthogonal then it may be advantageous to compute an equivalent basis {z 1,..., z N } of M of mutually orthogonal vectors and to express the projection as a linear combination of these new basis vectors. The process of finding the orthogonal basis {z j } equivalent to the basis {x j } is known as the Gram-Schmidt orthogonalization process and proceeds recursively as follows. We set z 1 = x 1. Assume that for j = 1,..., k 1 we have found orthogonal vectors {z j } span{x j }. Then we set k 1 z k = x k α j z j 31

where the α j are computed such that z k, z j = 0 for j = 1,..., k 1. Since z i, z j = 0 for i j and i, j < k this requires that α j = x k, z j z j, z j. When k = N we have generated N mutually orthogonal vectors, each of which is obtained as a combination of the basis vectors {x j }. Hence {z j } forms on orthogonal basis of M. Examples: 1) Let us find an orthogonal basis of M = span{(1, 2, 1, 2), (0, 1, 0, 1), (1, 0, 0, 1)} E 4. The notation E 4 implies that the inner product is the dot product. We set z 1 = (1, 2, 1, 2) and compute z 2 = (0, 1, 0, 1) α 1 (1, 2, 1, 2) where so that α 1 = (0, 1, 0, 1) (1, 2, 1, 2) (1, 2, 1, 2) (1, 2, 1, 2) = 4 10 z 2 = ( 2 / 5, 1 / 5, 2 / 5, 1 / 5 ). Since the span of a set remains unchanged if the vectors are scaled we can simplify the notation in our long-hand calculation by setting z 2 = ( 2, 1, 2, 1). Then z 3 = (1, 0, 0, 1) α 1 (1, 2, 1, 2) α 2 ( 2, 1, 2, 1) where α 1 = (1, 0, 0, 1) (1, 2, 1, 2) 10 = 1 10 α 2 = (1, 0, 0, 1) ( 2, 1, 2, 1) ( 2, 1, 2, 1) ( 2, 1, 2, 1) = 3 10 32

so that z 3 = ( 1 / 2, 1 / 2, 1 / 2, 1 / 2 ). Hence as an orthogonal basis of M we may take {(1, 2, 1, 2), ( 2, 1, 2, 1), (1, 1, 1, 1)}. The orthogonal projection of an arbitrary vector y E 4 onto M is then given by P y = 3 i=1 y, z i z i, z i z i. On occasion the vectors z j are scaled so that they are unit vectors in the norm z = z, z 1/2. The vectors {z j } are said to be orthonormal in this case. 2) Find an orthogonal basis of span{1, t, t 2 } L 2 ( 1, 1). Scale the functions such that they assume a value of 1 at t = 1. The notation indicates that the inner product is f, g = 1 1 f(t)g(t)dt. We now apply the Gram-Schmidt process to generate orthogonal functions {φ 0, φ 1, φ 2 }. We set φ 0 (t) 1 (which already satisfies φ(1) = 1) and compute φ 1 (t) = t α 0 1 where α 0 = t, 1 1, 1 = 0. Hence φ 1 (t) = t (which also is already properly scaled). Then φ 2 (t) = t 2 α 0 φ 0 (t) α 1 φ 1 (t) with α 0 = t2, 1 1, 1 = 1 3 33

and α 1 = t2, t t, t = 0. Hence a function orthogonal to φ 0 (t) and φ 1 (t) is c(t 2 1 / 3 ) for any constant c. To insure that φ 2 (1) = 1 we choose φ 2 (t) = 1 / 2 (3t 2 1). These three orthogonal polynomials are known as the first three Legendre polynomials which arise, for example, in the solution of the heat equation in spherical coordinates. Legendre polynomials of order up to N can be found by applying the Gram-Schmidt process to the linearly independent functions {1, t,..., t N } with the L 2 inner product. We shall conclude our discussion of projections and best approximations with a problem which, strictly speaking, leads neither to a projection nor a best approximation, but which has much the same flavor as the material presented above. Suppose we are in a vector space V with inner product, and its associated norm. Suppose further that we wish to find an element u V which satisfies the following N constraint equations u, x i = b i, i = 1,..., N where the vectors {x 1,.., x N } are assumed to be linearly independent in V. In general, the solution to this problem will not be unique. The solution does exist and is unique if we restrict it to lie in In this case u has to have the form M = span{x 1,..., x n }. u = α j x j. Substitution into the N equations shows that {α j } is a solution of the system A α = b where as before A ij = x j, x i. 34

Linear independence of the {x j } guarantees that A is invertible and that therefore the {α j } are uniquely defined. Theorem: Let y be any vector in V which satisfies the N constraint equations and let u be the specific solution which belongs to M then u y. Proof: For any y we can write y = y u + u. Then y 2 = y u + u, y u + u = y u, y u + 2Re y u, u + u, u. But Hence y u, u = α j y u, x = 0 because y, x j = b j = u, x j. y 2 = y u 2 + u 2 > u 2 for y u. Problems of this type arise in the theory of optimal controls for linear state equations and quadratic cost functionals. We shall not pursue this subject here but instead consider the following simpler geometric problem. Problem: Let x = (x, y, z) be a point in E 3. Find the point in the intersection of the planes x + 2y + 3z = 1 3x + 2y + z = 5 which is closest to the point (0, 1, 0). Answer: The geometry is clear. These are two planes which are not parallel to each other. Hence they intersect in a line. The problem then is to find the point on the line which is closest to (0, 1, 0). Since the setting is E 3, closest means closest in Euclidean distance. We shall examine three different approaches to solving this problem. i) If we define u = x (0, 1, 0) then we want the minimum norm u which satisfies u, (1, 2, 3) = x, (1, 2, 3) (0, 1, 0), (1, 2, 3) = 1 u, (3, 2, 1) = x, (3, 2, 1) (0, 1, 0), (3, 2, 1) = 3. 35

According to the last theorem the minimum norm solution is the uniquely defined solution belonging to M = span{(1, 2, 3), (3, 2, 1)}. It is computed as u = α 1 (1, 2, 3) + α 2 (3, 2, 1) where {α 1, α 2 } are found from the linear system ( ) ( ) (1, 2, 3), (1, 2, 3) (3, 2, 1), (1, 2, 3) α1 = (1, 2, 3), (3, 2, 1) (3, 2, 1), (3, 2, 1) α 2 Doing the arithmetic we find that ( ) ( ) 14 10 α1 = 10 14 α 2 ( ) 1. 3 (x, y, z) = ( 7 / 6, 7 / 6, 5 / 6 ). ii) We shall find the equation of the line of intersection. Subtracting the equation of one plane from the other we find that coordinates of points in the intersection must satisfy x z = 2. We set z = s. Then x = 2 + s and y = 1 / 2 2s. Hence the line is (x, y, z) = (2, 1 / 2, 0) + s(1, 2, 1) i.e., the line through (2, 1 / 2, 0) with direction (1, 2, 1). The square of the distance from (0, 1, 0) to a point on the line is g(s) = (2 + s) 2 + ( 3 / 2 + 2s) 2 + s 2. This function is minimized for s = 5 / 6 yielding (x, y, z) = ( 7 / 6, 7 / 6, 5 / 6 ). iii) The problem can be solved with Lagrange multipliers. We want to minimize g( x) = g(x, y, z) = x (0, 1, 0), x (0, 1, 0). subject to the constraints x, (1, 2, 3) 1 = 0 36

x, (3, 2, 1) 5 = 0. The Lagrangian is L = x 2 + (y 1) 2 + z 2 + λ 1 (x + 2y + 3z 1) + λ 2 (3x + 2y + z 5). The minimizer has to satisfy L x = L y = L z = 0 as well as the constraint equations. This leads to the linear system which again has the solution 2 0 0 1 3 x 0 0 2 0 2 2 y 2 0 0 2 3 1 = z = 0 1 2 3 0 0 λ 1 1 3 2 1 0 0 λ 2 5 (x, y, z) = ( 7 / 6, 7 / 6, 5 / 6 ). 37

Module 6 - Homework 1) Consider the vectors {(1, 1, 0), (1, 0, 1), (0, 1, 1), (1, 1, 1)}. i) Why can these vectors not be linearly independent? ii) Carry out the Gram-Schmidt process. How does the dependence of the vectors {x j } affect the calculation of the {z j }? 2) Let y = (1, 2, 3) E 3. Find three orthonormal vectors {u 1, u 2, u 3 } such that u 1 is parallel to y. 3) Let V be the set of all continuous real valued functions which are square integrable over (0, ) with respect to the weight function w(t) = e t, i.e., V = { f : f C 0 (0, ) and i) Show that V is a vector space over R. ii) Show that defines an inner product on V. f, g = iii) Show that M = span{1, t, t 2 } V. 0 0 } f(t) 2 e t dt < f(t)g(t)e t dt iv) Find an orthogonal basis of M. Scale the vectors such that they assume the value 1 at t = 0 (if you solve this problem correctly you will find the first three so-called Laguerre polynomials). 38