Further Mathematical Methods (Linear Algebra)

Similar documents
Further Mathematical Methods (Linear Algebra)

Further Mathematical Methods (Linear Algebra) 2002

Further Mathematical Methods (Linear Algebra) 2002

Math 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name:

Lecture 15, 16: Diagonalization

1. General Vector Spaces

Recall : Eigenvalues and Eigenvectors

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Linear Algebra Lecture Notes-II

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Further Mathematical Methods (Linear Algebra) 2002

Linear Algebra. and

Econ Slides from Lecture 7

1. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal

Quantum Computing Lecture 2. Review of Linear Algebra

Chapter 3. Matrices. 3.1 Matrices

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Repeated Eigenvalues and Symmetric Matrices

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

5 Compact linear operators

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

Math Linear Algebra II. 1. Inner Products and Norms

Linear Algebra Massoud Malek

Mathematical foundations - linear algebra

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

Math 310 Final Exam Solutions

COMP 558 lecture 18 Nov. 15, 2010

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

MTH 2032 SemesterII

Review of some mathematical tools

7 : APPENDIX. Vectors and Matrices

Review of Some Concepts from Linear Algebra: Part 2

Linear Algebra. Session 12

Linear algebra II Homework #1 due Thursday, Feb. 2 A =. 2 5 A = When writing up solutions, write legibly and coherently.

Transpose & Dot Product

2. Review of Linear Algebra

Mathematical Methods wk 2: Linear Operators

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Chapter 4 Euclid Space

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Solutions to Review Problems for Chapter 6 ( ), 7.1

Homework 11 Solutions. Math 110, Fall 2013.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

Typical Problem: Compute.

Transpose & Dot Product

1 Last time: least-squares problems

Conceptual Questions for Review

Symmetric and anti symmetric matrices

Diagonalization of Matrix

NOTES on LINEAR ALGEBRA 1

A Brief Outline of Math 355

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

MAT Linear Algebra Collection of sample exams

Chapter 3 Transformations

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Mathematics Department Stanford University Math 61CM/DM Inner products

Math 1553, Introduction to Linear Algebra

Linear Algebra: Matrix Eigenvalue Problems

The following definition is fundamental.

6 Inner Product Spaces

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

Designing Information Devices and Systems II

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

SUMMARY OF MATH 1600

7. Symmetric Matrices and Quadratic Forms

Review of Linear Algebra

There are two things that are particularly nice about the first basis

WI1403-LR Linear Algebra. Delft University of Technology

I. Multiple Choice Questions (Answer any eight)

Math Matrix Algebra

q n. Q T Q = I. Projections Least Squares best fit solution to Ax = b. Gram-Schmidt process for getting an orthonormal basis from any basis.

Spectral Theorem for Self-adjoint Linear Operators

Notes on Eigenvalues, Singular Values and QR

JUST THE MATHS UNIT NUMBER 9.8. MATRICES 8 (Characteristic properties) & (Similarity transformations) A.J.Hobson

GQE ALGEBRA PROBLEMS

MAC Module 12 Eigenvalues and Eigenvectors

Lecture notes: Applied linear algebra Part 1. Version 2

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Problem # Max points possible Actual score Total 120

Linear Algebra. Min Yan

Lecture 3: Review of Linear Algebra

Chapter 6: Orthogonality

Lecture 3: Review of Linear Algebra

Linear algebra II Homework #1 solutions A = This means that every eigenvector with eigenvalue λ = 1 must have the form

LINEAR ALGEBRA SUMMARY SHEET.

Linear Algebra Practice Problems

NATIONAL UNIVERSITY OF SINGAPORE DEPARTMENT OF MATHEMATICS SEMESTER 2 EXAMINATION, AY 2010/2011. Linear Algebra II. May 2011 Time allowed :

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Math 225 Linear Algebra II Lecture Notes. John C. Bowman University of Alberta Edmonton, Canada

The Singular Value Decomposition

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

ECE 275A Homework #3 Solutions

Transcription:

Further Mathematical Methods (Linear Algebra) Solutions For The 2 Examination Question (a) For a non-empty subset W of V to be a subspace of V we require that for all vectors x y W and all scalars α R: i. Closure under vector addition: x + y W. ii. Closure under scalar multiplication: αx W. To be an inner product on V a function x y which maps vectors x y V to R must be such that: i. Positivity: x x and x x = if and only if x =. ii. Symmetry: x y = y x. iii. Linearity: αx + βy z = α x z + β y z. for all vectors x y z V and all scalars α β R. (b) We consider the vector space which consists of all 2 2 real matrices given by {[ ] a b V = a b R} where the operations of vector addition and scalar multiplication are defined in the normal way. To show that the set W V defined by {[ ] a W = a R} is a subspace of V we start by noting that W is a subset of V since any matrix of the form [ ] a can be written as [ ] a b with b =. So to show that W is a subspace of V we take any two vectors in W say [ ] a u = and v = and any scalar α R and note that W is closed under: vector addition since [ ] a u + v = + and so u + v W too since a + a b + b d + d R. [ ] a c d [ ] [ ] a a + a c d = c + c d + d scalar multiplication since [ ] αa αu = αc αd and so αu W too since αa αb αd R.

as required. Clearly the matrix which plays the role of the additive identity (i.e. ) in this vector space is [ ] since [ ] a u + = + [ ] = [ ] a =. (c) We are asked to show that the function defined by [ ] [ ] a a c d = aa + cc + dd is an inner product on W. To do this we show that this formula satisfies all of the conditions given in part (a). Thus taking any three vectors [ ] [ ] [ ] a a a u = v = c d and w = c d in W and any two scalars α and β in R we have: i. u u = a 2 + c 2 + d 2 which is the sum of the squares of three real numbers and as such is real and non-negative. Further to show that u u = if and only if u = (where here is the 2 2 zero matrix) we note that: LTR: If u u = then a 2 + c 2 + d 2 =. But this is the sum of the squares of three real numbers and so it must be the case that a = c = d =. Thus u =. RTL: If u = then a = c = d =. Thus u u =. (as required). ii. Obviously u v = aa + cc + dd = a a + c c + d d = v u. iii. We note that the vector αu + βv is given by [ ] [ ] a a αu + βv = α + β c d = and so we have [ ] αa + βa αc + βc αd + βd αu + βv w = (αa + βa )a + (αc + βc )c + (αd + βd )d = α(aa + cc + dd ) + β(a a + b b + c c ) αu + βv w = α u w + β v w Consequently the formula given above does define an inner product on W (as required). (d) To find a matrix A such that where [M] S and [M] S and [M] S = A[M] S are the coordinate vectors of M W relative to the bases {[ ] [ ] [ ]} S = S = {[ ] [ ] [ ]} respectively we use the definition of coordinate vector. That is we use the fact that the equality [ ] [ ] [ ] [ ] [ ] [ ] a + b + c = a + b + c 2

holds if and only if where is the required matrix. a b c S a = A b c S A =

Question 2. The required Leslie matrix is L = /4 /2 and to find its unique real positive eigenvalue we solve the equation λ /4 λ /2 λ = = ( λ)λ2 + ( 4 λ = = λ λ 2 λ ) = 4 where we have simplified the determinant using a cofactor expansion on the third column. This can be factorised to give: ( λ λ ) ( λ + ) = 2 2 and so the eigenvalues are λ = 2 2. Thus the unique positive real eigenvalue is λ = 2 and this is clearly dominant since it is larger in magnitude than the other two (or alternatively since we have two successive fertile classes). To describe the long-term behaviour of this population we recall from the lectures that x (k) cλ k v and x (k) λ x (k ) for large k since we have a dominant eigenvalue. To find v an eigenvector corresponding to λ it is probably easiest to recall that: 8 v = b /λ = /6 and so we take v =. b b 2 λ 2 /8 Consequently in the long-term we have ( ) k 8 x (k) c and x (k) 2 2 x(k ) where the first result tells us that The proportion of the female population in each age-class becomes constant in the ratio 8::. whereas the second result tells us that The population in each age class increases by a factor of 2 time period (i.e. every twenty years in this case). (i.e. increases by 5%) every (b) The required Leslie matrix and the initial population distribution vector are given by: [ ] [ ] L = and x () 2 =. /4 2 So to find an exact formula for the population distribution vector at the end of the kth time period we note that x (k) = L k x () = PD k P x () where D is a diagonal matrix and the matrix P is invertible. To find these two matrices we can see that the eigenvalues of L are given by the equation: λ /4 l = = λ( λ) 4 = = λ2 λ (λ 4 = = ) ( λ + ) = 2 2 i.e. the eigenvalues of L are λ = 2 and λ = 2. The corresponding eigenvectors are then given by: 4

For λ = 2 we have [ ] [ ] /2 x /4 /2 y = = x + 6y = i.e. x = 6y for any y R and so a corresponding eigenvector would be [6 ] t. For λ = 2 we have [ ] [ ] /2 x /4 /2 y = = x + 2y = i.e. x = 2y for any y R and so a corresponding eigenvector would be [ 2 ] t. Consequently we can take our matrices P and D to be [ ] 6 2 P = and D = [ ] /2 /2 and so we have P = 8 [ ] 2 6 as well. Hence to calculate x (k) = PD k P x () we note that D k P x () = [ ] [ ] [ ] [ (/2) k 2 2 (/2) k 2(/2) 8 ( /2) k = k ] [ ] [ ] 4 (/2) k 6 2 ( /2) k ( /2) k = 4 4 5( /2) k and so x (k) = PD k P x () = 4 [ ] [ ] 6 2 (/2) k 5( /2) k = 4 [ 8(/2) k ( /2) k (/2) k + 5( /2) k ] is the required exact formula for x (k). The above expression for x (k) gives the population distribution vector at time t k = + 2k years so for the population distribution vector in the year 4 we want to evaluate this for k = 2. Thus [ x (2) 8(/2) = 4 2 ( /2) 2 ] [ ] [ ] 8 9 52 (/2) 2 + 5( /2) 2 = = 9 + 5 2 is the required population distribution vector. (c) As the females are now infertile the Leslie matrix for this population is [ ] L =. /4 Thus we can see that and so x () = Lx (2) = x (4) = Lx () = [ ] [ ] 52 /4 2 [ = [ ] [ ] /4 52/4 52/4 = [ ] i.e. the female population dies out after 4 years i.e. in the year 8. Or alternatively noting that the maximum lifespan of a female in this population is forty years it should be clear that (since no new females are being born after 4) they will all be dead within forty years i.e. by the year 8. ] 5

Question (a) We are given that Y and Z are subspaces of a vector space X and we are asked to show that the following two statements are equivalent: I. X = Y + Z and every x X can be written uniquely as x = y + z with y Y and z Z. II. X = Y + Z and Y Z = {}. That is we need to show that I is true if and only if II is true. So as this is an if and only if statement we have to show that it holds both ways : If I then II: We are given that X = Y + Z and every x X can be written uniquely as x = y + z with y Y and z Z and so noting that: We are allowed to assume that X = Y + Z. Taking any vector u Y Z we have u Y and u Z. Now as u X and is in both Y and Z (since they are subspaces of X) we can write u = + u = u + and so by uniqueness we have u =. Thus Y Z = {}. we have X = Y + Z and Y Z = {} (as required). If II then I: We are given that X = Y + Z and Y Z = {} and so noting that: We are allowed to assume that X = Y + Z. Consider any vectors y y Y and any vectors z z Z which are such that On re-arranging this gives x = y + z = y + z. y y = z z where y y Y and z z Z as Y and Z are subspaces of X. So as y y = z z the vector given by these differences must be in Y Z = {} i.e. y y = z z =. Thus we have y = y and z = z and so there is only one way of writing x X in terms of y Y and z Z. we have X = Y + Z and every x X can be written uniquely as x = y + z with y Y and z Z (as required). (b) Given that X = Y Z (i.e. the subspaces X Y and Z satisfy I or II as given above) we define the projection P from X onto Y parallel to Z to be a mapping P : X Y such that for any x X we have Px = y where x = y + z with y Y and z Z. To show that P is idempotent we note that: For any y Y X we can write y = y + and so Py = y. So for any x X we can write x = y + z with y Y and z Z and so we have Px = y = P 2 x = Py = P 2 x = y since Py = y. But Px = y and so this gives us P 2 x = Px for any x X. Thus P 2 = P and so P is idempotent. (as required). (c) We are given that A be a real matrix where R(A) and R(A) denote the range of A and the orthogonal complement of R(A) respectively. Now we let Ps be a projection of any vector s onto R(A) parallel to R(A) and this is clearly going to be the orthogonal projection of s onto R(A). So to show that this vector i.e. Ps is the vector in R(A) closest to s we note that: 6

Ps is clearly in R(A) since P projects vectors onto R(A). Taking r to be any vector in R(A) we can construct the vector u = Ps r and this is in R(A) too (since R(A) is a vector space). Now the distance between the vectors s and r is given by s r 2 = s Ps + u 2 = (I P)s + u 2 So as u R(A) and (I P)s R(A) we can apply the Generalised Theorem of Pythagoras to get s r 2 = (I P)s 2 + u 2 and this quantity is minimised when u =. Thus the vector in R(A) closest to s is given by u = i.e. r = Ps. (as required). (d) Hence or otherwise we are asked to find the least squares estimate of the parameters m and c when x and y are related by the expression y = mx + c and we are given the data: x - y - To start with we note that the relationship that the parameters that m and c should satisfy gives us a set of three [inconsistent] equations i.e. m + c = c = m + c = and we can write these in the form Ax = b using [ ] A = m x = c and b =. Now the hence or otherwise indicates that there are [at least] two methods and so let s look at them both: Hence The least squares estimate of these parameters are the values of m and c which minimise the quantity Ax b. But since Ax is just a vector in R(A) we have R(A) = Lin and so the vector we seek will be of the form m + c. Indeed we seek the values of m and c which make this vector closest to the vector [ ] t. But by inspection we can see that R(A) = Lin 2 (since the vector [ 2 ] t is orthogonal to both [ ] t and [ ] t ) and so by part (c) as = + + 2 }{{ } }{{} R(A) R(A) 7

we can see that orthogonally projecting the vector [ ] t onto R(A) using P yields i.e. x = [m c ] t = [ ] t. P = + Otherwise We know from lectures that a least squares solution to the matrix equation Ax = b is given by x = (A t A) A t b and so since and we have A t A = [ ] = A t b = [ ] 2 [ ] = [ ] m c = x = (A t A) A t b = 6 as the least squares estimate for m and c. = (A t A) = 6 [ 2 [ ] 2 ] [ ] 2 = [ ] [ ] 2 8

Question 4. (a) A complex matrix A is called anti-hermitian if A = A (where A denotes the complex conjugate transpose of A). We are asked to show that: The non-zero eigenvalues of an anti-hermitian matrix are all purely imaginary. Let λ be a non-zero eigenvalue of the anti-hermitian matrix A with corresponding eigenvector x i.e. Ax = λx. Multiplying through by x we get x Ax = λx x and taking the complex conjugate transpose of this yields (i) x A x = λ x x = x Ax = λ x x (ii) since A is anti-hermitian. Now adding (i) and (ii) together we get (λ + λ )x x = and so as x x = x 2 (since x is an eigenvector) we have λ = λ. Thus since λ λ is purely imaginary (as required). The eigenvectors of an anti-hermitian matrix corresponding to distinct eigenvalues are orthogonal. Let A be an anti-hermitian matrix and let x and y be eigenvectors corresponding to the distinct eigenvalues λ and µ respectively of A. (Without loss of generality we assume that µ.) Thus we have Ax = λx and Ay = µy which on multiplying both sides of these expressions by y and x respectively yields y Ax = λy x and x Ay = µx y. Now taking the complex conjugate transpose of the latter expression gives y A x = µ y x = y Ax = µ y x since A is anti-hermitian. So equating our two expressions for y Ax we get (λ + µ )y x = = (λ µ)y x = since µ is non-zero and hence purely imaginary. Thus we have y x = = y x = = x y = and so the eigenvectors x and y are orthogonal (as required). (b) A complex matrix A is normal if A A = AA. Clearly anti-hermitian matrices are normal since A A = A 2 = AA. (c) To find the spectral decomposition of the matrix i A = i where i 2 = we start by noting that this matrix is anti-hermitian and hence it is normal by part (b). So to find the eigenvalues of A we solve the determinant equation given by i λ λ i λ = = (i λ)[ λ(i λ) + ] + (i λ) = = (λ i)(λ2 iλ + 2) =. Factorising this gives (λ i)(λ 2i)(λ + i) = and so the eigenvalues are λ = ±i 2i. We now need to find an orthonormal set of eigenvectors corresponding to these eigenvalues i.e. 9

For λ = i a corresponding eigenvector [x y z] t is given by x i y = = z y = x + iy + z = y = = y = x + z = i.e. x = z for z R and y =. Thus a corresponding eigenvector is [ ] t. For λ = i a corresponding eigenvector [x y z] t is given by 2i x i y = = 2i z 2ix + y = x iy + z = y + 2iz = = 2ix + y = 2ix + 2y + 2iz = y + 2iz = i.e. y = 2ix and z = x for x R. Thus a corresponding eigenvector is [ 2i ] t. For λ = 2i a corresponding eigenvector [x y z] t is given by i x 2i y = = i z ix y = x + 2iy + z = y iz = = ix y = ix 2y + iz = y iz = i.e. y = ix and z = x for x R. Thus a corresponding eigenvector is [ i ] t. and so an orthonormal set of eigenvectors is 2i i 2 6 (since the eigenvectors are already mutually orthogonal). Thus A = i [ ] i 2i [ 2i ] + 2i i [ i ] 2 6 is the spectral decomposition of A. = i i 2i 2i 4 2i + 2i i i i. 2 6 2i i

Question 5. (a) To find the general solution to the system of coupled linear differential equations given by ẏ = y 2y 2 ẏ 2 = 2y + 2y ẏ = 2y 2 + y we re-write them as ẏ = Ay where y = [y y 2 y ] t and 2 A = 2 2 2 and diagonalise this matrix. So to find the eigenvalues of A we solve the determinant equation λ 2 2 λ 2 = = ( + λ)[ λ( λ) 4] 4( λ) = 2 λ which on simplifying yields λ( λ 2 ) + 4( + λ) 4( λ) = = λ(λ 2 9) = and so the eigenvalues are λ = ±. Then to find the eigenvectors we note that: For λ = a corresponding eigenvector [x y z] t is given by 2 x x + 2y = 2 2 y = = x z = 2 z 2y + z = i.e. x = z and y = z/2 for z R. Thus a corresponding eigenvector is [ /2 ] t or [2 2] t. For λ = a corresponding eigenvector [x y z] t is given by 4 2 x 2x + y = 2 2 y = = 2x + y 2z = 2 2 z 2y 2z = = 2x + y = y z = i.e. x = z/2 and y = z for z R. Thus a corresponding eigenvector is [ /2 ] t or [ 2 2] t. For λ = a corresponding eigenvector [x y z] t is given by 2 2 x x y = 2 2 y = = 2x + y + 2z = 2 4 z y + 2z = = x y = y + 2z = i.e. x = y and z = y/2 for y R. Thus a corresponding eigenvector is [ /2] t or [2 2 ] t Thus we have P AP = D where 2 2 P = 2 2 and D =. 2 2 So given our set of coupled linear differential equations we can use this to write ẏ = Ay = ẏ = PDP y = ż = Dz

where z = P y. Thus we proceed by solving the uncoupled linear differential equations given by ż z z A ż 2 = z 2 = z 2 Be t ż z z Ce t for arbitrary constants A B and C. Consequently using the fact that y = Pz we can see that 2 2 A 2A Be t 2Ce t y(t) = 2 2 Be t = A + 2Be t 2Ce t 2 2 Ce t 2A + 2Be t + Ce t is the required general solution to ẏ = Ay. The initial conditions for this system are restricted so that they satisfy the equation y () 2y 2 () 2y () =. and so noting that our general solution gives y () 2A B 2C y(t) = y 2 () = A + 2B 2C y () 2A + 2B + C we substitute these expressions for y () y 2 () and y () into the restriction to get (2A B 2C) 2( A + 2B 2C) 2(2A + 2B + C) = = B =. Thus substituting this into our general solution we find that 2A 2Ce t y(t) = A 2Ce t. 2A + Ce t is the general solution in the presence of this restriction. If the initial conditions are further restricted by stipulating that 2 2γ y() = 2γ 2 + γ for some constant γ we can see that at t = we have 2 2γ 2A 2C 2 2 2γ = A 2C = (A + ) + (C γ) 2 =. 2 + γ 2A + C 2 But the vectors [2 2] t and [ 2 2 ] t are clearly linearly independent and so we must have A = and C = γ. Thus in the presence of this new restriction the general solution becomes 2 2γe t 2 2 y(t) = 2γe t = + 2 e t 2 + γe t 2 and so in the long-term we can see that y(t) [ 2 2] t. (b) The steady states of the coupled non-linear differential equations ẏ = y (y y 2 4) ẏ 2 = y 2 (y 2 4y 6) 2

are given by the solutions of the simultaneous equations i.e. by (y y 2 ) = ( ) ( 6) (4 ) and ( 2 2). y (y y 2 4) = y 2 (y 2 4y 6) = To assess the stability of the steady state given by (y y 2 ) = ( 2 2) we evaluate the Jacobian for this system at ( 2 2) i.e. DF[(y y 2 )] = [ ] 2y y 2 4 y 4y 2 2y 2 4y 6 = DF[( 2 2)] = [ ] 2 6 8 2 and find the eigenvalues of this matrix. So solving 2 λ 6 8 2 λ = = (2 + λ)2 6 8 = = 2 + λ = ±4 we find that the eigenvalues are λ = 2 ± 4. According to a result given in the lectures since these are not both both real and negative (note that 4 > 2) the steady state (y y 2 ) = ( 2 2) is not asymptotically stable.

Question 6 (a) To test the set of functions { x x 2 } for linear independence we calculate the Wronskian as instructed i.e. f f 2 f W (x) = f f 2 f f = x x 2 2x 2 = 2 f 2 f and as W (x) for all x R this set of functions is linearly independent (as required). (b) We consider the inner product space formed by the vector space P [ 22] and the inner product f(x) g(x) = f( 2)g( 2) + f( )g( ) + f()g() + f()g() + f(2)g(2). To find an orthonormal basis of the space Lin{ x x 2 } we use the Gram-Schmidt procedure: We start with the vector and note that 2 = = + + + + = 5. Consequently we set e = / 5. We need to find a vector u 2 where But as we have u 2 = x. Now as we set e 2 = x/. u 2 = x x 5 = x 5 x 5 x = ( 2)() + ( )() + ()() + ()() + (2)() = x 2 = x x = ( 2)( 2) + ( )( ) + ()() + ()() + (2)(2) = Lastly we need to find a vector u where But as and u = x 2 x 2 we have u = x 2 2. Now as x x 2 5 = x 2 x2 x 5 x 2 x = (4)( 2) + ()( ) + ()() + ()() + (4)(2) = x 2 = (4)() + ()() + ()() + ()() + (4)() = x2 5 x 2 2 2 = x 2 2 x 2 2 = (2)(2) + ( )( ) + ( 2)( 2) + ( )( ) + (2)(2) = 4 we set e = (x 2 2)/ 4. Consequently the set { } x 5 x2 2 4 is an orthonormal basis for the space Lin{ x x 2 }. (c) We are given that {e e 2... e k } is an orthonormal basis of a subspace S of an inner product space V. Extending this to an orthonormal basis {e e 2... e k e k+... e n } of V we note that for any vector x V n x = α i e i. i= 4

Now for any j (where j n) we have n x e j = α i e i e j = i= n α i e i e j = α j i= since we are using an orthonormal basis. Thus we can write x = n x e i e i = i= k x e i e i + i= } {{ } in S n i=k+ x e i e i. } {{ } in S and so the orthogonal projection of V onto S [parallel to S ] is given by Px = k x e i e i i= for any x V (as required). Further since x y measures the distance between x and a vector y S this quantity is minimised since orthogonal projections give the vector Px S which is closest to x. (d) Using the results in (c) it should be clear that a least squares approximation to x in Lin{ x x 2 } will be given by Px. So using the inner product in (b) and the orthonormal basis for Lin{ x x 2 } which we found there we have: since Px = x e e + x e 2 e 2 + x e e = x 5 Px = 4 x. + x x x + x x 2 2 (x 2 2) 4 x = ( 8)() + ( )() + ()() + ()() + (8)() = x x = ( 8)( 2) + ( )( ) + ()() + ()() + (8)(2) = 4 x x 2 2 = ( 8)(2) + ( )( ) + ()() + ()( ) + (8)(2) =. Thus our least squares approximation to x is 7 5 x. 5