Math Real Analysis II

Similar documents
There are two things that are particularly nice about the first basis

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

MAT Linear Algebra Collection of sample exams

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Completion Date: Monday February 11, 2008

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Math 1180, Notes, 14 1 C. v 1 v n v 2. C A ; w n. A and w = v i w i : v w = i=1

Math 3191 Applied Linear Algebra

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Section 7.5 Inner Product Spaces

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Math Linear Algebra II. 1. Inner Products and Norms

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Math 416, Spring 2010 Gram-Schmidt, the QR-factorization, Orthogonal Matrices March 4, 2010 GRAM-SCHMIDT, THE QR-FACTORIZATION, ORTHOGONAL MATRICES

MA 1B ANALYTIC - HOMEWORK SET 7 SOLUTIONS

Math Real Analysis II

Math 113 Final Exam: Solutions

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

MATH 115A: SAMPLE FINAL SOLUTIONS

Projections and Least Square Solutions. Recall that given an inner product space V with subspace W and orthogonal basis for

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Vectors. Vectors and the scalar multiplication and vector addition operations:

NATIONAL UNIVERSITY OF SINGAPORE DEPARTMENT OF MATHEMATICS SEMESTER 2 EXAMINATION, AY 2010/2011. Linear Algebra II. May 2011 Time allowed :

Physics 505 Homework No. 1 Solutions S1-1

MTH 309Y 37. Inner product spaces. = a 1 b 1 + a 2 b a n b n

Recall that any inner product space V has an associated norm defined by

Functional Analysis Exercise Class

Math Linear Algebra

Further Mathematical Methods (Linear Algebra) 2002

NUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.

Review problems for MA 54, Fall 2004.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Linear Algebra. Session 12

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

MATH 22A: LINEAR ALGEBRA Chapter 4

MATH 167: APPLIED LINEAR ALGEBRA Chapter 3

Math 24 Spring 2012 Sample Homework Solutions Week 8

Mathematics Department Stanford University Math 61CM/DM Inner products

n=1 ( 2 3 )n (a n ) converges by direct comparison to

MATH Linear Algebra

The Gram Schmidt Process

The Gram Schmidt Process

Power Series Solutions to the Legendre Equation

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

LECTURE 14: REGULAR SINGULAR POINTS, EULER EQUATIONS

Math 61CM - Solutions to homework 6

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Chapter 6: Orthogonality

Kernel Method: Data Analysis with Positive Definite Kernels

Functional Analysis F3/F4/NVP (2005) Homework assignment 2

MIDTERM I LINEAR ALGEBRA. Friday February 16, Name PRACTICE EXAM SOLUTIONS

Recall: Dot product on R 2 : u v = (u 1, u 2 ) (v 1, v 2 ) = u 1 v 1 + u 2 v 2, u u = u u 2 2 = u 2. Geometric Meaning:

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

Reproducing Kernel Hilbert Spaces

Inner Product, Length, and Orthogonality

Orthogonal-Polynomials

Contents. Appendix D (Inner Product Spaces) W-51. Index W-63

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Methods of Mathematical Physics X1 Homework 2 - Solutions

Answer Keys For Math 225 Final Review Problem

Applied Linear Algebra in Geoscience Using MATLAB

Reduction to the associated homogeneous system via a particular solution

Functional Analysis (2006) Homework assignment 2

10 Orthogonality Orthogonal subspaces

A Brief Outline of Math 355

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

Review of Power Series

MATH 369 Linear Algebra

Linear Models Review

15B. Isometries and Functionals 1

235 Final exam review questions

Math 310 Final Exam Solutions

Linear algebra II Homework #1 due Thursday, Feb. 2 A =. 2 5 A = When writing up solutions, write legibly and coherently.

Archive of past papers, solutions and homeworks for. MATH 224, Linear Algebra 2, Spring 2013, Laurence Barker

Introduction to Signal Spaces

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

October 25, 2013 INNER PRODUCT SPACES

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

Math 205 A B Final Exam page 1 12/12/2012 Name

Honours Algebra 2, Assignment 8

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0

LINEAR ALGEBRA QUESTION BANK

Let p 2 ( t), (2 t k), we have the scaling relation,

Columbus State Community College Mathematics Department Public Syllabus

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra Massoud Malek

Further Mathematical Methods (Linear Algebra)

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

1. Subspaces A subset M of Hilbert space H is a subspace of it is closed under the operation of forming linear combinations;i.e.,

AA242B: MECHANICAL VIBRATIONS

MAT2342 : Introduction to Applied Linear Algebra Mike Newman, fall Projections. introduction

Solving Differential Equations Using Power Series

Solving Differential Equations Using Power Series

Chapter 4 Euclid Space

Transcription:

Math 4 - Real Analysis II Solutions to Homework due May Recall that a function f is called even if f( x) = f(x) and called odd if f( x) = f(x) for all x. We saw that these classes of functions had a particularly nice orthogonality relationship. More precisely we considered the inner product space and its inner product C([ ]) = {f : [ ] R f is continuous} f g = f(t)g(t) dt. Furthermore we considered the susbpace of even functions and odd functions given by In class we proved that E = O. E = {f C([ ]) f( x) = f(x)} O = {f C([ ]) f( x) = f(x)}. (a) Show that if f E O then f is equal to the zero function (i.e. the zero vector in this vector space). (b) Use (a) to show that C([ ]) = E O. To do this show that you can write any f C([ ]) uniquely as f = f e + f o where f e E and f o O. Note that you cannot use the decomposition theorem from class since that requires your subspace to be finite dimensional which this is not. (c) Show that if a function f is odd then f(0) = 0. (d) Show that the derivative of an even function is odd and the derivative of an odd function is even. (e) Let p(x) = a n x n + a n x n + + a x + a 0 be a polynomial. Show p is odd if and only if a k = 0 for all even k. As a hint take a derivative and evaluate the derivative at 0; continue in this pattern. (f) Similar to above show that the polynomial p is even if and only if a k = 0 for all odd k. (g) Consider the polynomial p(x) = 5x 7 7x 6 + x 4 x + x. Decompose p as an even function and an odd function. Use the above to verify that your your components are indeed even and odd. (h) Decompose f(x) = e x as an even plus odd function. The even and odd function components have special names; what are they? Solution. (a) If f E O then f(x) = f( x) = f(x) for all x. Since f(x) = f(x) for all x then f(x) = 0 for all x. Thus f is the zero constant function. (b) In class we learned that we can write any function on [ ] as f = f e + f o where f e (x) = f(x) + f( x) and f o (x) = f(x) f( x). Note that in particular f e E and f o O. Assume that we have another expression of f as an even plus and an odd function. Thus f = g e + g o for g e E and g o O. Then consider f e g e E. Notice that f e g e = (f f o ) (f g o ) = g o f o O. Thus f e g e E O. By above we have that f e g e = 0 the constant 0 function. Thus f e = g e. A similar computation shows that f o = g 0. Thus there is a unique way to decompose f as an even plus an odd function. So C([ ]) = E O.

(c) If f is odd then f( x) = f(x). When x = 0 we get that f(0) = f(0). The only way for this to be true is if f(0) = 0. (d) If f is even then f( x) = f(x). Differentiating we get that f ( x) = f (x) and thus f ( x) = f (x). Thus the derivative of an even function is odd. Now assume that g is odd. Then g( x) = g(x). Differentiating we get that g ( x) = g (x) and thus g ( x) = g (x). Thus g is even. (e) If a k = 0 for all even k then computing p( x) we get a constant term that can be factored out. Thus p( x) = p(x). Conversely assume that p is odd. Then plugging in 0 we get that p(0) = a 0. By the above we must have that p(0) = 0 thus a 0 = 0. Taking two derivatives we get that p (x) = n(n )a n x n + (n )(n )a n x n + + a. Since this is two derivatives of an odd function it is itself odd. Thus evaluating at 0 we get that p (0) = a = 0. Thus a = 0. Continuing in this way we get that a k = 0 for all even k. (f) If a k = 0 for all odd k then replacing x by x will not change the polynomial. Thus p( x) = p(x). Conversely assume that p is even. Then its derivative p (x) = na n x n + + a is an odd function. Thus evaluating at 0 should yield 0. Thus we get that 0 = p (0) = a. Taking two more derivatives we get another odd function and after evaluating at 0 we must have that a = 0. Continuing in this way we get that a k = 0 for all odd k. (g) Using the equation for p e and p o we get that p e (x) = ( 5x7 7x 6 + x 4 x + x ) + (5x 7 7x 6 + x 4 x x p o (x) = ( 5x7 7x 6 + x 4 x + x ) (5x 7 7x 6 + x 4 x x ) (h) Using the equation for f e and f o mentioned above we get that f e (x) = ex + e x and f o (x) = ex e x. The function f e is also known as cosh x and f o is known as sinh x. = 7x 6 + x 4 x ; = 5x 7 x. Question. Consider the finite dimensional vector space P n the space of polynomials with real coefficients of degree at most n along with its inner product given by f g = f(t)g(t) dt. (a) A common basis for P is { x}. Use the Gram-Schmidt Process to turn this into an orthonormal basis for P. (b) A common basis for P is { x x }. Use the Gram-Schmidt Process to turn this into an orthonormal basis for P. (c) Let U be the subspace of P of constant functions. Find U. (d) Compute the orthogonal complement of the U you computed in (c). Verify that (U ) = U. Solution.

(a) First we normalize the vector. Since = dx = then normalizing the vector we get that e =. To compute the next vector we first compute the dot product of e with x. Doing so we get that Thus To normalize this we first compute Thus e x = x x = x d = x u = x e x e = x. x dx = x e = x. Thus an orthonormal basis for P is { } x. (b) We can use our previous work to find the Gram-Schmidt process for P. First we compute the various components of u : x = x = x dx = x 6 =. ( ) x x x = x dx x = 0. Thus Normalizing we compute that x / x / = 5 9 x + 9 x x 5 = 0. =. u = x x e e x e e = x. (x /) dx = x 4 x + 9 dx = = 5 4 9 + 9 = 8 0 + 0 = 8. Thus 8 x / =. So normalizing we get that e = 8 (x /). Thus the orthonormal basis obtained via the Gram-Schmidt process is { x 8 (x /) }.

Question. In this question we will apply orthogonal projections to the space C([ ]) of continuous functions on [ ]. We will project onto the finite-dimensional subspaces P and P investigated in Question. In what follows using a computer program that can perform numerical integration (like Grapher Sage or Wolfram Alpha) is recommended as well as using decimal expansions. (a) Consider the function f(x) = e x C([ ]). Compute the orthogonal projection of this function onto the subspace P. In other words compute P P (e x ). Your orthonormal basis for P from Question a should be handy in this computation. (b) Compute e x P P (e x ). (c) Compare this to how good of an approximation the first-degree Taylor polynomial T (x) = + x is for e x by comparing the norm computation from (b) to e x ( + x). (d) Repeat steps (a) - (c) but with the subspace P. That is compute P P (e x ) and compare the values of the norms e x P P (e x ) and e x ( + x + x /). Solution. (a) To compute the orthogonal projection of e x we first compute the various inner products of e x with the basis elements: e x e x = dx.66 e x Thus the projection of e x onto P is x = (b) Computing the distance via the norm we get xex dx.90 P P (e x ) =.66 +.90.75 +.06x. x = P P (e x ) e x ) = (P P (e x ) e x ) dx.057 =.95. (c) The distance between e x to + x is given by e x ( + x)) = (e x x) dx. =.48. Thus the projection of e x onto the space of linear functions is better than the first-degree Taylor polynomial for e x. Question 4. For a finite dimensional inner produt space V the Riesz Representation Theorem says that given any linear function ϕ : V F there exists a vector x V such that ϕ(v) = v x. For each of the following ϕ show that it is indeed a linear functional and find the corresponding x V such that ϕ(v) = v x. 4

(a) ϕ : P R given by ϕ(p) = p (0). That is ϕ(p) is obtained by taking a derivative and evaluating at 0. (b) ϕ : P R given by ϕ(p) = p(0). (c) ϕ : P R given by ϕ(p) = p (0). Solution 4. { } (a) Consider the orthonormal basis x. for P. Applying ϕ to each of these we get that ϕ(/ ) = ( ) 0 and ϕ x =. Therefore the vector that corresponds to this functional is given by { (b) For the orthonormal basis ( ) x = 0 and ϕ ϕ is ϕ(/ ) / + ϕ (c) For the orthonormal basis ( ) x = ϕ is ϕ(/ ) / + ϕ ( 8 (x /) ( ) x x = x. } x 8 (x /) for P we apply ϕ and get ϕ(/ ) = / ) =. Thus the vector that represents the linear functional 5 8 ( ) ( ) x x + ϕ 8 (x /) 8 (x /) = / 5 8 (x /). { ( and ϕ 8 (x /) ϕ(/ ) / + ϕ } x 8 (x /) for P we apply ϕ and get ϕ(/ ) = 0 ) = 0. Therefore the vector that represents the linear functional ( ) ( ) x x + ϕ 8 (x /) x = x. Question 5. Let V be an n-dimensional inner product space and fix a vector x V. We saw that we can define a linear functional ϕ x V by ϕ x (v) = v x. Let U = span{x}. Notice that U is a -dimensional subspace if x 0. Furthermore recall that the kernel of a linear transformation ϕ is the set of all vectors that are sent to 0: which forms a subspace. (a) Show that ker(ϕ x ) = U. ker(ϕ) = {v V ϕ(v) = 0} (b) Use the Riesz Representation Theorem the Orthogonal Decomposition Theorem and (a) to show that the kernel of a linear functional is always a subspace of dimension n or n. When is it n-dimensional? (c) Let W be any (n )-dimensional subspace of V. Use (a) to show that W is the kernel of some linear functional. Be sure to define what that linear functional is. Solution 5. 5

(a) We will show that ker(ϕ x ) = U. First we will show that ker(ϕ x ) U. To do so let v ker(ϕ x ) and let αx U. Then since v ker(ϕ x ) we have that 0 = ϕ x (v) = v x. Thus v αx = α v x = α0 = 0. Thus v U. Next we will show that U ker(ϕ x ). Let v U. Then v is orthogonal to any element of U. In particular since x U v x = 0. Thus Thus v ker(ϕ x ). ϕ x (v) = x v = 0. (b) Let ϕ V be any linear functional of V. By the Riesz Representation Theorem there exists a unique x V such that ϕ(v) = v x. Let U = span{x}. By (a) ker(φ) = U. By the decomposition Theorem we have that V = U U = U ker(ϕ). If x 0 then U is -dimensional and thus ker(ϕ) is n -dimensional. If x = 0 then U is 0-dimensional and ker(ϕ) is n-dimensional. Thus ker(ϕ) has dimension n if and only if x = 0 if and only if ϕ is the zero functional. (c) Let W be an (n ) dimensional vector space and let W be its orthogonal complement. By the Decomposition Theorem V = W W and thus W is -dimensional. Let x be any non-zero vector in W. Then W = span{x}. Consider the linear functional ϕ x given by ϕ x (v) = v x. By (a) we have that ker(ϕ x ) = (W ) = W. 6