ON ANGLES BETWEEN SUBSPACES OF INNER PRODUCT SPACES

Similar documents
ON GENERALIZED n-inner PRODUCT SPACES

LINEAR ALGEBRA SUMMARY SHEET.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

BOUNDED LINEAR FUNCTIONALS ON THE n-normed SPACE OF p-summable SEQUENCES

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Chapter 6: Orthogonality

MTH 2032 SemesterII

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

y 2 . = x 1y 1 + x 2 y x + + x n y n 2 7 = 1(2) + 3(7) 5(4) = 3. x x = x x x2 n.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH Linear Algebra

Principal Component Analysis

2. Every linear system with the same number of equations as unknowns has a unique solution.

Diagonalizing Matrices

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

HOSTOS COMMUNITY COLLEGE DEPARTMENT OF MATHEMATICS

Ph.D. Katarína Bellová Page 1 Mathematics 2 (10-PHY-BIPMA2) EXAM - Solutions, 20 July 2017, 10:00 12:00 All answers to be justified.

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Principal Angles Between Subspaces and Their Tangents

Math Linear Algebra II. 1. Inner Products and Norms

Lecture 1: Review of linear algebra

Solutions to Review Problems for Chapter 6 ( ), 7.1

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Review problems for MA 54, Fall 2004.

Spectral Theorem for Self-adjoint Linear Operators

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Schur s Triangularization Theorem. Math 422

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

NORMS ON SPACE OF MATRICES

1 Quasi-definite matrix

Majorization for Changes in Ritz Values and Canonical Angles Between Subspaces (Part I and Part II)

Definitions for Quizzes

Linear Algebra Review. Vectors

Systems of Linear Equations

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

INNER PRODUCTS ON n-inner PRODUCT SPACES

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

Math 3191 Applied Linear Algebra

Chapter 3 Transformations

Linear Algebra: Matrix Eigenvalue Problems

Math 307 Learning Goals

Applied Linear Algebra in Geoscience Using MATLAB

p-summable SEQUENCE SPACES WITH INNER PRODUCTS

Math 302 Outcome Statements Winter 2013

Linear Algebra Primer

Quantum Computing Lecture 2. Review of Linear Algebra

1. General Vector Spaces

Final Exam Practice Problems Answers Math 24 Winter 2012

A Brief Outline of Math 355

Math 291-2: Lecture Notes Northwestern University, Winter 2016

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Final Exam. Linear Algebra Summer 2011 Math S2010X (3) Corrin Clarkson. August 10th, Solutions

LECTURE VI: SELF-ADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY

Introduction to Linear Algebra, Second Edition, Serge Lange

A geometric proof of the spectral theorem for real symmetric matrices

Math 307 Learning Goals. March 23, 2010

NOTES ON LINEAR ALGEBRA CLASS HANDOUT

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

Exercise Set 7.2. Skills

Some Results in Generalized n-inner Product Spaces

SYLLABUS. 1 Linear maps and matrices

Rigid Geometric Transformations

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Linear Algebra. Paul Yiu. 6D: 2-planes in R 4. Department of Mathematics Florida Atlantic University. Fall 2011

Solutions to practice questions for the final

Exercise Solutions for Introduction to 3D Game Programming with DirectX 10

No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.

SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Review of Linear Algebra

Math 4377/6308 Advanced Linear Algebra

Rigid Geometric Transformations

Further Mathematical Methods (Linear Algebra) 2002

We wish to solve a system of N simultaneous linear algebraic equations for the N unknowns x 1, x 2,...,x N, that are expressed in the general form

Homework 2. Solutions T =

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

MATHEMATICS COMPREHENSIVE EXAM: IN-CLASS COMPONENT

Review of some mathematical tools

Math 108b: Notes on the Spectral Theorem

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

Solving a system by back-substitution, checking consistency of a system (no rows of the form

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Lecture 1.4: Inner products and orthogonality

Eigenvalues and Eigenvectors

Math 21b. Review for Final Exam

arxiv: v2 [math.na] 27 Dec 2016

LINEAR ALGEBRA W W L CHEN

Chapter 5 Eigenvalues and Eigenvectors

The SVD-Fundamental Theorem of Linear Algebra

Introduction to Matrix Algebra

Elementary maths for GMT

Linear Algebra Highlights

Transcription:

ON ANGLES BETWEEN SUBSPACES OF INNER PRODUCT SPACES HENDRA GUNAWAN AND OKI NESWAN Abstract. We discuss the notion of angles between two subspaces of an inner product space, as introduced by Risteski and Trenčevski [11] and Gunawan, Neswan and Setya-Budhi [7], and study its connection with the so-called canonical angles. Our approach uses only elementary calculus and linear algebra. 1. Introduction The notion of angles between two subspaces of a Euclidean space R n has attracted many researchers since the 1950 s (see, for instance, [1,, 3]). These angles are known to statisticians and numerical analysts as canonical or principal angles. A few results on canonical angles can be found in, for example, [4, 8, 10, 1]. Recently, Risteski and Trenčevski [11] introduced an alternative definition of angles between two subspaces of R n and explained their connection with canonical angles. Let U = span{u 1,..., u p } and V = span{v 1,..., v q } be p- and q-dimensional subspaces of R n, where {u 1,..., u p } and {v 1,..., v q } are orthonormal, with 1 p q. Then, as is suggested by Risteski and Trenčevski [11], one can define the angle θ between the two subspaces U and V by cos θ := det(m T M) where M := [ u i, v k ] T is a q p matrix, M T is its transpose, and, denotes the usual inner product on R n. (Actually, Risteski and Trenčevski gave a more general formula for cos θ, but their formula is erroneous and is corrected by Gunawan, Neswan and Setya-Budhi in [7].) 000 Mathematics Subject Classification. 15A03, 51N0, 15A1, 65F30. Key words and phrases. canonical angles, angle between subspaces. The research is supported by QUE-Project V (003) Math-ITB. 1

Now let θ 1 θ θ p be the canonical angles between U and V, which are defined recursively by cos θ 1 := cos θ i+1 := max max u U i, u =1 max u, v = u 1, v 1 max u, v = u i+1, v i+1 v V i, v =1 where U i is the orthogonal complement of u i relative to U i 1 and V i is the orthogonal complement of v i relative to V i 1 (with U 0 = U and V 0 = V ). Then we have Theorem 1 (Risteski & Trenčevski). cos θ = p cos θ i. The proof of Theorem 1 uses matrix theoretic arguments (involving the notions of orthogonal, positive definite, symmetric, and diagonalizable matrices). In this note, we present an alternative approach towards Theorem 1. For the case p =, we reprove Theorem 1 by using only elementary calculus. For the general case, we prove a statement similar to Theorem 1 by using elementary calculus and linear algebra. As in [7], we replace R n by an inner product space of arbitrary dimension. Related results may be found in [9].. Main results In this section, let (X,, ) be an inner product space of dimension or higher (may be infinite), and U = span{u 1,..., u p } and V = span{v 1,..., v q } be two subspaces of X with 1 p q <. Assuming that {u 1,..., u p } and {v 1,..., v q } are orthonormal, let θ be the angle between U and V, given by cos θ := det(m T M) with M := [ u i, v k ] T being a q p matrix..1. The case p =. For p =, we have the following reformulation of Theorem 1. Fact. cos θ = max max u, v min max u, v. Proof. First note that, for each u X, u, v is maximized on {v V : v = 1} at v = proj V u and max u, v = proj u, V u = proj proj V u proj V u V u. Here of course proj V u = u, v k v k, denoting the orthogonal projection of u on V. Now consider the function

f(u) := proj V u, u U, u = 1. function g := f, given by g(u) = proj V u = To find its extreme values, we examine the u, v k, u U, u = 1. Writing u = (cos t)u 1 + (sin t)u, we can view g as a function of t [0, π], given by [ g(t) = cos t u1, v k + sin t u, v k ]. Expanding the series and using trigonometric identities, we can rewrite g as g(t) = 1 [ C + A + B cos(t α) ] where A = and tan α = B/A. [ u1, v k u, v k ], B = C = u 1, v k u, v k, [ u1, v k + u, v k ], From this expression, we see that g has the maximum value m := C+ A +B and the minimum value n := C A +B. (Note particularly that n is nonnegative.) Accordingly the extreme values of f are max Now we observe that max u, v = m and min max u, v = n. mn = C (A + B ) 4 [ q = u 1, v k u, v k u 1, v k u, v k = det(m T M) = cos θ. ] This completes the proof. Geometrically, the value of cos θ, where θ is the angle between U = span{u 1, u } and V = span{v 1,..., v q } with q <, represents the ratio between the area of the parallelogram spanned by proj V u 1 and proj V u and the area of the parallelogram spanned by u 1 and u. The area of the parallelogram spanned by two vectors x and y in X is given by [ x y x, y ] 1/ (see [6]). In particular, when {u1, u } and 3

{v 1,..., v q } are orthonormal, the area of the parallelogram spanned by proj V u 1 and [ q proj V u is u 1, v k q u, v k [ q u 1, v k u, v k ] ] 1/, while the area of the parallelogram spanned by u 1 and u is 1. Our result here says that, in statistical terminology, the value of cos θ is the product of the maximum and the minimum correlations between u U and its best predictor proj V u V. The maximum correlation is of course the cosine of the first canonical angle, while the minimum correlation is the cosine of the second canonical angle. The fact that the cosine of the last canonical angle is in general the minimum correlation between u U and proj V u V is justified in [10]... The general case. The main ideas of the proof of Theorem 1 are that the matrix M T M is diagonalizable so that its determinant is equal to the product of its eigenvalues, and that the eigenvalues are the square of the cosines of the canonical angles. The following statement is slightly less sharp than Theorem 1, but much easier to prove (in the sense that there are no sophisticated arguments needed). Theorem 3. cos θ is equal to the product of all critical values of f(u) := u U, u = 1. max u, v, Proof. For each u X, u, v is maximized on {v V : v = 1} at v = proj V u and proj V u max u, v = proj V u. Now consider the function g := f, given by g(u) = proj V u, u U, u = 1. Writing u = p t i u i, we have proj V u = u, v k v k = and so we can view g as g(t 1,..., t p ), given by p g(t 1,..., t p ) := t i u i, v k v k = p t i u i, v k v k, ( p, t i u i, v k ) where p t i = 1. In terms of the matrix M = [ u i, v k ] T, one may write g(t) = t T M T Mt = Mt q, t p = 1, where t = [t 1 t p ] T R p and p denotes the usual norm in R p. 4

To find its critical values, we use the Lagrange method, working with the function h λ (t) := Mt q + λ ( 1 t p), t R p. For the critical values, we must have h λ (t) = M T Mt λt = 0, whence (.1) M T Mt = λt. Multilpying both sides by t T on the left, we obtain λ = t T M T Mt = g(t), because t T t = t = 1. Hence any Lagrange multiplier λ is a critical value of g. But (.1) also tells us that λ must be an eigenvalue of M T M. From elementary linear algebra, we know that λ must satisfy det(m T M λi) = 0, which has in general p roots, λ 1 λ p 0, say. Since M T M is diagonalizable, we conclude that det(m T M) = This proves the assertion. p λ i. A geometric interpretation of the value of cos θ is provided below. Fact 4. cos θ is equal to the volume of the p-dimensional parallelepiped spanned by proj V u i, i = 1,..., p. Proof. Since we are assuming that {v 1,..., v q } is orthonormal, we have proj V u i = u i, v k v k, for each i = 1,..., p. Thus, for i, j = 1,..., p, we obtain proj V u i, proj V u j = whence u i, v k u j, v k, det[ proj V u i, proj V u j ] = det(m T M) = cos θ. But det[ proj V u i, proj V u j ] is the square of the volume of the p-dimensional parallelepiped spanned by proj V u i, i = 1,..., p (see [6]). 5

3. Concluding remarks Statisticians use canonical angles as measures of dependency of one set of random variables on another. The drawback is that finding these angles between two given subspaces is rather involved (see, for example, [, 4]). Many researchers often use only the first canonical angle for estimation purpose (see, for instance, [5]). Geometrically, however, the first canonical angle is not a good measurement for approximation (in R 3, for instance, the first canonical angle between two arbitrary subspaces is 0 even though the two subspaces do not coincide). The work of Risteski and Trenčevski [11] suggests that if we multiply the cosines of the canonical angles (instead of computing each of them separately), we get the cosine of some value θ that can be considered as the geometrical angle between the two given subspaces. In general, as is shown in [7], the value of cos θ represents the ratio between the volume of the parallelepiped spanned by the projection of the basis vectors of the lower dimension subspace on the higher dimension subspace and the volume of the parallelepiped spanned by the basis vectors of the lower dimension subspace. (Thus, the notion of angles between two subspaces developed here can be thought of as a generalization of the notion of angles in trigonometry taught at schools.) What is particularly nice here is that we have an explicit formula for cos θ in terms of the basis vectors of the two subspaces. References [1] T.W. Anderson, An Introduction to Multivariate Statistical Analysis, John Wiley & Sons, Inc., New York, 1958. [] Ȧ. Björck and G.H. Golub, Numerical methods for computing angles between linear subspaces, Math. Comp. 7 (1973), 579 594. [3] C. Davies and W. Kahan, The rotation of eigenvectors by a perturbation. III, SIAM J. Numer. Anal. 7 (1970), 1 46. [4] Z. Drmač, On principal angles between subspaces of Euclidean space, SIAM J. Matrix Anal. Appl. (000), 173 194 (electronic). [5] S. Fedorov, Angle between subspaces of analytic and antianalytic functions in weighted L space on a boundary of a multiply connected domain in Operator Theory, System Theory and Related Topics, Beer-Sheva/Rehovot, 1997, 9 56. [6] F.R. Gantmacher, The Theory of Matrices, Vol. I, Chelsea Publ. Co., New York (1960), 47 56. [7] H. Gunawan, O. Neswan and W. Setya-Budhi, A formula for angles between two subspaces of inner product spaces, submitted. [8] A.V. Knyazev and M.E. Argentati, Principal angles between subspaces in an A-based scalar product: algorithms and perturbation estimates, SIAM J. Sci. Comput. 3 (00), 008 040. 6

[9] J. Miao and A. Ben-Israel, Product cosines of angles between subspaces, Linear Algebra Appl. 37/38 (1996), 71 81. [10] V. Rakočević and H.K. Wimmer, A variational characterization of canonical angles between subspaces, J. Geom. 78 (003), 1 14. [11] I.B. Risteski and K.G. Trenčevski, Principal values and principal subspaces of two subspaces of vector spaces with inner product, Beiträge Algebra Geom. 4 (001), 89 300. [1] H.K. Wimmer, Canonical angles of unitary spaces and perturbations of direct complements, Linear Algebra Appl. 87 (1999), 373 379. Department of Mathematics, Bandung Institute of Technology, Bandung 4013, Indonesia E-mail address: hgunawan, oneswan@dns.math.itb.ac.id 7