Lecture # 3 Orthogonal Matrices and Matrix Norms. We repeat the definition an orthogonal set and orthornormal set.

Similar documents
12 CHAPTER 1. PRELIMINARIES Lemma 1.3 (Cauchy-Schwarz inequality) Let (; ) be an inner product in < n. Then for all x; y 2 < n we have j(x; y)j (x; x)

Part 1a: Inner product, Orthogonality, Vector/Matrix norm

Linear Algebra Massoud Malek

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

INNER PRODUCT SPACE. Definition 1

Inner products and Norms. Inner product of 2 vectors. Inner product of 2 vectors x and y in R n : x 1 y 1 + x 2 y x n y n in R n

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

Linear Algebra Review

Lecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Elementary linear algebra

Chapter 6 Inner product spaces

Lecture 1.4: Inner products and orthogonality

Functional Analysis Review

Review of Linear Algebra

Lecture 20: 6.1 Inner Products

Inner Product and Orthogonality

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Chapter 3 Transformations

Section 7.5 Inner Product Spaces

Lecture 23: 6.1 Inner Products

Section 3.9. Matrix Norm

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

Vector Spaces. Commutativity of +: u + v = v + u, u, v, V ; Associativity of +: u + (v + w) = (u + v) + w, u, v, w V ;

Linear Algebra Methods for Data Mining

Lecture Notes 1: Vector spaces

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

ORTHOGONALITY AND LEAST-SQUARES [CHAP. 6]

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Linear Analysis Lecture 5

(, ) : R n R n R. 1. It is bilinear, meaning it s linear in each argument: that is

Linear Algebra. Session 12

Definitions for Quizzes

Lecture 8 : Eigenvalues and Eigenvectors

Another consequence of the Cauchy Schwarz inequality is the continuity of the inner product.

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am

Projection Theorem 1

Inner Product Spaces 6.1 Length and Dot Product in R n

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

Hilbert Spaces: Infinite-Dimensional Vector Spaces

Math 108b: Notes on the Spectral Theorem

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

Linear Algebra Review. Vectors

Basic Calculus Review

Polar Form of a Complex Number (I)

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors

What is it we are looking for in these algorithms? We want algorithms that are

which has a check digit of 9. This is consistent with the first nine digits of the ISBN, since

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Lecture 1 Introduction

Inner product spaces. Layers of structure:

Matrix Algebra: Vectors

6.1. Inner Product, Length and Orthogonality

1. General Vector Spaces

Numerical Linear Algebra

Lecture 1: Review of linear algebra

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Linear algebra 2. Yoav Zemel. March 1, 2012

Lecture 3: Review of Linear Algebra

Functions of Several Variables

Linear Algebra: Matrix Eigenvalue Problems

Lecture 3: Review of Linear Algebra

Vectors. Vectors and the scalar multiplication and vector addition operations:

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Dot product. The dot product is an inner product on a coordinate vector space (Definition 1, Theorem

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

Vectors in Function Spaces

2. Review of Linear Algebra

Typical Problem: Compute.

Basic Concepts in Linear Algebra

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Lecture 6: Geometry of OLS Estimation of Linear Regession

Inner Product Spaces An inner product on a complex linear space X is a function x y from X X C such that. (1) (2) (3) x x > 0 for x 0.

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

Introduction to Numerical Linear Algebra II

Quantum Computing Lecture 2. Review of Linear Algebra

The Transpose of a Vector

NORMS ON SPACE OF MATRICES

FUNCTIONAL ANALYSIS-NORMED SPACE

Dot product and linear least squares problems

Elements of Convex Optimization Theory

FUNCTIONAL ANALYSIS LECTURE NOTES: ADJOINTS IN HILBERT SPACES

Review of Basic Concepts in Linear Algebra

Chapter 4 Euclid Space

LINEAR ALGEBRA W W L CHEN

Applied Linear Algebra in Geoscience Using MATLAB

Math 3191 Applied Linear Algebra

4 Linear Algebra Review

MATH 532: Linear Algebra

Real Analysis III. (MAT312β) Department of Mathematics University of Ruhuna. A.W.L. Pubudu Thilan

Theorem (4.11). Let M be a closed subspace of a Hilbert space H. For any x, y E, the parallelogram law applied to x/2 and y/2 gives.

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Advanced Digital Signal Processing -Introduction

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

Transcription:

Lecture # 3 Orthogonal Matrices and Matrix Norms We repeat the definition an orthogonal set and orthornormal set. Definition A set of k vectors {u, u 2,..., u k }, where each u i R n, is said to be an orthogonal with respect to the inner product (, ) if (u i, u j ) = 0 for i j. The set is said to be orthonormal if it is orthogonal and (u i, u i ) = for i =, 2,..., k The definition of an orthogonal matrix is related to the definition for vectors, but with a subtle difference. Definition 2 The matrix U = (u, u 2,..., u k ) R n k whose columns form an orthonormal set is said to be left orthogonal. If k = n, that is, U is square, then U is said to be an orthogonal matrix. Note that the columns of (left) orthogonal matrices are orthonormal, not merely orthogonal. Square complex matrices whose columns form an orthonormal set are called unitary. Example Here are some common 2 2 orthogonal matrices Let x R n then 0 U = 0 U = 0.5 0.8 0.6 U = 0.6 0.8 cos θ sin θ U = sin θ cos θ Ux 2 2 = (Ux) T (Ux) = x T U T Ux = x T x = x 2 2

So Ux 2 = x 2. This property is called orthogonal invariance, it is an important and useful property of the two norm and orthogonal transformations. That is, orthogonal transformations DO NOT AFFECT the two-norm, there is no comparable property for the one-norm or -norm. The Cauchy-Schwarz inequality given below is quite useful for all inner products. Lemma (Cauchy-Schwarz inequality) Let (, ) be an inner product in R n. Then for all x, y R n we have (x, y) (x, x) /2 (y, y) /2. () Moreover, equality in () holds if and only if x = αy for some α R. In the Euclidean inner product, this is written x T y x 2 y 2. This inequality leads to the following definition of the angle between two vectors relative to an inner product. Definition 3 The angle θ between the two nonzero vectors x, y R n with respect to an inner product (, ) is given by cos θ = (x, y). (2) (x, x) /2 /2 (y, y) With respect to the Euclidean inner product, this reads cos θ = xt y x 2 y 2. The one-norm and the -norm share two inequalities similar to the Cauchy Schwarz inequality. x T y x y, x T y x y, but these do not lead to any reasonable definition of angle between vectors. 2

Example 2 We have x = ( ), y = ( 3 ). Thus x 2 = 2, y 2 = 0, x T y = 2 cos θ = xt y = /5. x 2 y 2 θ =.07 R = 63.43 is the angle between the two vectors. product are x 2 y 2 = 20, x y = 2 3 = 6 x y = 4 = 4. The three upper bounds on the dot It is possible to come up with examples where each of these three is the smallest bound (or the largest one). Matrix Norms We will also need norms for matrices. Definition 4 A norm in R m n is a function mapping R m n into R satisfying the following three axioms. X 0; X = 0 if and only if X = 0, X R m n 2. αx = α X X R m n, α R 3. X + Y X + Y X, Y R m n. 3

This definition is isomorphic to the definition of a vector norm on R mn. For example, the Frobenius norm defined by m X F = /2 n x 2 ij i= j= is isomorphic to the two-norm on R mn. Since X represents a linear operator from R n to R m, it is appropriate to define the induced norm α on R m n by It is a simple matter to show that (3) Xy α X α = sup. (4) y 0 y α X α = max y α = Xy α. (5) Note that the maximum is taken over a closed, bounded set, thus we have that X α = Xy α (6) for some y such that y α =. The above definition leads to the very useful bound Xy α X α y α (7) where equality occurs for every vector of the form γy, γ R. For any induced norm, the identity matrix I n for R n n satisfies However, for the Frobenius norm I n =. (8) I n F = n, thus it is not an induced norm for any vector norm. For the one-norm and the -norm there are formulas for the corresponding matrix norms and for a vector y satisfying (6). The one-norm formula is m X = max x ij. (9) j n i= 4

If j max is the index of a column such that m X = x i,jmax i= then y = e jmax, the corresponding column of the identity matrix. The -norm formula is If i max is the index of a row such that n X = max x ij. (0) i m j= n X = x imax,j j= then the vector y = (y,..., y n) T with components y j = sign(x imax,j) satisfies (6). Note that X = X T. The matrix two-norm does not have a formula like (9) or (0) and all other formulations are really equivalent to (4). Moreover, computing the vector y in (6) is a nontrivial task that we will discuss later. The induced norms have a convenient property that is important in understanding matrix computations. For X R m n and Y R n s consider XY α. We have that Thus XY α = max z α = XY z α max z α = X α Y z α = X α max z α = Y z α = X α Y α. XY α X α Y α. () A norm α (or really family of norms) that satisfies the property () is said to be consistent. Since they are induced norms the two-norm, one-norm, and the -norm are all consistent. The Frobenius norm also satisfies (). An example of a matrix norm that is not consistent is given below. 5

Example 3 Consider the norm β on R m n given by X β = max (i,j) x ij. This is simply the -norm applied to X written out as vector in R mn. For m = n = 2, consider X = Y =. Note that XY = ( 2 2 2 2 and thus XY β = 2 > X β Y β =. Clearly, β is not consistent. Henceforth, we use only consistent norms. Now we give a numerical example with our four most used norms. Example 4 Consider It is easily verified that X = ) 3 2 0 0 6 3 25 X = 27, y = e 2 = X = 29. y =. 0 0,, X F = 3.70, X 2 = 25.46. The magic vector in the two-norm is (to the digits displayed) 0.8943 y2 = 0.97256 0.3508 This will not always be true, but notice that its sign pattern is the same as y and that its largest component corresponds to the non-zero component of y. 6