Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition. Name:

Similar documents
The Singular Value Decomposition

14 Singular Value Decomposition

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

7. Symmetric Matrices and Quadratic Forms

Assignment #9: Orthogonal Projections, Gram-Schmidt, and Least Squares. Name:

15 Singular Value Decomposition

1 Linearity and Linear Systems

Maths for Signals and Systems Linear Algebra in Engineering

Lecture 7: Positive Semidefinite Matrices

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4

Background Mathematics (2/2) 1. David Barber

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Problem # Max points possible Actual score Total 120

Review problems for MA 54, Fall 2004.

Properties of Matrices and Operations on Matrices

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

Math Matrix Algebra

Positive Definite Matrix

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

Solutions to Review Problems for Chapter 6 ( ), 7.1

UNIT 6: The singular value decomposition.

Linear Algebra (Review) Volker Tresp 2017

There are six more problems on the next two pages

Chapter 3 Transformations

Chapter 7: Symmetric Matrices and Quadratic Forms

Fall TMA4145 Linear Methods. Exercise set Given the matrix 1 2

Problem Set 1. Homeworks will graded based on content and clarity. Please show your work clearly for full credit.

Singular Value Decomposition

Linear Algebra - Part II

Introduction to SVD and Applications

Linear Algebra (Review) Volker Tresp 2018

7 Principal Component Analysis

Linear Least Squares. Using SVD Decomposition.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

Cheat Sheet for MATH461

Linear Algebra- Final Exam Review

Designing Information Devices and Systems II

1 Inner Product and Orthogonality

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Tutorial on Principal Component Analysis

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

The Singular Value Decomposition and Least Squares Problems

Symmetric Matrices and Eigendecomposition

1 Last time: least-squares problems

Linear Algebra Review. Vectors

COMP 558 lecture 18 Nov. 15, 2010

Review of Some Concepts from Linear Algebra: Part 2

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models

Large Scale Data Analysis Using Deep Learning

I. Multiple Choice Questions (Answer any eight)

Lecture 9: SVD, Low Rank Approximation

Econ Slides from Lecture 8

Maths for Signals and Systems Linear Algebra in Engineering

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Linear Algebra Solutions 1

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

EXAM. Exam 1. Math 5316, Fall December 2, 2012

Singular Value Decomposition

Linear Algebra. Shan-Hung Wu. Department of Computer Science, National Tsing Hua University, Taiwan. Large-Scale ML, Fall 2016

Announcements Monday, November 26

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Mathematical foundations - linear algebra

235 Final exam review questions

Announce Statistical Motivation Properties Spectral Theorem Other ODE Theory Spectral Embedding. Eigenproblems I

Lecture: Face Recognition and Feature Reduction

18.06 Problem Set 10 - Solutions Due Thursday, 29 November 2007 at 4 pm in

Singular value decomposition

Exercise Sheet 1.

Name: Final Exam MATH 3320

Least Squares Optimization

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

TBP MATH33A Review Sheet. November 24, 2018

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Chap 3. Linear Algebra

Basic Calculus Review

1. The Polar Decomposition

Jordan Normal Form and Singular Decomposition

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Gaussian Elimination without/with Pivoting and Cholesky Decomposition

Linear Algebra Review. Fei-Fei Li

Linear Algebra Primer

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

Mathematical foundations - linear algebra

Lecture 2: Linear Algebra

Lecture: Face Recognition and Feature Reduction

Linear Algebra and Eigenproblems

PCA, Kernel PCA, ICA

Singular Value Decomposition

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Review of similarity transformation and Singular Value Decomposition

1 Singular Value Decomposition and Principal Component

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Transcription:

Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition Due date: Friday, May 4, 2018 (1:35pm) Name: Section Number

Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition Due date: Friday, May 4, 2018 (1:35pm) For full credit you must show all of your work. 1. a a. (8 points) Let v = b be any arbitrary vector in  3 c i. Compute the matrix M = vv T ii. Verify that M is symmetric iii. Verify that rank(m) = 1 iv. Verify that v is an eigenvector of M: in other words, that Mv = lv. v. What is the associated eigenvalue l (either in terms of v or its components a, b, c)? x b. (6 points) Let w = y be any arbitrary vector in  3. z i. Verify that the matrix M = vv T projects w onto the subspace of  3 spanned by v. Specifically, show that: M(x) = av. ii. Show that a = (v w). c. (4 points) Show that for any m n matrix A, the matrix products A T A and AA T are symmetric. 1 4 2 2. (10 points) Orthogonally diagonalize A = 4 1 2. 2 2 2 1 4 8 3. (8 points) Suppose the symmetric matrix B = 4 7 4 can be orthogonally diagonalized as 8 4 1 B = 1 3 2 3 2 3 9 0 0 0 9 0 0 0 9 1 3 2 3 2 3 What is the spectral decomposition of B?

4. (12 points) Consider the quadratic form Q(x) = 5x1 2 + 20x1x2 + 2x2 2 + 16x1x3 4x2x3 + 11x3 2. a. Construct the matrix A of this quadratic form. b. Express the quadratic form of A in terms of a new variable y without using any cross-product terms. c. Classify this quadratic form. What are its principal axes? a 2 5. (5 points) For what values of a will the matrix A = be positive definite? positive semi- 2 4 definite? indefinite? Are there any values of a for which this matrix will be negative definite? 4 b 6. (5 points) For what values of b will the matrix B = be negative definite? Are there any b 2 values of b for which this matrix will be indefinite? Please explain. 7. (8 points) a. Prove that for any arbitrary m n matrix A, the matrix A T A is guaranteed to be positive semidefinite. b. Under what conditions will the product A T A of an arbitrary m n matrix A fail to be positive definite? (Characterize the matrices A for which this outcome will occur.) Please justify your answer. 1 4 8 8. (11 points) Consider the symmetric matrix B = 4 7 4 which can be orthogonally 8 4 1 diagonalized as: B = 1 3 2 3 2 3 9 0 0 0 9 0 0 0 9 a. What is the minimum of the quadratic form Q(x) of B, over all unit length vectors x? b. What is the maximum of the quadratic form Q(x) of B, over all unit length vectors x? c. For what unit vectors x will Q(x) take its maximum value? d. For what unit vectors x will Q(x) take its minimum value? 1 3 2 3 2 3 e. Find a unit vector x at which Q(x) = 0. Describe the set of unit vectors x at which Q(x) = 0.

9. (15 points) Working by hand, compute the singular value decomposition of the following matrices. (This will be good practice for the exam). "! a. A = 2 3 $ 1 1 % $ ' " 2 1 2 % # & b. B = 1 1 c. C = " 4 6 $ ' $ ' % $ # 1 1 ' # 2 2 1 & & 1 2 0 10. (8 points) Let A =. A is a linear transformation that maps vectors x in R 3 into vectors 0 2 1 b in R 2. Consider the set of all possible vectors b = Ax, where x is of unit length. What is the longest vector b in this set, and what unit length vector x is used to obtain it? 11. (10 points extra credit) The singular value decomposition of an m x n matrix A = USV T can be expressed as a sum of r distinct matrices B1.. Br, where r = min(m, n) and Bi = uisivi T, where ui is the i th column of U, si is the i th diagonal element of S and vi T is the i th row of V. If we treat an image as an m x n matrix, and decompose it using a singular value decomposition, then the sum of all of the components would be equal to the original image. But what is the result if we take the sum of only the first few components? This Matlab exercise provides an opportunity for you to explore that question. Please follow the steps below: a. Use the command A = imread('testpat1.png'); to load a test pattern image into a matrix A. A will have dimensions 512 x 512, and its entries will be 8-bit unsigned integers. Note that Matlab has many other images you could also use. b. Use the command [U, S, V] = svd(double(a)); to obtain the singular value decomposition of A, where A = USV T. Note that you have to convert the values of A from int to double before you can to apply Matlab s svd() function to it. c. In a loop of increasing values of k, for example k = 1 : 2 : 100, reconstruct a reduced version of A using only the first k components of the singular value decomposition. Specifically, you need to form the product a = usv T where u is a 512 x k matrix containing the first k columns of U, s is a k x k matrix containing the first k diagonal values in S, and v T is a k x 512 matrix containing the first k rows of V T. You can use a command like: u = U(:, i) to extract the i th column from U, and a command like: u = U(:, i:j) to extract a range of columns between the i th and the j th. Be careful to extract the correct portions of each component matrix. In your loop, use the commands figure and imshow(uint8(a)) to display the series of reconstructed images. (You need to convert the values of the matrix a back to 8-bit unsigned integers before they can be displayed as an image.) d. Report your impressions. How many components do you need to use before the image becomes recognizable? After how many components do you stop seeing much difference in the result? e. Extract and plot the singular values in matrix S. You can do this using the command: plot(1:length(diag(s)),diag(s)); Do you notice a relationship between the sizes of the singular values and the pace of improvement in the quality of the reduced image?

Please pass in: a printout of your code; your answers to the questions above, and one of the reduced images that you obtained, along with an annotation of the number of components you used to construct it (the value k).