The Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transfor

Similar documents
However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

Composite Hypotheses testing

[ ] λ λ λ. Multicollinearity. multicollinearity Ragnar Frisch (1934) perfect exact. collinearity. multicollinearity. exact

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

LECTURE 9 CANONICAL CORRELATION ANALYSIS

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1.

Multi-dimensional Central Limit Theorem

be a second-order and mean-value-zero vector-valued process, i.e., for t E

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Dynamic Systems on Graphs

Fisher Linear Discriminant Analysis

2.3 Nilpotent endomorphisms

= = = (a) Use the MATLAB command rref to solve the system. (b) Let A be the coefficient matrix and B be the right-hand side of the system.

Differentiating Gaussian Processes

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Statistical pattern recognition

Quantum Mechanics for Scientists and Engineers. David Miller

e - c o m p a n i o n

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Perron Vectors of an Irreducible Nonnegative Interval Matrix

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

Effects of Ignoring Correlations When Computing Sample Chi-Square. John W. Fowler February 26, 2012

PHYS 705: Classical Mechanics. Calculus of Variations II

Eigenvalues of Random Graphs

Quantum Mechanics I - Session 4

On the symmetric character of the thermal conductivity tensor

Radar Trackers. Study Guide. All chapters, problems, examples and page numbers refer to Applied Optimal Estimation, A. Gelb, Ed.

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Lecture 6/7 (February 10/12, 2014) DIRAC EQUATION. The non-relativistic Schrödinger equation was obtained by noting that the Hamiltonian 2

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

Some Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to

Unified Subspace Analysis for Face Recognition

Solutions Homework 4 March 5, 2018

The Order Relation and Trace Inequalities for. Hermitian Operators

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Lecture 10 Support Vector Machines II

MATH 241B FUNCTIONAL ANALYSIS - NOTES EXAMPLES OF C ALGEBRAS

Singular Value Decomposition: Theory and Applications

MATH Sensitivity of Eigenvalue Problems

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

MIMO Systems and Channel Capacity

APPENDIX A Some Linear Algebra

The exam is closed book, closed notes except your one-page cheat sheet.

Linear Approximation with Regularization and Moving Least Squares

A Local Variational Problem of Second Order for a Class of Optimal Control Problems with Nonsmooth Objective Function

CRITICAL POINT ANALYSIS OF JOINT DIAGONALIZATION CRITERIA. Gen Hori and Jonathan H. Manton

Multi-dimensional Central Limit Argument

COS 521: Advanced Algorithms Game Theory and Linear Programming

arxiv: v2 [quant-ph] 29 Jun 2018

This model contains two bonds per unit cell (one along the x-direction and the other along y). So we can rewrite the Hamiltonian as:

Solutions to Problem Set 6

Chapter 7 Channel Capacity and Coding

Non-linear Canonical Correlation Analysis Using a RBF Network

The lower and upper bounds on Perron root of nonnegative irreducible matrices

1 Convex Optimization

p(z) = 1 a e z/a 1(z 0) yi a i x (1/a) exp y i a i x a i=1 n i=1 (y i a i x) inf 1 (y Ax) inf Ax y (1 ν) y if A (1 ν) = 0 otherwise

Lecture 3. Ax x i a i. i i

Supporting Information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

ψ = i c i u i c i a i b i u i = i b 0 0 b 0 0

Statistics and Probability Theory in Civil, Surveying and Environmental Engineering

Automatic Object Trajectory- Based Motion Recognition Using Gaussian Mixture Models

Feb 14: Spatial analysis of data fields

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)

ECE559VV Project Report

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

6) Derivatives, gradients and Hessian matrices

Tensor Analysis. For orthogonal curvilinear coordinates, ˆ ˆ (98) Expanding the derivative, we have, ˆ. h q. . h q h q

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

CSCE 790S Background Results

Propagation of error for multivariable function

Ph 219a/CS 219a. Exercises Due: Wednesday 23 October 2013

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

MATH Homework #2

Machine Learning for Signal Processing

A Comparative Study for Estimation Parameters in Panel Data Model

PHYS 215C: Quantum Mechanics (Spring 2017) Problem Set 3 Solutions

DISCRIMINANTS AND RAMIFIED PRIMES. 1. Introduction A prime number p is said to be ramified in a number field K if the prime ideal factorization

Chapter Newton s Method

Chapter 7 Channel Capacity and Coding

(A and B must have the same dmensons to be able to add them together.) Addton s commutatve and assocatve, just lke regular addton. A matrx A multpled

Lecture 3: Probability Distributions

Outline. Multivariate Parametric Methods. Multivariate Data. Basic Multivariate Statistics. Steven J Zeil

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

Error Bars in both X and Y

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

Probability Theory (revisited)

Errata to Invariant Theory with Applications January 28, 2017

14 Lagrange Multipliers

Transcription:

Prncpal Component Transform Multvarate Random Sgnals A real tme sgnal x(t can be consdered as a random process and ts samples x m (m =0; ;N, 1 a random vector: The mean vector of X s X =[x0; ;x N,1] T M X = E(X =[E(x0; ;E(x N,1] T =[0; ; N,1] T The covarance matrx of X s X = E[(X, M X (X, M X T ]=E(XX T, MM T = :: j :: where j = E(x x j, j s the covarance of two random varables x and x j. When = j, j becomes the varance of x, = E(x,. The correlaton matrx of X s R X = E(XX T = :: r j :: where r j = j + j. Both X and R X are symmetrc matrces (Hermtan f X s complex. A sgnal vector X can always be easly converted nto a zero-mean vector X 0 = X, M X wth all of ts nformaton (or dynamc energy conserved. In the followng, wthout loss of generalty, we wll assume M X = 0 and therefore X = R X. 7 7 1

The Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transform. Let and be the th egenvector and egenvalue of the correlaton matrx R X : R X = ( =0; ;N, 1 We can construct an N N matrx =[0; ; N,1] Snce the columns of are the egenvectors of a symmetrc (Hermtan f X s complex matrx R X,s orthogonal (untary: T =I.e., and we have,1 = T R X = where =dag(0; ; N,1. Or, we have,1 R X = T R X = We can now dene the orthogonal (untary f X s complex Prncpal Component Transform of X by ( Y = T X X =Y The th component of the forward transform Y = T X s the projecton of X on : y =( ;X= T X and the nverse transform X = Y represents X n the N-dmensonal space spanned by ( =0; 1; ;N, 1: N,1 X X = y

KLT Completely Decorrelates the Sgnal KLT s the optmal orthogonal transform n the followng sense: KLT completely decorrelates the sgnal KLT optmally compacts the energy (nformaton contaned n the sgnal. The rst property s smply due to the denton of KLT, and the second property s due to the fact that KLT redstrbutes the energy among the N components n such a way that most of the energy s contaned n a small number of components of Y = T X. To see the rst property, consder the correlaton matrx R Y of Y : R Y = E(YY T =E[ T X( T X T ] = E[ T (XX T ]= T E(XX T = T R X = We see that after KLT, the correlaton matrx of the sgnal s dagonalzed,.e., the correlaton r j =0between any two components x and x j s always zero. In other words, the sgnal s completely decorrelated.

KLT Optmally Compacts the Energy Consder a general orthogonal transform par dened as ( Y = A T X X = AY where X and Y arenby1vectors and A s an arbtrary N by N orthogonal matrx A,1 = A T. We represent A by ts column vectors A ; ( =0; ;N, 1 as or A =[A0; ;A N,1] A T = A T 0 : : A T N,1 Now the th component of Y can be wrtten as 7 y = A T X As we assume the mean vector of X s zero M X = 0 (and obvously we also have M Y = A T M x = 0, we have X = R X, and the varance of the th element nbothx and Y are x = E(x = E(e x and y = E(y = E(e y where e x = x and e y = y represent the energy contaned n the th component of X and Y, respectvely. In order words, the trace of X (the sum of all the dagonal elements of the matrx represents the expectaton of the total amount of energy contaned n the sgnal X N,1 X Total energy contaned n X = tr X = N,1 X x = E(x N,1 X =E( e x

Snce an orthogonal transform A does not change the length of a vector X,.e., k Y k=k AX k=k X k, where vu u X vu u X k X k= t N,1 x = t N,1 the total energy contaned n the sgnal vector X s conserved after the orthogonal transform. (Ths concluson can also be obtaned from the fact that orthogonal transforms do not change the trace of a matrx. We next dene S m (A = E(y = y = E(e y where m N. S m (A s a functon of the transform matrx A and represents the amount of energy contaned n the rst m components of Y = A T X. Snce the total energy s conserved, S m (A also represents the percentage of energy contaned n the rst m components. In the followng we wll show that S m (A s maxmzed f and only f the transform A s the KLT: S m (A = S m ( S m (A.e., KLT optmally compacts energy nto a few components of the sgnal. Consder = = e x E(y = E[A T X(AT XT ] E[A T X(X T A ] = A T E(XXT A A T R XA (1 Now we need to nd a transform matrx A so that ( Sm (A! max subject to A T j A j =1 (j =0; ;m, 1 The constrant A T j A j = 1 s to guarantee that the column vectors n A are normalzed. Ths constraned optmzaton problem can be solved by Lagrange multpler method as shown below.

We let @ @A [S m (A, j (A T j A j, 1] = 0 j=0 = @ [ (A T j @A R XA j, j A T j A j + j ] j=0 = @ [A T @A R XA, A T A ] = R x A, A =0 (* the last equal sgn s due to explanaton n the handout of revew of lnear algebra. We see that the column vectors of A must be the egenvectors of R X : R X A = A ( =0; ;m, 1.e., the transform matrx must be A =[A0; ;A N,1] ==[0; ; N,1] Thus we have proved that the optmal transform s ndeed KLT, and S m ( = T R X = where the th egenvalue of R X s also the average (expectaton energy contaned n the th component of the sgnal. If we choose those 0 s that correspond to the m largest egenvalues of R X : 0 1 m N,1, then S m ( wll acheve maxmum.