Correlation Matrices and the Perron-Frobenius Theorem

Size: px
Start display at page:

Download "Correlation Matrices and the Perron-Frobenius Theorem"

Transcription

1 Wilfrid Laurier University July

2 Acknowledgements Thanks to David Melkuev, Johnew Zhang, Bill Zhang, Ronnnie Feng Jeyan Thangaraj for research assistance.

3 Outline The - theorem extensions to negative elements

4 1

5 This is preliminary work. Comments welcome. Markowitz efficient portfolios selected by investors These portfolios have desirable properties Mean variance efficient Sharpe used these ideas to develop the CAPM Equilibrium model relating expected return to risk Market portfolio is mean variance efficient

6 Compatible Σ, Ü (m),µ Σ covariance matrix; µ return vector, Ü (m) market weights These three entities have to be compatible since Ü (m) on frontier Best Grauer (1985) Assume Σ known ΣÜ (m) = γ 1 µ+γ 2 (1) Assume- for now- we have a way to find Ü (m) Task is to find compatible µ

7 Picking the market portfolio Origins of idea from Sharpe Ross (APT) Dominant common factor that influences stock returns PCA used to identify this factor Principal eigenvector of the correlation matrix Trzcinka (1986), Laloux et al(1999), Avellaneda Lee (2010) & Allez Bouchaud (2011) Market portfolio should have positive weights When will principal eigenvector have positive weights?

8 1 2

9 The - (-) A real n n matrix,, with positive entries has a unique largest real eigenvalue the corresponding eigenvector has strictly positive components. Provides sufficient conditions Result can be weakened Matrix can have some negative elements retain the - (PF) property. There are sometimes negative correlations between stock returns. Hence we are interested in correlation matrices which have the PF property

10 1 2 3

11 Empirical experiment We obtain CRSP daily returns for S&P1500 components from Divide data to five-year periods Select 10,000 rom samples of 50 stocks, compute C All matrices positive-definite For non-positive matrices, we study the distribution of elements Test whether the type can be determined based on simple rules for the elements

12 Changes in correlation through time Breakdown of Matrices by Type Non-PF PF w negative Positive Proportion (%) Period

13 Visualizing changes in correlation through time Daily Return sp1500_90_ Daily Return sp1500_94_ Daily Return sp1500_99_ Daily Return sp1500_04_

14 Distribution of elements Fatter left tails for Non-PF correlation matrices. E.g : Non-PF Density of elements PF

15 Count of negative elements Estimated density for the count of negative elements. E.g : Non-PF PF Count of negative elements

16 Characteristics of stock correlations Which stocks are most prevalent in Non-PF matrices? - rank stocks by average correlation with the rest (ascending) - take the top three; these are low correlation stocks Low correlation stocks appear as (almost) full rows of negative elements When these stocks are selected we end up with Non-PF matrices - Consistent with proposition on strictly negative rows (discussed later)

17 Element distribution: empirical vs simulated Statistics of negative elements Empirical Simulated Stat PF Non-PF PF Non-PF Count* N/A 612 Mean N/A Std dev N/A Min N/A % N/A % N/A % N/A Max N/A * average number per matrix

18 Visualizing negative rows Daily Return sp1500_94_ Figure: matrix visualization. 100 stocks daily returns,

19 Visualizing negative rows Different sampling frequency (e.g. weekly) can sometimes reduce the number of negative correlation Daily Return sp1500_94_ Weekly Return sp1500_94_ Figure: matrix visualization. Daily (left) returns vs weekly (right) returns

20

21 Notation PF(n) is the set of n n correlation matrices possessing the strong - property PF + (n) is the set of matrices in PF(n) that have only positive elements PF (n) is the set of matrices in PF(n) that have at least one negative element PD(n) is the set of n n positive-definite correlation matrices PF PD (n) is the set PF(n) PD(n)

22 Analytic result for three by three case Consider the 3 3 correlation matrix 1 ρ 12 ρ 13 C = ρ 12 1 ρ 23 ρ 13 ρ 23 1 C PF(3) when the following condition holds ρ 12 +ρ 13 > 0 ρ 12 +ρ 23 > 0 ρ 13 +ρ 23 > 0 When A,B PF PD (3), then convex combination always preserve the above condition PF PD (3) is convex

23 Negative row Assume C PF(n), λ υ are dominant eigenpair. Then λ > 1 υ i > 0,i = 1,...,n. For each row i, υ i + j i υ j ρ i,j = λυ i (2) υ 1 (λ 1) = j i υ j ρ i,j (3) The LHS of (3) is positive. Thus, ρ i,j cannot be all negative. In other words, PF matrix cannot have rows of only negative off diagonals!

24 Constant correlation matrices Suppose P is a correlation matrix. P has all off-diagonal entries equal to ρ Whenever ρ > 1 n 1, P PD(n) Furthermore, P PF PD (n) if only if ρ > 0

25

26 Simulate 10,000 rom positive-definite correlation matrices using Harry Joe s method (2006) Study the distribution of Non-PF PF matrices among various dimensions Within the set of PF matrices, we test convexity properties Better underst eventually positive condition

27 Simulated proportions by type Dimension PF + (n) PF (n) Non-PF % 10.04% 75.18% % 9.31% 87.40% % 5.59% 93.96% % 3.02% 96.90% % 1.55% 98.44% % 0.79% 99.21% Table: Proportion of sample correlation matrices by type from dimension 3 to 8 The set of PF matrices shrinks while matrix dimension increases The decreasing portion of positive matrices proposes a limitation of simulation method - low likelihood of getting positive matrices

28 Eventually positive matrices Definition An n n matrix A is said to be eventually positive if there exists a positive integer k 0 such that A k > 0 for all k > k 0. (Noutsos) For any symmetric n n matrix A the following properties are equivalent. 1 A possesses the strong - property. 2 A is eventually positive

29

30 Summary Interest in finding compatible Ü (m) µ Market portfolio has positive weights Proxied by dominant eigenvector of correlation matrix When does dominant eigenvector have positive weights - property Explored this question in three ways 1 Using 2 analysis 3 simulation

31 Conclusions Empirical Negative correlation has declined during last 20 years Negative correlations tend to occur in rows Has implications for PF property Analytical A row of negative correlations destroys the PF property Obtained a simple characterization of 3 3 matrices Constant correlation matrices have simple classification Simulation PF matrices rare in high dimensions for rom matrices Failure of PF related to negative elements Related to number, size position of negative elements

Identifying Financial Risk Factors

Identifying Financial Risk Factors Identifying Financial Risk Factors with a Low-Rank Sparse Decomposition Lisa Goldberg Alex Shkolnik Berkeley Columbia Meeting in Engineering and Statistics 24 March 2016 Outline 1 A Brief History of Factor

More information

Optimal Investment Strategies: A Constrained Optimization Approach

Optimal Investment Strategies: A Constrained Optimization Approach Optimal Investment Strategies: A Constrained Optimization Approach Janet L Waldrop Mississippi State University jlc3@ramsstateedu Faculty Advisor: Michael Pearson Pearson@mathmsstateedu Contents Introduction

More information

R = µ + Bf Arbitrage Pricing Model, APM

R = µ + Bf Arbitrage Pricing Model, APM 4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.

More information

Estimation of large dimensional sparse covariance matrices

Estimation of large dimensional sparse covariance matrices Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture

More information

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM. 4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage?

Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage? Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage? Stefan Thurner www.complex-systems.meduniwien.ac.at www.santafe.edu London July 28 Foundations of theory of financial

More information

Threading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection. A Physics & Pizza on Wednesdays presentation

Threading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection. A Physics & Pizza on Wednesdays presentation Threading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection A Physics & Pizza on Wednesdays presentation M. Hossein Partovi Department of Physics & Astronomy, California

More information

Estimating Covariance Using Factorial Hidden Markov Models

Estimating Covariance Using Factorial Hidden Markov Models Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide

More information

Descriptive Statistics

Descriptive Statistics Descriptive Statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Descriptive statistics Techniques to visualize

More information

Multivariate GARCH models.

Multivariate GARCH models. Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 208 LECTURE 3 need for pivoting we saw that under proper circumstances, we can write A LU where 0 0 0 u u 2 u n l 2 0 0 0 u 22 u 2n L l 3 l 32, U 0 0 0 l n l

More information

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R, Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations

More information

Theory and Applications of High Dimensional Covariance Matrix Estimation

Theory and Applications of High Dimensional Covariance Matrix Estimation 1 / 44 Theory and Applications of High Dimensional Covariance Matrix Estimation Yuan Liao Princeton University Joint work with Jianqing Fan and Martina Mincheva December 14, 2011 2 / 44 Outline 1 Applications

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Econ671 Factor Models: Principal Components

Econ671 Factor Models: Principal Components Econ671 Factor Models: Principal Components Jun YU April 8, 2016 Jun YU () Econ671 Factor Models: Principal Components April 8, 2016 1 / 59 Factor Models: Principal Components Learning Objectives 1. Show

More information

Peter S. Karlsson Jönköping University SE Jönköping Sweden

Peter S. Karlsson Jönköping University SE Jönköping Sweden Peter S. Karlsson Jönköping University SE-55 Jönköping Sweden Abstract: The Arbitrage Pricing Theory provides a theory to quantify risk and the reward for taking it. While the theory itself is sound from

More information

Modern Portfolio Theory with Homogeneous Risk Measures

Modern Portfolio Theory with Homogeneous Risk Measures Modern Portfolio Theory with Homogeneous Risk Measures Dirk Tasche Zentrum Mathematik Technische Universität München http://www.ma.tum.de/stat/ Rotterdam February 8, 2001 Abstract The Modern Portfolio

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

Mathematics for Economics and Finance

Mathematics for Economics and Finance Mathematics for Economics and Finance Michael Harrison and Patrick Waldron B 375482 Routledge Taylor & Francis Croup LONDON AND NEW YORK Contents List of figures ix List of tables xi Foreword xiii Preface

More information

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications

Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Machine Learning CSE546 Carlos Guestrin University of Washington November 13, 2014 1 E.M.: The General Case E.M. widely used beyond mixtures of Gaussians The recipe is the same

More information

Econophysics III: Financial Correlations and Portfolio Optimization

Econophysics III: Financial Correlations and Portfolio Optimization FAKULTÄT FÜR PHYSIK Econophysics III: Financial Correlations and Portfolio Optimization Thomas Guhr Let s Face Chaos through Nonlinear Dynamics, Maribor 21 Outline Portfolio optimization is a key issue

More information

1 Principal component analysis and dimensional reduction

1 Principal component analysis and dimensional reduction Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an

More information

= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E.

= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E. 3.3 Diagonalization Let A = 4. Then and are eigenvectors of A, with corresponding eigenvalues 2 and 6 respectively (check). This means 4 = 2, 4 = 6. 2 2 2 2 Thus 4 = 2 2 6 2 = 2 6 4 2 We have 4 = 2 0 0

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

Regularization of Portfolio Allocation

Regularization of Portfolio Allocation Regularization of Portfolio Allocation Benjamin Bruder Quantitative Research Lyxor Asset Management, Paris benjamin.bruder@lyxor.com Jean-Charles Richard Quantitative Research Lyxor Asset Management, Paris

More information

The Cross-Section of Positively Weighted Portfolios

The Cross-Section of Positively Weighted Portfolios The Cross-Section of Positively Weighted Portfolios Daniel Niedermayer Heinz Zimmermann January 11, 2007 Abstract This paper examines properties of mean-variance inefficient proxies with respect to producing

More information

Principal component analysis (PCA) for clustering gene expression data

Principal component analysis (PCA) for clustering gene expression data Principal component analysis (PCA) for clustering gene expression data Ka Yee Yeung Walter L. Ruzzo Bioinformatics, v17 #9 (2001) pp 763-774 1 Outline of talk Background and motivation Design of our empirical

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Jeffrey D. Ullman Stanford University

Jeffrey D. Ullman Stanford University Jeffrey D. Ullman Stanford University 2 Often, our data can be represented by an m-by-n matrix. And this matrix can be closely approximated by the product of two matrices that share a small common dimension

More information

Can Random Matrix Theory resolve Markowitz Optimization Enigma? The impact of noise filtered covariance matrix on portfolio selection.

Can Random Matrix Theory resolve Markowitz Optimization Enigma? The impact of noise filtered covariance matrix on portfolio selection. Can Random Matrix Theory resolve Markowitz Optimization Enigma? The impact of noise filtered covariance matrix on portfolio selection. by Kim Wah Ng A Dissertation submitted to the Graduate School-Newark

More information

This paper develops a test of the asymptotic arbitrage pricing theory (APT) via the maximum squared Sharpe

This paper develops a test of the asymptotic arbitrage pricing theory (APT) via the maximum squared Sharpe MANAGEMENT SCIENCE Vol. 55, No. 7, July 2009, pp. 1255 1266 issn 0025-1909 eissn 1526-5501 09 5507 1255 informs doi 10.1287/mnsc.1090.1004 2009 INFORMS Testing the APT with the Maximum Sharpe Ratio of

More information

Unsupervised Learning: Dimensionality Reduction

Unsupervised Learning: Dimensionality Reduction Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,

More information

Dimension. Eigenvalue and eigenvector

Dimension. Eigenvalue and eigenvector Dimension. Eigenvalue and eigenvector Math 112, week 9 Goals: Bases, dimension, rank-nullity theorem. Eigenvalue and eigenvector. Suggested Textbook Readings: Sections 4.5, 4.6, 5.1, 5.2 Week 9: Dimension,

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits

More information

Sparse PCA with applications in finance

Sparse PCA with applications in finance Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction

More information

Linear Programming: Chapter 1 Introduction

Linear Programming: Chapter 1 Introduction Linear Programming: Chapter 1 Introduction Robert J. Vanderbei October 17, 2007 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544 http://www.princeton.edu/ rvdb Resource

More information

A New Method to Find the Eigenvalues of Convex. Matrices with Application in Web Page Rating

A New Method to Find the Eigenvalues of Convex. Matrices with Application in Web Page Rating Applied Mathematical Sciences, Vol. 4, 200, no. 9, 905-9 A New Method to Find the Eigenvalues of Convex Matrices with Application in Web Page Rating F. Soleymani Department of Mathematics, Islamic Azad

More information

Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations

Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations Sahand Rabbani Introduction Modern portfolio theory deals in part with the efficient allocation of investments

More information

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = =

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = = 18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013 1. Consider a bivariate random variable: [ ] X X = 1 X 2 with mean and co [ ] variance: [ ] [ α1 Σ 1,1 Σ 1,2 σ 2 ρσ 1 σ E[X] =, and Cov[X]

More information

Financial Econometrics and Volatility Models Copulas

Financial Econometrics and Volatility Models Copulas Financial Econometrics and Volatility Models Copulas Eric Zivot Updated: May 10, 2010 Reading MFTS, chapter 19 FMUND, chapters 6 and 7 Introduction Capturing co-movement between financial asset returns

More information

Theorems. Least squares regression

Theorems. Least squares regression Theorems In this assignment we are trying to classify AML and ALL samples by use of penalized logistic regression. Before we indulge on the adventure of classification we should first explain the most

More information

Math 315: Linear Algebra Solutions to Assignment 7

Math 315: Linear Algebra Solutions to Assignment 7 Math 5: Linear Algebra s to Assignment 7 # Find the eigenvalues of the following matrices. (a.) 4 0 0 0 (b.) 0 0 9 5 4. (a.) The characteristic polynomial det(λi A) = (λ )(λ )(λ ), so the eigenvalues are

More information

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis .. December 20, 2013 Todays lecture. (PCA) (PLS-R) (LDA) . (PCA) is a method often used to reduce the dimension of a large dataset to one of a more manageble size. The new dataset can then be used to make

More information

On-line Variance Minimization

On-line Variance Minimization On-line Variance Minimization Manfred Warmuth Dima Kuzmin University of California - Santa Cruz 19th Annual Conference on Learning Theory M. Warmuth, D. Kuzmin (UCSC) On-line Variance Minimization COLT06

More information

Introduction Eigen Values and Eigen Vectors An Application Matrix Calculus Optimal Portfolio. Portfolios. Christopher Ting.

Introduction Eigen Values and Eigen Vectors An Application Matrix Calculus Optimal Portfolio. Portfolios. Christopher Ting. Portfolios Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 November 4, 2016 Christopher Ting QF 101 Week 12 November 4,

More information

Applications of Random Matrix Theory to Economics, Finance and Political Science

Applications of Random Matrix Theory to Economics, Finance and Political Science Outline Applications of Random Matrix Theory to Economics, Finance and Political Science Matthew C. 1 1 Department of Economics, MIT Institute for Quantitative Social Science, Harvard University SEA 06

More information

Math Matrix Algebra

Math Matrix Algebra Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI? Section 5. The Characteristic Polynomial Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI? Property The eigenvalues

More information

Eigen Portfolio Selection: A Robust Approach to Sharpe. Ratio Maximization

Eigen Portfolio Selection: A Robust Approach to Sharpe. Ratio Maximization Eigen Portfolio Selection: A Robust Approach to Sharpe Ratio Maximization Danqiao Guo 1, Phelim Boyle 2, Chengguo Weng 1, and Tony Wirjanto 1 1 Department of Statistics & Actuarial Science, University

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Prior Information: Shrinkage and Black-Litterman. Prof. Daniel P. Palomar

Prior Information: Shrinkage and Black-Litterman. Prof. Daniel P. Palomar Prior Information: Shrinkage and Black-Litterman Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics

More information

Learning representations

Learning representations Learning representations Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 4/11/2016 General problem For a dataset of n signals X := [ x 1 x

More information

FINM 331: MULTIVARIATE DATA ANALYSIS FALL 2017 PROBLEM SET 3

FINM 331: MULTIVARIATE DATA ANALYSIS FALL 2017 PROBLEM SET 3 FINM 331: MULTIVARIATE DATA ANALYSIS FALL 2017 PROBLEM SET 3 The required files for all problems can be found in: http://www.stat.uchicago.edu/~lekheng/courses/331/hw3/ The file name indicates which problem

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

Math 180B Problem Set 3

Math 180B Problem Set 3 Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P

More information

First of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if

First of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if 5 Matrices in Different Coordinates In this section we discuss finding matrices of linear maps in different coordinates Earlier in the class was the matrix that multiplied by x to give ( x) in standard

More information

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1 Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of

More information

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes Week 2 Based in part on slides from textbook, slides of Susan Holmes Part I Other datatypes, preprocessing October 3, 2012 1 / 1 2 / 1 Other datatypes Other datatypes Document data You might start with

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Manifold Learning for Signal and Visual Processing Lecture 9: Probabilistic PCA (PPCA), Factor Analysis, Mixtures of PPCA

Manifold Learning for Signal and Visual Processing Lecture 9: Probabilistic PCA (PPCA), Factor Analysis, Mixtures of PPCA Manifold Learning for Signal and Visual Processing Lecture 9: Probabilistic PCA (PPCA), Factor Analysis, Mixtures of PPCA Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/

More information

LECTURE 16: PCA AND SVD

LECTURE 16: PCA AND SVD Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal

More information

Section 4.4 Reduction to Symmetric Tridiagonal Form

Section 4.4 Reduction to Symmetric Tridiagonal Form Section 4.4 Reduction to Symmetric Tridiagonal Form Key terms Symmetric matrix conditioning Tridiagonal matrix Similarity transformation Orthogonal matrix Orthogonal similarity transformation properties

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors 5 Eigenvalues and Eigenvectors 5.2 THE CHARACTERISTIC EQUATION DETERMINANATS n n Let A be an matrix, let U be any echelon form obtained from A by row replacements and row interchanges (without scaling),

More information

Financial Econometrics Lecture 6: Testing the CAPM model

Financial Econometrics Lecture 6: Testing the CAPM model Financial Econometrics Lecture 6: Testing the CAPM model Richard G. Pierse 1 Introduction The capital asset pricing model has some strong implications which are testable. The restrictions that can be tested

More information

Markov Chains Handout for Stat 110

Markov Chains Handout for Stat 110 Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of

More information

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory

More information

Feature Engineering, Model Evaluations

Feature Engineering, Model Evaluations Feature Engineering, Model Evaluations Giri Iyengar Cornell University gi43@cornell.edu Feb 5, 2018 Giri Iyengar (Cornell Tech) Feature Engineering Feb 5, 2018 1 / 35 Overview 1 ETL 2 Feature Engineering

More information

Definition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices).

Definition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices). Matrices (general theory). Definition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices). Examples. 1 2 1 1 0 2 A= 0 0 7 B= 0 1 3 4 5 0 Terminology and Notations. Each

More information

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can

More information

Numerical Methods I Singular Value Decomposition

Numerical Methods I Singular Value Decomposition Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)

More information

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics Diagonalization MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Motivation Today we consider two fundamental questions: Given an n n matrix A, does there exist a basis

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

Optimal Covariance Matrix with Parameter Uncertainty for the Tangency Portfolio

Optimal Covariance Matrix with Parameter Uncertainty for the Tangency Portfolio Optimal Covariance Matrix with Parameter Uncertainty for the Tangency Portfolio Ghislain Yanou June 3, 2009 Abstract We provide a model for understanding the impact of the sample size neglect when an investor,

More information

Analysis of Spectral Kernel Design based Semi-supervised Learning

Analysis of Spectral Kernel Design based Semi-supervised Learning Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,

More information

1 Principal Components Analysis

1 Principal Components Analysis Lecture 3 and 4 Sept. 18 and Sept.20-2006 Data Visualization STAT 442 / 890, CM 462 Lecture: Ali Ghodsi 1 Principal Components Analysis Principal components analysis (PCA) is a very popular technique for

More information

ARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES

ARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES Advances in Complex Systems, Vol. 11, No. 5 (2008) 655 668 c World Scientific Publishing Company ARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES RICARDO COELHO,PETERRICHMONDandSTEFANHUTZLER

More information

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and 7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =

More information

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38 Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters

More information

Linear Programming: Chapter 1 Introduction

Linear Programming: Chapter 1 Introduction Linear Programming: Chapter 1 Introduction Robert J. Vanderbei September 16, 2010 Slides last edited on October 5, 2010 Operations Research and Financial Engineering Princeton University Princeton, NJ

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University

More information

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to: System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

CSE 554 Lecture 7: Alignment

CSE 554 Lecture 7: Alignment CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex

More information

Study of Topological Surfaces

Study of Topological Surfaces Study of Topological Surfaces Jamil Mortada Abstract This research project involves studying different topological surfaces, in particular, continuous, invertible functions from a surface to itself. Dealing

More information