Correlation Matrices and the Perron-Frobenius Theorem
|
|
- Elinor Shelton
- 5 years ago
- Views:
Transcription
1 Wilfrid Laurier University July
2 Acknowledgements Thanks to David Melkuev, Johnew Zhang, Bill Zhang, Ronnnie Feng Jeyan Thangaraj for research assistance.
3 Outline The - theorem extensions to negative elements
4 1
5 This is preliminary work. Comments welcome. Markowitz efficient portfolios selected by investors These portfolios have desirable properties Mean variance efficient Sharpe used these ideas to develop the CAPM Equilibrium model relating expected return to risk Market portfolio is mean variance efficient
6 Compatible Σ, Ü (m),µ Σ covariance matrix; µ return vector, Ü (m) market weights These three entities have to be compatible since Ü (m) on frontier Best Grauer (1985) Assume Σ known ΣÜ (m) = γ 1 µ+γ 2 (1) Assume- for now- we have a way to find Ü (m) Task is to find compatible µ
7 Picking the market portfolio Origins of idea from Sharpe Ross (APT) Dominant common factor that influences stock returns PCA used to identify this factor Principal eigenvector of the correlation matrix Trzcinka (1986), Laloux et al(1999), Avellaneda Lee (2010) & Allez Bouchaud (2011) Market portfolio should have positive weights When will principal eigenvector have positive weights?
8 1 2
9 The - (-) A real n n matrix,, with positive entries has a unique largest real eigenvalue the corresponding eigenvector has strictly positive components. Provides sufficient conditions Result can be weakened Matrix can have some negative elements retain the - (PF) property. There are sometimes negative correlations between stock returns. Hence we are interested in correlation matrices which have the PF property
10 1 2 3
11 Empirical experiment We obtain CRSP daily returns for S&P1500 components from Divide data to five-year periods Select 10,000 rom samples of 50 stocks, compute C All matrices positive-definite For non-positive matrices, we study the distribution of elements Test whether the type can be determined based on simple rules for the elements
12 Changes in correlation through time Breakdown of Matrices by Type Non-PF PF w negative Positive Proportion (%) Period
13 Visualizing changes in correlation through time Daily Return sp1500_90_ Daily Return sp1500_94_ Daily Return sp1500_99_ Daily Return sp1500_04_
14 Distribution of elements Fatter left tails for Non-PF correlation matrices. E.g : Non-PF Density of elements PF
15 Count of negative elements Estimated density for the count of negative elements. E.g : Non-PF PF Count of negative elements
16 Characteristics of stock correlations Which stocks are most prevalent in Non-PF matrices? - rank stocks by average correlation with the rest (ascending) - take the top three; these are low correlation stocks Low correlation stocks appear as (almost) full rows of negative elements When these stocks are selected we end up with Non-PF matrices - Consistent with proposition on strictly negative rows (discussed later)
17 Element distribution: empirical vs simulated Statistics of negative elements Empirical Simulated Stat PF Non-PF PF Non-PF Count* N/A 612 Mean N/A Std dev N/A Min N/A % N/A % N/A % N/A Max N/A * average number per matrix
18 Visualizing negative rows Daily Return sp1500_94_ Figure: matrix visualization. 100 stocks daily returns,
19 Visualizing negative rows Different sampling frequency (e.g. weekly) can sometimes reduce the number of negative correlation Daily Return sp1500_94_ Weekly Return sp1500_94_ Figure: matrix visualization. Daily (left) returns vs weekly (right) returns
20
21 Notation PF(n) is the set of n n correlation matrices possessing the strong - property PF + (n) is the set of matrices in PF(n) that have only positive elements PF (n) is the set of matrices in PF(n) that have at least one negative element PD(n) is the set of n n positive-definite correlation matrices PF PD (n) is the set PF(n) PD(n)
22 Analytic result for three by three case Consider the 3 3 correlation matrix 1 ρ 12 ρ 13 C = ρ 12 1 ρ 23 ρ 13 ρ 23 1 C PF(3) when the following condition holds ρ 12 +ρ 13 > 0 ρ 12 +ρ 23 > 0 ρ 13 +ρ 23 > 0 When A,B PF PD (3), then convex combination always preserve the above condition PF PD (3) is convex
23 Negative row Assume C PF(n), λ υ are dominant eigenpair. Then λ > 1 υ i > 0,i = 1,...,n. For each row i, υ i + j i υ j ρ i,j = λυ i (2) υ 1 (λ 1) = j i υ j ρ i,j (3) The LHS of (3) is positive. Thus, ρ i,j cannot be all negative. In other words, PF matrix cannot have rows of only negative off diagonals!
24 Constant correlation matrices Suppose P is a correlation matrix. P has all off-diagonal entries equal to ρ Whenever ρ > 1 n 1, P PD(n) Furthermore, P PF PD (n) if only if ρ > 0
25
26 Simulate 10,000 rom positive-definite correlation matrices using Harry Joe s method (2006) Study the distribution of Non-PF PF matrices among various dimensions Within the set of PF matrices, we test convexity properties Better underst eventually positive condition
27 Simulated proportions by type Dimension PF + (n) PF (n) Non-PF % 10.04% 75.18% % 9.31% 87.40% % 5.59% 93.96% % 3.02% 96.90% % 1.55% 98.44% % 0.79% 99.21% Table: Proportion of sample correlation matrices by type from dimension 3 to 8 The set of PF matrices shrinks while matrix dimension increases The decreasing portion of positive matrices proposes a limitation of simulation method - low likelihood of getting positive matrices
28 Eventually positive matrices Definition An n n matrix A is said to be eventually positive if there exists a positive integer k 0 such that A k > 0 for all k > k 0. (Noutsos) For any symmetric n n matrix A the following properties are equivalent. 1 A possesses the strong - property. 2 A is eventually positive
29
30 Summary Interest in finding compatible Ü (m) µ Market portfolio has positive weights Proxied by dominant eigenvector of correlation matrix When does dominant eigenvector have positive weights - property Explored this question in three ways 1 Using 2 analysis 3 simulation
31 Conclusions Empirical Negative correlation has declined during last 20 years Negative correlations tend to occur in rows Has implications for PF property Analytical A row of negative correlations destroys the PF property Obtained a simple characterization of 3 3 matrices Constant correlation matrices have simple classification Simulation PF matrices rare in high dimensions for rom matrices Failure of PF related to negative elements Related to number, size position of negative elements
Identifying Financial Risk Factors
Identifying Financial Risk Factors with a Low-Rank Sparse Decomposition Lisa Goldberg Alex Shkolnik Berkeley Columbia Meeting in Engineering and Statistics 24 March 2016 Outline 1 A Brief History of Factor
More informationOptimal Investment Strategies: A Constrained Optimization Approach
Optimal Investment Strategies: A Constrained Optimization Approach Janet L Waldrop Mississippi State University jlc3@ramsstateedu Faculty Advisor: Michael Pearson Pearson@mathmsstateedu Contents Introduction
More informationR = µ + Bf Arbitrage Pricing Model, APM
4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.
More informationEstimation of large dimensional sparse covariance matrices
Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional
More informationLecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides
Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture
More informationRoss (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.
4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationEigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage?
Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage? Stefan Thurner www.complex-systems.meduniwien.ac.at www.santafe.edu London July 28 Foundations of theory of financial
More informationThreading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection. A Physics & Pizza on Wednesdays presentation
Threading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection A Physics & Pizza on Wednesdays presentation M. Hossein Partovi Department of Physics & Astronomy, California
More informationEstimating Covariance Using Factorial Hidden Markov Models
Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationEcon 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis
Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide
More informationDescriptive Statistics
Descriptive Statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Descriptive statistics Techniques to visualize
More informationMultivariate GARCH models.
Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains
More informationPrincipal Components Analysis (PCA)
Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13
STAT 309: MATHEMATICAL COMPUTATIONS I FALL 208 LECTURE 3 need for pivoting we saw that under proper circumstances, we can write A LU where 0 0 0 u u 2 u n l 2 0 0 0 u 22 u 2n L l 3 l 32, U 0 0 0 l n l
More informationPrincipal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,
Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations
More informationTheory and Applications of High Dimensional Covariance Matrix Estimation
1 / 44 Theory and Applications of High Dimensional Covariance Matrix Estimation Yuan Liao Princeton University Joint work with Jianqing Fan and Martina Mincheva December 14, 2011 2 / 44 Outline 1 Applications
More informationData Preprocessing Tasks
Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can
More informationEcon671 Factor Models: Principal Components
Econ671 Factor Models: Principal Components Jun YU April 8, 2016 Jun YU () Econ671 Factor Models: Principal Components April 8, 2016 1 / 59 Factor Models: Principal Components Learning Objectives 1. Show
More informationPeter S. Karlsson Jönköping University SE Jönköping Sweden
Peter S. Karlsson Jönköping University SE-55 Jönköping Sweden Abstract: The Arbitrage Pricing Theory provides a theory to quantify risk and the reward for taking it. While the theory itself is sound from
More informationModern Portfolio Theory with Homogeneous Risk Measures
Modern Portfolio Theory with Homogeneous Risk Measures Dirk Tasche Zentrum Mathematik Technische Universität München http://www.ma.tum.de/stat/ Rotterdam February 8, 2001 Abstract The Modern Portfolio
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More informationMathematics for Economics and Finance
Mathematics for Economics and Finance Michael Harrison and Patrick Waldron B 375482 Routledge Taylor & Francis Croup LONDON AND NEW YORK Contents List of figures ix List of tables xi Foreword xiii Preface
More informationRandom Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications
Random Correlation Matrices, Top Eigenvalue with Heavy Tails and Financial Applications J.P Bouchaud with: M. Potters, G. Biroli, L. Laloux, M. A. Miceli http://www.cfm.fr Portfolio theory: Basics Portfolio
More informationExpectation Maximization
Expectation Maximization Machine Learning CSE546 Carlos Guestrin University of Washington November 13, 2014 1 E.M.: The General Case E.M. widely used beyond mixtures of Gaussians The recipe is the same
More informationEconophysics III: Financial Correlations and Portfolio Optimization
FAKULTÄT FÜR PHYSIK Econophysics III: Financial Correlations and Portfolio Optimization Thomas Guhr Let s Face Chaos through Nonlinear Dynamics, Maribor 21 Outline Portfolio optimization is a key issue
More information1 Principal component analysis and dimensional reduction
Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an
More information= main diagonal, in the order in which their corresponding eigenvectors appear as columns of E.
3.3 Diagonalization Let A = 4. Then and are eigenvectors of A, with corresponding eigenvalues 2 and 6 respectively (check). This means 4 = 2, 4 = 6. 2 2 2 2 Thus 4 = 2 2 6 2 = 2 6 4 2 We have 4 = 2 0 0
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationRegularization of Portfolio Allocation
Regularization of Portfolio Allocation Benjamin Bruder Quantitative Research Lyxor Asset Management, Paris benjamin.bruder@lyxor.com Jean-Charles Richard Quantitative Research Lyxor Asset Management, Paris
More informationThe Cross-Section of Positively Weighted Portfolios
The Cross-Section of Positively Weighted Portfolios Daniel Niedermayer Heinz Zimmermann January 11, 2007 Abstract This paper examines properties of mean-variance inefficient proxies with respect to producing
More informationPrincipal component analysis (PCA) for clustering gene expression data
Principal component analysis (PCA) for clustering gene expression data Ka Yee Yeung Walter L. Ruzzo Bioinformatics, v17 #9 (2001) pp 763-774 1 Outline of talk Background and motivation Design of our empirical
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationJeffrey D. Ullman Stanford University
Jeffrey D. Ullman Stanford University 2 Often, our data can be represented by an m-by-n matrix. And this matrix can be closely approximated by the product of two matrices that share a small common dimension
More informationCan Random Matrix Theory resolve Markowitz Optimization Enigma? The impact of noise filtered covariance matrix on portfolio selection.
Can Random Matrix Theory resolve Markowitz Optimization Enigma? The impact of noise filtered covariance matrix on portfolio selection. by Kim Wah Ng A Dissertation submitted to the Graduate School-Newark
More informationThis paper develops a test of the asymptotic arbitrage pricing theory (APT) via the maximum squared Sharpe
MANAGEMENT SCIENCE Vol. 55, No. 7, July 2009, pp. 1255 1266 issn 0025-1909 eissn 1526-5501 09 5507 1255 informs doi 10.1287/mnsc.1090.1004 2009 INFORMS Testing the APT with the Maximum Sharpe Ratio of
More informationUnsupervised Learning: Dimensionality Reduction
Unsupervised Learning: Dimensionality Reduction CMPSCI 689 Fall 2015 Sridhar Mahadevan Lecture 3 Outline In this lecture, we set about to solve the problem posed in the previous lecture Given a dataset,
More informationDimension. Eigenvalue and eigenvector
Dimension. Eigenvalue and eigenvector Math 112, week 9 Goals: Bases, dimension, rank-nullity theorem. Eigenvalue and eigenvector. Suggested Textbook Readings: Sections 4.5, 4.6, 5.1, 5.2 Week 9: Dimension,
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationLEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach
LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits
More informationSparse PCA with applications in finance
Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction
More informationLinear Programming: Chapter 1 Introduction
Linear Programming: Chapter 1 Introduction Robert J. Vanderbei October 17, 2007 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544 http://www.princeton.edu/ rvdb Resource
More informationA New Method to Find the Eigenvalues of Convex. Matrices with Application in Web Page Rating
Applied Mathematical Sciences, Vol. 4, 200, no. 9, 905-9 A New Method to Find the Eigenvalues of Convex Matrices with Application in Web Page Rating F. Soleymani Department of Mathematics, Islamic Azad
More informationMarkowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations
Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations Sahand Rabbani Introduction Modern portfolio theory deals in part with the efficient allocation of investments
More information18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = =
18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013 1. Consider a bivariate random variable: [ ] X X = 1 X 2 with mean and co [ ] variance: [ ] [ α1 Σ 1,1 Σ 1,2 σ 2 ρσ 1 σ E[X] =, and Cov[X]
More informationFinancial Econometrics and Volatility Models Copulas
Financial Econometrics and Volatility Models Copulas Eric Zivot Updated: May 10, 2010 Reading MFTS, chapter 19 FMUND, chapters 6 and 7 Introduction Capturing co-movement between financial asset returns
More informationTheorems. Least squares regression
Theorems In this assignment we are trying to classify AML and ALL samples by use of penalized logistic regression. Before we indulge on the adventure of classification we should first explain the most
More informationMath 315: Linear Algebra Solutions to Assignment 7
Math 5: Linear Algebra s to Assignment 7 # Find the eigenvalues of the following matrices. (a.) 4 0 0 0 (b.) 0 0 9 5 4. (a.) The characteristic polynomial det(λi A) = (λ )(λ )(λ ), so the eigenvalues are
More informationDecember 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis
.. December 20, 2013 Todays lecture. (PCA) (PLS-R) (LDA) . (PCA) is a method often used to reduce the dimension of a large dataset to one of a more manageble size. The new dataset can then be used to make
More informationOn-line Variance Minimization
On-line Variance Minimization Manfred Warmuth Dima Kuzmin University of California - Santa Cruz 19th Annual Conference on Learning Theory M. Warmuth, D. Kuzmin (UCSC) On-line Variance Minimization COLT06
More informationIntroduction Eigen Values and Eigen Vectors An Application Matrix Calculus Optimal Portfolio. Portfolios. Christopher Ting.
Portfolios Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 November 4, 2016 Christopher Ting QF 101 Week 12 November 4,
More informationApplications of Random Matrix Theory to Economics, Finance and Political Science
Outline Applications of Random Matrix Theory to Economics, Finance and Political Science Matthew C. 1 1 Department of Economics, MIT Institute for Quantitative Social Science, Harvard University SEA 06
More informationMath Matrix Algebra
Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationQuestion: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?
Section 5. The Characteristic Polynomial Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI? Property The eigenvalues
More informationEigen Portfolio Selection: A Robust Approach to Sharpe. Ratio Maximization
Eigen Portfolio Selection: A Robust Approach to Sharpe Ratio Maximization Danqiao Guo 1, Phelim Boyle 2, Chengguo Weng 1, and Tony Wirjanto 1 1 Department of Statistics & Actuarial Science, University
More informationPCA, Kernel PCA, ICA
PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per
More informationPrior Information: Shrinkage and Black-Litterman. Prof. Daniel P. Palomar
Prior Information: Shrinkage and Black-Litterman Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics
More informationLearning representations
Learning representations Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 4/11/2016 General problem For a dataset of n signals X := [ x 1 x
More informationFINM 331: MULTIVARIATE DATA ANALYSIS FALL 2017 PROBLEM SET 3
FINM 331: MULTIVARIATE DATA ANALYSIS FALL 2017 PROBLEM SET 3 The required files for all problems can be found in: http://www.stat.uchicago.edu/~lekheng/courses/331/hw3/ The file name indicates which problem
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationMath 180B Problem Set 3
Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P
More informationFirst of all, the notion of linearity does not depend on which coordinates are used. Recall that a map T : R n R m is linear if
5 Matrices in Different Coordinates In this section we discuss finding matrices of linear maps in different coordinates Earlier in the class was the matrix that multiplied by x to give ( x) in standard
More informationStatistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1
Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of
More informationPart I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes
Week 2 Based in part on slides from textbook, slides of Susan Holmes Part I Other datatypes, preprocessing October 3, 2012 1 / 1 2 / 1 Other datatypes Other datatypes Document data You might start with
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationManifold Learning for Signal and Visual Processing Lecture 9: Probabilistic PCA (PPCA), Factor Analysis, Mixtures of PPCA
Manifold Learning for Signal and Visual Processing Lecture 9: Probabilistic PCA (PPCA), Factor Analysis, Mixtures of PPCA Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inria.fr http://perception.inrialpes.fr/
More informationLECTURE 16: PCA AND SVD
Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal
More informationSection 4.4 Reduction to Symmetric Tridiagonal Form
Section 4.4 Reduction to Symmetric Tridiagonal Form Key terms Symmetric matrix conditioning Tridiagonal matrix Similarity transformation Orthogonal matrix Orthogonal similarity transformation properties
More informationEigenvalues, Eigenvectors, and an Intro to PCA
Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.
More informationEigenvalues and Eigenvectors
5 Eigenvalues and Eigenvectors 5.2 THE CHARACTERISTIC EQUATION DETERMINANATS n n Let A be an matrix, let U be any echelon form obtained from A by row replacements and row interchanges (without scaling),
More informationFinancial Econometrics Lecture 6: Testing the CAPM model
Financial Econometrics Lecture 6: Testing the CAPM model Richard G. Pierse 1 Introduction The capital asset pricing model has some strong implications which are testable. The restrictions that can be tested
More informationMarkov Chains Handout for Stat 110
Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of
More informationThe Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility
The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory
More informationFeature Engineering, Model Evaluations
Feature Engineering, Model Evaluations Giri Iyengar Cornell University gi43@cornell.edu Feb 5, 2018 Giri Iyengar (Cornell Tech) Feature Engineering Feb 5, 2018 1 / 35 Overview 1 ETL 2 Feature Engineering
More informationDefinition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices).
Matrices (general theory). Definition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices). Examples. 1 2 1 1 0 2 A= 0 0 7 B= 0 1 3 4 5 0 Terminology and Notations. Each
More informationDimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining
Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can
More informationNumerical Methods I Singular Value Decomposition
Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)
More informationDiagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics
Diagonalization MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Motivation Today we consider two fundamental questions: Given an n n matrix A, does there exist a basis
More informationMATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018
MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S
More informationOptimal Covariance Matrix with Parameter Uncertainty for the Tangency Portfolio
Optimal Covariance Matrix with Parameter Uncertainty for the Tangency Portfolio Ghislain Yanou June 3, 2009 Abstract We provide a model for understanding the impact of the sample size neglect when an investor,
More informationAnalysis of Spectral Kernel Design based Semi-supervised Learning
Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,
More information1 Principal Components Analysis
Lecture 3 and 4 Sept. 18 and Sept.20-2006 Data Visualization STAT 442 / 890, CM 462 Lecture: Ali Ghodsi 1 Principal Components Analysis Principal components analysis (PCA) is a very popular technique for
More informationARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES
Advances in Complex Systems, Vol. 11, No. 5 (2008) 655 668 c World Scientific Publishing Company ARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES RICARDO COELHO,PETERRICHMONDandSTEFANHUTZLER
More informationA A x i x j i j (i, j) (j, i) Let. Compute the value of for and
7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =
More informationPreliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38
Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters
More informationLinear Programming: Chapter 1 Introduction
Linear Programming: Chapter 1 Introduction Robert J. Vanderbei September 16, 2010 Slides last edited on October 5, 2010 Operations Research and Financial Engineering Princeton University Princeton, NJ
More informationLinear Algebra Methods for Data Mining
Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University
More informationSystem 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:
System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationCSE 554 Lecture 7: Alignment
CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex
More informationStudy of Topological Surfaces
Study of Topological Surfaces Jamil Mortada Abstract This research project involves studying different topological surfaces, in particular, continuous, invertible functions from a surface to itself. Dealing
More information