Appendix to Portfolio Construction by Mitigating Error Amplification: The Bounded-Noise Portfolio

Similar documents
1 Planar rotations. Math Abstract Linear Algebra Fall 2011, section E1 Orthogonal matrices and rotations

Homework 2. Solutions T =

NORMS ON SPACE OF MATRICES

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

This appendix provides a very basic introduction to linear algebra concepts.

Introduction to quantum information processing

Quantum Nonlocality Pt. 4: More on the CHSH Inequality

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002

Large-scale eigenvalue problems

THE EIGENVALUE PROBLEM

CHAPTER 6. Representations of compact groups

MTH 2032 SemesterII

Chapter 8 Gradient Methods

15.S24 Sample Exam Solutions

Example Linear Algebra Competency Test

Machine Learning. Support Vector Machines. Fabio Vandin November 20, 2017

Robust Network Codes for Unicast Connections: A Case Study

Review of similarity transformation and Singular Value Decomposition

0.1 Schrödinger Equation in 2-dimensional system

Lecture 19: Polar and singular value decompositions; generalized eigenspaces; the decomposition theorem (1)

Quantum Computing Lecture 2. Review of Linear Algebra

Lecture 19: Polar and singular value decompositions; generalized eigenspaces; the decomposition theorem (1)

Contents. 0.1 Notation... 3

MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Linear Algebra and Dirac Notation, Pt. 3

Multivariate Distributions

Find all of the real numbers x that satisfy the algebraic equation:

Applied Linear Algebra in Geoscience Using MATLAB

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Lecture Notes 1: Vector spaces

Exercise Set 7.2. Skills

Math 408 Advanced Linear Algebra

arxiv: v5 [math.na] 16 Nov 2017

Stability of an abstract wave equation with delay and a Kelvin Voigt damping

Principal Component Analysis

Linear Algebra & Geometry why is linear algebra useful in computer vision?

LINEAR ALGEBRA QUESTION BANK

arxiv: v1 [cs.lg] 26 Jul 2017

Corner. Corners are the intersections of two edges of sufficiently different orientations.

Math Linear Algebra

Linear Algebra 2 Spectral Notes

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Separation of Variables in Linear PDE: One-Dimensional Problems

Recall the convention that, for us, all vectors are column vectors.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Full-State Feedback Design for a Multi-Input System

Physics 505 Homework No. 1 Solutions S1-1

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

CS 246 Review of Linear Algebra 01/17/19

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Family Feud Review. Linear Algebra. October 22, 2013

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

DS-GA 1002 Lecture notes 10 November 23, Linear models

A Robust Form of Kruskal s Identifiability Theorem

Multi-Robotic Systems

Optimal spectral shrinkage and PCA with heteroscedastic noise

Zhaoxing Gao and Ruey S Tsay Booth School of Business, University of Chicago. August 23, 2018

Adjusted Empirical Likelihood for Long-memory Time Series Models

1. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal

Reasoning with Probabilities. Eric Pacuit Joshua Sack. Outline. Basic probability logic. Probabilistic Epistemic Logic.

Singular Value Decomposition

Notes on SU(3) and the Quark Model

2. Matrix Algebra and Random Vectors

SOLUTION KEY TO THE LINEAR ALGEBRA FINAL EXAM 1 2 ( 2) ( 1) c a = 1 0

STAT 200C: High-dimensional Statistics

Lecture 19 October 28, 2015

Homework 11 Solutions. Math 110, Fall 2013.

MATH 220: INNER PRODUCT SPACES, SYMMETRIC OPERATORS, ORTHOGONALITY

Lebesgue-Radon-Nikodym Theorem

Singular Value Decomposition (SVD)

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

MATH 211 review notes

The Principles of Quantum Mechanics: Pt. 1

Calculating determinants for larger matrices

Review problems for MA 54, Fall 2004.

Sparse PCA in High Dimensions

1 Directional Derivatives and Differentiability

Recall that any inner product space V has an associated norm defined by

Missing dependent variables in panel data models

Consensus building: How to persuade a group

Quantum Physics II (8.05) Fall 2002 Assignment 3

MULTI-AGENT TRACKING OF A HIGH-DIMENSIONAL ACTIVE LEADER WITH SWITCHING TOPOLOGY

Plug-in Measure-Transformed Quasi Likelihood Ratio Test for Random Signal Detection

Lecture : Probabilistic Machine Learning

Statistical Data Analysis

The Cartan Dieudonné Theorem

Lecture 3: Review of Linear Algebra

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

Math 108b: Notes on the Spectral Theorem

1 Last time: least-squares problems

A Short Course on Frame Theory

Lecture 12 : Graph Laplacians and Cheeger s Inequality

LOCAL AND GLOBAL STABILITY OF FUSION FRAMES

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

Transcription:

Appendix to Portfolio Construction by Mitigating Error Amplification: The Bounded-Noise Portfolio Long Zhao, Deepayan Chakrabarti, and Kumar Muthuraman McCombs School of Business, University of Texas, Austin, TX, USA, long.zhao@mccombs.utexas.edu, deepayan.chakrabarti@mccombs.utexas.edu, kumar.muthuraman@mccombs.utexas.edu Appendix A: Absolute Versus Relative Differences in Eigenvalues Lemma 3.2 shows that the bottom eigenvectors could not be well estimated because they are not well separated (i.e., the absolute differences between their eigenvalues are small). This observation might suggest that the corresponding eigenvectors are almost interchangeable and that these errors have a limited effect on the performance of the aggressive noise-only portfolio. However, this intuition is false, because the optimal portfolio depends on the relative differences between eigenvalues, which can still be large. We demonstrate this understanding with some examples. Example. Suppose the eigenvalues vary as a (heavy-tailed) power-law, with λ i = ξ β p i for some ξ > 0 and β >. The absolute difference between consecutive eigenvalues is λ i λ i+ = ξ β p i ( /β), which decreases with increasing i and is at most ξ for the last two eigenvalues. Thus, with a large enough β and a small enough ξ, every consecutive pair of eigenvectors is well separated, except for the bottom two eigenvectors. Under these conditions, for some number of samples n, we can expect the top p 2 eigenvectors to be well estimated, and the last two to be poorly estimated. Let us also assume for simplicity that v p = v p = ρ 0. In general, the bottom eigenvalues are poorly estimated, as in Figure. However, let us consider the best-case scenario for estimation: Suppose that the top p 2 eigenvectors are estimated perfectly, as are all eigenvalues. Let us take Ŝ to be the span of the first p 2 sample eigenvectors. Because these eigenvectors are perfectly estimated, we have Ŝ = S. Let ˆN and N denote the spans of the last two sample eigenvectors and true eigenvectors, respectively. Note that ˆN = N because each is simply the space orthogonal to Ŝ = S. Thus, the only error is in the orientation of the bottom two eigenvectors, ˆv p and ˆv p.

2 In other words, ˆv p = v p cos θ v p sin θ ˆv p = v p sin θ + v p cos θ for some random angle θ. Because we can always reverse the direction of these four eigenvectors without loss of generality, we confine θ to [0, π]. Then, we can show: RV (w ξ N) = () ρ 2 (β + ) RV (ŵ N) EV (ŵ N) ( + (β ) (sin θ) 2) for β (2) RV (ŵ N) RV (w β (sin θ) 2 N) for β. (3) (cos θ + sin θ) 2 The approximations hold when θ 3π, which is true with probability. Thus, the aggressive 4 noise-only portfolio is considered to be a far better portfolio than it actually is (EV (ŵ N) RV (ŵ N)) and performs poorer than the optimal portfolio from the noise space (RV (ŵ N) RV (w N)). Recall that the individual bottom eigenvectors might be poorly estimated, but the span of these eigenvectors is well estimated (i.e., space ˆN itself is well estimated). This reasoning is that ˆN is the space that is orthogonal to Ŝ, which is well estimated. Thus, we can expect good performance from a portfolio that depends only on space ˆN while being invariant to the precise configuration of the eigenvectors in ˆN. The following example illustrates the case. Example 2 (An extension of Example ). From Equation (6), we know that the projected equal-weighted portfolio on the noise space weights the bottom two eigenvectors as follows: p w EW j=p N = (ˆv j)ˆv j p j=p (ˆv j). 2 Note that this definition does not refer to eigenvalues at all. One property of this portfolio is that it is invariant to θ: w EW N = ˆv p (ˆv p ) 2 + (ˆv p) 2 ˆv p + ˆv p (ˆv p ) 2 + (ˆv p) 2 ˆv p ρ(cos θ sin θ) = ρ 2 (cos θ sin θ) 2 + ρ 2 (cos θ + sin θ) (v 2 p cos θ v p sin θ) ρ(cos θ + sin θ) + ρ 2 (cos θ sin θ) 2 + ρ 2 (cos θ + sin θ) (v 2 p sin θ + v p cos θ) = ( vp (cos 2 θ + sin 2 θ) + v p (sin 2 θ + cos 2 θ) ) 2ρ = 2ρ (v p + v p ) = v p (v p ) 2 + (v p) v v p 2 p + (v p ) 2 + (v p) v p. 2

3 Using Equation (), we find that: RV (w EW N ) = 4ρ 2 ( ξ + ξ ) RV (w β N) β 4 for β. Again, the realized variance of w EW N is invariant to θ. We find that w EW N is comparable to the aggressive noise-only portfolio (Equation 3) in terms of realized variance, and that it can, in fact, be better than ŵ N when π 4 θ 3π 4. This finding makes sense because π 4 θ 3π 4 is closer to v p than v p. means that ˆv p Appendix B: Proofs Proposition 3.(Eigenvalue Concentration) By Weyl s inequality, λ i ˆλ i Σ ˆΣ op. Dividing both sides by λ i proves the proposition. Lemma 3.3 (Portfolio Decomposition) Using the Lagrangian multiplier method, we can easily find: w = Σ Σ = where we use Σ = i (/λ i)v i v i. Similarly, we have: i i v i λ i v i, (v i )2 λ i w S = k j= k j= v j v j (v j )2, RV (w S) =, k (v j )2 j= RV (w S )w S = k j= v j. (4) Repeat this process for w N, and some algebraic manipulations yield Equation (2). Proposition 4.(Bounding Realized Variance of any Portfolio from the Noise Space) Because the noise space ˆN is spanned by ˆvk+,..., ˆv p, any vector w from ˆN can be presented as a linear combination of this basis, namely, a w = (ˆv k+,..., ˆv p ). From the orthonormality of eigenvectors, we have: a n = ˆNa. w 2 2 = w w = a ˆN ˆNa = a a = a 2 2. (5) Meanwhile, the definition of the noise bound, m, guarantees that the following inequality holds for any vector b R n such that b 2 =, b ( ˆN (Σ ) ˆΣ) ˆN b m. (6)

4 Plugging a/ a 2 into the previous inequality, we have ( ) a ( ˆN (Σ a 2 Rearranging, we get: ) ( a ˆΣ) ˆN a 2 ) m. ( ˆNa) Σ( ˆNa) ( ˆNa) ˆΣ( ˆNa) + m a 2 2. Substituting w = ˆNa and w 2 2 = a 2 2 into the preceding inequality proves the proposition. Lemma 6. (Projection Portfolios) Clearly, w S and w N as defined in Equation (2) satisfy w S Ŝ, w S = and w N ˆN, w N =. Combining ŜŜ + ˆN ˆN = I with w =, we have = w = w (ŜŜ + ˆN ˆN ) = θ + w ˆN ˆN, which implies θ = w ˆN ˆN. Plugging this equation into the right-hand side of Equation (), RHS = ŜŜ w + ˆN ˆN w = w = LHS. In this way, we prove that Equation (2) gives one solution. Assume that there is another solution, Then, we have w = θ w S + ( θ) w N. θw S θ w S = ( θ)w N + ( θ) w N. The left-hand side belongs to Ŝ while the right-hand side belongs to ˆN. Because Ŝ ˆN = 0, both sides are 0. However, w S = w S =. Therefore, the following holds: 0 = 0 = (θw S θ w S ) = θ θ. The equation implies that w S = w S and w N = w N. Lemma 6.2 (The Solution to the Robust Optimization) Because w ˆN, we have w = ˆNa. Thus, max Ψ U w Ψw = max Ψ U a ˆN Ψ ˆNa = ba I n k+ a. The last equality holds because of the definition of the uncertainty set. Then Equation (4) becomes min a ba I n k+ a, subject to a ( ˆN ) =.

5 Its solution is a = ˆN ˆN ˆN, which implies that the solution to the robust optimization is ˆNa = ˆN ˆN ˆN ˆN. From Equation (2), the projection portfolio of the equal-weighted portfolio on ˆN is: w EW N = ˆN ˆN (/p) (/p) ˆN ˆN = ˆN ˆN ˆN ˆN = ˆNa.