Lectures 6 7 : Marchenko-Pastur Law

Similar documents
Lectures 2 3 : Wigner s semicircle law

Lectures 2 3 : Wigner s semicircle law

Random Matrices: Beyond Wigner and Marchenko-Pastur Laws

1 Tridiagonal matrices

A Generalization of Wigner s Law

Wigner s semicircle law

Determinantal point processes and random matrix theory in a nutshell

Distribution of Eigenvalues of Weighted, Structured Matrix Ensembles

Exponential tail inequalities for eigenvalues of random matrices

Fluctuations from the Semicircle Law Lecture 1

Markov operators, classical orthogonal polynomial ensembles, and random matrices

Lecture I: Asymptotics for large GUE random matrices

1 Intro to RMT (Gene)

A new method to bound rate of convergence

BULK SPECTRUM: WIGNER AND MARCHENKO-PASTUR THEOREMS

RANDOM MATRIX THEORY AND TOEPLITZ DETERMINANTS

On the concentration of eigenvalues of random symmetric matrices

Random Matrix: From Wigner to Quantum Chaos

Painlevé Representations for Distribution Functions for Next-Largest, Next-Next-Largest, etc., Eigenvalues of GOE, GUE and GSE

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Eigenvalue variance bounds for Wigner and covariance random matrices

Convergence of spectral measures and eigenvalue rigidity

DISTRIBUTION OF EIGENVALUES OF REAL SYMMETRIC PALINDROMIC TOEPLITZ MATRICES AND CIRCULANT MATRICES

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

Ma/CS 6b Class 20: Spectral Graph Theory

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Universality of local spectral statistics of random matrices

Universality for random matrices and log-gases

GROUP THEORY PRIMER. New terms: so(2n), so(2n+1), symplectic algebra sp(2n)

The norm of polynomials in large random matrices

Triangular matrices and biorthogonal ensembles

Ma/CS 6b Class 20: Spectral Graph Theory

The Singular Value Decomposition

ACO Comprehensive Exam October 18 and 19, Analysis of Algorithms

A Note on the Central Limit Theorem for the Eigenvalue Counting Function of Wigner and Covariance Matrices

arxiv: v1 [math-ph] 19 Oct 2018

ON THE CONVERGENCE OF THE NEAREST NEIGHBOUR EIGENVALUE SPACING DISTRIBUTION FOR ORTHOGONAL AND SYMPLECTIC ENSEMBLES

(x 1 +x 2 )(x 1 x 2 )+(x 2 +x 3 )(x 2 x 3 )+(x 3 +x 1 )(x 3 x 1 ).

MTH 2032 Semester II

From the mesoscopic to microscopic scale in random matrix theory

Random Matrix Theory Lecture 1 Introduction, Ensembles and Basic Laws. Symeon Chatzinotas February 11, 2013 Luxembourg

Assessing the dependence of high-dimensional time series via sample autocovariances and correlations

Fluctuations from the Semicircle Law Lecture 4

Freeness and the Transpose

Bulk scaling limits, open questions

Comparison Method in Random Matrix Theory

A sequence of triangle-free pseudorandom graphs

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Math 108b: Notes on the Spectral Theorem

Local Kesten McKay law for random regular graphs

Operator norm convergence for sequence of matrices and application to QIT

Universality of distribution functions in random matrix theory Arno Kuijlaars Katholieke Universiteit Leuven, Belgium

Concentration Inequalities for Random Matrices

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Second Order Freeness and Random Orthogonal Matrices

Random matrix pencils and level crossings

A Brief Outline of Math 355

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices

Meixner matrix ensembles

Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage?

Quantum Chaos and Nonunitary Dynamics

MATH 505b Project Random Matrices

Algebraic Methods in Combinatorics

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

1 Last time: least-squares problems

The Matrix Dyson Equation in random matrix theory

On the Limiting Distribution of Eigenvalues of Large Random Regular Graphs with Weighted Edges

8.1 Concentration inequality for Gaussian random matrix (cont d)

Brown University Analysis Seminar

1 Inner Product and Orthogonality

Quantum chaos in composite systems

Quantum Computing Lecture 2. Review of Linear Algebra

Solution Set 7, Fall '12

1 Positive definiteness and semidefiniteness

. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in

The Free Central Limit Theorem: A Combinatorial Approach

Lecture 13: Spectral Graph Theory

Clustering compiled by Alvin Wan from Professor Benjamin Recht s lecture, Samaneh s discussion

On the algebraic structure of Dyson s Circular Ensembles. Jonathan D.H. SMITH. January 29, 2012

Distributions on matrix moment spaces

Coloring Vertices and Edges of a Path by Nonempty Subsets of a Set

Sudoku and Matrices. Merciadri Luca. June 28, 2011

Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples

Random Matrices: Invertibility, Structure, and Applications

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0

ORIE 6334 Spectral Graph Theory September 8, Lecture 6. In order to do the first proof, we need to use the following fact.

CS 246 Review of Linear Algebra 01/17/19

On corrections of classical multivariate tests for high-dimensional data

Density of States for Random Band Matrices in d = 2

PHYS 705: Classical Mechanics. Rigid Body Motion Introduction + Math Review

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR

arxiv: v2 [cond-mat.dis-nn] 9 Feb 2011

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

ICALP g Mingji Xia

5.6. PSEUDOINVERSES 101. A H w.

On singular values distribution of a matrix large auto-covariance in the ultra-dimensional regime. Title

Random Toeplitz Matrices

9. Numerical linear algebra background

Transcription:

Fall 2009 MATH 833 Random Matrices B. Valkó Lectures 6 7 : Marchenko-Pastur Law Notes prepared by: A. Ganguly We will now turn our attention to rectangular matrices. Let X = (X 1, X 2,..., X n ) R p n where X ij are iid, E(X ij ) = 0, E(Xij 2 ) = 1 and p = p(n). Define S n = 1 n XXT R p p and let denote the eigenvalues of the matrix S n. Define the random spectral measure by λ 1 λ 2... λ p µ n = 1 p p We are now ready to state the Marchenko-Pastur law Theorem 1. Let S n, µ n be as above. Assume that p/n n y (0, 1]. Then we have δ λi µ n (, ω) µ a.s where µ is a deterministic measure whose density is given by Here a and b are functions of y given by dµ dx = 1 (b x)(x a)1(a x b) (1) 2πxy a(y) = (1 y) 2, b(y) = (1 + y) 2 Remark 2. If y > 1 then since rank(s) = p n we will have roughly n(y 1) zero eigenvalues. Since µ n = 1 p p δ λ i we see that there will be a mass of (1 y 1 ) at 0 in the limiting measure. Since the nonzero eigenvalues of XX T and X T X are same we can say that in this case the limiting distribution is (1 y 1 )δ 0 + µ where µ satisfies (1) Remark 3. Observe that if y = 1, then a = 0, b = 4, and thus dµ dx = 1 (4 x)x1(0 x 4) 2πx In this case µ is the image of semicircle distribution under the mapping x x 2 1

Proof. : We now begin the proof of Marchenko Pastur Law. Since the support of µ is compact, µ is uniquely determined by its moments. So as in the Wigners case it is enough to show x k dµ n x k dµ Again following Wigner s case, Borel Cantelli lemma says it is enough to show the following 1. E x k dµ n x k dµ 2. Var( x k dµ n ) C k n 2 Computation of the second integral in 1 will show that x k dµ = k 1 r=0 y r r + 1 ( k r )( ) k 1 r Now notice that E x k dµ n = 1 p p E( λ p i ) = 1 p E[T r(xxt /n) k ] = 1 pn k E[ X i1 j 1 X i2 j 1 X i2 j 2 X i3 j 2... X ik j k X i1 j k ] 1 pn k E(I, J) I,J I,J where I [p] k and J [n] k. Now this corresponds to a directed loop on a bipartite graph. For example if k = 4 then for typical {i 1, i 2, i 3, i 4 } and {j 1, j 2, j 3, j 4 } we have the following picture. As in the Wigner s case we see that each edge must appear at least twice, otherwise E(I, J) = 0. Now we have 2k steps in the directed loop. Thus we see that we have at most k edges in the 2

skeleton, hence at most k + 1 vertices in the skeleton. Next assume that number of vertices = m k. Let m = a + b where a = # of I vertices and b = # of J vertices. Then the total number of ways choosing a I vertices and b J vertices Cp a n b, where C is a constant independent of n. The contribution of these terms in the expectation C p a n b /pn k 0 as n. Thus we need to look at loops which have exactly k + 1 vertices and k edges. These are exactly the double trees. Reshuffle them to get the following structure. Start with an I vertex. Vertices that can be reached in even steps are the I vertices, the rest are the J vertices. Next we ask the question: How many double trees are there for a given shape? Here by the shape of a tree we mean the vertices numbered in order of appearence. For example (2 3 4 5 4 6 7 6 8 6 4 3 9 3 10 11 10 3 2), (12 13 14 15 16 17 16 18 16 14 13 19 13 6 7 6 13 12) will give us the same shape, because after renumbering in order of appearance both will give us the following double tree (1 2 3 4 3 5 6 5 7 5 3 2 8 2 9 10 9 2 1) and all of them look like the figure above. Thats is we have to choose r + 1 I vertices from [p] and k r J vertices from [n]. This can be done in P (p, r + 1)P (n, k r) where P (n, k) = n(n 1)... (n k + 1) is permutation of k objects from n distinct objects. Notice that P (p, r + 1)P (n, k r) = np k y r n(1 + O(n 1 )), where y n = p/n Thus E( x k dµ n ) = 1 k 1 pn k E(I, J) = yn(1+o(n r 1 )) #{double tree shapes with r+1 I and k r J I,J r=0 vertices} 3

Since y n y, as n, its clear that all we need now is to show that #{double tree shapes with r + 1 I and k r J vertices} = 1 ( )( ) k k 1 r + 1 r r Towards this end we try to correspond each double tree shape with the following type of path/sequence of 2k steps. 1. If i is odd then s i { 1, 0} 2. If i is even then s i {0, 1}, s 2k = 0 3. For any l = 1, 2,..., 2k, we have l s i 0. That is the path is never below 0. 4. #{i : s i = 1} = #{i : s i = 1} = r. Thats is there are exactly r upsteps and r downsteps 5. 2k s i = 0. That is we return to 0 at the end. Given any such sequence {s i } 2k, clearly we can construct a tree as following: Suppose i is odd. If s i = 1 then we go down the double tree, if s i = 0 then we go up from an I vertex but we will return Suppose now i is even. If s i = 1 then we go one step up in the double tree. If s i = 0 then we go one step down Next given a double tree shape we construct such a sequence {s i } 2k First for each I vertex we mark the first edge leading to it and the last edge leaving it. After marking the previous double will look like the following. The circled vertices are the I vertices. Now put s i = 1 if the i-th edge is marked and its going up, s i = 1 if i-th edge is marked and going down, s i = 0 otherwise. For example the above double tree will give the following path. 4

We have to verify this allocation of 1, 0, 1 would still make {s i } 2k satisfy condition (3). Suppose if possible we have a first l such that 2l 1 s i = 1, hence 2l s i = 0, and s 2l 1 = 1 Then the other part tells us that we can construct a double tree with vertices {1, 2..., 2l}, and since s 2l 1 = 1, the second bullet from the other part says that we are not moving up to a new vertex,but going down to an old vertex in {1, 2..., 2l}. But this destroys the double tree shape giving us a contradiction. Hence we see that if we allocate 1, 0, 1 by the above rule then we indeed get a sequence {s i } 2k satisfying required conditions. Thus the set of the double tree shapes is in bijection with the set of sequence {s i } 2k, so all we need to do now is count such sequences {s i } 2k. Since s 2k +1, not considering condition (3) for the moment we see that out of k 1 positions for +1 and k positions for 1 we have to choose r each. Therefore the number of such sequences is ( )( k 1 k ) r r. Lets now count the number of sequences which fail condition (3). Since those paths hit 1 there exists a first l, such that 2l 1 s i = 1 (By construction of the sequence s k can be 1 only when k is odd.). We now construct a new sequence {s i }2k by reflection. Put For l i k 1, put s i = s i, for i = 1,..., 2l 1, s 2k = s 2k = 0 (s 2i, s 2i+1) = (1, 1) if (s 2i, s 2i+1 ) = (1, 1) = (0, 0) if (s 2i, s 2i+1 ) = (0, 0) = (1, 0) if (s 2i, s 2i+1 ) = (0, 1) = (0, 1) if (s 2i, s 2i+1 ) = (1, 0) Clearly the set of all sequences {s i } 2k, which fail condition (3) is in bijection with the set of sequences {s i }2k But to count the number of such sequences {s i }2k we just have to count the number of ways we can choose r 1 +1 from k 1 of them, and r + 1 1 from k of them. This can be done in ( )( k 1 k ) r 1 r+1. So the total number of sequences {si } 2k which satisfies condition (1-4) is given by ( )( ) ( )( ) k 1 k k 1 k = 1 ( )( ) k 1 k r r r 1 r + 1 r + 1 r r This proves the fact about expectation and the proof of the variance bound is similar to that of Wigner Matrix. 5

We now move to some particular type of random matrices, namely the Gaussian Ensembles. Gaussian Orthogonal Ensemble (GOE): Here we look at matrices M n of the form M n = [X i,j ] n i,j=1 where X i,j = X j,i, X i,j iid N(0, 1), i < j, and X i,i 2N(0, 1) and they are all independent. We can construct them in following way. Take a matrix A = [Y i,j ] n i,j=1, where Y i,j iid N(0, 1). Then is a GOE. M n = (A + A T )/ 2 Gaussian Unitary Ensemble (GUE): M n of the form M n = [X i,j ] n i,j=1 where These are very similar to GOE. Here we look at matrices X i,j = X j,i, X i,j N(0, 1/2) + in(0, 1/2), i < j, and X i,i N(0, 1) and they are all independent. We can construct them in them in the following way. Take a matrix A = [Y i,j ] n i,j=1, where Y i,j iid N(0, 1/2) + in(0, 1/2) ] Then is GUE M n = (A + A )/ 2 Gaussian Symplectic Ensemble (GSE) Define Z as the following block diagonal matrix Z 2n 2n = diag(a, A..., A), where [ ] 0 1 A = 1 0 Call a matrix M C 2n 2n symplectic if Z = MZM T We next define the space of quaternions. Define the following 2 2 matrices [ ] [ ] [ ] [ i 0 0 1 0 i 1 0 e 1 =, e 2 =, e 3 =, 1 = 0 i 1 0 i 0 0 1 ] Note that e 1 e 2 = e 2 e 1 = e 3, e 1 2 = e 2 2 = e 3 2 = 1 6

The conjugation rule is as follows 1 = 1, ē 2 = e 2, ē 3 = e 3, ē 4 = e 4 The vector space [ generated ] by {e 1, e 2, e 3, 1} over C is called the space of quaternions. a b A quaternion is real if q (1), q (2), q (3), q (4) are real where c d [ ] a b = q (1) 1 + q (2) e 2 + q (3) e 3 + q (4) e 4 c d A random quaternion [ a c b d ] is called real standard quaternion if its real and if q (1), q (2), q (3), q (4) iid N(0, 1/4) Let Q i,j = q (1) i,j 1 + q(2) i,j e 2 + q (3) i,j e 3 + q (4) i,j e 4 A GSE is defined by M n = [Q i,j ] n i,j=1 where for i < j, Q i,j are iid standard quaternions, Q i,j = Q j,i, and on the diagonal i = j we have q (0) i,i N(0, 1/2) We can construct such a matrix as follows. Let A = [Y i,j ] n i,j=1 where Y i,j are iid real standard quaternions. Then M n = (A + A )/ 2 is GSE. Let dm be the reference lebesgue measure, based on the determining entries. Define the density function w.r.t dm as 1 Z n,β exp( β 4 T r(m 2 )) Then this defines the density of GOE, GUE and GSE for β = 1, 2, 4 respectively. 7