Random matrices: A Survey. Van H. Vu. Department of Mathematics Rutgers University

Similar documents
Concentration and Anti-concentration. Van H. Vu. Department of Mathematics Yale University

Inverse Theorems in Probability. Van H. Vu. Department of Mathematics Yale University

Random matrices: Distribution of the least singular value (via Property Testing)

Invertibility of random matrices

, then the ESD of the matrix converges to µ cir almost surely as n tends to.

The rank of random regular digraphs of constant degree

Small Ball Probability, Arithmetic Structure and Random Matrices

RANDOM MATRICES: LAW OF THE DETERMINANT

Random regular digraphs: singularity and spectrum

Invertibility of symmetric random matrices

EIGENVECTORS OF RANDOM MATRICES OF SYMMETRIC ENTRY DISTRIBUTIONS. 1. introduction

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Anti-concentration Inequalities

Inverse Littlewood-Offord theorems and the condition number of random discrete matrices

Random Matrices: Invertibility, Structure, and Applications

Coloring Simple Hypergraphs

arxiv: v1 [math.pr] 22 May 2008

BALANCING GAUSSIAN VECTORS. 1. Introduction

1 Last time: least-squares problems

RANDOM MATRICES: TAIL BOUNDS FOR GAPS BETWEEN EIGENVALUES. 1. Introduction

Coloring Simple Hypergraphs

Least singular value of random matrices. Lewis Memorial Lecture / DIMACS minicourse March 18, Terence Tao (UCLA)

BILINEAR AND QUADRATIC VARIANTS ON THE LITTLEWOOD-OFFORD PROBLEM

Eigenvalues and Eigenvectors

Linear Algebra Review. Vectors

REPORT DOCUMENTATION PAGE

Singular value decomposition (SVD) of large random matrices. India, 2010

I. Multiple Choice Questions (Answer any eight)

Online Exercises for Linear Algebra XM511

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

Ma/CS 6b Class 20: Spectral Graph Theory

Conceptual Questions for Review

A Generalization of Komlós s Theorem on Random Matrices

arxiv: v5 [math.na] 16 Nov 2017

Lecture 1 and 2: Random Spanning Trees

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

Preface to the Second Edition...vii Preface to the First Edition... ix

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Eigenvalues and Eigenvectors

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

CONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES. Alexander Barvinok

Recall the convention that, for us, all vectors are column vectors.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Sample Final Exam: Solutions

1 Intro to RMT (Gene)

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

MAT Linear Algebra Collection of sample exams

Introduction to Machine Learning CMU-10701

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Concentration Inequalities for Random Matrices

Chapter 3 Transformations

Commuting birth-and-death processes

Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability. COMPSTAT 2010 Paris, August 23, 2010

Uniformly X Intersecting Families. Noga Alon and Eyal Lubetzky

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

arxiv: v2 [math.fa] 7 Apr 2010

arxiv: v3 [math.pr] 8 Oct 2017

RANDOM MATRICES: OVERCROWDING ESTIMATES FOR THE SPECTRUM. 1. introduction

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Ma/CS 6b Class 20: Spectral Graph Theory

Symmetric and self-adjoint matrices

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

arxiv:math/ v1 [math.pr] 11 Mar 2007

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Extreme eigenvalues of Erdős-Rényi random graphs

Review problems for MA 54, Fall 2004.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

Algebraic Methods in Combinatorics

Fall 2016 MATH*1160 Final Exam

This section is an introduction to the basic themes of the course.

arxiv: v1 [math.co] 4 May 2009

Markov Chains, Stochastic Processes, and Matrix Decompositions

Probabilistic Method. Benny Sudakov. Princeton University

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Eigenvalues and Eigenvectors A =

Cheat Sheet for MATH461

Lecture 7. Econ August 18

1 Tridiagonal matrices

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Chapter 6: Orthogonality

MTH 464: Computational Linear Algebra

Review of some mathematical tools

Generalized eigenvector - Wikipedia, the free encyclopedia

arxiv: v3 [math.pr] 9 Nov 2015

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

Example Linear Algebra Competency Test

Math Matrix Algebra

Spectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Common-Knowledge / Cheat Sheet

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods

LINEAR ALGEBRA REVIEW

Control Systems. Linear Algebra topics. L. Lanari

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Transcription:

Random matrices: A Survey Van H. Vu Department of Mathematics Rutgers University

Basic models of random matrices Let ξ be a real or complex-valued random variable with mean 0 and variance 1. Examples. real gaussian, complex gaussian, Bernoulli (±1 with probability 1/2).

Basic models of random matrices Let ξ be a real or complex-valued random variable with mean 0 and variance 1. Examples. real gaussian, complex gaussian, Bernoulli (±1 with probability 1/2). Non-symmetric model. M n (ξ) denotes the random n n matrix whose entries are i.i.d. copies of ξ.

Basic models of random matrices Let ξ be a real or complex-valued random variable with mean 0 and variance 1. Examples. real gaussian, complex gaussian, Bernoulli (±1 with probability 1/2). Non-symmetric model. M n (ξ) denotes the random n n matrix whose entries are i.i.d. copies of ξ. Symmetric model. Mn sym denotes the random symmetric n n matrix whose upper triangular entries are i.i.d. copies of ξ.

Basic models of random matrices Let ξ be a real or complex-valued random variable with mean 0 and variance 1. Examples. real gaussian, complex gaussian, Bernoulli (±1 with probability 1/2). Non-symmetric model. M n (ξ) denotes the random n n matrix whose entries are i.i.d. copies of ξ. Symmetric model. Mn sym denotes the random symmetric n n matrix whose upper triangular entries are i.i.d. copies of ξ. Remark. For the symmetric model, one can often have a different distribution for the diagonal entries.

Basic models of random matrices Let ξ be a real or complex-valued random variable with mean 0 and variance 1. Examples. real gaussian, complex gaussian, Bernoulli (±1 with probability 1/2). Non-symmetric model. M n (ξ) denotes the random n n matrix whose entries are i.i.d. copies of ξ. Symmetric model. Mn sym denotes the random symmetric n n matrix whose upper triangular entries are i.i.d. copies of ξ. Remark. For the symmetric model, one can often have a different distribution for the diagonal entries. Related models. Adjacency matrix of a random graphs (Erdös-Rény G(n, p), random regular graphs, etc).

Some Main Problems Statistics, Numerical Analysis. Spectral decomposition (Hwang, Wishart 1920s) Complexity of a computational problem involving a random matrix (von Neumann-Goldstine 1940s).

Some Main Problems Statistics, Numerical Analysis. Spectral decomposition (Hwang, Wishart 1920s) Complexity of a computational problem involving a random matrix (von Neumann-Goldstine 1940s). Mathematical Physics/Probability. Distributions of eigenvalues (global and local statistics) (Wigner 1950s).

Some Main Problems Statistics, Numerical Analysis. Spectral decomposition (Hwang, Wishart 1920s) Complexity of a computational problem involving a random matrix (von Neumann-Goldstine 1940s). Mathematical Physics/Probability. Distributions of eigenvalues (global and local statistics) (Wigner 1950s). Combinatorics. Various combinatorial problems (Komlós 1960s). There is a wonderful interaction between the theory of random matrices and other areas of mathematics (number theory, additive combinatorics, theoretical computer science etc).

Combinatorial Problems Singularity. How often is a random matrix singular?

Combinatorial Problems Singularity. How often is a random matrix singular? Determinant. What is the typical value of the determinant? How is it distributed?

Combinatorial Problems Singularity. How often is a random matrix singular? Determinant. What is the typical value of the determinant? How is it distributed? Permanent. What is the typical value of the determinant? How is it distributed?

Combinatorial Problems Singularity. How often is a random matrix singular? Determinant. What is the typical value of the determinant? How is it distributed? Permanent. What is the typical value of the determinant? How is it distributed? Eigenvectors. How does a typical eigenvector look like?

The singularity problem: non-symmetric case Let ξ be Bernoulli (so we consider random ±1 matrices). Question. What is p n, the probability that M n is singular? Conjecture. (folklore/notorious) p n = (1/2 + o(1)) n. The lower bound is obvious: P(there are two equal rows/columns) = (1 + o(1))n 2 2 n. Upper bound: o(1) (Komlós 67).

Short proof of Komlós theorem n 1 p n P(X i+1 Span(X 1,..., X i )). i=1 Fact. A subspace V of dim d contains at most 2 d Bernoulli vectors (as any vector in V is determined by a set of d coordinates). So p n n 1 i=1 2 i 2 n =

Short proof of Komlós theorem n 1 p n P(X i+1 Span(X 1,..., X i )). i=1 Fact. A subspace V of dim d contains at most 2 d Bernoulli vectors (as any vector in V is determined by a set of d coordinates). So p n n 1 i=1 2 i 2 n = 1 2 2 n.

Short proof of Komlós theorem n 1 p n P(X i+1 Span(X 1,..., X i )). i=1 Fact. A subspace V of dim d contains at most 2 d Bernoulli vectors (as any vector in V is determined by a set of d coordinates). So p n n 1 i=1 2 i 2 n = 1 2 2 n. It is enough to show that the contribution of the last k := log log n terms is o(1). We will show P(X n Span(X 1,..., X n 1 )) 1 log 1/3 n.

A m n matrix is l-universal if for any set of l indices i 1,..., i l and any set of signs ɛ 1,..., ɛ l, there is a row X where the i j th entry of X has sign ɛ j, for all 1 j l. Fact. l = log n. A random n n Bernoulli matrix is l-universal with probability at least 1 1 n. Proof. P(fails) ( ) n l 2 l (1 1 ) n exp(2l log n 2 l n) n 1. 2 l So with probability 1 1 n, any vector v orthogonal to X 1,..., X n 1 should have at least l non-zero coordinate. Then P(X n Span(X 1,..., X n 1 )) P(X n v = 0) = O(l 1/2 ) < Lemma (Littlewood-Offord-Erdős, 1940s) If a 1,..., a l are non zero numbers, then P(a 1 ξ 1 + + a l ξ l = 0) = O(l 1/2 ). 1 log 1/3 n.

Further developments

Further developments O(n 1/2 ) (Komlós 77),

Further developments O(n 1/2 ) (Komlós 77),.999 n (Kahn-Komlós -Szemerédi 95),

Further developments O(n 1/2 ) (Komlós 77),.999 n (Kahn-Komlós -Szemerédi 95), (3/4 + o(1)) n (Tao-V. 06),

Further developments O(n 1/2 ) (Komlós 77),.999 n (Kahn-Komlós -Szemerédi 95), (3/4 + o(1)) n (Tao-V. 06), (1/ 2 + o(1)) n (Bourgain-V-Wood 09).

Tools Dominating principle (Halász, KKSZ). Inverse Littlewood-Offord theory (TV) cos x 3 4 + 1 cos 2x. 4

Tools Dominating principle (Halász, KKSZ). Inverse Littlewood-Offord theory (TV) Fractional dimension (BVW) cos x 3 4 + 1 cos 2x. 4 General theorem (BVW 09). cos x 2 1 2 + 1 cos 2x. 2

Littlewood-Offord theorems Improvements of LOE lemma with extra assumptions on the a i. For instance, if the a i are different, then (Erdős-Moser, Sárközy-Szemerédi 1960s) showed P(a 1 ξ 1 + + a l ξ l = 0) = O(l 3/2 ). Stanley showed the extremal set is an arithmetic progression. Kleitman, Katona, Frankl-Füredi, Halaśz, etc.

Inverse Littlewood-Offord theorems Tao-V. 05: If the probability in question is large, then {a 1,..., a n } can be characterized. If P n A, then (most of) a i belong to an AP of lenght n B. The relation between A and B is of importance. A near optimal bound was obtained by Tao-V. (07), that lead to the establishment of the Circular Law Conjecture concerning the eigenvalues of M n (Budapest 08, Bulletin AMS 09).

Inverse Littlewood-Offord theorems Using a different approach, Nguyen-V. (09+) obtained the optimal relation. As a corollary, one obtains all forward LOE results such as Sárkozy-Szemerédi theorem. One can also obtain an asymptotic, stable, version of Stanley s result (algebra-free). See: Hoi Nguyen s talk (December).

The singularity problem: symmetric case Let ξ be Bernoulli (so we consider random ±1 matrices). Question. What is pn sym, the probability that Mn sym is singular? Conjecture. (B. Weiss 1980s) p sym n = o(1). This is the symmetric version of Komlós 1967 theorem. Theorem (Costello-Tao-V. 2005) p sym = O(n 1/4 ).

Quadratic Littlewood-Offord Recently, Kevin Costello (2009) improved the bound to n 1/2+ɛ, which seems to be the limit of the method. Lemma (Costello 09) Consider the quadratic form Q = 1 i,j n a ijξ i ξ j with a ij 0. Then P(Q = 0) n 1/2+ɛ. Question. Higher degree polynomials?

Quadratic Littlewood-Offord Recently, Kevin Costello (2009) improved the bound to n 1/2+ɛ, which seems to be the limit of the method. Lemma (Costello 09) Consider the quadratic form Q = 1 i,j n a ijξ i ξ j with a ij 0. Then P(Q = 0) n 1/2+ɛ. Question. Higher degree polynomials? Conjecture. (V. 2006) p sym (n) = (1/2 + o(1)) n.

Rank of random graphs The Costello-Tao-V. result also holds for A(n, p), the adjacency matrix of G(n, p) with constant p.

Rank of random graphs The Costello-Tao-V. result also holds for A(n, p), the adjacency matrix of G(n, p) with constant p. Question. What about non-constant p? If p < (1 ɛ) log n/n, A(n, p) is almost surely singular, as the graph has isolated vertices. Theorem (Costello-V. 06) For p > (1 + ɛ) log n/n, A(n, p) is a.s. non-singular. Theorem (Costello-V. 07) For p = c log n/n with 0 < c < 1, the co-rank of A(n, p) (a.s.) comes from small local obstructions. A set S is a local obstruction if the number of neighbors of S is less than S. Small means S K(c). For p = c/n, Bordenare and Lelarge (09) computed the asymptotics of the rank. I wonder if one can characterize the co-rank.

Random regular graphs Conjecture. [V. 2006] A(n, d) (adjacency matrix of a random regular graphs of degree d) is a.s non-singular for all 3 d n/2.

The second eigenvalue Conjecture. A(n, d) (adjacency matrix of a random regular graphs of degree d) has second eigenvalue O( d) for all 3 d n/2. Known for d = O(1) (Freedman, Kahn-Szemerédi 1989). Recently Freedman showed λ = (2 + o(1)) d 1. The KSz argument extends to cover up to d n 1/2 (Frieze et. al.). For d = n/2, the best current bound is O( n log n).

Determinant: non-symmetric case Question. What is the typical value of det M n? Turán (1940s): By linearity of expectation, E(det M 2 n) = n! = n (1+o(1)n.

Determinant: non-symmetric case Question. What is the typical value of det M n? Turán (1940s): By linearity of expectation, E(det M 2 n) = n! = n (1+o(1)n. Conjecture. A.s. det M n = n (1/2+o(1))n.

Determinant: non-symmetric case Question. What is the typical value of det M n? Turán (1940s): By linearity of expectation, E(det M 2 n) = n! = n (1+o(1)n. Conjecture. A.s. det M n = n (1/2+o(1))n. Theorem (Tao-V. 2004) A.s. det M n = n (1/2+o(1))n.

Ideas Determinant is the volume of the parallelepiped spanned by the row vectors. Use the height times base formula. Most of the distances are strongly concentrated. Lemma (Tao-V. 04) The distance from a random vector in R n to a fixed subspace of dimension d is, with high prob, n d + small error. The remaining few distances are not too small!.

Determinant: Distribution Question. We know P(det = 0) = exp( Θ(n)). What about P(det = z), for integer z 0? I think P(det = z) = exp( ω(n)), perhaps n cn for some constant c > 0. Question. Limiting distribution of det log det M n 1 2 log(n 1)! c log n (Girko (???) 80s). N(0, 1).

Determinant: Concentration For the next discussion, consider a slightly more general model: Let M be a deterministic matrix with entries 0 < c < m ij < C. Let ξ be a random variable with mean 0 and variance one and M n be the random matrix with entries m ij ξ ij. (So the entries have different variances.) Question. How strongly is det M n concentrated?

Determinant: Concentration For the next discussion, consider a slightly more general model: Let M be a deterministic matrix with entries 0 < c < m ij < C. Let ξ be a random variable with mean 0 and variance one and M n be the random matrix with entries m ij ξ ij. (So the entries have different variances.) Question. How strongly is det M n concentrated? Motivation. Estimating permanent using random determinant. Given a matrix A, define m ij := a ij, then Per A := E det M n 2.

Markov chain: Jerrum-Sinclair, J-S-Vigoda (00). Random determinant: Barvinok (00): exp(cn) with ξ guassian, Friedland-Rider-Zeitouni (04) exp(ɛn) with ξ guassian (under boundedness). Theorem (Costello-V. 07) With high probability and ξ guassian or Bernoulli det M n /E det M n exp(n 2/3+o(1) ).

Markov chain: Jerrum-Sinclair, J-S-Vigoda (00). Random determinant: Barvinok (00): exp(cn) with ξ guassian, Friedland-Rider-Zeitouni (04) exp(ɛn) with ξ guassian (under boundedness). Theorem (Costello-V. 07) With high probability and ξ guassian or Bernoulli det M n /E det M n exp(n 2/3+o(1) ). Conjecture. (Kostello-V. 07) With high probability, det M n /E det M n n O(1).

Permanent: non-symmetric case Question. What is the typical value of PerM n? Turán s : E( PerM 2 n) = n!.

Permanent: non-symmetric case Question. What is the typical value of PerM n? Turán s : E( PerM 2 n) = n!. Conjecture. A.s. PerM n = n (1/2+o(1))n. It had been a long standing open conjecture that a.s PerM n > 0 (the permanent version of Komlós 1967 theorem). Theorem (Tao-V. 2008) A.s. Per M n = n (1/2+o(1))n. Question. What about limiting distribution and concentration? (not known even for gaussian case).

Permanent: non-symmetric case Question. What is the typical value of PerM n? Turán s : E( PerM 2 n) = n!. Conjecture. A.s. PerM n = n (1/2+o(1))n. It had been a long standing open conjecture that a.s PerM n > 0 (the permanent version of Komlós 1967 theorem). Theorem (Tao-V. 2008) A.s. Per M n = n (1/2+o(1))n. Question. What about limiting distribution and concentration? (not known even for gaussian case). Question. What is q n, the probability that the permanent is zero?

Permanent: non-symmetric case Question. What is the typical value of PerM n? Turán s : E( PerM 2 n) = n!. Conjecture. A.s. PerM n = n (1/2+o(1))n. It had been a long standing open conjecture that a.s PerM n > 0 (the permanent version of Komlós 1967 theorem). Theorem (Tao-V. 2008) A.s. Per M n = n (1/2+o(1))n. Question. What about limiting distribution and concentration? (not known even for gaussian case). Question. What is q n, the probability that the permanent is zero? Current bounds. q n n b. The truth may be n bn.

Determinant-Permanent: Symmetric case Still by linearity of expectation E det Mn sym 2 = n (1+o(1)n E PerM sym n 2 = n (1+o(1))n. Conjecture. A.s. det M sym n = n (1/2+o(1))n. Conjecture. A.s. PerM sym n = n (1/2+o(1))n.

Determinant-Permanent: Symmetric case Still by linearity of expectation E det Mn sym 2 = n (1+o(1)n E PerM sym n 2 = n (1+o(1))n. Conjecture. A.s. det M sym n = n (1/2+o(1))n. Conjecture. A.s. PerM sym n = n (1/2+o(1))n. The major difficulty here is that the rows are no longer independent.

Determinant-Permanent: Symmetric case Still by linearity of expectation E det Mn sym 2 = n (1+o(1)n E PerM sym n 2 = n (1+o(1))n. Conjecture. A.s. det M sym n = n (1/2+o(1))n. Conjecture. A.s. PerM sym n = n (1/2+o(1))n. The major difficulty here is that the rows are no longer independent., Recently, Tao-V. (2009) confirmed the first conjecture, the second is till open.

Eigenvectors Consider Mn sym. Its eigenvectors form an orthonormal system v 1,..., v n, v i = 1. Question. How do the v i look like? sub-question. v i =?(Linial) Theorem (Tao-V. 2009) With high probability max v i = n 1/2+o(1). i