Spectral Statistics of Erdős-Rényi Graphs II: Eigenvalue Spacing and the Extreme Eigenvalues

Similar documents
Semicircle law on short scales and delocalization for Wigner random matrices

Local semicircle law, Wegner estimate and level repulsion for Wigner random matrices

arxiv: v3 [math-ph] 21 Jun 2012

Random Matrix: From Wigner to Quantum Chaos

On the principal components of sample covariance matrices

Isotropic local laws for random matrices

Comparison Method in Random Matrix Theory

The Isotropic Semicircle Law and Deformation of Wigner Matrices

arxiv: v4 [math.pr] 10 Sep 2018

The Matrix Dyson Equation in random matrix theory

Local Kesten McKay law for random regular graphs

Local Semicircle Law and Complete Delocalization for Wigner Random Matrices

Universality of sine-kernel for Wigner matrices with a small Gaussian perturbation

RENORMALIZATION OF DYSON S VECTOR-VALUED HIERARCHICAL MODEL AT LOW TEMPERATURES

From the mesoscopic to microscopic scale in random matrix theory

Universality of local spectral statistics of random matrices

Wigner s semicircle law

Universality of Random Matrices and Dyson Brownian Motion

Dynamical approach to random matrix theory

FRG Workshop in Cambridge MA, May

Universality for random matrices and log-gases

Eigenvalue variance bounds for Wigner and covariance random matrices

Random regular digraphs: singularity and spectrum

2 A Model, Harmonic Map, Problem

Exponential tail inequalities for eigenvalues of random matrices

Random Matrices: Invertibility, Structure, and Applications

Local law of addition of random matrices

Mesoscopic eigenvalue statistics of Wigner matrices

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

1 Tridiagonal matrices

DISTRIBUTION OF EIGENVALUES OF REAL SYMMETRIC PALINDROMIC TOEPLITZ MATRICES AND CIRCULANT MATRICES

Bulk scaling limits, open questions

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

On the concentration of eigenvalues of random symmetric matrices

MATH 205C: STATIONARY PHASE LEMMA

Global Maxwellians over All Space and Their Relation to Conserved Quantites of Classical Kinetic Equations

Extreme eigenvalues of Erdős-Rényi random graphs

A determinantal formula for the GOE Tracy-Widom distribution

Measurable functions are approximately nice, even if look terrible.

be the set of complex valued 2π-periodic functions f on R such that

On semilinear elliptic equations with measure data

Markov operators, classical orthogonal polynomial ensembles, and random matrices

arxiv: v5 [math.na] 16 Nov 2017

Applications of random matrix theory to principal component analysis(pca)

RANDOM MATRICES: TAIL BOUNDS FOR GAPS BETWEEN EIGENVALUES. 1. Introduction

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Lecture 4 Lebesgue spaces and inequalities

Lectures 6 7 : Marchenko-Pastur Law

Universality for a class of random band matrices. 1 Introduction

F (z) =f(z). f(z) = a n (z z 0 ) n. F (z) = a n (z z 0 ) n

3 Integration and Expectation

Central Limit Theorem for discrete log gases

Random Fermionic Systems

1 Math 241A-B Homework Problem List for F2015 and W2016

The Canonical Gaussian Measure on R

Determinantal point processes and random matrix theory in a nutshell

The Altshuler-Shklovskii Formulas for Random Band Matrices I: the Unimodular Case

Determinantal point processes and random matrix theory in a nutshell

Eigenvector Statistics of Sparse Random Matrices. Jiaoyang Huang. Harvard University

Lecture Notes on the Matrix Dyson Equation and its Applications for Random Matrices

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Valerio Cappellini. References

Density of States for Random Band Matrices in d = 2

A Generalization of Wigner s Law

Kernel Method: Data Analysis with Positive Definite Kernels

Inhomogeneous circular laws for random matrices with non-identically distributed entries

Quadratic vector equations on complex upper half-plane

III. Quantum ergodicity on graphs, perspectives

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

A strongly polynomial algorithm for linear systems having a binary solution

arxiv: v2 [math.ag] 24 Jun 2015

MODULUS OF CONTINUITY OF THE DIRICHLET SOLUTIONS

Concentration Inequalities for Random Matrices

On rational approximation of algebraic functions. Julius Borcea. Rikard Bøgvad & Boris Shapiro

EIGENFUNCTIONS OF DIRAC OPERATORS AT THE THRESHOLD ENERGIES

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Linear Algebra Massoud Malek

Homogenization of the Dyson Brownian Motion

where r n = dn+1 x(t)

A Sharpened Hausdorff-Young Inequality

SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS. Massimo Grosi Filomena Pacella S. L. Yadava. 1. Introduction

Lecture I: Asymptotics for large GUE random matrices

Random matrices: Distribution of the least singular value (via Property Testing)

Derivatives of Harmonic Bergman and Bloch Functions on the Ball

High-dimensional distributions with convexity properties

DELOCALIZATION OF EIGENVECTORS OF RANDOM MATRICES

Least singular value of random matrices. Lewis Memorial Lecture / DIMACS minicourse March 18, Terence Tao (UCLA)

Reflected Brownian Motion

Part II Probability and Measure

WEYL S LEMMA, ONE OF MANY. Daniel W. Stroock

Positive entries of stable matrices

The expansion of random regular graphs

2. Function spaces and approximation

arxiv: v2 [math.pr] 13 Jul 2018

Non white sample covariance matrices.

Applications and fundamental results on random Vandermon

A Randomized Algorithm for the Approximation of Matrices

Wiener Measure and Brownian Motion

Invertibility of symmetric random matrices

Transcription:

Spectral Statistics of Erdős-Rényi Graphs II: Eigenvalue Spacing and the Extreme Eigenvalues László Erdős Antti Knowles 2 Horng-Tzer Yau 2 Jun Yin 2 Institute of Mathematics, University of Munich, Theresienstrasse 39, D-80333 Munich, Germany lerdos@math.lmu.de Department of Mathematics, Harvard University Cambridge MA 0238, USA knowles@math.harvard.edu htyau@math.harvard.edu jyin@math.harvard.edu 2 July 26, 202 We consider the ensemble of adjacency matrices of Erdős-Rényi random graphs, i.e. graphs on vertices where every edge is chosen independently and with probability p p. We rescale the matrix so that its bulk eigenvalues are of order one. Under the assumption p 2/3, we prove the universality of eigenvalue distributions both in the bulk and at the edge of the spectrum. More precisely, we prove that the eigenvalue spacing of the Erdős-Rényi graph in the bulk of the spectrum has the same distribution as that of the Gaussian orthogonal ensemble; and 2 that the second largest eigenvalue of the Erdős-Rényi graph has the same distribution as the largest eigenvalue of the Gaussian orthogonal ensemble. As an application of our method, we prove the bulk universality of generalized Wigner matrices under the assumption that the matrix entries have at least 4 + ε moments. AMS Subject Classification 200: 5B52, 82B44 Keywords: Erdős-Rényi graphs, universality, Dyson Brownian motion. Partially supported by SFB-TR 2 Grant of the German Research Council Partially supported by SF grant DMS-0757425 Partially supported by SF grants DMS-0757425, 0804279 Partially supported by SF grant DMS-00655

. Introduction The Erdős-Rényi ensemble [20, 2] is a law of a random graph on vertices, in which each edge is chosen independently with probability p p. The corresponding adjacency matrix is called the Erdős-Rényi matrix. Since each row and column has typically p nonzero entries, the matrix is sparse as long as p. We shall refer to p as the sparseness parameter of the matrix. In the companion paper [], we established the local semicircle law for the Erdős-Rényi matrix for p log C, i.e. we showed that, assuming p log C, the eigenvalue density is given by the Wigner semicircle law in any spectral window containing on average at least log C eigenvalues. In this paper, we use this result to prove both the bulk and edge universalities for the Erdős-Rényi matrix under the restriction that the sparseness parameter satisfies p 2/3.. More precisely, assuming that p satisfies., we prove that the eigenvalue spacing of the Erdős-Rényi graph in the bulk of the spectrum has the same distribution as that of the Gaussian orthogonal ensemble GOE. In order to outline the statement of the edge universality for the Erdős-Rényi graph, we observe that, since the matrix elements of the Erdős-Rényi ensemble are either 0 or, they do not satisfy the mean zero condition which typically appears in the random matrix literature. In particular, the largest eigenvalue of the Erdős-Rényi matrix is very large and lies far away from the rest of the spectrum. We normalize the Erdős-Rényi matrix so that the bulk of its spectrum lies in the interval [ 2, 2]. By the edge universality of the Erdős-Rényi ensemble, we therefore mean that its second largest eigenvalue has the same distribution as the largest eigenvalue of the GOE, which is the well-known Tracy-Widom distribution. We prove the edge universality under the assumption.. eglecting the mean zero condition, the Erdős-Rényi matrix becomes a Wigner random matrix with a Bernoulli distribution when 0 < p < is a constant independent of. Thus for p we can view the Erdős-Rényi matrix, up to a shift in the expectation of the matrix entries, as a singular Wigner matrix for which the probability distributions of the matrix elements are highly concentrated at zero. Indeed, the probability for a single entry to be zero is p. Alternatively, we can express the singular nature of the Erdős-Rényi ensemble by the fact that the k-th moment of a matrix entry is bounded by p k 2/2..2 For p this decay in k is much slower than in the case of Wigner matrices. There has been spectacular progress in the understanding of the universality of eigenvalue distributions for invariant random matrix ensembles [5, 7, 8, 27, 28]. The Wigner and Erdős-Rényi matrices are not invariant ensembles, however. The moment method [3, 33, 32] is a powerful means for establishing edge universality. In the context of sparse matrices, it was applied in [32] to prove edge universality for the zero mean version of the d-regular graph, where the matrix entries take on the values and instead of 0 and. The need for this restriction can be ascribed to the two following facts. First, the moment method is suitable for treating the largest and smallest eigenvalues. But in the case of the Erdős-Rényi matrix, it is the second largest eigenvalue, not the largest one, which behaves like the largest eigenvalue of the GOE. Second, the modification of the moment method to matrices with non-symmetric distributions poses a serious technical challenge. A general approach to proving the universality of Wigner matrices was recently developed in the series of papers [2, 3, 4, 5, 6, 7, 8, 9]. In this paper, we further extend this method to cover sparse matrices such as the Erdős-Rényi matrix in the range.. Our approach is based on the following three 2

ingredients. A local semicircle law a precise estimate of the local eigenvalue density down to energy scales containing around log C eigenvalues. 2 Establishing universality of the eigenvalue distribution of Gaussian divisible ensembles, via an estimate on the rate of decay to local equilibrium of the Dyson Brownian motion [9]. 3 A density argument which shows that for any probability distribution of the matrix entries there exists a Gaussian divisible distribution such that the two associated Wigner ensembles have identical local eigenvalue statistics down to the scale /. In the case of Wigner matrices, the edge universality can also be obtained by a modification of and 3 [9]. The class of ensembles to which this method applies is extremely general. So far it includes all generalized Wigner matrices under the sole assumption that the distributions of the matrix elements have a uniform subexponential decay. In this paper we extend this method to the Erdős-Rényi matrix, which in fact represents a generalization in two unrelated directions: a the law of the matrix entries is much more singular, and b the matrix elements have nonzero mean. As an application of the local semicircle law for sparse matrices proved in [], we also prove the bulk universality for generalized Wigner matrices under the sole assumption that the matrix entries have 4 + ε moments. This relaxes the subexponential decay condition on the tail of the distributions assumed in [7, 8, 9]. Moreover, we prove the edge universality of Wigner matrices under the assumption that the matrix entries have 2 + ε moments. These results on Wigner matrices are stated and proved in Section 7 below. We note that in [3] it was proved that the distributions of the largest eigenvalues are Poisson if the entries have at most 4 ε moments. umerical results [4] predict that the existence of four moments corresponds to a sharp transition point, where the transition is from the Poisson process to the determinantal point process with Airy kernel. We remark that the bulk universality for Hermitian Wigner matrices was also obtained in [34], partly by using the result of [22] and the local semicircle law from Step. For real symmetric Wigner matrices, the bulk universality in [34] requires that the first four moments of every matrix element coincide with those of the standard Gaussian random variable. In particular, this restriction rules out the real Bernoulli Wigner matrices, which may be regarded as the simplest kind of an Erdős-Rényi matrix again neglecting additional difficulties arising from the nonzero mean of the entries. As a first step in our general strategy to prove universality, we proved, in the companion paper [], a local semicircle law stating that the eigenvalue distribution of the Erdős-Rényi ensemble in any spectral window which on average contains at least log C eigenvalues is given by the Wigner semicircle law. As a corollary, we proved that the eigenvalue locations are equal to those predicted by the semicircle law, up to an error of order p. The second step of the strategy outlined above for Wigner matrices is to estimate the local relaxation time of the Dyson Brownian motion [5, 6]. This is achieved by constructing a pseudo-equilibrium measure and estimating the global relaxation time to this measure. For models with nonzero mean, such as the Erdős-Rényi matrix, the largest eigenvalue is located very far from its equilibrium position, and moves rapidly under the Dyson Brownian motion. Hence a uniform approach to equilibrium is impossible. We overcome this problem by integrating out the largest eigenvalue from the joint probability distribution of the eigenvalues, and consider the flow of the marginal distribution of the remaining eigenvalues. This enables us to establish bulk universality for sparse matrices with nonzero mean under the restriction.. This approach trivially also applies to Wigner matrices whose entries have nonzero mean. Since the eigenvalue locations are only established with accuracy p, the local relaxation time for the Dyson Brownian motion with the initial data given by the Erdős-Rényi ensemble is only shown to be less than /p 2 /. For Wigner ensembles, it was proved in [9] that the local relaxation time is of order /. Moreover, the slow decay of the third moment of the Erdős-Rényi matrix entries, as given in.2, makes the approximation in Step 3 above less effective. These two effects impose the restriction. in our proof of bulk universality. At the end of Section 2 we give a more detailed account of how this 3

restriction arises. The reason for the same restriction s being needed for the edge universality is different; see Section 6.3. We note, however, that both the bulk and edge universalities are expected to hold without this restriction, as long as the graphs are not too sparse in the sense that p log ; for d-regular graphs this condition is conjectured to be the weaker p [30]. A discussion of related problems on d-regular graphs can be found in [26]. Acknowledgement. We thank P. Sarnak for bringing the problem of universality of sparse matrices to our attention. 2. Definitions and results We begin this section by introducing a class of sparse random matrices A A. Here is a large parameter. Throughout the following we shall often refrain from explicitly indicating -dependence. The motivating example is the Erdős-Rényi matrix, or the adjacency matrix of the Erdős-Rényi random graph. Its entries are independent up to the constraint that the matrix be symmetric, and equal to with probability p and 0 with probability p. For our purposes it is convenient to replace p with the new parameter q q, defined through p = q 2 /. Moreover, we rescale the matrix in such a way that its bulk eigenvalues typically lie in an interval of size of order one. Thus we are led to the following definition. Let A = a ij be the symmetric matrix whose entries a ij are independent up to the symmetry constraint a ij = a ji and each element is distributed according to a ij = γ q { with probability q2 0 with probability q2. 2. Here γ.= q 2 / /2 is a scaling introduced for convenience. The parameter q /2 expresses the sparseness of the matrix; it may depend on. Since A typically has q 2 nonvanishing entries, we find that if q /2 then the matrix is sparse. We extract the mean of each matrix entry and write A = H + γq e e, where the entries of H given by h ij = a ij γq/ have mean zero, and we defined the vector e e.=,..., T. 2.2 Here we use the notation e e to denote the orthogonal projection onto e, i.e. e e ij.=. One readily finds that the matrix elements of H satisfy the moment bounds Eh 2 ij =, E hij p, 2.3 qp 2 where p 2. More generally, we consider the following class of random matrices with non-centred entries characterized by two parameters q and f, which may be -dependent. The parameter q expresses how singular the distribution of h ij is; in particular, it expresses the sparseness of A for the special case 2.. The parameter f determines the nonzero expectation value of the matrix elements. 4

Definition 2. H. We consider random matrices H = h ij whose entries are real and independent up to the symmetry constraint h ij = h ji. We assume that the elements of H satisfy the moment conditions Eh ij = 0, E h ij 2 =, E h ij p Cp q p 2 2.4 for i, j and 2 p log 0 log log, where C is a positive constant. Here q q satisfies for some positive constant C. log 5 log log q C /2 2.5 Definition 2.2 A. Let H satisfy Definition 2.. Define the matrix A = a ij through where f f is a deterministic number that satisfies for some constants ε 0 > 0 and C. A.= H + f e e, 2.6 + ε 0 f C, 2.7 Remark 2.3. For definiteness, and bearing the Erdős-Rényi matrix in mind, we restrict ourselves to real symmetric matrices satisfying Definition 2.2. However, our proof applies equally to complex Hermitian sparse matrices. Remark 2.4. As observed in [], Remark 2.5, we may take H to be a Wigner matrix whose entries have subexponential decay E h ij p Cp θp p/2 by choosing q = /2 log 5θ log log. We shall use C and c to denote generic positive constants which may only depend on the constants in assumptions such as 2.4. Typically, C denotes a large constant and c a small constant. ote that the fundamental large parameter of our model is, and the notations,, O, o always refer to the limit. Here a b means a = ob. We write a b for C a b Ca. After these preparations, we may now state our results. They concern the distribution of the eigenvalues of A, which we order in a nondecreasing fashion and denote by µ µ. We shall only consider the distribution of the first eigenvalues µ,..., µ. The largest eigenvalue µ lies far removed from the others, and its distribution is known to be normal with mean f + f and variance /2 ; see [], Theorem 6.2, for more details. First, we establish the bulk universality of eigenvalue correlations. Let pµ,..., µ be the probability density of the ordered eigenvalues µ µ of A. Introduce the marginal density p µ,..., µ.=! σ S dµ pµ σ,..., µ σ, µ. In other words, p is the symmetrized probability density of the first eigenvalues of H. For n we define the n-point correlation function marginal through Similarly, we denote by p n GOE matrix. p n µ,..., µ n.= GOE, dµ n+ dµ p µ,..., µ. 2.8 the n-point correlation function of the symmetrized eigenvalue density of an ote that we use the density of the law of the eigenvalue density for simplicity of notation, but our results remain valid when no such density exists. 5

Theorem 2.5 Bulk universality. Suppose that A satisfies Definition 2.2 with q φ for some φ satisfying 0 < φ /2, and that f additionally satisfies f C /2 for some C > 0. Let β > 0 and assume that φ > 3 + β 6. 2.9 Let E 2, 2 and take a sequence b satisfying ε β b E 2 /2 for some ε > 0. Let n and O : R n R be compactly supported and continuous. Then E+b de lim E b 2b dα dα n Oα,..., α n n p ϱ sc E n pn GOE, E + α ϱ sc E,..., E + α n ϱ sc E = 0, where we abbreviated for the density of the semicircle law. ϱ sc E.= [4 E2 ] + 2.0 2π Remark 2.6. Theorem 2.5 implies bulk universality for sparse matrices provided that /3 < φ /2. See the end of this section for an account on the origin of the condition 2.9. We also prove the universality of the extreme eigenvalues. Theorem 2.7 Edge universality. Suppose that A satisfies Definition 2.2 with q φ for some φ satisfying /3 < φ /2. Let V be an GOE matrix whose eigenvalues we denote by λ V λ V. Then there is a δ > 0 such that for any s we have P V 2/3 λ V 2 s δ δ as well as P A 2/3 µ 2 s P V 2/3 λ V 2 s + δ + δ 2. P V 2/3 λ V + 2 s δ δ P A 2/3 µ + 2 s P V 2/3 λ V + 2 s + δ + δ, 2.2 for 0, where 0 is independent of s. Here P V denotes the law of the GOE matrix V, and P A the law of the sparse matrix A. Remark 2.8. Theorem 6.4 can be easily extended to correlation functions of a finite collection of extreme eigenvalues. Remark 2.9. The GOE distribution function F s.= lim P V 2/3 λ V 2 s of the largest eigenvalue of V has been identified by Tracy and Widom [36, 37], and can be computed in terms of Painlevé equations. A similar result holds for the smallest eigenvalue λ V of V. Remark 2.0. A result analogous to Theorem 2.7 holds for the extreme eigenvalues of the centred sparse matrix H; see 6.5 below. We conclude this section by giving a sketch of the origin of the restriction φ > /3 in Theorem 2.5. To simplify the outline of the argument, we set β = 0 in Theorem 2.5 and ignore any powers of ε. The proof of Theorem 2.5 is based on an analysis of the local relaxation properties of the marginal Dyson Brownian 6

motion, obtained from the usual Dyson Brownian motion by integrating out the largest eigenvalue µ. As an input, we need the bound Q.= E µ α γ α 2 4φ, 2.3 α= where γ α denotes the classical location of the α-th eigenvalue see 3.5 below. The bound 2.3 was proved in []. In that paper we prove, roughly, that µ α γ α q 2 2φ, from which 2.3 follows. The precise form is given in 3.6. We then take an arbitrary initial sparse matrix ensemble A 0 and evolve it according to the Dyson Brownian motion up to a time τ = ρ, for some ρ > 0. We prove that the local spectral statistics, in the first eigenvalues, of the evolved ensemble A τ at time τ coincide with those of a GOE matrix V, provided that Qτ = Q ρ. 2.4 The precise statement is given in 4.9. This gives us the condition 4φ + ρ < 0. 2.5 ext, we compare the local spectral statistics of a given Erdős-Rényi matrix A with those of the time-evolved ensemble A τ by constructing an appropriate initial A 0, chosen so that the first four moments of A and A τ are close. More precisely, by comparing Green functions, we prove that the local spectral statistics of A and A τ coincide if the first three moments of the entries of A and A τ coincide and their fourth moments differ by at most 2 δ for some δ > 0. See Proposition 5.2. Given A we find, by explicit construction, a sparse matrix A 0 such that the first three moments of the entries of A τ are equal to those of A, and their fourth moments differ by at most 2φ τ = 2φ ρ ; see 5.6. Thus the local spectral statistics of A and A τ coincide provided that 2φ ρ < 0. 2.6 From the two conditions 2.5 and 2.6 we find that the local spectral statistics of A and V coincide provided that φ > /3. 3. The strong local semicircle law and eigenvalue locations In this preliminary section we collect the main notations and tools from the companion paper [] that we shall need for the proofs. Throughout this paper we shall make use of the parameter ξ ξ.= 5 log log, 3. which will keep track of powers of log and probabilities of high-probability events. ote that in [], ξ was a free parameter. In this paper we choose the special form 3. for simplicity. We introduce the spectral parameter z = E + iη where E R and η > 0. Let Σ 3 be a fixed but arbitrary constant and define the domain D L.= { z C. E Σ, log L η 3 }, 3.2 7

with a parameter L L that always satisfies L 8ξ. 3.3 For Im z > 0 we define the Stieltjes transform of the local semicircle law ϱ sc x m sc z.= dx, 3.4 x z where the density ϱ sc was defined in 2.0. The Stieltjes transform m sc z m sc may also be characterized as the unique solution of m sc + = 0 3.5 z + m sc satisfying Im m sc z > 0 for Im z > 0. This implies that m sc z = z + z 2 4 2 R, 3.6 where the square root is chosen so that m sc z z as z. We define the resolvent of A through Gz.= A z, as well as the Stieltjes transform of the empirical eigenvalue density mz.= Tr Gz. For x R we define the distance κ x to the spectral edge through κ x.= x 2. 3.7 At this point we warn the reader that we depart from our conventions in []. In that paper, the quantities Gz and mz defined above in terms of A bore a tilde to distinguish them from the same quantities defined in terms of H. In this paper we drop the tilde, as we shall not need resolvents defined in terms of H. We shall frequently have to deal with events of very high probability, for which the following definition is useful. It is characterized by two positive parameters, ξ and ν, where ξ is given by 3.. Definition 3. High probability events. We say that an -dependent event Ω holds with ξ, ν-high probability if PΩ c e νlog ξ 3.8 for 0 ν. Similarly, for a given event Ω 0, we say that Ω holds with ξ, ν-high probability on Ω 0 if for 0 ν. PΩ 0 Ω c νlog ξ e Remark 3.2. In the following we shall not keep track of the explicit value of ν; in fact we allow ν to decrease from one line to another without introducing a new notation. All of our results will hold for ν ν 0, where ν 0 depends only on the constants C in Definition 2. and the parameter Σ in 3.2. 8

Theorem 3.3 Local semicircle law []. Suppose that A satisfies Definition 2.2 with the condition 2.7 replaced with 0 f C. 3.9 Moreover, assume that q log 20ξ, 3.0 L 20ξ. 3. Then there is a constant ν > 0, depending on Σ and the constants C in 2.4 and 2.5, such that the following holds. We have the local semicircle law: the event { mz { mscz log 40ξ log 40ξ min κe + η q 2, } + } 3.2 q η z D L holds with ξ, ν-high probability. Moreover, we have the following estimate on the individual matrix elements of G. If instead of 3.9 f satisfies 0 f C 0 /2, 3.3 for some constant C 0, then the event { max Gij z δ ij m sc z log 40ξ i,j q + Im m sc z + } 3.4 η η z D L holds with ξ, ν-high probability. ext, we recall that the first eigenvalues of A are close the their classical locations predicted by the semicircle law. Let n sc E.= E ϱ scx dx denote the integrated density of the local semicircle law. Denote by γ α the classical location of the α-th eigenvalue, defined through n sc γ α = α for α =,...,. 3.5 The following theorem compares the locations of the eigenvalues µ,..., µ to their classical locations γ,..., γ. Theorem 3.4 Eigenvalue locations []. Suppose that A satisfies Definition 2.2, and let φ be an exponent satisfying 0 < φ /2, and set q = φ. Then there is a constant ν > 0 depending on Σ and the constants C in 2.4, 2.5, and 2.7 as well as a constant C > 0 such that the following holds. We have with ξ, ν-high probability that µ α γ α 2 log Cξ 4φ + 4/3 8φ. 3.6 α= Moreover, for all α =,..., we have with ξ, ν-high probability that µ α γ α log Cξ 2/3[ α ] /3 + α log Cξ + 3φ + 2/3 4φ α 2/3 + 2φ, where we abbreviated α.= min{α, α}. 3.7 9

Remark 3.5. Under the assumption φ /3 the estimate 3.7 simplifies to which holds with ξ, ν-high probability. µ α γ α log Cξ 2/3 α /3 + 2φ, 3.8 Finally, we record two basic results from [] for later reference. From [], Lemmas 4.4 and 6., we get, with ξ, ν-high probability, max λ α 2 + log Cξ q 2 + 2/3, α max µ α 2 + log Cξ q 2 + 2/3. 3.9 α Moreover, from [], Theorem 6.2, we get, with ξ, ν-high probability, µ = f + f + o. 3.20 In particular, using 2.7 we get, with ξ, ν-high probability, 2 + σ µ C, 3.2 where σ > 0 is a constant spectral gap depending only on the constant ε 0 from 2.7. 4. Local ergodicity of the marginal Dyson Brownian motion In Sections 4 and 5 we give the proof of Theorem 2.5. Throughout Sections 4 and 5 it is convenient to adopt a slightly different notation for the eigenvalues of A. In these two sections we shall consistently use x x to denote the ordered eigenvalues of A, instead of µ µ used in the rest of this paper. We abbreviate the collection of eigenvalues by x = x,..., x. The main tool in the proof of Theorem 2.5 is the marginal Dyson Brownian motion, obtained from the usual Dyson Brownian motion of the eigenvalues x by integrating out the largest eigenvalue x. In this section we establish the local ergodicity of the marginal Dyson Brownian and derive an upper bound on its local relaxation time. Let A 0 = a ij,0 ij be a matrix satisfying Definition 2.2 with constants q 0 φ and f 0 + ε 0. Let B ij,t ij be a symmetric matrix of independent Brownian motions, whose off-diagonal entries have variance t and diagonal entries variance 2t. Let the matrix A t = a ij,t ij satisfy the stochastic differential equation da ij = db ij 2 a ij dt. 4. It is easy to check that the distribution of A t is equal to the distribution of e t/2 A 0 + e t /2 V, 4.2 where V is a GOE matrix independent of A 0. Let ρ be a constant satisfying 0 < ρ < to be chosen later. In the following we shall consider times t in the interval [t 0, τ], where t 0.= ρ, τ.= ρ. 0

One readily checks that, for any fixed ρ as above, the matrix A t satisfies Definition 2.2, with constants f t = f + O δ0 + ε 0 2, q t q 0 φ, where all estimates are uniform for t [t 0, τ]. Denoting by x,t the largest eigenvalue of A t, we get in particular from 3.2 that P t [t 0, τ]. x,t / [2 + σ, C ] e νlog ξ 4.3 for some σ > 0 and C > 0. From now on we shall never use the symbols f t and q t in their above sense. The only information we shall need about x is 4.3. In this section we shall not use any information about q t, and in Section 5 we shall only need that q t c φ uniformly in t. Throughout this section f t will denote the joint eigenvalue density evolved under the Dyson Brownian motion. See Definition 4. below. It is well known that the eigenvalues x t of A t satisfy the stochastic differential equation Dyson Brownian motion dx i = db i + 4 x i + 2 j i x i x j where B,..., B is a family of independent standard Brownian motions. In order to describe the law of V, we define the equilibrium Hamiltonian Hx.= i and denote the associated probability measure by 4 x2 i dt for i =,...,, 4.4 log x i x j 4.5 i<j µ dx µdx.= Z e Hx dx, 4.6 where Z is a normalization. We shall always consider the restriction of µ to the domain Σ.= {x. x < < x }, i.e. a factor x Σ is understood in expressions like the right-hand side of 4.6; we shall usually omit it. The law of the ordered eigenvalues of the GOE matrix V is µ. Define the Dirichlet form D µ and the associated generator L through D µ f = flf dµ.= 2 where f is a smooth function of compact support on Σ. One may easily check that L = 2 2 i + 4 x i + i, 2 x i i i x j j i f 2 dµ, 4.7 and that L is the generator of the Dyson Brownian motion 4.4. More precisely, the law of x t is given by f t x µdx, where f t solves t f t = Lf t and f 0 xµdx is the law of x 0.

Definition 4.. Let f t to denote the solution of t f t = Lf t satisfying f t t=0 = f 0. It is well known that this solution exists and is unique, and that Σ is invariant under the Dyson Brownian motion, i.e. if f 0 is supported in Σ, so is f t for all t 0. For a precise formulation of these statements and their proofs, see e.g. Appendices A and B in [6]. In Appendix A, we present a new, simpler and more general, proof. Theorem 4.2. Fix n and let m = m,..., m n n be an increasing family of indices. Let G. R n R be a continuous function of compact support and set G i,m x.= G x i x i+m, x i+m x i+m2,..., x i+mn x i+mn. Let γ,..., γ denote the classical locations of the first eigenvalues, as defined in 3.5, and set i J Q.= sup t [t 0,τ] i= i J x i γ i 2 f t dµ. 4.8 Choose an ε > 0. Then for any ρ satisfying 0 < ρ < there exists a τ [τ/2, τ] such that, for any J {, 2,..., m n }, we have G i,m f τ dµ G i,m dµ C ε +ρ Q + ρ 4.9 for all 0 ρ. Here µ is the equilibrium measure of eigenvalues GOE. ote that, by definition, the observables G i,m in 4.9 only depend on the eigenvalues x,..., x. The rest of this section is devoted to the proof of Theorem 4.2. We begin by introducing a pseudo equilibrium measure. Abbreviate R.= τ ε = ρ/2 ε/2 and define W x.= i= 2R 2 x i γ i 2. Here we set γ.= 2+σ for convenience, but one may easily check that the proof remains valid for any larger choice of γ. Define the probability measure ωdx.= ψx µdx where ψx.= Z Z e W x. ext, we consider marginal quantities obtained by integrating out the largest eigenvalue x. To that end we write x = x, x, x = x,..., x and denote by ωd x the marginal measure of ω obtained by integrating out x. By a slight abuse of notation, we sometimes make use of functions µ, ω, and ω, defined as the densities with respect to Lebesgue measure of their respective measures. Thus, µx = Z e Hx, ωx = Z e Hx W x, ω x = x ω x, x dx. 2

For any function hx we introduce the conditional expectation h x.= E ω [h x] = x h x, x ω x, x dx ω x. Throughout the following, we write g t.= f t /ψ. In order to avoid pathological behaviour of the extreme eigenvalues, we introduce cutoffs. Let σ be the spectral gap from 4.3, and choose θ, θ 2, θ 3 [0, ] to be smooth functions that satisfy { θ x = 0 if x 4 if x 3, { θ 2 x = if x 2 + σ 5 0 if x 2 + 2σ, 5 { θ 3 x = 0 if x 2 + 3σ 5 if x 2 + 4σ 5. Define θ θx, x, x = θ x θ 2 x θ 3 x. One easily finds that θ 2 σ C 4 x 3 + C θ 2 x 2 2σ 3σ + C 5 5 x 2 4σ 5 where the left-hand side is understood to vanish outside the support of θ. Define the density h t.= θg t, Ẑ t.= θg t dω. Ẑ t, 4.0 If ν is a probability measure and q a density such that qν is also a probability measure, we define the entropy S ν q.= q log q dν. The following result is our main tool for controlling the local ergodicity of the marginal Dyson Brownian motion. Proposition 4.3. Suppose that i S µ f t0 C, ii iii sup t [t 0,τ] sup t [t 0,τ] [ x 3 + Then for t [t 0, τ] we have 4. x 2 + σ + x 2 + 4σ ] f t dµ e νlog ξ, 4.2 5 5 sup θ θ 2 x log θgt x 2 C. 4.3 x Σ t S ω h D ω h + S ω h e clog ξ + CQR 4 + C. 4.4 3

Proof. First we note that Ẑ t = θf t dµ = O e νlog ξ 4.5 uniformly for t [t 0, τ], by 4.2. Dropping the time index to avoid cluttering the notation, we find θg t S ω h = t Ẑ log θg d ω t log Ẑ = Ẑ t θg log θg dω + log Ẑ + S ω h t log Ẑ. We find that t Ẑ = θlf dµ = 2 θ f dµ /2 θ 2 f dµ D µ f /2. Bounding the Dirichlet form in terms of the entropy see e.g. [0], Theorem 3.2, we find that by 4.. Using 4.0 we therefore find D µ f t 2 t S µf t0 C, 4.6 t Ẑ C e clog ξ. 4.7 Thus we have t S ω h 2 t θg log θg dω + + S ω h C e clog ξ. 4.8 We therefore need to estimate t θg log θg dω = θlf log θg dµ + θg t θg d ω. 4.9 θg The second term of 4.9 is given by t θg d ω = θlf dµ = t Ẑ. Therefore 4.8 yields t S ω h 2 θlf log θg dµ + + S ω h C e clog ξ. 4.20 The first term of 4.20 is given by f θ log θg dµ = θf log θg dµ + E + E 2, 4.2 where we defined E.= θ log θg f dµ, E 2.= θ f log θg dµ. 4

ext, we estimate the error terms E and E 2. Using 4.0 we get E = θ θ log θg θf dµ e νlog ξ θg 2 θg 2 where we used 4.5. Similarly, we find Using 4.0, 4.3, and 4.6 we therefore get E 2 θ 2 /2 θg 2 /2 f dµ θg 2 θf dµ θ /2 θg d ω e clog ξ + e clog ξ D ω h, θ /2 2 log θg 2 f 2 /2 f dµ dµ. f θ E 2 C 2 θ log θg /2 2 f dµ C e clog ξ. θ Having dealt with the error terms E and E 2, we compute the first term on the right-hand side of 4.2, θf log θg dµ = x θg x log θg ψ dµ x log ψ x log θg θgψ dµ. 4.22 The second term of 4.22 is bounded by η x log ψ 2 f dµ + η θg 2 θg 2 θg d ω η R 4 i= Q ηr 4 + 4ηD ω h, x i γ i 2 f dµ + 4ηD ω h where η > 0. The first term of 4.22 is equal to x θg x log θg d ω. A simple calculation shows that x θg = x θg θg x log ω + θg x log ω, so that the first term of 4.22 becomes x θg x log θg d ω + θg x log ω θg x log ω x log θg d ω 4 ηd ω h + η θg x log ω θg x log ω 2 d ω. θg 5

Using the Cauchy-Schwarz inequality ab 2 a 2 b 2 we find that the second term is bounded by θg x log ω x log ω 2 d ω θg x log ω x log ω 2 d ω η θg η = x log ω x log ω 2 θf dµ η Thus, we have to estimate E 3.= η i= x x i = η i= 2 θf dµ, E 4.= η x x i Since x x i σ/5 on the support of θfµ, one easily gets from 3.9 that In order to estimate E 4, we write where x x i = E 3 C η. i= dx x x i w i x, dx w i x w i x.= x x e 4 x2 2R 2 x γ 2 j i, x x i x x i x x j. 2 θf dµ. 2 θf dµ. We now claim that on the support of θ, in particular for 4 x < x 2 + 2σ/5, we have dx x x i w i x c γ, 4.23 dx w i x uniformly for x Σ. Indeed, writing γ.= γ + R 2, we have on the support of θ dx x x i w i x dx x γ w i x γ /2 +. dx w i x dx w i x Moreover, the second term is nonnegative: dx x γ w i x = C x dx e R x x 2 x γ 2 x x j j i, = C x e R 2 x γ 2 x x j 0, j i, + C x dx e R 2 x γ 2 x k i,j j i,k, x x j 6

where C x is nonnegative. This proves 4.23. Using 4.23 we get E 4 C γ 2 η f dµ = C η. Summarizing, we have proved that t S ω h 4 8η e clog ξ D ω h + + S ω h e clog ξ + Q ηr 4 + C η. Choosing η small enough completes the proof. ext, we derive a logarithmic convexity bound for the marginal measure ω. Lemma 4.4. We have that ω x = Z e Ĥ x, where Ĥ x = log x i x j + V x, 4.24 i<j< and 2 V x R 2. Proof. Write H x, x = H x + H x, x where H x.= log x i x j, H x, x.= log x x i + i i<j< i< 2R 2 x i γ i 2. By definition, we have ω x = Z e H x x e H x,x dx. The main tool in our proof is the Brascamp-Lieb inequality [6]. In order to apply it, we need to extend the integration over x to R and replace the singular logarithm with a C 2 -function. To that end, we introduce the approximation parameter δ > 0 and define, for x Σ, V δ x.= [ ] log exp log δ x x i 2R 2 x i γ i 2 dx, where we defined i< log δ x.= x δ log x + x < δ log δ + x δ δ It is easy to check that log δ C 2 R, is concave, and satisfies lim δ 0 log δx = { log x if x > 0 if x 0. i x δ2. 2δ2 7

Thus we find that V δ C 2 Σ and that we have the pointwise convergence, for all x Σ, lim V δ x = V x.= δ 0 log e H x,x dx, x where V C 2 Σ satisfies 4.24. ext, we claim that if ϕ = ϕx, y satisfies 2 ϕx, y K then ψx, defined by e ψx.= e ϕx,y dy, satisfies 2 ψx K. In order to prove the claim, we use subscripts to denote partial derivatives and recall the Brascamp-Lieb inequality for log-concave functions Equation 4.7 in [6] ψ xx ϕxx ϕ xy ϕ yy ϕ yx e ϕ dy. e ϕ dy Then the claim follows from ϕxx ϕ xy ϕ yz ϕ yy K = ϕxx ϕ xy ϕ yy ϕ yx K. Using this claim, we find that 2 V δ x R 2 for all x Σ. In order to prove that 2 V x R 2 and hence complete the proof it suffices to consider directional derivatives and prove the following claim. If ζ δ δ>0 is a family of functions on a neighbourhood U that converges pointwise to a C 2 -function ζ as δ 0, and if ζ δ x K for all δ > 0 and x U, then ζ x K for all x U. Indeed, taking δ 0 in ζ δ x + h + ζ δ x h 2ζ δ x = h 0 ζ δ x + ξ + ζ δ x ξ h ξ dξ Kh 2 yields ζx + h + ζx h 2ζx h 2 K, from which the claim follows by taking the limit h 0. As a first consequence of Lemma 4.4, we derive an estimate on the expectation of observables depending only on eigenvalue differences. Proposition 4.5. Let q L d ω be probability density. Then for any J {, 2,..., m n } and any t > 0 we have G i,m q d ω G i,m d ω C D ω q t + C S ω q e ct/r2. i J i J Proof. Using Lemma 4.4, the proof of Theorem 4.3 in [6] applies with merely cosmetic changes. Another, standard, consequence of Lemma 4.4 is the logarithmic Sobolev inequality S ω q CR 2 D ω q. 4.25 Using 4.25 and Proposition 4.3, we get the following estimate on the Dirichlet form. 8

Proposition 4.6. Under the assumptions of Proposition 4.3, there exists a τ [τ/2, τ] such that S ω h τ CR 2 Q + CR 2, D ω h τ CR 4 Q + C. Proof. Combining 4.25 with 4.4 yields t S ω h t CR 2 S ω h t + CQR 4 + C, 4.26 which we integrate from t 0 to t to get S ω h t e CR 2 t t 0 S ω h t0 + CQR 2 + CR 2. 4.27 Moreover, 4.5 yields νlog ξ S ω h t0 CS ω g t0 + e νlog ξ CS ω g t0 + e = CS µ f t0 C log ψ f t0 dµ + e νlog ξ, where the second inequality follows from the fact that taking marginals reduces the relative entropy; see the proof of Lemma 4.7 below for more details. Thus we get Thus 4.27 yields S ω h t0 C + R 2 Q C. S ω h t C e CR 2 t t 0 + CR 2 Q + CR 2 4.28 for t [t 0, τ]. Integrating 4.4 from τ/2 to τ therefore gives 2 τ D ω h t dt CR 4 Q + C, τ τ/2 and the claim follows. We may finally complete the proof of Theorem 4.2. Proof of Theorem 4.2. The assumptions of Proposition 4.3 are verified in Subsection 4. below. Hence Propositions 4.5 and 4.6 yield G i,m h τ dω G i,m dω ε C +ρ Q i J i J + C 2φ ρ. Using 4.5 and 4.2 we get G i,m f τ dµ G i,m dω i J i J C ε +ρ Q + C 2φ ρ. 4.29 9

In order to compare the measures ω and µ, we define the density { qx.= Z exp 4 x2 i + 2R 2 x i γ i 2 } log x x i, i< i< i< where Z is a normalization chosen so that θq dω is a probability measure. It is easy to see that i J q dω = dµ dg, where dg = Ce 4 x2 2R 2 x γ 2 dx is a Gaussian measure. Similarly to Proposition 4.5, we have G i,m θq dω G i,m dω C D ω θqτ + C S ω θq e cτ/r2. Thus we have to estimate D ω θq C log q 2 θq dω + C θ 2 q dω θ C 4 x2 i + 2 R 4 x i γ i 2 + i< i J C + R 4 i< x i γ i 2 dµ x x i 2 θq dω + where the second inequality follows from standard large deviation results for GOE. Since i< x i γ i 2 dµ C +ε for arbitrary ε is known to hold for GOE see [9] where this is proved for more general Wigner matrices, we find G i,m θq dω G i,m dω C ρ + C ρ+2ε+ε. i J i J The cutoff θ can be easily removed using the standard properties of dµ. Choosing ε = ε, replacing ε with ε/2, and recalling 4.29 completes the proof. 4.. Verifying the assumptions of Proposition 4.3. The estimate 4. is an immediate consequence of the following lemma. Lemma 4.7. Let the entries of A 0 have the distribution ζ 0. Then for any t > 0 we have where m 2 ζ 0 is the second moment of ζ 0. S µ f t 2 m 2 ζ 0 log e t, Proof. Recall that the relative entropy is defined, for ν µ, as Sν µ.= log dν dµ dν. If ν and µ are marginals of ν and µ with respect to the same variable, it is easy to check that S ν µ Sν µ. Therefore S µ f t = Sf t µ µ SA t V = 2 Sζ t g 2/, 20

where ζ t denotes the law of the off-diagonal entries of A t, and g λ is a standard Gaussian with variance λ the diagonal entries are dealt with similarly. Setting γ = e t, we find from 4.2 that ζ t has probability density ϱ γ g 2γ/, where ϱ γ is the probability density of γ /2 ζ 0. Therefore Jensen s inequality yields Sζ t g 2/ = S dy ϱ γ y g 2γ/ y g 2/ dy ϱ γ ys g 2γ/ y g 2/. By explicit computation one finds S g 2γ/ y g 2/ = 2 2 y2 log γ + γ. Therefore and the claim follows. Sζ t g 2/ m 2 ζ 0 log γ, The estimate 4.2 follows from 4.3 and 3.9. It only remains to verify 4.3. Lemma 4.8. For any t [t 0, τ] we have θ θ 2 x log θg t x 2 C. 4.30 Proof. Let ζ t be the law of an off-diagonal entry a of A t the diagonal entries are treated similarly. From 4.2 we find ζ t = ϱ γ g 2γ/, where γ = e t, ϱ γ is the law of γ /2 ζ 0, and g λ is a standard Gaussian with variance λ. Using da to denote Lebesgue measure, we find by explicit calculation that which gives e C C a 2 dζ t da C e 4 a2, e C C a 2 dζ t dg 2γ/ e C. Therefore, the density F t A of the law of A with respect to the GOE measure satisfies e C C Tr A 2 F t A e C. Parametrizing A = Ax, v using the eigenvalues x and eigenvectors v, the GOE measure can be written in the factorized form µdxp dv, where µ is defined in 4.6 and P is a probability measure. Thus we get that the density f t x = F t x, v P dv 2

satisfies e C C i x2 i f t x e C. 4.3 ext, it is easy to see that Using 4.32 we may now derive an upper bound on θ 3 g t : e C C i x2 i ψx e C. 4.32 θ 3 g t x = dx θ 3 x f t x, x µ x, x dx ψ x, x µ x, x e C + C i< x2 i dx µ x, x dx e C x 2 µ x, x. Since dx e C x 2 µ x, x dx µ x, x by a straightforward calculation, we get = x dx e C x 2 i< x x i e 4 x2 x dx i< x x i e 4 x2 e C C i< x2 i 4.33 θ 3 g t x e C + C i< x2 i. We now derive a lower bound on θ 3 g t. Using 4.32 and 4.3 we find θ 3 g t x e C dx θ 3 x f t x, x µ x, x dx µ x, x 2+σ/2 dx e C x 2 µ x, x e C C i< x2 i x dx µ x, x e C C i< x2 i, by a calculation similar to 4.33. The claim follows from θ θ 2 x log θg t x 2 2θ θ 2 x log θ θ 2 2 + 2θ θ 2 x log θ 3 g t x 2 2 + C. 5. Bulk universality: proof of Theorem 2.5 Similarly to 2.8, we define p t, x,..., x as the probability density obtained by symmetrizing in the variables x,..., x the function dx f t x µx, and set, for n, p n t, x,..., x n.= dx n+ dx p t, x,..., x. We begin with a universality result for sparse matrices with a small Gaussian convolution. 22

Theorem 5.. Let E [ 2 + κ, 2 κ] for some κ > 0 and let b b satisfy b κ/2. Pick ε, β > 0, and set τ.= 2α+β, where { α αφ.= min 2φ 2, 4φ 2 }. 5. 3 Let n and O : R n R be compactly supported and continuous. Then there is a τ [τ/2, τ] such that E+b E b de 2b dα dα n Oα,..., α n ϱ sc E n p n τ, pn GOE, E + α ϱ sc E,..., E + α n ϱ sc E C n ε[ b 2φ + b /2 β/2]. 5.2 Proof. The claim follows from Theorem 4.2 and Theorem 3.4, similarly to the proof of Theorem 2. in [6]. We use that Q log Cξ 2α, as follows from 3.6; the contribution of the low probability complement event to 3.6 may be easily estimated using Cauchy-Schwarz and the estimate i Et x 4 i = Et Tr A 4 C, uniformly for t 0. The assumption IV of [6] is a straightforward consequence of the local semicircle law, Theorem 3.3. Proposition 5.2. Let A = a ij and A2 = a 2 ij be sparse random matrices, both satisfying Definition 2.2 with q q 2 φ in self-explanatory notation. Suppose that, for each i, j, the first three moments of a ij and a 2 ij same, and that the fourth moments satisfy are the E a 4 2 4 ij E a ij 2 δ, 5.3 for some δ > 0. Let n and let F C 5 C n. We assume that, for any multi-index α n with α 5 and any sufficiently small ε > 0, we have { max α F x,..., x n. } x i ε i C0ε, { max α F x,..., x n. } x i 2 i C0, where C 0 is a constant. Let κ > 0 be arbitrary. Choose a sequence of positive integers k,..., k n and real parameters Ej m [ 2 + κ, 2 κ], where m =,..., n and j =,..., k m. Let ε > 0 be arbitrary and choose η with ε η. Set zj m.= Ej m ± iη with an arbitrary choice of the ± signs. Then, abbreviating G l z.= A l z, we have EF Tr k G z k j,..., j= kn Tr k n j= G zj n EF G G 2 C 3φ+Cε + C δ+cε. 23

Proof. The proof of Theorem 2.3 in [7] may be reproduced almost verbatim; the rest term in the Green function expansion is estimated by an L -L bound using E a l ij 5 C 3φ. As in [7] Theorem 6.4, Proposition 5.2 readily implies the following correlation function comparison theorem. Theorem 5.3. Suppose the assumptions of Proposition 5.2 hold. Let p n, and pn 2, be n-point correlation functions of the eigenvalues of A and A 2 respectively. Then for any E < 2, any n and any compactly supported test function O : R n R we have lim dα dα n Oα,..., α n p n, pn 2, E + α,..., E + α n We may now complete the proof of Theorem 2.5. Proof of Theorem 2.5. In order to invoke Theorems 5. and 5.3, we construct a sparse matrix A 0, satisfying Definition 2.2, such that its time evolution A τ is close to A in the sense of the assumptions of Proposition 5.2. For definiteness, we concentrate on off-diagonal elements the diagonal elements are dealt with similarly. For the following we fix i < j; all constants in the following are uniform in i, j, and. Let ξ, ξ, ξ 0 be random variables equal in distribution to a ij, a τ ij, a 0 ij respectively. For any random variable X we use the notation X.= X EX. Abbreviating γ.= e τ, we have ξ = γ ξ 0 + γ g, where g is a centred Gaussian with variance /, independent of ξ 0. We shall construct a random variable ξ 0, supported on at most three points, such that A 0 satisfies Definition 2.2 and the first four moments of ξ are sufficiently close to those of ξ. For k =, 2,... we denote by m k X the k-th moment of a random variable X. We set ξ 0 = = 0. f γ + ξ 0, 5.4 where m ξ 0 = 0 and m 2 ξ 0 =. It is easy to see that m k ξ = m k ξ for k =, 2. We take the law of ξ 0 to be of the form pδ a + qδ b + p qδ 0 where a, b, p, q 0 are parameters satisfying p + q. The conditions m ξ 0 = 0 and m 2 ξ 0 = imply p = aa + b, q = ba + b. Thus, we parametrize ξ 0 using a and b; the condition p + q reads ab. Our aim is to determine a and b so that ξ 0 satisfies 2.4, and so that the third and fourth moments of ξ and ξ are close. By explicit computation we find m 3 ξ 0 = a b, m 4 ξ 0 = m 3 ξ 0 2 + ab. 5.5 24

ow we require that a and b be chosen so that ab and m 3 ξ 0 = γ 3/2 m 3 ξ, m 4 ξ 0 = m 3 ξ 0 2 + m 4 ξ m 3 ξ 2. Using 5.5, it is easy to see that such a pair a, b exists provided that m 4 ξ m 3 ξ 2 2. This latter estimate is generally valid for any random variable with m = 0; it follows from the elementary inequality m 4 m 2 m 2 3 m 3 2 valid whenever m = 0. ext, using 5.5 and the estimates m 3 ξ = O φ, m 4 ξ = O 2φ, we find a b = O φ, ab = O 2φ, which implies a, b = O φ. We have hence proved that A 0 satisfies Definition 2.2. One readily finds that m 3 ξ = m 3 ξ. Moreover, using we find Summarizing, we have proved m 4 ξ 0 m 4 ξ = m 3 ξ 2[ γ 3 ] = O 2φ γ, m 4 ξ m 4 ξ = γ 2 m 4 ξ 0 + 6γ 2 3γ2 2 m 4 ξ = O 2φ γ. m k ξ = m k ξ k =, 2, 3, m 4 ξ m 4 ξ C 2φ τ. 5.6 The claim follows now by setting δ = 2αφ + 2φ β in 5.3, and invoking Theorems 5. and 5.3. 6. Edge universality: proof of Theorem 2.7 6.. Rank-one perturbations of the GOE. We begin by deriving a simple, entirely deterministic, result on the eigenvalues of rank-one perturbations of matrices. We choose the perturbation to be proportional to e e, but all results of this subsection hold trivially if e is replaced with an arbitrary l 2 -normalized vector. Lemma 6. Monotonicity and interlacing. Let H be a symmetric matrix. For f 0 we set Af.= H + f e e. Denote by λ λ the eigenvalues of H, and by µ f µ f the eigenvalues of Af. Then for all α =,..., and f 0 the function µ α f is nondecreasing, satisfies µ α 0 = λ α, and has the interlacing property λ α µ α f λ α+. 6. Proof. From [], Equation 6.3, we find that µ is an eigenvalue of H + f e e if and only if α u α, e 2 µ λ α = f, 6.2 where u α is the eigenvector of H associated with the eigenvalue λ α. The right-hand side of 6.2 has singularities at λ,..., λ, away from which it is decreasing. All claims now follow easily. 25

ext, we establish the following eigenvalue sticking property for GOE. Let α label an eigenvalue close to the right say spectral edge. Roughly we prove that, in the case where H = V is a GOE matrix and f >, the eigenvalue µ α of V + f e e sticks to λ α+ with a precision log Cξ. This behaviour can be interpreted as a form of long-distance level repulsion, in which the eigenvalues µ β, β < α, repel the eigenvalue µ α and push it close to its maximum possible value, λ α+. Lemma 6.2 Eigenvalue sticking. Let V be an GOE matrix. Suppose moreover that ξ satisfies 3.0 and that f satisfies f + ε 0. Then there is a δ δε 0 > 0 such that for all α satisfying δ α we have with ξ, ν-high probability λ α+ µ α log Cξ Similarly, if α instead satisfies α δ we have with ξ, ν-high probability µ α λ α log Cξ. 6.3. 6.4 For the proof of Lemma 6.2 we shall need the following result about Wigner matrices, proved in [9]. Lemma 6.3. Let H be a Wigner matrix with eigenvalues λ λ and associated eigenvectors u,..., u. Assume that ξ is given by 3.. Then the following two statements hold with ξ, ν-high probability: and max α u α log Cξ, 6.5 λ α γ α log Cξ 2/3 min{α, + α} /3. 6.6 Moreover, let L satisfy 3. and write G H ij z.= [ H z ]. Then we have, with ξ, ν-high probability, ij z D L { max i,j where D L was defined in 3.2 G H ij z δ ij m sc z log Cξ Im m sc z η + }, 6.7 η Proof of Lemma 6.2. We only prove 6.3; the proof of 6.4 is analogous. By orthogonal invariance of V, we may replace e with the vector, 0,..., 0. Let us abbreviate ζ β.= u β 2. ote that 6.5 implies with ξ, ν-high probability. ow from we 6.2 we get max β ζ β log Cξ 6.8 ζ α + µ α λ α+ β α+ ζ β µ α λ β = f, which yields λ α+ µ α = ζ α β α+ ζ β + λ β µ α f. 6.9 26

We estimate from below, introducing an arbitrary η > 0, β α+ ζ β = λ β µ α β<α+ β<α+ ζ β µ α λ β β>α+ ζ β λ β µ α ζ β µ α λ β µ α λ β 2 + η 2 = Re G V µ α + iη + Re G V µ α + iη β>α+ β>α+ β>α+ ζ β λ β µ α ζ β λ β µ α λ β µ α 2 + η 2 β>α+ ζ β λ β µ α ζ β η 2 λ β µ α 3, 6.0 where in the third step we used that λ α+ µ α by 6.. We now choose η = log C log log. For C large enough, we get from 6.7 that G V µ α + iη = m sc µ α + iη + o. Therefore 3.6 yields Re G V µ α + iη 2 2 µ α + o. 6. From 6.6 and 6. we get that µ α γ α log Cξ 2/3 with ξ, ν-high probability. Moreover, the definition 3.5 and α δ imply γ α 2 Cδ 2/3. Thus we get, with ξ, ν-high probability, that 2 µ α = o + Cδ 2/3. Therefore 6. yields, with ξ, ν-high probability, Re G V µ α + iη + o Cδ /3. Recalling 6.8, we therefore get from 6.0, with ξ, ν-high probability, ζ β λ β µ α + o mlog Cδ/3 Cξ log Cξ 3 λ α+ µ α 3 3 β α+ β>α+m λ β µ α 3, 6.2 for any m. ext, from 6.6 we find that, provided C 2 is large enough, m.= log C2ξ, and β > α + m, then we have with ξ, ν-high probability λ β λ α+ γ β γ α+ Then for C 2 large enough we have, with ξ, ν-high probability, β>α+m β α+ λ β µ α 3 C log Cξ 2/3 + β /3 c γ β γ α+. β>α+m γ β γ α+ 3 C 3 log 3C2ξ. Thus we get from 6.2, with ξ, ν-high probability, ζ β λ β µ α + o log Cξ Cδ/3 3 λ α+ µ α 3. 27

Plugging this into 6.9 and recalling that f + ε 0 > yields, with ξ, ν-high probability, log Cξ λ α+ µ α ε 0 Cδ /3 log Cξ o 3 λ α+ µ α 2, from which the claim follows. 6.2. Proof of Theorem 2.7. In this section we prove Theorem 2.7 by establishing the following comparison result for sparse matrices. Throughout the following we shall abbreviate the lower bound in 2.7 by f.= + ε 0. 6.3 Proposition 6.4. Let P v and P w be laws on the symmetric random matrices H, each satisfying Definition 2. with q φ for some φ satisfying /3 < φ /2. In particular, we have the moment matching condition E v h ij = E w h ij = 0, E v h 2 ij = E w h 2 ij =. 6.4 Set f.= f in Definition 2.2: A Af = a ij.= H+f e e. As usual, we denote the ordered eigenvalues of H by λ λ and the ordered eigenvalues of A by µ µ. Then there is a δ > 0 such that for any s R we have P v 2/3 λ 2 s δ δ as well as P v 2/3 µ 2 s δ δ P w 2/3 λ 2 s P v 2/3 λ 2 s + δ + δ 6.5 P w 2/3 µ 2 s P v 2/3 µ 2 s + δ + δ 6.6 for 0 sufficiently large, where 0 is independent of s. Assuming Proposition 6.4 is proved, we may easily complete the proof of Theorem 2.7 using the results of Section 6.. Proof of Theorem 2.7. Choose P v to be the law of GOE see Remark 2.4, and choose P w to be the law of a sparse matrix satisfying Definition 2. with q φ. We prove 2.; the proof of 2.2 is similar. For the following we write µ α f µ α to emphasize the f-dependence of the eigenvalues of Af. Using first 6. and then 6.5 we get P w 2/3 µ f 2 s P w 2/3 λ 2 s P v 2/3 λ 2 s δ δ, for some δ > 0. ext, using first the monotonicity of µ α f from Lemma 6., then 6.6, and finally 6.3, we get P w 2/3 µ f 2 s P w 2/3 µ f 2 s P v 2/3 µ f 2 s + δ + δ P v 2/3 λ 2 s + 2 δ + 2 δ, for some δ > 0. This concludes the proof of 2., after a renaming of δ. 28