Eigenvalues and Singular Values of Random Matrices: A Tutorial Introduction

Similar documents
Random Matrix Theory Lecture 1 Introduction, Ensembles and Basic Laws. Symeon Chatzinotas February 11, 2013 Luxembourg

Theory of Large Dimensional Random Matrices for Engineers (Part I)

Supelec Randomness in Wireless Networks: how to deal with it?

Random Matrices and Wireless Communications

Free Probability, Sample Covariance Matrices and Stochastic Eigen-Inference

Estimation of the Global Minimum Variance Portfolio in High Dimensions

1 Intro to RMT (Gene)

Random Matrix Theory Lecture 3 Free Probability Theory. Symeon Chatzinotas March 4, 2013 Luxembourg

INVERSE EIGENVALUE STATISTICS FOR RAYLEIGH AND RICIAN MIMO CHANNELS

Large sample covariance matrices and the T 2 statistic

Applications and fundamental results on random Vandermon

Random Matrices: Invertibility, Structure, and Applications

Theory of Large Dimensional Random Matrices for Engineers

Determinantal point processes and random matrix theory in a nutshell

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

arxiv: v1 [math-ph] 19 Oct 2018

Exponential tail inequalities for eigenvalues of random matrices

Random Matrix: From Wigner to Quantum Chaos

Non white sample covariance matrices.

Free Probability Theory and Random Matrices. Roland Speicher Queen s University Kingston, Canada

Fluctuations from the Semicircle Law Lecture 4

Design of MMSE Multiuser Detectors using Random Matrix Techniques

Isotropic local laws for random matrices

Tight Lower Bounds on the Ergodic Capacity of Rayleigh Fading MIMO Channels

Local semicircle law, Wegner estimate and level repulsion for Wigner random matrices

WE study the capacity of peak-power limited, single-antenna,

The norm of polynomials in large random matrices

Ergodic and Outage Capacity of Narrowband MIMO Gaussian Channels

COMPLEX HERMITE POLYNOMIALS: FROM THE SEMI-CIRCULAR LAW TO THE CIRCULAR LAW

1. Introduction. Let an n m complex Gaussian random matrix A be distributed

Ergodic and Outage Capacity of Narrowband MIMO Gaussian Channels

Lecture 7 MIMO Communica2ons

Free Meixner distributions and random matrices

Insights into Large Complex Systems via Random Matrix Theory

Eigenvalue variance bounds for Wigner and covariance random matrices

Multiuser Receivers, Random Matrices and Free Probability

Free deconvolution for signal processing applications

Universality of distribution functions in random matrix theory Arno Kuijlaars Katholieke Universiteit Leuven, Belgium

Mutual Information and Eigenvalue Distribution of MIMO Ricean Channels

Random matrices: Distribution of the least singular value (via Property Testing)

Invertibility of random matrices

Preface to the Second Edition...vii Preface to the First Edition... ix

Concentration Inequalities for Random Matrices

LARGE DIMENSIONAL RANDOM MATRIX THEORY FOR SIGNAL DETECTION AND ESTIMATION IN ARRAY PROCESSING

Distribution of Eigenvalues of Weighted, Structured Matrix Ensembles

On corrections of classical multivariate tests for high-dimensional data. Jian-feng. Yao Université de Rennes 1, IRMAR

Inhomogeneous circular laws for random matrices with non-identically distributed entries

Triangular matrices and biorthogonal ensembles

Operator-Valued Free Probability Theory and Block Random Matrices. Roland Speicher Queen s University Kingston

Universality of local spectral statistics of random matrices

Selfadjoint Polynomials in Independent Random Matrices. Roland Speicher Universität des Saarlandes Saarbrücken

Progress in the method of Ghosts and Shadows for Beta Ensembles

EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME. Xavier Mestre 1, Pascal Vallet 2

Parallel Additive Gaussian Channels

Operator norm convergence for sequence of matrices and application to QIT

MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran

Strong Convergence of the Empirical Distribution of Eigenvalues of Large Dimensional Random Matrices

Markov operators, classical orthogonal polynomial ensembles, and random matrices

An introduction to G-estimation with sample covariance matrices

c 2005 Society for Industrial and Applied Mathematics

Fluctuations of Random Matrices and Second Order Freeness

From random matrices to free groups, through non-crossing partitions. Michael Anshelevich

Matrix-valued stochastic processes

arxiv: v1 [cs.na] 6 Jan 2017

On corrections of classical multivariate tests for high-dimensional data

Open Issues on the Statistical Spectrum Characterization of Random Vandermonde Matrices

The Maximum Entropy Principle and Applications to MIMO Channel Modeling

Random Matrices: Beyond Wigner and Marchenko-Pastur Laws

1 Tridiagonal matrices

The Matrix Dyson Equation in random matrix theory

RANDOM MATRIX THEORY AND TOEPLITZ DETERMINANTS

The Free Central Limit Theorem: A Combinatorial Approach

Statistical Inference and Random Matrices

A note on a Marčenko-Pastur type theorem for time series. Jianfeng. Yao

On the Optimum Asymptotic Multiuser Efficiency of Randomly Spread CDMA

Assessing the dependence of high-dimensional time series via sample autocovariances and correlations

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

Random matrices: A Survey. Van H. Vu. Department of Mathematics Rutgers University

SEA s workshop- MIT - July 10-14

Alan Edelman. Why are random matrix eigenvalues cool? MAA Mathfest 2002 Thursday, August 1. MIT: Dept of Mathematics, Lab for Computer Science

Random Matrices in Machine Learning

Covariance estimation using random matrix theory

Complex singular Wishart matrices and applications

STAT C206A / MATH C223A : Stein s method and applications 1. Lecture 31

Advances in Random Matrix Theory: Let there be tools

Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage?

ELEMENTS OF PROBABILITY THEORY

Free probability and quantum information

Lecture 2. Spring Quarter Statistical Optics. Lecture 2. Characteristic Functions. Transformation of RVs. Sums of RVs

Full text available at: Random Matrix Theory and Wireless Communications

Universality for random matrices and log-gases

Freeness and the Transpose

Local law of addition of random matrices

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata

On the Fluctuation of Mutual Information in Double Scattering MIMO Channels

On the distinguishability of random quantum states

Products of Rectangular Gaussian Matrices

MIMO Capacities : Eigenvalue Computation through Representation Theory

Random Fermionic Systems

Transcription:

Random Matrix Theory and its applications to Statistics and Wireless Communications Eigenvalues and Singular Values of Random Matrices: A Tutorial Introduction Sergio Verdú Princeton University National University of Singapore Institute for Mathematical Sciences February 27, 2006

Applications of Random Matrices Condensed Matter Physics Statistical Physics String Theory Quantum Chaos Disordered Systems Number Theory 1

Applications of Random Matrices Riemann Hypothesis von Neumann and C*-algebra theory Multivariate Statistics Stochastic Differential Equations Numerical Linear Algebra Economics 2

Engineering Applications of Random Matrices Information Theory Wireless Communications Signal Processing Neural Networks Small-World Networks 3

Typical Random Matrix Questions Distribution of λ(h) Distribution of λ(h H) Distribution of λ max (H) E [ det ( H k)] E[det(I + γw)] Joint distribution of λ 1 (H)... λ N (H) Distribution of the spacings between adjacent eigenvalues Distribution of H H Distribution of the matrix of eigenvectors of H H 4

Our Motivation Vector noisy channels of the form where y = Hx + n (1) x is the K-dimensional input vector, y is the N-dimensional output vector, n is the circularly symmetric N-dimensional vector Gaussian noise. 5

Why Singular Values? Shannon Capacity = 1 N N log ( 1 + SNR λ i (HH ) ) i=1 6

Why Singular Values? Minimum Mean Square Error = 1 K K i=1 1 1 + SNR λ i (H H) 7

Asymptotics K N K N β 8

Asymptotics Shannon Capacity = 1 N N log ( 1 + SNR λ i (HH ) ) i=1 4 4 3 3 2 2 1 1 0 0 2 4 6 8 10 0 0 2 4 6 8 10 N = 3 SNR N = 5 SNR 4 4 3 3 2 2 1 1 0 0 2 4 6 8 10 0 0 2 4 6 8 10 SNR N = 15 N = 50 SNR Figure 1: β = 1 for sizes: N = 3, 5, 15, 50 9

The Birth of (Nonasymptotic) Random Matrix Theory: (Wishart, 1928) J. Wishart, The generalized product moment distribution in samples from a normal multivariate population, Biometrika, vol. 20 A, pp. 32 52, 1928. Probability density function of the matrix: v 1 v 1 +... + v nv n where v i are iid Gaussian vectors. 10

Wishart Matrices Definition 1. The m m random matrix A = HH is a (central) real/complex Wishart matrix with n degrees of freedom and covariance matrix Σ, (A W m (n, Σ)), if the columns of the m n matrix H are zeromean independent real/complex Gaussian vectors with covariance matrix Σ. 1 The p.d.f. of a complex Wishart matrix A W m (n, Σ) for n m is f A (B) = π m(m 1)/2 detσ n m i=1 (n i)! exp [ tr { Σ 1 B }] detb n m. (2) 1 If the entries of H have nonzero mean, HH is a non-central Wishart matrix. 11

Wishart Matrices: Eigenvectors Theorem 1. The matrix of eigenvectors of Wishart matrices is uniformly distributed on the manifold of unitary matrices ( Haar measure) 12

The Birth of Asymptotic Random Matrix Theory E. Wigner, Characteristic vectors of bordered matrices with infinite dimensions, The Annals of Mathematics, vol. 62, pp. 546 564, 1955. 1 n 0 +1 +1 1 1 +1 +1 0 1 1 +1 +1 +1 1 0 +1 +1 1 1 1 +1 0 +1 +1 1 +1 +1 +1 0 1 +1 +1 1 +1 1 0 As the matrix dimension n, the histogram of the eigenvalues converges to...? Motivation: bypass the Schrödinger equation and explain the statistics of experimentally measured atomic energy levels in terms of the limiting spectrum of those random matrices. 13

Wigner Matrices Definition 2. An n n Hermitian matrix W is a Wigner matrix if its upper-triangular entries are independent zero-mean random variables with identical variance. If the variance is 1 n, then W is a standard Wigner matrix. 14

Wigner Matrices: The Semicircle Law E. Wigner, On the distribution of roots of certain symmetric matrices, The Annals of Mathematics, vol. 67, pp. 325 327, 1958. Theorem 2. Consider an N N standard Wigner matrix W such that, for some constant κ, and sufficiently large N max E [ W i,j 4] κ 1 i j N N 2 (3) Then, the empirical distribution of W converges almost surely to the semicircle law whose density is with x 2. w(x) = 1 4 x 2 (4) 2π 15

Wigner Matrices: The Semicircle Law 0.3 0.25 0.2 0.15 0.1 0.05 0 2.5 2 1.5 1 0.5 0 0.5 1 1.5 2 2.5 Figure 2: The semicircle law density function (4) compared with the histogram of the average of 100 empirical density functions for a Wigner matrix of size n = 100. 16

Wigner Matrices: The Semicircle Law Wigner s original proof of the convergence to the semicircle law: the empirical moments 1 N tr { W 2k} the Catalan numbers: lim N 1 N tr { W 2k} = = 2 2 1 k + 1 x 2k w(x) dx ( 2k k ). (5) 17

Asymptotics Distribution Insensitivity: The asymptotic eigenvalue distribution does not depend on the distribution with which the independent matrix coefficients are generated. Ergodicity : The eigenvalue histogram of one matrix realization converges almost surely to the asymptotic eigenvalue distribution. Speed of Convergence: 8 =. Gaussian case: Nonasymptotic joint distribution of eigenvalues known. 18

Gaussian Wigner Matrices: Nonasymptotic M. L. Mehta and M. Gaudin, On the density of the eigenvalues of a random matrix, Nuclear Physics, vol. 18, pp. 420 427, 1960. Theorem 3. Let W be an n n Wigner matrix whose entries are i.i.d. zeromean Gaussian with unit variance. Then, its p.d.f. is 2 n/2 π n2 /2 exp [ ] tr{w2 } 2 while the joint p.d.f. of its ordered eigenvalues λ 1... λ n is (6) 1 ni=1 (2π) n/2 e 1 2 λ 2 i n 1 i=1 1 i! n (λ i λ j ) 2. (7) i<j 19

Gaussian Wigner Matrices: Nonasymptotic E. Wigner, Distribution laws for the roots of a random Hermitian matrix, in Statistical Theories of Spectra: Fluctuations, (C. E. Porter, ed.), New York: Academic, 1965. Theorem 4. Let W be an n n Gaussian Wigner matrix. The marginal p.d.f. of the unordered eigenvalues is 1 n n 1 i=0 1 2 i i! 2π ( e x2 4 Hi (x)) 2 (8) with H i ( ) the ith Hermite polynomial. 20

Square matrix of iid coefficients 1 n 1 +1 +1 1 1 +1 1 +1 1 1 +1 +1 +1 +1 +1 +1 +1 1 +1 1 +1 1 +1 +1 1 1 +1 1 1 1 1 +1 +1 +1 +1 1 1.5 1 0.5 0 0.5 1 1.5 1.5 1 0.5 0 0.5 1 1.5 Figure 3: The full-circle law and the eigenvalues of a realization of a 500 500 matrix 21

Full Circle Law V. L. Girko, Circular law, Theory Prob. Appl., vol. 29, pp. 694 706, 1984. Z. D. Bai, The circle law, The Annals of Probability, pp. 494 529, 1997. Theorem 5. Let H be an N N complex random matrix whose entries are independent random variables with identical mean and variance and finite kth moments for k 4. Assume that the joint distributions of the real and imaginary parts of the entries have uniformly bounded densities. Then, the asymptotic spectrum of H converges almost surely to the circular law, namely the uniform distribution over the unit disk on the complex plane {ζ C : ζ 1} whose density is given by f c (ζ) = 1 π ζ 1 (9) (also holds for real matrices replacing the assumption on the joint distribution of real and imaginary parts with the one-dimensional distribution of the real-valued entries.) 22

Singular Values: Fisher-Hsu-Girshick-Roy R. A. Fisher, The sampling distribution of some statistics obtained from non-linear equations, The Annals of Eugenics, vol. 9, pp. 238 249, 1939. M. A. Girshick, On the sampling theory of roots of determinantal equations, The Annals of Math. Statistics, vol. 10, pp. 203 204, 1939. P. L. Hsu, On the distribution of roots of certain determinantal equations, The Annals of Eugenics, vol. 9, pp. 250 258, 1939. S. N. Roy, p-statistics or some generalizations in the analysis of variance appropriate to multivariate problems, Sankhya, vol. 4, pp. 381 396, 1939. The joint p.d.f. of the ordered strictly positive eigenvalues of the Wishart matrix HH, where the entries of H are i.i.d. complex Gaussian with zero mean and unit variance. 23

Singular Values 2 : Fisher-Hsu-Girshick-Roy Theorem 6. Let the entries of H be i.i.d. complex Gaussian with zero mean and unit variance. The joint p.d.f. of the ordered strictly positive eigenvalues of the Wishart matrix HH, λ 1... λ t, equals e t i=1 λ i t i=1 λ r t i (t i)! (r i)! t (λ i λ j ) 2 (10) i<j where t and r are the minimum and maximum of the dimensions of H. The marginal p.d.f. of the unordered eigenvalues is g r,t (λ) = 1 t t 1 k=0 k! (k + r t)! [ L r t k (λ) ] 2 λ r t e λ (11) where the Laguerre polynomials are L n k(λ) = eλ k!λ n d k dλ k ( e λ λ n+k). (12) 24

Singular Values 2 : Fisher-Hsu-Girshick-Roy 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0 5 4 3 2 1 0 0 1 2 3 4 5 Figure 4: Joint p.d.f. of the unordered positive eigenvalues of the Wishart matrix HH with n = 3 and m = 2. (Scaled version of (10).) 25

Asymptotic Distribution of Singular Values: Quarter circle law Consider an N N matrix H whose entries are independent zero-mean complex (or real) random variables with variance 1 N, the asymptotic distribution of the singular values converges to q(x) = 1 π 4 x2, 0 x 2 (13) 26

Asymptotic Distribution of Singular Values: Quarter circle law 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 0.5 1 1.5 2 2.5 Figure 5: The quarter circle law compared a histogram of the average of 100 empirical singular value density functions of a matrix of size 100 100. 27

Minimum Singular Value of Gaussian Matrix A. Edelman, Eigenvalues and condition number of random matrices. PhD thesis, Dept. Mathematics, MIT, Cambridge, MA, 1989. J. Shen, On the singular values of Gaussian random matrices, Linear Algebra and its Applications, vol. 326, no. 1-3, pp. 1 14, 2001. Theorem 7. The minimum singular value of an N N standard complex Gaussian matrix H satisfies lim P [Nσ min x] = e x x2 /2. (14) N 28

The Mar cenko-pastur Law V. A. Mar cenko and L. A. Pastur, Distributions of eigenvalues for some sets of random matrices, Math USSR-Sbornik, vol. 1, pp. 457 483, 1967. Theorem 8. Consider an N K matrix H whose entries are independent zero-mean complex (or real) random variables with variance 1 N and fourth moments of order O( 1 ). As K, N with K N 2 N β, the empirical distribution of H H converges almost surely to a nonrandom limiting distribution with density f β (x) = ( 1 1 β) + δ(x) + (x a)+ (b x) + 2πβx (15) where a = (1 β) 2 b = (1 + β) 2. 29

The Mar cenko-pastur Law 2 1.8 1.6 1.4 β= 1 1.2 1 0.5 0.8 0.6 0.2 0.4 0.2 0 0 0.5 1 1.5 2 2.5 Figure 6: The Mar cenko-pastur density function for β = 1, 0.5, 0.2. 30

Rediscovering/Strenghtening the Mar cenko-pastur Law U. Grenander and J. W. Silverstein, Spectral analysis of networks with random topologies, SIAM J. of Applied Mathematics, vol. 32, pp. 449 519, 1977. K. W. Wachter, The strong limits of random matrix spectra for sample matrices of independent elements, The Annals of Probability, vol. 6, no. 1, pp. 1 18, 1978. J. W. Silverstein and Z. D. Bai, On the empirical distribution of eigenvalues of a class of large dimensional random matrices, J. of Multivariate Analysis, vol. 54, pp. 175 192, 1995. Y. L. Cun, I. Kanter, and S. A. Solla, Eigenvalues of covariance matrices: Application to neural-network learning, Physical Review Letters, vol. 66, pp. 2396 2399, 1991. 31

Generalizations needed! W = HTH W = W 0 + HH W = W 0 + HTH 32

Transforms 1. Stieltjes transform 2. η transform 3. Shannon transform 4. Mellin transform 5. R-transform 6. S-transform 33

Stieltjes Transform Let X be a real-valued random variable with distribution F X ( ). Its Stieltjes transform is defined for complex arguments as [ ] 1 S X (z) = E = X z 1 λ z df X(λ). (16) 34

Stieltjes Transform and Moments S X (z) = 1 z k=0 E[X k ] z k. (17) 35

Inverse Stieltjes Transform T. J. Stieltjes, Recherches sur les fractions continues, Annales de la Faculte des Sciences de Toulouse, vol. 8 (9), no. A (J), pp. 1 47 (1 122), 1894 (1895). [ ] 1 f X (λ) = lim ω 0 + π Im S X (λ + j ω). (18) 36

Stieltjes Transform of Semicircle law w(x) = { 1 2π 4 x 2 if x 2 0 if x > 2 (19) S w (z) = 1 2π 2 2 4 λ 2 λ z dλ = 1 2 [ z ± z 2 4 ]. (20) 37

Stieltjes Transform of Mar cenko-pastur law f β (x) = ( 1 1 β) + δ(x) + (x a)+ (b x) + 2πβx (21) S fβ (z) = b a 1 λ z f β(λ) dλ = 1 β z ± z 2 2(β + 1)z + (β 1) 2. (22) 2βz 38

η transform S. Verdú, Large random matrices and wireless communications, 2002 MSRI Information Theory Workshop, Feb 25 Mar 1, 2002. Definition 3. where γ 0. The η-transform of a nonnegative random variable X is [ ] 1 η X (γ) = E 1 + γx (23) Note: 0 < η X (γ) 1. 39

Why Singular Values? Minimum Mean Square Error = 1 K K i=1 1 1 + SNR λ i (H H) 40

η transform and Stieltjes transform η X (γ) = S X( 1 γ ) γ (24) η X (γ) = ( γ) k E[X k ], (25) k=0 41

η-transform of Mar cenko-pastur law Example: The η-transform of the Mar cenko-pastur law is η(γ) = 1 F(γ, β) 4 β γ (26) with F(x, z) = ( x(1 + z) 2 + 1 x(1 z) 2 + 1) 2. (27) 42

Random Quadratic Forms Z. D. Bai and J. W. Silverstein, No eigenvalues outside the support of the limiting spectral distribution of large dimensional sample covariance matrices, The Annals of Probability, vol. 26, pp. 316 345, 1998. Theorem 9. Let the components of the N-dimensional vector x be zeromean and independent with variance 1 N. For any N N nonnegative definite random matrix B independent of x whose spectrum converges almost surely, lim N x (I + γb) 1 x = η B (γ) a.s. (28) lim N x (B zi) 1 x = S B (z) a.s. (29) 43

Shannon transform S. Verdú, Random matrices in wireless communication, proposal to the National Science Foundation, Feb. 1999. Definition 4. is defined as The Shannon transform of a nonnegative random variable X V X (γ) = E[log(1 + γx)]. (30) where γ 0 44

Why Singular Values? Shannon Capacity = 1 N N log ( 1 + SNR λ i (HH ) ) i=1 45

Stieltjes, Shannon and η γ log e d dγ V X(γ) = 1 1 γ S X ( 1 ) γ = 1 η X (γ) (31) 46

Shannon transform of Mar cenko-pastur law Example: The Shannon transform of the Mar cenko-pastur law is V(γ) = log (1 + γ 14 ) F (γ, β) + 1 β log ( 1 + γβ 1 4 F (γ, β) ) log e F (γ, β) (32) 4 β γ 47

Mellin transform Definition 5. by The Mellin transform of a positive random variable X is given M X (z) = E[X z 1 ] (33) where z belongs to a strip of the complex plane where the expectation is finite. 48

Mellin and Shannon transforms R. Janaswamy, Analytical expressions for the ergodic capacities of certain MIMO systems by the Mellin transform, Proc. of IEEE Global Telecomm. Conf., vol. 1, pp. 287 291, Dec. 2003. Theorem 10. where M 1 Υ V X (γ) = M 1 Υ (γ) (34) is the inverse Mellin transform of Υ(z) = z 1 Γ(z)Γ(1 z)m X (1 z) (35) 49

R-transform D. Voiculescu, Addition of certain non-commuting random variables, J. Funct. Analysis, vol. 66, pp. 323 346, 1986. Definition 6. R X (z) = S 1 X ( z) 1 z. (36) 50

R-transform and η-transform η X (γ) = 1 1 + γ R X ( γ η X (γ)) (37) 51

R-transform of the semicircle law R(z) = z. (38) 52

R-transform of the Mar cenko-pastur law R(z) = 1 1 βz. (39) 53

R-transform: Key theorem D. Voiculescu, Addition of certain non-commuting random variables, J. Funct. Analysis, vol. 66, pp. 323 346, 1986. Theorem 11. If A and B are asymptotically free random matrices, then the R-transform of their sum satisfies R A+B (z) = R A (z) + R B (z) (40) free analog of the log-moment generating function 54

S-transform D. Voiculescu, Multiplication of certain non-commuting random variables, J. Operator Theory, vol. 18, pp. 223 235, 1987. Definition 7. The S-transform of a nonnegative random variable X is Σ X (x) = x + 1 x η 1 X (1 + x), (41) which maps ( 1, 0) onto the positive real line. 55

S-transform of the Mar cenko-pastur law Σ(x) = 1 1 + βx. (42) 56

S-transform: Key Theorem D. Voiculescu, Multiplication of certain non-commuting random variables, J. Operator Theory, vol. 18, pp. 223 235, 1987. Theorem 12. Let A and B be nonnegative asymptotically free random matrices. The S-transform of their product satisfies Σ AB (x) = Σ A (x)σ B (x) (43) 57

Generalizations needed! W = HTH W = W 0 + HH W = W 0 + HTH 58

Asymptotic Spectrum of HTH V. A. Mar cenko and L. A. Pastur, Distributions of eigenvalues for some sets of random matrices, Math USSR-Sbornik, vol. 1, pp. 457 483, 1967. Theorem 13. Let H be an N K matrix whose entries are i.i.d. variables with variance 1 N. complex random Let T be a K K Hermitian nonnegative random matrix, independent of H, whose empirical eigenvalue distribution converges almost surely to a nonrandom limit. The empirical eigenvalue distribution of HTH converges almost surely, as K, N with K N β, with η HTH (γ) = η the solution of: β = 1 η 1 η T (γη) (44) 59

Example Further, if T = I, we have η T (γ) = 1 1+γ, and (44) becomes: η = 1 β + β 1 + γη 60

Asymptotic Spectrum of HTH Theorem 14. Let H be an N K matrix whose entries are i.i.d. variables with variance 1 N. complex random Let T be a K K Hermitian nonnegative random matrix, independent of H, whose empirical eigenvalue distribution converges almost surely to a nonrandom limit. The empirical eigenvalue distribution of HTH converges almost surely, as K, N with K N β, with Shannon transform V HTH (γ) = βv T (ηγ) + log 1 η + (η 1) log e (45) 61

Asymptotic Spectrum of W 0 + HTH A. Tulino and S. Verdú, Random Matrix Theory and Wireless Communications, Now Publishers, 2004 Theorem 15. Let H be an N K matrix whose entries are i.i.d. variables with zero-mean and variance 1 N. complex random Let T be a K K positive definite random matrix whose empirical eigenvalue distribution converges almost surely to a nonrandom limit. Let W 0 be an N N nonnegative definite diagonal random matrix with empirical eigenvalue distribution converging almost surely to a nonrandom limit. H, T, and W 0 are independent 62

The empirical eigenvalue distribution of W = W 0 + HTH (46) converges almost surely, as K, N with K N β, to a nonrandom limiting distribution whose η-transform is the solution of the following pair of equations: γ η = ϕ η 0 (ϕ) (47) η = η 0 (ϕ) β (1 η T (γ η)) (48) with η 0 and η T the η-transforms of W 0 and T respectively. 63