Error Bound for Compound Wishart Matrices

Size: px
Start display at page:

Download "Error Bound for Compound Wishart Matrices"

Transcription

1 Electronic Journal of Statistics Vol. 0 (2006) 1 8 ISSN: Error Bound for Compound Wishart Matrices Ilya Soloveychik, The Rachel and Selim Benin School of Computer Science and Engineering, The Hebrew University of Jerusalem, Givat Ram, Jerusalem, ilya.soloveychik@mail.huji.ac.il Abstract: In this paper we consider the non-asymptotic behavior of the real compound Wishart matrices. This family of random matrices generalizes the classical real Wishart distribution. In particular, we consider matrices of the form XBX T, where X is a p n matrix with centered Gaussian elements and B is an arbitrary n n matrix; and sequences of such matrices for varying n. We derive high-probability error bounds on the deviations of the compound Wishart matrices from their mean. MSC 2010 subject classifications: Primary 62E99. Keywords and phrases: Compound Wishart distribution, correlated sample covariance matrix, concentration of Gaussian measure, sample complexity. 1. Introduction For a real p n matrix X consisting of normally distributed elements the p p Wishart matrix, defined as W = XX T, was introduced by Wishart [1], who also derived the law of its distribution. As evidenced by the wide interest among the scientists and engineers, the Wishart law is of primary importance to statistics, see e.g. [2, 3]. In particular, Wishart law describes exactly the distribution of the sample covariance matrix in the Gaussian populations. As a natural generalization, the compound Wishart matrices were introduced by Speicher [4]. Definition 1. (Compound Wishart Matrix) Let X i N (0, Θ), i = 1,..., n, be independent and belong to R p, where Θ is a p p real positive definite matrix, and B - be an arbitrary real n n matrix. We say that a random p p matrix W is compound Wishart with shape parameter B and scale parameter Θ if where X = [X 1,..., X n ]. We write W W(Θ, B). W = XBX T, (1.1) Remark 1. It will sometimes be instructive to use the representation W = Θ 1/2 Y BX T Θ 1/2, where Y = [Y 1,..., Y n ] and Y i N (0, I). This work was partially supported Kaete Klausner Scholarship, the Hebrew University of Jerusalem 1

2 I. Soloveychik/Error Bound for Compound Wishart Matrices 2 When B is positive definite, [5] interprets W as a scaled sample covariance under correlated sampling. Similar interpretation for the complex case appears in [6]. The usual real Wishart matrices correspond to the choice B = I. Wishart distribution and its generalization to the case of positive definite matrix B are widely used in economics, in particular in portfolio allocation theory [7]. When applying the representation theoretic machinery to the study of symmetries of normally distributed random vectors (see [8] for an example of such setting) we encountered the case of compound Wishart distribution with a skew symmetric B of the form (2.2) and this problem motivated the current study. Most of the literature concerning Wishart distribution deals with the asymptotic, n, and the double asymptotic n, p regimes. In particular, a version of Marchenko-Pastur law was generalized to this case, see [9] and references therein for a wide survey on the asymptotic behavior of the compound Wishat matrices. In a different line of research in the recent years there have been observed a number of prominent achievements in the random matrix theory concerning the non-asymptotic regime. In particular, Bernstein inequality and other concentration inequalities [10, 11] were generalized to the matrix case using the Stein s method of exchangeable pairs and Lieb s Theorem. The finite sample performance of the sample covariance matrix was also profoundly investigated for a large class of distributions, see [12, 13, 14] and references therein. Closely related to the Wishart family of distributions is partial estimation of covariance matrices by the sample covariance in Gaussian models. In particular, matrices of the form M XX T, where denotes the Hadamard (entry-wise) product of matrices, became attractive due to the advances in structured covariance estimation [15, 16, 17, 18]. The matrix M represents the a priory knowledge about the structure of the true covariance matrix in the form of a mask. The most widespread examples of such assumptions are banding, tapering and thresholding, which assume the elements of the mask M belong to the interval [0, 1]. The non-asymptotic behavior of such masked sample covariance matrices was investigated in [19, 20]. It has been shown that the sample complexity (the minimal number of samples need to achieve some predefined accuracy with stated probability) is proportional to mlog c (2p), where m is a sparsity parameter of M and c 1. The main purpose of this contribution is to demonstrate the non-asymptotic results concerning the average deviations of the compound Wishart matrices from their mean. We actually consider two closely related problems. A Single Compound Wishart matrix. The first one can be roughly described as following: given the dimension p the and shape matrix B, what accuracy can one obtain on average when approximating the expected value of the corresponding Wishart matrix by its empirical measurement. A Sequence of Compound Wishart matrices. Another setting arise from a different approach to the problem: assume we are given a sequence of matrices B n satisfying some assumptions stated below and the dimen-

3 I. Soloveychik/Error Bound for Compound Wishart Matrices 3 sion p is fixed. This data provides us with a sequence of Wishart matrices W n = XB n X T, where the number of columns in X changes appropriately (an exact definition provided below). Assume in addition that all these Wishart matrices have a common expectation W 0. The natural question is how many measurements n does one need to collect in order to estimate the mean value W 0 accurately. Both of these problems arise in different areas of research and non-asymptotic analysis is often required. The problem becomes especially critical when the values p and n are of the same order of magnitude. Below we provide a theorem answering the two posed questions. Although the result obtained in Corollary 2 is related to the case of fixed dimension p, it can be extended to a sequence W p n, where the dimension p = p(n) varies with n, while keeping the spectral properties of the sequence of corresponding covariance matrices Θ n controlled. In particular, a partial answer to the second problem can be formulated as following: the number of samples proportional to p ln 2 p is needed to accurately estimate the expectation of the compound Wishart matrix. The rest of the text is organized as following. After the notations section we provide additional definitions and the statements of the results. Then a few examples demonstrating the applications are given. The proof of the theorem concludes the paper. Notations Capital last letters of the English alphabet (W, X, Y, Z) denote random entities, all the other letters stand for deterministic entities. For an arbitrary rectangular matrix A, A denotes its spectral norm and A F = Tr (AA T ) stands for its Frobenius norm. For two vectors v, u laying in a Euclidean space, (u, v) denotes their scalar product and v 2 the corresponding length. I m stands for the m m identity matrix, when the dimension is obvious from the context the subscript is omitted. By P(p) we denote the open cone of p p positive definite matrices. For brevity we write [p] = {1,..., p}, where p N. 2. Problem Formulation and the Main Results In addition to the definitions given above we define a notion of sequence of Wishart matrices, corresponding to the second question stated in the Introduction. Definition 2. (Sequence of Wishart Matrices) Consider a sequence {B n } n1 n=n 0 of real n n deterministic matrices, where n 0 N and n 1 N { }. Let Θ P(p) and for every n = n 0,..., n 1 : X i N (0, Θ), i = 1,..., n be independent. Then W n = XB n X T, n = n 0,..., n 1, (2.1) where X = [X 1,..., X n ], is a sequence of compound Wishart matrices.

4 I. Soloveychik/Error Bound for Compound Wishart Matrices 4 The same Remark 1 as above applies here, as well. Note also that the dimension of X depends on n, but this is not reflected by an additional subscript. To make this definition useful and meaningful we will have to make some assumptions on the sequence {W n } n S. In particular, we first want to answer the following question: what properties should we require from the sequence {B n } n S to make the sequence {W n } n S interesting to investigate. Below we fix the dimension p and refer to vectors X i, i = 1,..., n as measurements. So what actually changes from matrix to another in the sequence W n is the underlying matrix B n and the respective number of measurements. The examples of sequences {B n } n S are the following: The most widely used is the sequence of diagonal matrices B n = diag {b 1,..., b n }, n S. When Tr (B) = n the expectation of the Wishart sequence coincides with the covariance matrix Θ as shown below. Another common example is a sequence of skew-symmetric matrices of the form B n = ( 0 In/2 I n/2 0 ), (2.2) where n is assumed even. We encountered this case when investigating the group symmetry properties of sample covariance matrices. In order to generalize the properties of {B n } n S we encountered in the application we state an additional auxiliary result that we did not found in the literature. Lemma 1. Let B be a real n n matrix and X i N (0, I), i = 1,..., n, be independent, then for X = [X 1,..., X n ], W = XBX T, E(W ) = Tr (B) I. (2.3) Proof. Denote the expectation E(W ) by W 0 and consider the elements of W 0 : Wij 0 = E X ik B kl X jl. (2.4) k,l=1 As all X ik, i = 1,..., p, k = 1,..., n are independent we get immediately that Wij 0 = B kl E (X ik X jl ) = 0, i j, (2.5) W 0 ii = k,l=1 B kl E (X ik X il ) = k,l=1 And the statement follows. B kk E ( Xik) 2 = Tr (B). (2.6) Corollary 1. Let B be a real n n matrix, Θ P(p), and X i N (0, Θ), i = 1,..., n be independent, then for X = [X 1,..., X n ], W = XBX T, k E(W ) = Tr (B) Θ. (2.7)

5 I. Soloveychik/Error Bound for Compound Wishart Matrices 5 In particular, Lemma 1 implies that if n 1 =, then to ensure the sequence {W n } is consistent we should require Tr (B n ) β R. We actually make a stronger Assumption 1. The traces Tr (B n ) are equal Tr (B n ) = β, n = n 0,..., n 1. (2.8) Since we can scale the sequence {W n } n1 n=n 0, without loss of generality set Tr (B n ) = The Main Results In this section we state the main results of this paper. Theorem 1. Let Θ P(p) and X i N (0, Θ), i = 1,..., n, be independent. Let B be an arbitrary real n n matrix and denote κ = B F B, σ = B, then E W W 0 24 ln 2p 2 p(4σ + κ π) n Proof. The proof is covered by Sections?? and??. Θ. Corollary 2. Let Θ P(p) and {B n } n S R n n, where S N is ordered. For every n = n 0,..., n 1 let X i N (0, Θ), i = 1,..., n, be independent. Assume that Tr (B n ) = 1, n = n 0,..., n 1, and denote κ = max n0 n n 1 B F and σ = max n0 n n 1 B n, then 4. Preliminaries 4.1. Proof outline E W n W 0 24 ln 2p 2 p(4σ + κ π) n Θ. In the rest of this paper we prove Theorems 1 and 2. We shall observe that for a compound Wishart matrix W, the quadratic form (W x, y) is a Gaussian chaos (defined below) for fixed unit vectors x and y on the unit sphere S p 1. We control the chaos uniformly for all x, y by establishing concentration inequalities depending on the sparsity of x and y. We do so by applying the techniques of decoupling, conditioning, and applying concentration bounds for Gaussian measure. After this we make use of covering arguments to measure the number of sparse vectors x, y on the sphere S p 1. The general layout of the proof goes after [19], while we modify and generalize their results.

6 4.2. Decoupling I. Soloveychik/Error Bound for Compound Wishart Matrices 6 We start by considering bilinear forms in normally distributed vectors. The following definition will be useful is the sequel. Definition 3. Let Z R p be a centered Gaussian random vector and B a square p p matrix, then the bilinear form (BZ, Z) is called a quadratic Gaussian chaos. Lemma 2. (Decoupling of Gaussian Chaos) Let Z R p be a centered normal random vector and Z its independent copy. Let also B P(p), then E sup (BZ, Z) E(BZ, Z) E sup (BZ, Z ). (4.1) Proof. Without loss of generality assume Z is standard, otherwise replace B with Θ 1/2 BΘ 1/2, where Θ is the covariance matrix of Z, and follow the same reasoning. E := E Z sup (BZ, Z) E(BZ, Z) = E Z sup (BZ, Z) E Z (BZ, Z ) E Z,Z sup (BZ, Z) (BZ, Z ), where the equality is due to the fact that the distributions of Z and Z are identical and the inequality is due to Jensen. In the calculation above we emphasized explicitly the variables of integration in the expectations to make the transitions clear. For an arbitrary B note the identity (BZ, Z) (BZ, Z ) = (B Z + Z 2, Z Z 2 ) + (B Z Z, Z + ) Z. (4.2) 2 2 ( ) Z+Z By the rotation invariance of the standard Gaussian measure, the pair 2, Z Z 2 is distributed identically to (Z, Z ), hence we conclude that E E Z,Z sup (BZ, Z ) + (BZ, Z) E Z,Z sup (BZ, Z ) + E Z,Z sup (BZ, Z) = 2E Z,Z sup (BZ, Z), (4.3) and the statement follows. Lemma 3. Let X 1,..., X n, X 1,..., X n N (0, Θ), Θ P(p), be all independent. Consider the compound Wishart matrix and its decoupled counterpart Denote W 0 = E(W ) = Tr (B) Θ, then W = XBX T, W = X BX T. (4.4) E W W 0 2E W. (4.5)

7 I. Soloveychik/Error Bound for Compound Wishart Matrices 7 Proof. Using the definition of spectral norm we obtain Rewrite the inner product as E W W 0 = E sup (W x, y) E(W x, y). (4.6) x,y S p 1 (W x, y) = x T W T y = Tr ( XB T X T yx T ) = vec (X) T [ B T (yx T ) ] vec (X) = ([ B T (yx T ) ] vec (X), vec (X) ), (4.7) where vec (X) is obtained from the matrix X by stacking the columns into a single vector of height pn and denotes the Kronecker product. Since vec (X) is Gaussian, the right-hand side of the last equality is a quadratic Gaussian chaos and Lemma 2 applies Concentration Lemma 4. Let Z = (Z 1,..., Z p ) T N (0, Θ), Θ P(p) and let a = (a 1,..., a p ) T R p. Then (a, Z) is a centered normal variable with standard deviation Θ 1/2 a 2 Θ 1/2 a 2. Next we state an auxiliary result from the Concentration of the Gaussian Measure theory. Such concentration results are usually stated in terms of the standard normal distribution, but they can be easily generalized to an arbitrary normal distribution as follows. Lemma 5. [21] Let f : R p R be a Lipschitz function with respect to the Euclidean metric with constant L = f L and Z N (0, Θ), Θ P(p), then P(f(Z) Ef(Z)) 1 ( ) 2 exp t2 2L 2, t 0. Θ 4.4. Discretization Recall that given a Euclidean p dimensional space and an operator over it with the matrix A in some orthonormal basis, the spectral norm of a this operator and of matrix A is defined as A = max x,y Sp 1(Ax, y). We approximate the spectral norm of matrices by using ε-nets in the following way: Lemma 6. [14] Let A be a p p matrix and N be a δ-net of the sphere S p 1 in the Euclidean space for some δ [1, 0), then A 1 (1 δ) 2 max (Ax, y). x,y N

8 I. Soloveychik/Error Bound for Compound Wishart Matrices 8 Following [19], we introduce the notion of coordinate-wise sparse regular vectors. Definition 4. For s [p], the subset of regular vectors of the unit sphere S p 1 is defined as Reg p (s) = {x = (x 1,..., x p ) T S p 1 x 2 i {0, 1/s}}, i = 1,..., p}, Reg p = Reg p (s). s [p] Lemma 7. [19] Let A be a p p matrix, then A 12 ln 2p 2 max (Ax, y). x,y Reg p Proof. The proof can be found in [19], it uses the regular vectors to construct a specific δ-net and obtain the bound given in the statement. 5. The Proofs of Theorems 1 and 2 We partition the proof into a few parts Decoupling and Conditioning Using Remark 1 we can rescale the random vectors and assume without loss of generality that Θ = I. Now by Lemma 3 it suffices to estimate E W. From Lemma 7 we get that P( W t) P(12 ln 2p 2 max (W x, y) t). (5.1) x,y Reg p Write the inner product coordinate-wise and rearrange the summands to obtain p p (W x, y) = b lm x j X jm y i X il. (5.2) l=1 i=1 m=1 j=1 We now fix x and y and condition on the variables X jm, j = 1,..., p, m = 1,..., n so that expression (5.2) defines a centered normal random variable. We want to estimate its standard deviation with the help of Lemma 4. Since we have assumed Θ = I, the covariance matrix of vec (X ) is also the identity matrix. Then Lemma 4 implies that (5.2) is centered normal with standard deviation at most σ x (X) y, where σ x (X) = l=1 i=1 p m=1 j=1 2 1/2 p b lm x j X jm = p We need to bound this quantity uniformly with respect to x. l=1 m=1 j=1 2 1/2 p b lm x j X jm.

9 5.2. Concentration I. Soloveychik/Error Bound for Compound Wishart Matrices 9 Let x Reg p, we estimate σ x (X) using concentration in Gauss space, Lemma 5. Due to Jensen s inequality Eσ x (X) (Eσ x (X) 2 ) 1/2 = p E = p p l=1 m=1 j=1 b 2 lmx 2 j 1/2 l=1 = p m=1 j=1 2 1/2 p b lm x j X jm p b 2 lm x 2 j l=1 m=1 j=1 1/2 = p B F. (5.3) Now we consider σ x : R pn R as a function of vec (X) and compute its Lipschitz constant with respect to the Euclidean measure on R pn. Note that the Euclidean norm on this space coincides with Frobenius norm on the linear space of p n matrices. σ x (X) = p l=1 m=1 j=1 2 1/2 p b lm x j X jm = p p l=1 n x j j=1 m=1 b lm X jm (5.4) = p BX T x 2 p B X p B X F, (5.5) for the Lipschitz constant Lemma 5 now implies that x Reg p and t 0 σ x (X) L p B. (5.6) P (σ x (X) p B F + t) 1 2 exp ( t2 2p B 2 ). (5.7) 2 1/ Union bounds We return to the estimation of the random variable (W x, y). Let us fix u 1, then x Reg p we consider the event E x = {σ x (X) p B F + u p B }. By (5.7) we have P(E x ) exp( u2 /2). (5.8) Note that σ x (X) and, thus E x, are independent of X. Let now x Reg p (r) and y Reg p (s). As we have observed above, conditioned on a realization of X

10 I. Soloveychik/Error Bound for Compound Wishart Matrices 10 satisfying E x, the random variable (W x, y) is distributed as a centered normal random variable h, whose standard deviation is bounded by p p B σ x (X) y B s F + u =: σ. s Then by the usual tail estimate for Gaussian random variables, we have Choose ε = uσ to obtain P((W x, y) ε E x ) 1 2 exp( ε2 /2σ 2 ). P((W x, y) ε E x ) 1 2 exp( u2 /2), x Reg p (r), y Reg p (s). We would like to take the union bound in this estimate over all y Reg p (s) for a fixed s. Note that ( ) p Reg p (s) = 2 s exp(s ln (2ep/s)), (5.9) s as there are exactly ( p s) possibilities to choose the support and 2 s ways to choose the signs of the coefficients of a vector in Reg p (s), thus ( ) P max (W x, y) ε E x 1 y Reg p (s) 2 exp(s ln (2ep/s) u2 /2), in order for this bound to be not trivial we assume u 2s ln (2ep/s). Now, using (5.8), we obtain ( ) ( ) P max (W x, y) ε P max (W x, y) ε E x + P(Ex) c y Reg p (s) y Reg p (s) 1 2 exp(s ln (2ep/s) u2 /2) exp( u2 /2) exp(s ln (2ep/s) u 2 /2). (5.10) 5.4. Gathering the bounds We continue with formula (5.1): P( W t) P(12 ln 2p 2 max (W x, y) t) P(12 ln 2p 2 max x,y Reg p max r,s [p] x Reg p (r) y Reg p (s) With the help of (5.10) and the bound (5.9) on the number of points in Reg p (r) we obtain P( max x Reg p (r) y Reg p (s) (W x, y) ε) exp(r ln (2ep/r) + s ln (2ep/s) u 2 /2), (W x, y) t).

11 I. Soloveychik/Error Bound for Compound Wishart Matrices 11 for u 2s ln (2ep/s). The function x ln (2ep/x) increases monotonically on the interval [1, p], so k ln (2ep/k) p ln (2ep/p) = p ln (2e) 2p, k p. Choose u 3 p to get the bound P( max max (W x, y) ε) exp( u 2 /4). r,s, [p] x Reg p (r) y Reg p (s) Finally, replace t with 12 ln 2p 2 ε to obtain where Integration of (5.11) yields By Lemma 3 we obtain P( W 12 ln 2p 2 ε) exp( u 2 /4), (5.11) ɛ = u p B F + u 2 p B, u 3 p. (5.12) E W 12 ln 2p 2 p(4σ + κ π). E W W 0 24 ln 2p 2 p(4σ + κ π). Now multiply W by Θ 1/2 from left and right to scale the matrices and get the statement of Theorem 1. For Corollary 2 assume we are given a sequence {B n } n S R n n such that Tr (B n ) = n, then for every corresponding Wishart matrix E W W 0 24 ln 2p 2 p(4σn + κ n π). n By bounding the values κ n κ and σ n σ from above we get the desired inequality E W W 0 24 ln 2p 2 p(4σ + κ π). n 6. Acknowledgment The author is grateful to Dmitry Trushin for discussions of the proof and useful suggestions. References [1] J. Wishart. The generalised product moment distribution in samples from a normal multivariate population. Biometrika, 20(1/2):32 52, [2] T. W. Anderson. An introduction to multivariate statistical analysis. 2, [3] R. J. Muirhead. Aspects of multivariate statistical theory. 197, 2009.

12 I. Soloveychik/Error Bound for Compound Wishart Matrices 12 [4] R. Speicher. Combinatorial theory of the free product with amalgamation and operator-valued free probability theory. 627, [5] Z. Burda, A. Jarosz, M. A. Nowak, J. Jurkiewicz, G. Papp, and I. Zahed. Applying free random variables to random matrix analysis of financial data. Part I: The Gaussian case. Quantitative Finance, 11(7): , [6] C.-N. Chuah, D. N. C. Tse, and R. A. Kahn, J. M.and Valenzuela. Capacity scaling in MIMO wireless systems under correlated fading. IEEE Transactions on Information Theory, 48(3): , [7] B. Collins, D. McDonald, and N. Saad. Compound Wishart matrices and noisy covariance matrices: Risk underestimation. arxiv preprint arxiv: , [8] P. Shah and V. Chandrasekaran. Group symmetry and covariance regularization. 46th Annual Conference on Information Sciences and Systems (CISS), pages 1 6, [9] W. Bryc. Compound real Wishart and q-wishart matrices. arxiv preprint arxiv: , [10] J. A. Tropp. User-friendly tail bounds for sums of random matrices. Foundations of Computational Mathematics, 12(4): , [11] L. Mackey, M. I. Jordan, R. Y. Chen, B. Farrell, and J. A. Tropp. Matrix concentration inequalities via the method of exchangeable pairs. arxiv preprint arxiv: , [12] R. Vershynin. How close is the sample covariance matrix to the actual covariance matrix? Journal of Theoretical Probability, 25(3): , [13] N. Srivastava and R. Vershynin. Covariance estimation for distributions with 2+ε moments. arxiv preprint arxiv: , [14] R. Vershynin. Introduction to the non-asymptotic analysis of random matrices. Compressed sensing: theory and applications, edited by Y.C. Eldar and G. Kutyniok, [15] P. J. Bickel and E. Levina. Covariance regularization by thresholding. The Annals of Statistics, 36(6): , [16] P. J. Bickel and E. Levina. Regularized estimation of large covariance matrices. The Annals of Statistics, pages , [17] A. J. Rothman, E. Levina, and J. Zhu. Generalized thresholding of large covariance matrices. Journal of the American Statistical Association, 104(485): , [18] T. T. Cai, C.-H. Zhang, and H. H. Zhou. Optimal rates of convergence for covariance matrix estimation. The Annals of Statistics, 38(4): , [19] E. Levina and R. Vershynin. Partial estimation of covariance matrices. Probability Theory and Related Fields, 153(3-4): , [20] R. Y. Chen, A. Gittens, and J. A. Tropp. The masked sample covariance estimator: an analysis via the matrix Laplace transform. DTIC Document, technical report, [21] M. Ledoux and M. Talagrand. Probability in banach spaces: isoperimetry and processes. 23, 1991.

Stein s Method for Matrix Concentration

Stein s Method for Matrix Concentration Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Invertibility of random matrices

Invertibility of random matrices University of Michigan February 2011, Princeton University Origins of Random Matrix Theory Statistics (Wishart matrices) PCA of a multivariate Gaussian distribution. [Gaël Varoquaux s blog gael-varoquaux.info]

More information

Boolean Inner-Product Spaces and Boolean Matrices

Boolean Inner-Product Spaces and Boolean Matrices Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver

More information

arxiv: v1 [math.pr] 22 Dec 2018

arxiv: v1 [math.pr] 22 Dec 2018 arxiv:1812.09618v1 [math.pr] 22 Dec 2018 Operator norm upper bound for sub-gaussian tailed random matrices Eric Benhamou Jamal Atif Rida Laraki December 27, 2018 Abstract This paper investigates an upper

More information

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES Volume 10 (2009), Issue 4, Article 91, 5 pp. ON THE HÖLDER CONTINUITY O MATRIX UNCTIONS OR NORMAL MATRICES THOMAS P. WIHLER MATHEMATICS INSTITUTE UNIVERSITY O BERN SIDLERSTRASSE 5, CH-3012 BERN SWITZERLAND.

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

Stable Process. 2. Multivariate Stable Distributions. July, 2006

Stable Process. 2. Multivariate Stable Distributions. July, 2006 Stable Process 2. Multivariate Stable Distributions July, 2006 1. Stable random vectors. 2. Characteristic functions. 3. Strictly stable and symmetric stable random vectors. 4. Sub-Gaussian random vectors.

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

A matrix over a field F is a rectangular array of elements from F. The symbol

A matrix over a field F is a rectangular array of elements from F. The symbol Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted

More information

Matrix concentration inequalities

Matrix concentration inequalities ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random

More information

On the exponential convergence of. the Kaczmarz algorithm

On the exponential convergence of. the Kaczmarz algorithm On the exponential convergence of the Kaczmarz algorithm Liang Dai and Thomas B. Schön Department of Information Technology, Uppsala University, arxiv:4.407v [cs.sy] 0 Mar 05 75 05 Uppsala, Sweden. E-mail:

More information

Concentration Inequalities for Random Matrices

Concentration Inequalities for Random Matrices Concentration Inequalities for Random Matrices M. Ledoux Institut de Mathématiques de Toulouse, France exponential tail inequalities classical theme in probability and statistics quantify the asymptotic

More information

Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas

Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas Department of Mathematics Department of Statistical Science Cornell University London, January 7, 2016 Joint work

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information

Least squares under convex constraint

Least squares under convex constraint Stanford University Questions Let Z be an n-dimensional standard Gaussian random vector. Let µ be a point in R n and let Y = Z + µ. We are interested in estimating µ from the data vector Y, under the assumption

More information

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Feng Wei 2 University of Michigan July 29, 2016 1 This presentation is based a project under the supervision of M. Rudelson.

More information

Faster Johnson-Lindenstrauss style reductions

Faster Johnson-Lindenstrauss style reductions Faster Johnson-Lindenstrauss style reductions Aditya Menon August 23, 2007 Outline 1 Introduction Dimensionality reduction The Johnson-Lindenstrauss Lemma Speeding up computation 2 The Fast Johnson-Lindenstrauss

More information

Intertibility and spectrum of the multiplication operator on the space of square-summable sequences

Intertibility and spectrum of the multiplication operator on the space of square-summable sequences Intertibility and spectrum of the multiplication operator on the space of square-summable sequences Objectives Establish an invertibility criterion and calculate the spectrum of the multiplication operator

More information

Linear Ordinary Differential Equations

Linear Ordinary Differential Equations MTH.B402; Sect. 1 20180703) 2 Linear Ordinary Differential Equations Preliminaries: Matrix Norms. Denote by M n R) the set of n n matrix with real components, which can be identified the vector space R

More information

Random Matrices: Invertibility, Structure, and Applications

Random Matrices: Invertibility, Structure, and Applications Random Matrices: Invertibility, Structure, and Applications Roman Vershynin University of Michigan Colloquium, October 11, 2011 Roman Vershynin (University of Michigan) Random Matrices Colloquium 1 / 37

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Using Friendly Tail Bounds for Sums of Random Matrices

Using Friendly Tail Bounds for Sums of Random Matrices Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,

More information

Exponential tail inequalities for eigenvalues of random matrices

Exponential tail inequalities for eigenvalues of random matrices Exponential tail inequalities for eigenvalues of random matrices M. Ledoux Institut de Mathématiques de Toulouse, France exponential tail inequalities classical theme in probability and statistics quantify

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

Lecture Notes on Metric Spaces

Lecture Notes on Metric Spaces Lecture Notes on Metric Spaces Math 117: Summer 2007 John Douglas Moore Our goal of these notes is to explain a few facts regarding metric spaces not included in the first few chapters of the text [1],

More information

Estimation of large dimensional sparse covariance matrices

Estimation of large dimensional sparse covariance matrices Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)

More information

RATE-OPTIMAL GRAPHON ESTIMATION. By Chao Gao, Yu Lu and Harrison H. Zhou Yale University

RATE-OPTIMAL GRAPHON ESTIMATION. By Chao Gao, Yu Lu and Harrison H. Zhou Yale University Submitted to the Annals of Statistics arxiv: arxiv:0000.0000 RATE-OPTIMAL GRAPHON ESTIMATION By Chao Gao, Yu Lu and Harrison H. Zhou Yale University Network analysis is becoming one of the most active

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance

The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance Marina Meilă University of Washington Department of Statistics Box 354322 Seattle, WA

More information

On Expected Gaussian Random Determinants

On Expected Gaussian Random Determinants On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose

More information

arxiv: v1 [math.pr] 11 Feb 2019

arxiv: v1 [math.pr] 11 Feb 2019 A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke

More information

On the singular values of random matrices

On the singular values of random matrices On the singular values of random matrices Shahar Mendelson Grigoris Paouris Abstract We present an approach that allows one to bound the largest and smallest singular values of an N n random matrix with

More information

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania Submitted to the Annals of Statistics DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING By T. Tony Cai and Linjun Zhang University of Pennsylvania We would like to congratulate the

More information

Tuning Parameter Selection in Regularized Estimations of Large Covariance Matrices

Tuning Parameter Selection in Regularized Estimations of Large Covariance Matrices Tuning Parameter Selection in Regularized Estimations of Large Covariance Matrices arxiv:1308.3416v1 [stat.me] 15 Aug 2013 Yixin Fang 1, Binhuan Wang 1, and Yang Feng 2 1 New York University and 2 Columbia

More information

Linear algebra. S. Richard

Linear algebra. S. Richard Linear algebra S. Richard Fall Semester 2014 and Spring Semester 2015 2 Contents Introduction 5 0.1 Motivation.................................. 5 1 Geometric setting 7 1.1 The Euclidean space R n..........................

More information

ELEMENTARY LINEAR ALGEBRA

ELEMENTARY LINEAR ALGEBRA ELEMENTARY LINEAR ALGEBRA K R MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND First Printing, 99 Chapter LINEAR EQUATIONS Introduction to linear equations A linear equation in n unknowns x,

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

High-dimensional covariance estimation based on Gaussian graphical models

High-dimensional covariance estimation based on Gaussian graphical models High-dimensional covariance estimation based on Gaussian graphical models Shuheng Zhou Department of Statistics, The University of Michigan, Ann Arbor IMA workshop on High Dimensional Phenomena Sept. 26,

More information

The deterministic Lasso

The deterministic Lasso The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality

More information

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases 2558 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 48, NO 9, SEPTEMBER 2002 A Generalized Uncertainty Principle Sparse Representation in Pairs of Bases Michael Elad Alfred M Bruckstein Abstract An elementary

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES. By Sara van de Geer and Johannes Lederer. ETH Zürich

THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES. By Sara van de Geer and Johannes Lederer. ETH Zürich Submitted to the Annals of Applied Statistics arxiv: math.pr/0000000 THE LASSO, CORRELATED DESIGN, AND IMPROVED ORACLE INEQUALITIES By Sara van de Geer and Johannes Lederer ETH Zürich We study high-dimensional

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

Preface to the Second Edition...vii Preface to the First Edition... ix

Preface to the Second Edition...vii Preface to the First Edition... ix Contents Preface to the Second Edition...vii Preface to the First Edition........... ix 1 Introduction.............................................. 1 1.1 Large Dimensional Data Analysis.........................

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

A priori bounds on the condition numbers in interior-point methods

A priori bounds on the condition numbers in interior-point methods A priori bounds on the condition numbers in interior-point methods Florian Jarre, Mathematisches Institut, Heinrich-Heine Universität Düsseldorf, Germany. Abstract Interior-point methods are known to be

More information

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY OLIVIER GUÉDON AND ROMAN VERSHYNIN Abstract. We present a simple and flexible method to prove consistency of semidefinite optimization

More information

ON SUM OF SQUARES DECOMPOSITION FOR A BIQUADRATIC MATRIX FUNCTION

ON SUM OF SQUARES DECOMPOSITION FOR A BIQUADRATIC MATRIX FUNCTION Annales Univ. Sci. Budapest., Sect. Comp. 33 (2010) 273-284 ON SUM OF SQUARES DECOMPOSITION FOR A BIQUADRATIC MATRIX FUNCTION L. László (Budapest, Hungary) Dedicated to Professor Ferenc Schipp on his 70th

More information

Optimum Power Allocation in Fading MIMO Multiple Access Channels with Partial CSI at the Transmitters

Optimum Power Allocation in Fading MIMO Multiple Access Channels with Partial CSI at the Transmitters Optimum Power Allocation in Fading MIMO Multiple Access Channels with Partial CSI at the Transmitters Alkan Soysal Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland,

More information

Convex Feasibility Problems

Convex Feasibility Problems Laureate Prof. Jonathan Borwein with Matthew Tam http://carma.newcastle.edu.au/drmethods/paseky.html Spring School on Variational Analysis VI Paseky nad Jizerou, April 19 25, 2015 Last Revised: May 6,

More information

Estimation of the Global Minimum Variance Portfolio in High Dimensions

Estimation of the Global Minimum Variance Portfolio in High Dimensions Estimation of the Global Minimum Variance Portfolio in High Dimensions Taras Bodnar, Nestor Parolya and Wolfgang Schmid 07.FEBRUARY 2014 1 / 25 Outline Introduction Random Matrix Theory: Preliminary Results

More information

arxiv: v5 [math.na] 16 Nov 2017

arxiv: v5 [math.na] 16 Nov 2017 RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem

More information

Random Methods for Linear Algebra

Random Methods for Linear Algebra Gittens gittens@acm.caltech.edu Applied and Computational Mathematics California Institue of Technology October 2, 2009 Outline The Johnson-Lindenstrauss Transform 1 The Johnson-Lindenstrauss Transform

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Information-Theoretic Limits of Matrix Completion

Information-Theoretic Limits of Matrix Completion Information-Theoretic Limits of Matrix Completion Erwin Riegler, David Stotz, and Helmut Bölcskei Dept. IT & EE, ETH Zurich, Switzerland Email: {eriegler, dstotz, boelcskei}@nari.ee.ethz.ch Abstract We

More information

Invertibility of symmetric random matrices

Invertibility of symmetric random matrices Invertibility of symmetric random matrices Roman Vershynin University of Michigan romanv@umich.edu February 1, 2011; last revised March 16, 2012 Abstract We study n n symmetric random matrices H, possibly

More information

William P. Thurston. The Geometry and Topology of Three-Manifolds

William P. Thurston. The Geometry and Topology of Three-Manifolds William P. Thurston The Geometry and Topology of Three-Manifolds Electronic version 1.1 - March 00 http://www.msri.org/publications/books/gt3m/ This is an electronic edition of the 1980 notes distributed

More information

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University

More information

Matrix Concentration. Nick Harvey University of British Columbia

Matrix Concentration. Nick Harvey University of British Columbia Matrix Concentration Nick Harvey University of British Columbia The Problem Given any random nxn, symmetric matrices Y 1,,Y k. Show that i Y i is probably close to E[ i Y i ]. Why? A matrix generalization

More information

Supremum of simple stochastic processes

Supremum of simple stochastic processes Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there

More information

BALANCING GAUSSIAN VECTORS. 1. Introduction

BALANCING GAUSSIAN VECTORS. 1. Introduction BALANCING GAUSSIAN VECTORS KEVIN P. COSTELLO Abstract. Let x 1,... x n be independent normally distributed vectors on R d. We determine the distribution function of the minimum norm of the 2 n vectors

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Stein s Method and Characteristic Functions

Stein s Method and Characteristic Functions Stein s Method and Characteristic Functions Alexander Tikhomirov Komi Science Center of Ural Division of RAS, Syktyvkar, Russia; Singapore, NUS, 18-29 May 2015 Workshop New Directions in Stein s method

More information

The Canonical Gaussian Measure on R

The Canonical Gaussian Measure on R The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where

More information

Random matrices: Distribution of the least singular value (via Property Testing)

Random matrices: Distribution of the least singular value (via Property Testing) Random matrices: Distribution of the least singular value (via Property Testing) Van H. Vu Department of Mathematics Rutgers vanvu@math.rutgers.edu (joint work with T. Tao, UCLA) 1 Let ξ be a real or complex-valued

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses

More information

IEOR 265 Lecture 3 Sparse Linear Regression

IEOR 265 Lecture 3 Sparse Linear Regression IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M

More information

Analysis-3 lecture schemes

Analysis-3 lecture schemes Analysis-3 lecture schemes (with Homeworks) 1 Csörgő István November, 2015 1 A jegyzet az ELTE Informatikai Kar 2015. évi Jegyzetpályázatának támogatásával készült Contents 1. Lesson 1 4 1.1. The Space

More information

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 11 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,, a n, b are given real

More information

Introduction to Real Analysis Alternative Chapter 1

Introduction to Real Analysis Alternative Chapter 1 Christopher Heil Introduction to Real Analysis Alternative Chapter 1 A Primer on Norms and Banach Spaces Last Updated: March 10, 2018 c 2018 by Christopher Heil Chapter 1 A Primer on Norms and Banach Spaces

More information

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1 Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of

More information

Efron Stein Inequalities for Random Matrices

Efron Stein Inequalities for Random Matrices Efron Stein Inequalities for Random Matrices Daniel Paulin, Lester Mackey and Joel A. Tropp D. Paulin Department of Statistics and Applied Probability National University of Singapore 6 Science Drive,

More information

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes Week 2 Based in part on slides from textbook, slides of Susan Holmes Part I Other datatypes, preprocessing October 3, 2012 1 / 1 2 / 1 Other datatypes Other datatypes Document data You might start with

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Higher order moments of the estimated tangency portfolio weights

Higher order moments of the estimated tangency portfolio weights WORKING PAPER 0/07 Higher order moments of the estimated tangency portfolio weights Farrukh Javed, Stepan Mazur, Edward Ngailo Statistics ISSN 403-0586 http://www.oru.se/institutioner/handelshogskolan-vid-orebro-universitet/forskning/publikationer/working-papers/

More information

Randomized sparsification in NP-hard norms

Randomized sparsification in NP-hard norms 58 Chapter 3 Randomized sparsification in NP-hard norms Massive matrices are ubiquitous in modern data processing. Classical dense matrix algorithms are poorly suited to such problems because their running

More information

Intrinsic Volumes of Convex Cones Theory and Applications

Intrinsic Volumes of Convex Cones Theory and Applications Intrinsic Volumes of Convex Cones Theory and Applications M\cr NA Manchester Numerical Analysis Martin Lotz School of Mathematics The University of Manchester with the collaboration of Dennis Amelunxen,

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Optimisation Combinatoire et Convexe.

Optimisation Combinatoire et Convexe. Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix

More information

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA) The circular law Lewis Memorial Lecture / DIMACS minicourse March 19, 2008 Terence Tao (UCLA) 1 Eigenvalue distributions Let M = (a ij ) 1 i n;1 j n be a square matrix. Then one has n (generalised) eigenvalues

More information

Bounds for all eigenvalues of sums of Hermitian random matrices

Bounds for all eigenvalues of sums of Hermitian random matrices 20 Chapter 2 Bounds for all eigenvalues of sums of Hermitian random matrices 2.1 Introduction The classical tools of nonasymptotic random matrix theory can sometimes give quite sharp estimates of the extreme

More information

Multivariate Normal-Laplace Distribution and Processes

Multivariate Normal-Laplace Distribution and Processes CHAPTER 4 Multivariate Normal-Laplace Distribution and Processes The normal-laplace distribution, which results from the convolution of independent normal and Laplace random variables is introduced by

More information

A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices

A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices A Multiple Testing Approach to the Regularisation of Large Sample Correlation Matrices Natalia Bailey 1 M. Hashem Pesaran 2 L. Vanessa Smith 3 1 Department of Econometrics & Business Statistics, Monash

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

TAMS39 Lecture 2 Multivariate normal distribution

TAMS39 Lecture 2 Multivariate normal distribution TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution

More information

A Brief Introduction to Tensors

A Brief Introduction to Tensors A Brief Introduction to Tensors Jay R Walton Fall 2013 1 Preliminaries In general, a tensor is a multilinear transformation defined over an underlying finite dimensional vector space In this brief introduction,

More information

EXPLICIT EVALUATIONS OF SOME WEIL SUMS. 1. Introduction In this article we will explicitly evaluate exponential sums of the form

EXPLICIT EVALUATIONS OF SOME WEIL SUMS. 1. Introduction In this article we will explicitly evaluate exponential sums of the form EXPLICIT EVALUATIONS OF SOME WEIL SUMS ROBERT S. COULTER 1. Introduction In this article we will explicitly evaluate exponential sums of the form χax p α +1 ) where χ is a non-trivial additive character

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Concentration of Measures by Bounded Couplings

Concentration of Measures by Bounded Couplings Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional

More information

DATA MINING AND MACHINE LEARNING

DATA MINING AND MACHINE LEARNING DATA MINING AND MACHINE LEARNING Lecture 5: Regularization and loss functions Lecturer: Simone Scardapane Academic Year 2016/2017 Table of contents Loss functions Loss functions for regression problems

More information

The norm of polynomials in large random matrices

The norm of polynomials in large random matrices The norm of polynomials in large random matrices Camille Mâle, École Normale Supérieure de Lyon, Ph.D. Student under the direction of Alice Guionnet. with a significant contribution by Dimitri Shlyakhtenko.

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information