On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators

Size: px
Start display at page:

Download "On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators"

Transcription

1 On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators Stanislav Minsker sminsker@math.duke.edu Abstract: We present some extensions of Bernstein s inequality for random self-adjoint operators. The main feature of these results is that they can be applied in the infinitedimensional setting. In particular, our inequalities refine the previous results of D. Hsu, S. M. Kakade and T. Zhang [4]. 1. Introduction Theoretical analysis of many problems, such as low-rank matrix recovery, covariance estimation and approximate matrix multiplication, is built upon exponential bounds for P X i > t where {X i } is a finite sequence of self-adjoint random matrices and is the operator norm. Starting with the pioneering work of R. Ahlswede and A. Winter [1], the moment-generating function technique was used to produce generalizations of Chernoff, Bernstein and Friedman inequalities to the non-commutative case; see [14],[13],[8] and references therein for thorough treatment and applications. Besides the definition of a correct variance parameter, the main difference in the tail bounds between the scalar version and the d d matrix version of the inequalities is the multiplicative factor d. While being sufficient for most problems, explicit dependence on the dimension of the matrix does not allow straightforward application of these results in the high-dimensional or infinite-dimensional setting. The main purpose of this note is to provide a dimension-free version of Bernstein s inequality for a sequence of independent random matrices as well as for the case of martingale differences. Some interesting results in this direction were previously obtained in [4], but with a slightly suboptimal tail. We show that in many cases the tail behavior can be improved through the modification of original Tropp s method [14] based on Lieb s concavity theorem. The trace quantity appearing in our bounds never exceeds the dimension of a matrix, therefore presented results can be seen as a generalization of previously existing bounds. We proceed by stating the main theorems followed by applications to vector-valued Bernstein inequalities and estimation of integral operators. i. Bernstein s inequality for independent random matrices.1. Definitions and notations Everywhere below, stands for the operator norm A := max λi A A, where λ i A A are the eigenvalues of a self-adjoint operator A A, and Y is the usual Euclidean norm or a vector Department of Mathematics, Duke University i 1

2 S. Minsker/Extensions of Bernstein s inequality Y C d. Let ξ be a real-valued random variable and ψ be a nondecreasing convex function such that ψ0 = 0. The Orlicz norm of ξ is defined as { } ξ ξ ψ = inf C > 0 : Eψ 1. C We will be mostly interested in the case ψ 1 x := e x 1 which corresponds to sub exponential random variables. Given a random matrix X, EX will stand for its expectation taken elementwise and E i [ ] denotes the conditional expectation E[ X 1,..., X i ]. For two self-adjoint matrices A and B, we will write A B iff A B is nonnegative definite, A B 0. We will often use the following simple fact: let A be a self-adjoint matrix with eigenvalue decomposition A = T ΛT, let f 1, f : R R be such that f 1 λ f λ for any eigenvalue λ = λa. Then f 1 A f A, where f i A = T f i ΛT and f i Λ is formed by applying f i to every element on the diagonal. Finally, we will define Ψ σ t := t /. This function often appears as an exponent in Bernsteintype concentration σ +t/3 results... Main results We are ready to state our first result, a version of Bernstein s inequality for the sequence of independent self-adjoint random matrices. Theorem.1. Let X 1,..., X n C d d be a sequence of independent self-adjoint random matrices such that EX i = 0 and σ EXi. Assume that X i U a.s., 1 i n. Then, for any t > 0 tr EXi P X i > t σ exp Ψ σ t r σ t.1 where r σ t = t log 1+t/σ. Remark 1. Note that 1 1. tr EX σ i d in fact, if EXi is approximately low rank, i.e. has many small 1 eigenvalues, tr can be much smaller than d. σ EXi. r σ t is decreasing, so in the range of t when the inequality becomes nontrivial, r σ t can be replaced by a constant. For example, when t σ we have r σ t σ log 1+1/σ 1.5 for σ Presented bounds easily extend to the case of rectangular matrices via the application of Paulsen dilation [8], see section.6 of [14] and Corollary 5.1 below for details. Next, we present a version of inequality that allows X i to be unbounded but requires exponential integrability instead. Such modifications of Bernstein s inequality are well known in scalar case; the noncommutative version appears in [5] see Theorem.7.

3 S. Minsker/Extensions of Bernstein s inequality 3 Theorem.. Let X 1,..., X n C d d be a sequence of independent self-adjoint random matrices such that EX i = 0 and σ EXi. Assume that U max Xi ψ1. Then, for any t > 0 4 P X i > t tr 4 tr EXi σ EXi σ where Kσ, d = log 16 d + log nu. σ 1 i n t exp r σ 1+1/d σ t, exp t 4UKσ,d r σ t, t σ 1+1/d U Kσ,d, t > σ 1+1/d U Kσ,d, Remark. The bound above is not dimension-free since Kσ, d increases with d, however it still gives certain improvement over existing versions. In particular, integrating inequality. tr EXi and setting Intdim := d, we get σ E X i C 1 σ log 4 Intdim U Kσ, d log 4 Intdim. This should be compared to an estimate implied by Theorem.7 of [5]: E X i C σ logd U log nu logd. Improvement can be significant whenever Intdim d. Proof of Theorem.1. Proof of.1 follows the lines of [14], where the key role is played by Lieb s concavity theorem [7]: Theorem.3 Lieb. Given a fixed self-adjoint matrix H, the function is concave on a positive definite cone. A tr exp H + log A In [14], section 4.8, advantages of this approach over the classic method of Ahlswede and Winter based on the Golden-Thompson inequality [1], [3] are discussed. Let φθ = e θ θ 1. Note that φ is nonnegative and increasing on 0,. Denote S n := n σ. X i and note that σ = ES n. First, we reduce the bounds on probability to the bounds on moment generating functions through a chain of simple inequalities. Let θ > 0; we have P λ max S n > t = P λ max θs n > θt = P λ max φθs n > φθt The following semidefinite relation is straightforward: Etr φθs n..3 φθt log Ee θx i φθex i..4

4 S. Minsker/Extensions of Bernstein s inequality 4 Indeed, writing the series expansion for e θx i and using that EX i = 0, we obtain [ Ee θx i 1 = I d + E θ Xi! θx i k 1 ] +... k + 1! 1 I d + θ EXi! θk X i k k + 1! +... = = I d + θ EXi e θ Xi θ X i 1 θ X i I d + EX i φθ, where in the last line we used the assumption that X i 1 and the fact that u eu u 1 is u increasing. It remains to apply the inequality I + A e A which holds for self-adjoint A := EXi. Next, since ES n = 0 d, Lieb s concavity theorem and Jensen s inequality for conditional expectation imply Etr φθs n = Etr expθs n 1 + log e θxn I d = = EE n 1 tr expθs n 1 + log e θxn I d E tr expθs n 1 + log Ee θxn I d. Iterating this argument, we get Etr φθs n tr exp log Ee θx i I d, which together with.4 gives Note that, since ES n is nonnegative definite, we can write Etr φθs n tr expφθes n I d..5 exp φθesn I d = = φθ ESn 1/ 1 + 1! φθes n φθes n 1 ES n! n / n φθesn 1 + 1! φθ ES n φθ ES n! n n = = ES nφθ expφθσ 1 σ φθ ES n σ expφθσ..6 Combining.6 with.3, we get P λ max S n > t tr ES n σ exp φθσ exp θt expθt φθt.

5 S. Minsker/Extensions of Bernstein s inequality 5 Note that for y > 0, e y φy = y e y y y y / + y 3 / y..7 Choose θ := log 1 + t σ to minimize expφθσ θt. Together with the well-known estimate 1 + y log1 + y y y / 1 + y/3, y 0.8 and.7, this concludes the proof of one-sided inequality. It remains to repeat the argument with X i s replaced by X i s to obtain a bound for the operator norm. Proof of Theorem.. We will now outline the main changes needed to obtain the second bound.. We use a truncation argument similar to the aforementioned result in [5]. To this end, we need to bound log Ee θx i in a different way. For any γ > 0, [ ] Ee θx i = I d + θ e θ Xi θ X i 1 E I d + θ EX i X i e θγ θγ 1 θ γ θ X i + θ I d E [ X i ] e θ Xi θ X i 1 θ X i I { X i > γ}. For θ 1 U, by monotonicity of u eu u 1 and Hölder s inequality, u ] E [ X i e θ Xi θ X i 1 θ X i I { X i > γ} U Xi Eφ I{ X i > γ} U U P 1/ X i γ E 1/ e X U. By definition of U, E 1/ e X U, while P X i γ exp γ U by the properties of Orlicz norms see section. of [15]. This gives Ee θx i I d + θ EX i e θγ θγ 1 θ γ + 8 θ U exp γ I d..9 U Now choose γ := U log 16 d nu and assume that θγ σ 1. For such θ, we have Ee θx i I d + EX θ i + σ nd I d, log Ee θx i EX θ i + σ nd I d..10 Repeating the steps of the previous proof with.10 instead of.4, we get θ σ Etr φθs n tr exp d I d + ESn I d.

6 = θ Q exp S. Minsker/Extensions of Bernstein s inequality 6 Following.6 and denoting Q := ESn + σ d I d, we further obtain θ exp Q I d = = θ Q1/ θ! Q θ n 1 n! Q +... Q 1/ θ Q θ! Q θ n 1 n! Q +... = θ Q 1 θ θ Q Q Q exp Q..11 Note that Q = σ 1 + d 1 and trq = σ + tr ESn tr ES n. Combine.11 with.3 and.7 to get P λ max S n > t trq θ expθt Q exp Q θt tr ES n θ φθt σ exp Q θt r σ t. Choose θ = Whenever t If t > t σ 1+1/d σ 1+ d 1 U log 16 d nu σ σ 1+ d 1 U log 16 d nu θ to minimize θ Q θt; recall that by our assumptions Uθ log 16 d nu 1..1 σ,.1 holds with θ = θ and gives subgaussian-type tail behavior., set θ := 1 σ U log 16 d nu σ to get subexponential part of the bound. 3. Concentration inequality for the sums of martingale differences Our next goal is to obtain a version of Friedman s inequality for the sums of matrix-valued martingale differences. Although we get a slightly weaker bound compared to the previous inequality, it still improves the multiplicative dimension factor. For t R, define pt := min t, 1. Note that 1. pt is concave;. gt := e t 1 + pt is non-negative for all t and increasing for t > 0. Recall the following useful result: Proposition 3.1 Peierls inequality, []. Let f : R R be a convex function and {u 1,..., u n } - any orthonormal basis of C n. For any self-adjoint A C n n f u i, Au i tr fa

7 S. Minsker/Extensions of Bernstein s inequality 7 An immediate corollary of this fact is that A tr fa is convex for a convex real-valued f and self-adjoint A: to show that A + B tr f 1 tr fa + tr fb, it is enough to apply Peierls inequality to the orthonormal system given by the eigenvectors of A + B. In particular, since pt is concave, it follows from Jensen s inequality that for any random selfadjoint matrix Y such that EY is well-defined We are ready to state and prove the main result of this section: E tr py tr pey. 3.1 Theorem 3.1. Let X 1,..., X n be a sequence of martingale differences with values in the set of d d self-adjoint matrices and such that X i 1 a.s. Denote W n := n E i 1 Xi. Then, for any t > 0 [ P X i > t, λ max W n σ tr p t ] σ EW n exp Ψ σ t v σ t, where v σ t = Ψ σ t. Remark 3. Note that 1. tr p t σ EW n d for all t > 0;. v σ t is decreasing, so whenever Ψ σ t 1, v σ t can be replaced by a constant. For example, for t σ, v σ t σ 44 whenever σ 1. Proof. Recall that φθ = e θ θ 1. Denote S n := n X i. Let θ be such that θt φθσ > 0 and define an event E by Note that triangle inequality implies We proceed by bounding P E: E := { λ max θs n φθw n θt φθσ }. E { λ max S n t, λ max W n σ }. P E = P λ max gθs n φθw n gθt φθσ tr E gθs n φθw n exp φθσ θt expθt φθσ gθt φθσ 3. The second term in the product, expφθσ θt, is minimized for θ := log1 + t σ and expφθ σ θ t exp Ψ σ t 3.3

8 S. Minsker/Extensions of Bernstein s inequality 8 by.8. To bound the first term in the product 3., note that by Lieb s theorem Y k := tr expθs k φθw k is a supermartingale with initial value d which can be shown similar by repeating conditioning argument of Theorem.1, or see [13] for details, so that Together with 3.1, this gives E tr exp θs n φθw n d. tr E gθs n φθw n = tr E expθs n φθw n I d + pθs n φθw n E tr p θs n φθw n tr p θes n φθw n = tr p φθew n. 3.4 Since EW n is nonnegative definite and due to the obvious estimate φθ e θ 1, θ 0 applied for θ = θ, bound 3.4 becomes tr EgθS n φθ W n tr p t σ EW n. 3.5 Finally, by.8 θ t φθ σ = 1 σ 1 + t/σ log1 + t/σ t/σ t / σ + t/3 > 0. Since φy = gy for y 0, the third term in the product 3. can be estimated by.7: expθ t φθ σ gθ t φθ σ Ψ σ t, 3.6 where, as before, Ψ σ t = t /. Combination of bounds 3.3,3.5,3.6 concludes the proof. σ +t/3 Remark 4. Expression tr p t EW σ n which replaces the dimension factor in our bound has a very simple meaning: acting on the cone of nonnegative definite operators, the function A p A just truncates the eigenvalues of A on the unit level. It is easy to see that if the eigenvalues of EW n decay polynomially, i.e., λ i EW n σ i, p > 1, then p tr p t σ EW n mind, ct 1/p. In particular, this gives an improvement over the multiplicative factor t appearing in the bounds of [4].

9 S. Minsker/Extensions of Bernstein s inequality 9 4. Extensions to the case of self-adjoint operators Clearly, both Theorem.1 and Theorem 3.1 apply to the case when {X i } i 1 is a sequence of selfadjoint Hilbert-Schmidt operators X i : H H acting on a separable Hilbert space H,, H, such that kerex i = H here, EX is defined as an operator such that EXz 1, z H = E Xz 1, z H for any z 1, z H. We will formally show how to extend results of Theorem.1; similar argument applies to Theorem 3.1. Let L 1 L... be a nested sequence of finite dimensional subspaces of H such that L j = j H, and let P Lj be an orthogonal projector on L j. For any fixed j, we will apply Theorem.1 to a sequence of finite dimensional operators { P Lj X i P Lj }i 1 mapping L j to itself. Note that Xi P Lj X i P Lj j 0 almost surely, hence P X i > t lim inf j P P Lj X i P Lj > t. 4.1 Note that, since A B implies SAS SBS, taking A = PL j = P Lj, B = I and S = P Lj X gives P Lj XP Lj P Lj X P Lj, thus tr E P Lj X i P Lj tr E P Lj Xi P L j lim inf j lim inf λ max EP Lj X i P Lj j tr EXi = lim sup λ max EP Lj X i P Lj j 4. λ max EP Lj X i P Lj tr EXi λ max EX i where in the last step we used a simple bound X i P Lj X i P Lj = X i X P Lj X i P Lj + X i P Lj X i P Lj P Lj X i P Lj X i X i P Lj X i P Lj 0 almost surely. 4.3 Since X i 1 a.s., 4.3 implies by dominated convergence that EP Lj X i P Lj E EXi E P Lj X i P Lj, EX i Theorem.1 applied to the right-hand side of 4.1, combined with 4. and 4.4, yields the result. 5. Applications Many relevant applications of dimension-free bounds, such as covariance estimation and approximate matrix multiplication, were demonstrated in [4]. Our bounds apply in those examples as well and yield slightly sharper tails. We demonstrate other immediate corollaries of our results.

10 5.1. Vector-valued Bernstein s inequality S. Minsker/Extensions of Bernstein s inequality 10 Corollary 5.1. Let Y 1,..., Y n C d be a sequence of independent random vectors such that EY i = 0, Y i 1 almost surely, and let Q i = EY i Yi be the associated covariance matrices. Denote σ := n E Y i = n trq i. Then for all t > 0 P Y i > t 4 exp Ψ σ t r σ t. Remark 5. Note that the inequality is dimension-free. It should be compared to result in [6] see formula 6.13 obtained by a combination of classical martingale methods and a trick of V. Yurinskii. Proof. The proof is based on a combination of Theorem.1 with the Paulsen dilation trick [9]: 0 Y define a linear mapping L : C d C d+1 d+1 by LY =. Set X Y 0 i := LY i and note that X i are self-adjoint random matrices with Xi = Yi 0 0 Y i Yi. Note that LY = Y, which gives X i = Y i 1 almost surely. We also have EXi = E Y i 0 = E Y i = trq i 0 Q i and, similarly, tr X i s. EXi = n trq i. Result now follows from Theorem.1 applied to It is straightforward to extend this result to separable Hilbert space - valued random variables by applying Corollary 5.1 to finite-dimensional projections and taking the limit, as was done in the previous section for operators. It is also easy to derive vector-valued Bernstein s inequality for the - norm rather than the norm by mapping a d-dimensional vector to a diagonal d d matrix; this version might sometimes be more useful than the scalar Bernstein s inequality coupled with the union bound: Corollary 5.. Let Y 1,..., Y n C d be a sequence of independent random vectors such that for all i, EY i = 0, Y i 1 almost surely, and let Q i = EY i Yi be the associated covariance matrices. Denote σ := max E 1 j d Y i, e j, where {e j, 1 j d} is the canonical basis. Then tr Q i P Y i > t σ exp Ψ σ t r σ t.

11 5.. Estimation of the integral operators S. Minsker/Extensions of Bernstein s inequality 11 Let S, Π be a measurable space, with Π being a probability measure. Let K, be a symmetric continuous positive definite kernel with κ := sup Kx, x < and let H K,, HK be the x S corresponding reproducing kernel Hilbert space. For x S, let K x := K, x. Define the integral operator L K : H k H K by L K fx := Kx, yfydπy = Kx, y K y, f HK dπy S where the second equality follows from the reproducing property. Note that L K is self-adjoint with tr L K = EKX, X. In many problems, Π is unknown and L K is approximated by its empirical version L K,n : L K,n fx := 1 f, K Xi n HK K Xi x, where X 1,..., X n is an iid sample from Π. The natural question to ask is: what is the degree of approximation provided by L K,n, measured in the operator norm? Note that the spectrum of L K,n can be easily evaluated in practice since the sets of non-zero eigenvalues of L K,n and a matrix K = 1 n KX i, X j n are equal, see Proposition 9 in [10]. Theorem.1 gives an answer to this i,j=1 question. To apply the theorem, define operator-valued random variables ξ i :=, K Xi HK K Xi L K and note that ξ i s are iid with mean zero. Setting u i = S K X i K Xi HK, we have, K Xi HK K Xi = K Xi H K, u i HK u i K Xi H K = KX i, X i κ, hence ξ i κ. At the same time, since U i :=, u i HK u i is a projector, it satisfies Ui Eξi, E K Xi HK K Xi = E K Xi H K, K Xi HK K Xi κ E, K Xi HK K Xi = κ L K. = U i and Note that L K EKX, X in particular, it can be much smaller than κ. Applying Theorem.1 and Remark 1 which is possible due to observations made in Section 4, we get κ L Corollary 5.3. Under our assumptions on the kernel, for t K n 1 n P L K,n L K > t 5 tr Eξ1 Eξ1 exp nt. κ L K + t/3 This can be used together with the fact that L K,n L K sup j 1 λ j L K λ j L K,n where the eigenvalues of λ j L K and λ j L K,n are ordered increasingly. In particular, in some cases the bound of Corollary 5.3 improves upon the estimate of Proposition 1 in [11].

12 S. Minsker/Extensions of Bernstein s inequality 1 References [1] R. Ahlswede and A. Winter, Strong converse for identification via quantum channels, IEEE Trans. Inform. Theory 48 00, no. 3, MR d:94069 [] E. Carlen, Trace inequalities and quantum entropy: an introductory course., Lecture notes 009, Available at: [3] S. Golden, Lower bounds for the Helmholtz function, Phys. Rev , B117 B118. MR #7113 [4] D. Hsu, S. M. Kakade, and T. Zhang, Tail inequalities for sums of random matrices that depend on the intrinsic dimension, Electron. Commun. Probab , no. 14, 13. MR [5] V. Koltchinskii, Oracle inequalities in empirical risk minimization and sparse recovery problems, Lecture Notes in Mathematics, vol. 033, Springer, 011, Lectures from the 38th Probability Summer School held in Saint-Flour, 008, École d Été de Probabilités de Saint-Flour. MR i:91165 [6] M. Ledoux and M. Talagrand, Probability in Banach spaces, vol. 3, Springer-Verlag, Berlin, 1991, Isoperimetry and processes. MR c:60001 [7] E. H. Lieb, Convex trace functions and the Wigner-Yanase-Dyson conjecture, Advances in Math , MR #10407 [8] R. I. Oliveira, Concentration of the adjacency matrix and of the laplacian in random graphs with independent edges, Preprint 010, [9] V. Paulsen, Completely bounded maps and operator algebras, Cambridge Studies in Advanced Mathematics, vol. 78, Cambridge University Press, Cambridge, 00. MR c:46118 [10] L. Rosasco, M. Belkin, and E. De Vito, On learning with integral operators, J. Mach. Learn. Res , MR h:687 [11] S. Smale and D.-X. Zhou, Geometry on probability spaces, Constr. Approx , no. 3, MR c:60006 [1] C. J. Thompson, Inequality with applications in statistical mechanics, J. Mathematical Phys , MR #7110 [13] J. A. Tropp, Freedman s inequality for matrix martingales, Electron. Commun. Probab , MR k:6004 [14], User-friendly tail bounds for sums of random matrices, Found. Comput. Math. 1 01, no. 4, MR [15] A. W. van der Vaart and J. A. Wellner, Weak convergence and empirical processes, Springer Series in Statistics, Springer-Verlag, New York, MR g:60035

Matrix concentration inequalities

Matrix concentration inequalities ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random

More information

Stein s Method for Matrix Concentration

Stein s Method for Matrix Concentration Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology

More information

Using Friendly Tail Bounds for Sums of Random Matrices

Using Friendly Tail Bounds for Sums of Random Matrices Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,

More information

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER 1. The method Ashwelde and Winter [1] proposed a new approach to deviation inequalities for sums of independent random matrices. The

More information

5 Compact linear operators

5 Compact linear operators 5 Compact linear operators One of the most important results of Linear Algebra is that for every selfadjoint linear map A on a finite-dimensional space, there exists a basis consisting of eigenvectors.

More information

arxiv: v1 [math.pr] 11 Feb 2019

arxiv: v1 [math.pr] 11 Feb 2019 A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke

More information

Geometry on Probability Spaces

Geometry on Probability Spaces Geometry on Probability Spaces Steve Smale Toyota Technological Institute at Chicago 427 East 60th Street, Chicago, IL 60637, USA E-mail: smale@math.berkeley.edu Ding-Xuan Zhou Department of Mathematics,

More information

Matrix Concentration. Nick Harvey University of British Columbia

Matrix Concentration. Nick Harvey University of British Columbia Matrix Concentration Nick Harvey University of British Columbia The Problem Given any random nxn, symmetric matrices Y 1,,Y k. Show that i Y i is probably close to E[ i Y i ]. Why? A matrix generalization

More information

On the Spectra of General Random Graphs

On the Spectra of General Random Graphs On the Spectra of General Random Graphs Fan Chung Mary Radcliffe University of California, San Diego La Jolla, CA 92093 Abstract We consider random graphs such that each edge is determined by an independent

More information

Bounds for all eigenvalues of sums of Hermitian random matrices

Bounds for all eigenvalues of sums of Hermitian random matrices 20 Chapter 2 Bounds for all eigenvalues of sums of Hermitian random matrices 2.1 Introduction The classical tools of nonasymptotic random matrix theory can sometimes give quite sharp estimates of the extreme

More information

User-Friendly Tail Bounds for Sums of Random Matrices

User-Friendly Tail Bounds for Sums of Random Matrices DOI 10.1007/s10208-011-9099-z User-Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Received: 16 January 2011 / Accepted: 13 June 2011 The Authors 2011. This article is published with open

More information

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms (February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops

More information

Technical Report No January 2011

Technical Report No January 2011 USER-FRIENDLY TAIL BOUNDS FOR MATRIX MARTINGALES JOEL A. TRO! Technical Report No. 2011-01 January 2011 A L I E D & C O M U T A T I O N A L M A T H E M A T I C S C A L I F O R N I A I N S T I T U T E O

More information

Linear Algebra and Dirac Notation, Pt. 2

Linear Algebra and Dirac Notation, Pt. 2 Linear Algebra and Dirac Notation, Pt. 2 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 2 February 1, 2017 1 / 14

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

A Matrix Expander Chernoff Bound

A Matrix Expander Chernoff Bound A Matrix Expander Chernoff Bound Ankit Garg, Yin Tat Lee, Zhao Song, Nikhil Srivastava UC Berkeley Vanilla Chernoff Bound Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

Shunlong Luo. Academy of Mathematics and Systems Science Chinese Academy of Sciences

Shunlong Luo. Academy of Mathematics and Systems Science Chinese Academy of Sciences Superadditivity of Fisher Information: Classical vs. Quantum Shunlong Luo Academy of Mathematics and Systems Science Chinese Academy of Sciences luosl@amt.ac.cn Information Geometry and its Applications

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

A spectral clustering algorithm based on Gram operators

A spectral clustering algorithm based on Gram operators A spectral clustering algorithm based on Gram operators Ilaria Giulini De partement de Mathe matiques et Applications ENS, Paris Joint work with Olivier Catoni 1 july 2015 Clustering task of grouping

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

Advanced Probability

Advanced Probability Advanced Probability Perla Sousi October 10, 2011 Contents 1 Conditional expectation 1 1.1 Discrete case.................................. 3 1.2 Existence and uniqueness............................ 3 1

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

arxiv: v7 [math.pr] 15 Jun 2011

arxiv: v7 [math.pr] 15 Jun 2011 USER-FRIENDLY TAIL BOUNDS FOR SUMS OF RANDOM MATRICES JOEL A. TRO arxiv:1004.4389v7 [math.r] 15 Jun 2011 Abstract. This paper presents new probability inequalities for sums of independent, random, selfadjoint

More information

1 Quantum states and von Neumann entropy

1 Quantum states and von Neumann entropy Lecture 9: Quantum entropy maximization CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: February 15, 2016 1 Quantum states and von Neumann entropy Recall that S sym n n

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

Empirical Processes and random projections

Empirical Processes and random projections Empirical Processes and random projections B. Klartag, S. Mendelson School of Mathematics, Institute for Advanced Study, Princeton, NJ 08540, USA. Institute of Advanced Studies, The Australian National

More information

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms.

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms. Vector Spaces Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms. For each two vectors a, b ν there exists a summation procedure: a +

More information

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T 1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.

More information

Effective Dimension and Generalization of Kernel Learning

Effective Dimension and Generalization of Kernel Learning Effective Dimension and Generalization of Kernel Learning Tong Zhang IBM T.J. Watson Research Center Yorktown Heights, Y 10598 tzhang@watson.ibm.com Abstract We investigate the generalization performance

More information

Efron Stein Inequalities for Random Matrices

Efron Stein Inequalities for Random Matrices Efron Stein Inequalities for Random Matrices Daniel Paulin, Lester Mackey and Joel A. Tropp D. Paulin Department of Statistics and Applied Probability National University of Singapore 6 Science Drive,

More information

Multivariate trace inequalities. David Sutter, Mario Berta, Marco Tomamichel

Multivariate trace inequalities. David Sutter, Mario Berta, Marco Tomamichel Multivariate trace inequalities David Sutter, Mario Berta, Marco Tomamichel What are trace inequalities and why we should care. Main difference between classical and quantum world are complementarity and

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

Entropy and Ergodic Theory Lecture 15: A first look at concentration

Entropy and Ergodic Theory Lecture 15: A first look at concentration Entropy and Ergodic Theory Lecture 15: A first look at concentration 1 Introduction to concentration Let X 1, X 2,... be i.i.d. R-valued RVs with common distribution µ, and suppose for simplicity that

More information

arxiv: v1 [math.pr] 22 May 2008

arxiv: v1 [math.pr] 22 May 2008 THE LEAST SINGULAR VALUE OF A RANDOM SQUARE MATRIX IS O(n 1/2 ) arxiv:0805.3407v1 [math.pr] 22 May 2008 MARK RUDELSON AND ROMAN VERSHYNIN Abstract. Let A be a matrix whose entries are real i.i.d. centered

More information

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011 LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Sparsification by Effective Resistance Sampling

Sparsification by Effective Resistance Sampling Spectral raph Theory Lecture 17 Sparsification by Effective Resistance Sampling Daniel A. Spielman November 2, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened

More information

Concentration of Measures by Bounded Couplings

Concentration of Measures by Bounded Couplings Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional

More information

Chapter 7. Basic Probability Theory

Chapter 7. Basic Probability Theory Chapter 7. Basic Probability Theory I-Liang Chern October 20, 2016 1 / 49 What s kind of matrices satisfying RIP Random matrices with iid Gaussian entries iid Bernoulli entries (+/ 1) iid subgaussian entries

More information

1 Definition and Basic Properties of Compa Operator

1 Definition and Basic Properties of Compa Operator 1 Definition and Basic Properties of Compa Operator 1.1 Let X be a infinite dimensional Banach space. Show that if A C(X ), A does not have bounded inverse. Proof. Denote the unit ball of X by B and the

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Concentration behavior of the penalized least squares estimator

Concentration behavior of the penalized least squares estimator Concentration behavior of the penalized least squares estimator Penalized least squares behavior arxiv:1511.08698v2 [math.st] 19 Oct 2016 Alan Muro and Sara van de Geer {muro,geer}@stat.math.ethz.ch Seminar

More information

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES Volume 10 (2009), Issue 4, Article 91, 5 pp. ON THE HÖLDER CONTINUITY O MATRIX UNCTIONS OR NORMAL MATRICES THOMAS P. WIHLER MATHEMATICS INSTITUTE UNIVERSITY O BERN SIDLERSTRASSE 5, CH-3012 BERN SWITZERLAND.

More information

The spectral zeta function

The spectral zeta function The spectral zeta function Bernd Ammann June 4, 215 Abstract In this talk we introduce spectral zeta functions. The spectral zeta function of the Laplace-Beltrami operator was already introduced by Minakshisundaram

More information

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Hilbert Spaces Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Vector Space. Vector space, ν, over the field of complex numbers,

More information

Clarkson Inequalities With Several Operators

Clarkson Inequalities With Several Operators isid/ms/2003/23 August 14, 2003 http://www.isid.ac.in/ statmath/eprints Clarkson Inequalities With Several Operators Rajendra Bhatia Fuad Kittaneh Indian Statistical Institute, Delhi Centre 7, SJSS Marg,

More information

If Y and Y 0 satisfy (1-2), then Y = Y 0 a.s.

If Y and Y 0 satisfy (1-2), then Y = Y 0 a.s. 20 6. CONDITIONAL EXPECTATION Having discussed at length the limit theory for sums of independent random variables we will now move on to deal with dependent random variables. An important tool in this

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

NON POSITIVELY CURVED METRIC IN THE SPACE OF POSITIVE DEFINITE INFINITE MATRICES

NON POSITIVELY CURVED METRIC IN THE SPACE OF POSITIVE DEFINITE INFINITE MATRICES REVISTA DE LA UNIÓN MATEMÁTICA ARGENTINA Volumen 48, Número 1, 2007, Páginas 7 15 NON POSITIVELY CURVED METRIC IN THE SPACE OF POSITIVE DEFINITE INFINITE MATRICES ESTEBAN ANDRUCHOW AND ALEJANDRO VARELA

More information

approximation algorithms I

approximation algorithms I SUM-OF-SQUARES method and approximation algorithms I David Steurer Cornell Cargese Workshop, 201 meta-task encoded as low-degree polynomial in R x example: f(x) = i,j n w ij x i x j 2 given: functions

More information

Different quantum f-divergences and the reversibility of quantum operations

Different quantum f-divergences and the reversibility of quantum operations Different quantum f-divergences and the reversibility of quantum operations Fumio Hiai Tohoku University 2016, September (at Budapest) Joint work with Milán Mosonyi 1 1 F.H. and M. Mosonyi, Different quantum

More information

EXPLICIT UPPER BOUNDS FOR THE SPECTRAL DISTANCE OF TWO TRACE CLASS OPERATORS

EXPLICIT UPPER BOUNDS FOR THE SPECTRAL DISTANCE OF TWO TRACE CLASS OPERATORS EXPLICIT UPPER BOUNDS FOR THE SPECTRAL DISTANCE OF TWO TRACE CLASS OPERATORS OSCAR F. BANDTLOW AND AYŞE GÜVEN Abstract. Given two trace class operators A and B on a separable Hilbert space we provide an

More information

Open Questions in Quantum Information Geometry

Open Questions in Quantum Information Geometry Open Questions in Quantum Information Geometry M. R. Grasselli Department of Mathematics and Statistics McMaster University Imperial College, June 15, 2005 1. Classical Parametric Information Geometry

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2

Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2 Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2 arxiv:quant-ph/0410063v1 8 Oct 2004 Nilanjana Datta Statistical Laboratory Centre for Mathematical Sciences University of Cambridge

More information

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities.

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities. Chapter 6 Quantum entropy There is a notion of entropy which quantifies the amount of uncertainty contained in an ensemble of Qbits. This is the von Neumann entropy that we introduce in this chapter. In

More information

Fredholm Theory. April 25, 2018

Fredholm Theory. April 25, 2018 Fredholm Theory April 25, 208 Roughly speaking, Fredholm theory consists of the study of operators of the form I + A where A is compact. From this point on, we will also refer to I + A as Fredholm operators.

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Calculus in Gauss Space

Calculus in Gauss Space Calculus in Gauss Space 1. The Gradient Operator The -dimensional Lebesgue space is the measurable space (E (E )) where E =[0 1) or E = R endowed with the Lebesgue measure, and the calculus of functions

More information

arxiv: v5 [math.na] 16 Nov 2017

arxiv: v5 [math.na] 16 Nov 2017 RANDOM PERTURBATION OF LOW RANK MATRICES: IMPROVING CLASSICAL BOUNDS arxiv:3.657v5 [math.na] 6 Nov 07 SEAN O ROURKE, VAN VU, AND KE WANG Abstract. Matrix perturbation inequalities, such as Weyl s theorem

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

The norm of polynomials in large random matrices

The norm of polynomials in large random matrices The norm of polynomials in large random matrices Camille Mâle, École Normale Supérieure de Lyon, Ph.D. Student under the direction of Alice Guionnet. with a significant contribution by Dimitri Shlyakhtenko.

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

On the concentration of eigenvalues of random symmetric matrices

On the concentration of eigenvalues of random symmetric matrices On the concentration of eigenvalues of random symmetric matrices Noga Alon Michael Krivelevich Van H. Vu April 23, 2012 Abstract It is shown that for every 1 s n, the probability that the s-th largest

More information

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS G. RAMESH Contents Introduction 1 1. Bounded Operators 1 1.3. Examples 3 2. Compact Operators 5 2.1. Properties 6 3. The Spectral Theorem 9 3.3. Self-adjoint

More information

On positive maps in quantum information.

On positive maps in quantum information. On positive maps in quantum information. Wladyslaw Adam Majewski Instytut Fizyki Teoretycznej i Astrofizyki, UG ul. Wita Stwosza 57, 80-952 Gdańsk, Poland e-mail: fizwam@univ.gda.pl IFTiA Gdańsk University

More information

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk Instructor: Victor F. Araman December 4, 2003 Theory and Applications of Stochastic Systems Lecture 0 B60.432.0 Exponential Martingale for Random Walk Let (S n : n 0) be a random walk with i.i.d. increments

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Stein s Method: Distributional Approximation and Concentration of Measure

Stein s Method: Distributional Approximation and Concentration of Measure Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Concentration of Measure Distributional

More information

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures S.G. Bobkov School of Mathematics, University of Minnesota, 127 Vincent Hall, 26 Church St. S.E., Minneapolis, MN 55455,

More information

Linear Algebra and Dirac Notation, Pt. 1

Linear Algebra and Dirac Notation, Pt. 1 Linear Algebra and Dirac Notation, Pt. 1 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 1 February 1, 2017 1 / 13

More information

Linear Algebra 2 Spectral Notes

Linear Algebra 2 Spectral Notes Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex

More information

A Note on Hilbertian Elliptically Contoured Distributions

A Note on Hilbertian Elliptically Contoured Distributions A Note on Hilbertian Elliptically Contoured Distributions Yehua Li Department of Statistics, University of Georgia, Athens, GA 30602, USA Abstract. In this paper, we discuss elliptically contoured distribution

More information

Real Variables # 10 : Hilbert Spaces II

Real Variables # 10 : Hilbert Spaces II randon ehring Real Variables # 0 : Hilbert Spaces II Exercise 20 For any sequence {f n } in H with f n = for all n, there exists f H and a subsequence {f nk } such that for all g H, one has lim (f n k,

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Approximation Theoretical Questions for SVMs

Approximation Theoretical Questions for SVMs Ingo Steinwart LA-UR 07-7056 October 20, 2007 Statistical Learning Theory: an Overview Support Vector Machines Informal Description of the Learning Goal X space of input samples Y space of labels, usually

More information

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song

STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song STAT 992 Paper Review: Sure Independence Screening in Generalized Linear Models with NP-Dimensionality J.Fan and R.Song Presenter: Jiwei Zhao Department of Statistics University of Wisconsin Madison April

More information

MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1

MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1 The Annals of Probability 014, Vol. 4, No. 3, 906 945 DOI: 10.114/13-AOP89 Institute of Mathematical Statistics, 014 MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1 BY LESTER MACKEY,MICHAEL

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution

More information

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Feng Wei 2 University of Michigan July 29, 2016 1 This presentation is based a project under the supervision of M. Rudelson.

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

SPECTRAL THEORY EVAN JENKINS

SPECTRAL THEORY EVAN JENKINS SPECTRAL THEORY EVAN JENKINS Abstract. These are notes from two lectures given in MATH 27200, Basic Functional Analysis, at the University of Chicago in March 2010. The proof of the spectral theorem for

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Worst-Case Bounds for Gaussian Process Models

Worst-Case Bounds for Gaussian Process Models Worst-Case Bounds for Gaussian Process Models Sham M. Kakade University of Pennsylvania Matthias W. Seeger UC Berkeley Abstract Dean P. Foster University of Pennsylvania We present a competitive analysis

More information

Weak and strong moments of l r -norms of log-concave vectors

Weak and strong moments of l r -norms of log-concave vectors Weak and strong moments of l r -norms of log-concave vectors Rafał Latała based on the joint work with Marta Strzelecka) University of Warsaw Minneapolis, April 14 2015 Log-concave measures/vectors A measure

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Linear algebra and applications to graphs Part 1

Linear algebra and applications to graphs Part 1 Linear algebra and applications to graphs Part 1 Written up by Mikhail Belkin and Moon Duchin Instructor: Laszlo Babai June 17, 2001 1 Basic Linear Algebra Exercise 1.1 Let V and W be linear subspaces

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Kernel-Based Contrast Functions for Sufficient Dimension Reduction

Kernel-Based Contrast Functions for Sufficient Dimension Reduction Kernel-Based Contrast Functions for Sufficient Dimension Reduction Michael I. Jordan Departments of Statistics and EECS University of California, Berkeley Joint work with Kenji Fukumizu and Francis Bach

More information

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure. Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.

More information

Linear Algebra using Dirac Notation: Pt. 2

Linear Algebra using Dirac Notation: Pt. 2 Linear Algebra using Dirac Notation: Pt. 2 PHYS 476Q - Southern Illinois University February 6, 2018 PHYS 476Q - Southern Illinois University Linear Algebra using Dirac Notation: Pt. 2 February 6, 2018

More information

Convergence of Eigenspaces in Kernel Principal Component Analysis

Convergence of Eigenspaces in Kernel Principal Component Analysis Convergence of Eigenspaces in Kernel Principal Component Analysis Shixin Wang Advanced machine learning April 19, 2016 Shixin Wang Convergence of Eigenspaces April 19, 2016 1 / 18 Outline 1 Motivation

More information

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator

More information