Linear Ridge Estimator of High-Dimensional Precision Matrix Using Random Matrix Theory

Size: px
Start display at page:

Download "Linear Ridge Estimator of High-Dimensional Precision Matrix Using Random Matrix Theory"

Transcription

1 CIRJE-F-995 Linear Ridge Estimator of High-Dimensional Precision Matrix Using Random Matrix Theory Tsubasa Ito Graduate School of Economics, The University of Tokyo Tatsuya Kubokawa The University of Tokyo November 205 CIRJE Discussion Paers can be downloaded without charge from: htt:// Discussion Paers are a series of manuscrits in their draft form. They are not intended for circulation or distribution excet as indicated by the author. For that reason Discussion Paers may not be reroduced or distributed without the written consent of the author.

2 Linear Ridge Estimator of High-Dimensional Precision Matrix Using Random Matrix Theory Tsubasa Ito and Tatsuya Kubokawa November 4, 205 Abstract In estimation of the large recision matrix, this aer suggests a new shrinkage estimator, called the linear ridge estimator. This estimator is motivated from a Bayesian asect for a sike and slab rior distribution of the recision matrix, and has a form of convex combination of the ridge estimator and the identity matrix multilied by scalar. The otimal arameters in the linear ridge estimator are derived in terms of minimizing a Frobenius loss function and estimated in closed forms based on the random matrix theory. Finally, the erformance of the linear ridge estimator is numerically investigated and comared with some existing estimators. Key words and hrases: Large-dimensional asymtotics, nonlinear shrinkage, recision matrix, random matrix theory, ridge tye estimator, rotation-equivariant estimators. Introduction The estimation of a large covariance matrix has been actively studied in the literature in recent years. Analysis of high-dimensional data is an imortant and modern toic in genetics in bioscience, ortfolio in financial economics and others. However, the samle covariance matrix is not invertible when the dimension of the variables is larger than the samle size N. When is large and close to N, the inverse of the samle covariance matrix may be ill-conditioned even if N >. To address this roblem, many aroaches have been considered in the literature. Under some model structures such as sarsity or ordering, enalized methods have been roosed and the oulation covariance matrix can be estimated consistently. However, the true model structures are generally unknown and the estimates become inconsistent in the case that the model structure is missecified. In the absence of such a rior information on the structure of the covariance matrix, the shrinkage method is a useful aroach. Since the eigenvalues of the samle covariance matrix diverge more widely than the eigenvalues of the oulation covariance matrix, it is reasonable to shrink the samle eigenvalues in the direction of their center. Linear shrinkage estimators of the covariance matrix have been suggested in Ledoit and Wolf (2004), Srivastava and Kubokawa (2007), Touloumis (205) and others. A nonlinear shrinkage aroach Graduate School of Economics, University of Tokyo, 7-3- Hongo, Bunkyo-ku, Tokyo , JAPAN. tsubasa ito.070@gmail.com Faculty of Economics, University of Tokyo, 7-3- Hongo, Bunkyo-ku, Tokyo , JAPAN. tatsuya@e.u-tokyo.ac.j

3 based on the random matrix theory is one of recent hot issues in statistics, and Ledoit and Wolf (202, 5) develoed some fundamental results and more efficient nonlinear shrinkage estimators. For the related works, see also Wang, Pan, Tong and Zhu (205), Bodnar, Guta and Parolya (205) and the references therein. Although the estimation of the covariance matrix has received more attention than that of the recision matrix, the estimation of the recision is more imortant in multivariate analysis, since for examle, the recision matrix aears in the Fisher linear discriminant analysis, confidence intervals based on the Mahalanobis distance and generalized least squares estimators in multivariate linear regression models. As an aroach, one can use the inverse matrix of estimators of the covariance matrix. However, the inverse of the covariance matrix estimator is not necessarily otimal in the estimation of the recision matrix. Further, two estimation roblems of the covariance and recision matrices are drastically different when > N. Thus, in this aer, we address the roblem of estimating the recision matrix directly. To exlain the roblem more secifically, let Σ be a covariance matrix, and let S be the samle covariance matrix such that E[S ] = Σ. Using the random matrix theory, Ledoit and Wolf (202, 5) derived asymtotic otimal shrinkage coefficients in the general class of rotationequivariant estimators, and roosed the nonlinear shrinkage estimator for the recision matrix. The nonlinear shrinkage estimator erforms well for large N, but the erformance dwindles when N or N/ is small, because the consistent estimators of the eigenvalues of Σ are not stable for small N or N/. Bodnar, et al. (205) suggested the linear shrinkage estimator { Ω linear αs + β I = if N > αs + + β I if N <. This estimator has an aealing form of shrinking S or S + towards βi. However, a drawback is that under their settings the arameters α and β cannot be estimated in the case of > N, without assuming that Σ = σ 2 I for scalar σ 2. Wang, et al. (205) considered the ridge-tye estimator Ω ridge = α(s + γi ), and rovided estimators of α and γ using the random matrix theory. Although this estimator is stable for any and N, it shrinks all the samle eigenvalues toward the same direction, which is not necessarily desirable in the estimation of the recision matrix. In this aer, we suggest a new tye of estimator given by Ω LR = α(s + γi ) + βi, and we call it the linear ridge estimator. This estimator is stable for any and N, and re-shrinks the shrunk samle eigenvalues toward the constant β. By combining the linear and the ridge estimators, the linear ridge estimator can comensate their weakness. Using the random matrix theory, we can derive the limits of the otimal α and β for fixed γ. Based on the limits, we can rovide the estimators of the otimal α and β in closed forms. It is also interesting to oint out that the linear ridge estimator can be motivated from a Bayesian asect for a sike and slab rior distribution of Σ. This motivation is useful when we need to restrict the estimators of α, β and γ. 2

4 The aer is organized as follows: The basic framework of the random matrix theory is described in Section 2. The loss function based on the Frobenius norm and the class of rotationequivariant estimators are given. In Section 3, we rovided the Ledoit-Wolf tye estimator and the suggested linear ridge estimator. The otimal estimators of the arameters in these estimators are derived based on the random matrix theory. The Bayesian motivation of the linear ridge estimator is also given. In Section 4, we investigate the finite-samle erformance of the roosed estimator through simulation and emirical studies. Concluding remarks are given in Section 5 and the technical roofs are given in the Aendix. 2 Basic Framework 2. Preliminary results in the random matrix theory We begin by stating the basic assumtions which are common in estimation of the highdimensional covariance matrix based on the random matrix theory. Throughout the aer, R and C denote the saces of real and comlex numbers, resectively. Also, C + denotes the half-lane of comlex numbers with strictly ositive imaginary art. The real and imaginary arts of z C are denoted by R(z) and I(z), resectively. (A) Let N and (N) denote the samle size and the number of variables resectively. In the large-dimensional asymtotics, goes to infinity, as N does, but it is assumed that the ratio /N converges to a limit y (0, ) (, + ), as N +. The case y = is excluded for technical reason. (A2) The oulation covariance matrix Σ is a non-random -dimensional ositive definite matrix, where the subscrit denotes the dimension of the matrix. Let X = (x,,..., x,n ) T be an N random matrix, where x,,..., x,n are mutually indeendently and identically distributed as E[x,j ] = 0 and Cov (x,j ) = I. It is assumed that there exist fourth moments of all elements of x,j. Let Y = (y,,..., y,n ) T, where y,j = Σ /2 x,j for the squared symmetric root matrix Σ /2 such that Σ = (Σ /2 ) 2. It is assumed that Y is observable. (A3) Let t = (t,,, t, ) T be a system of eigenvalues of Σ, sorted in decreasing order. The emirical sectral distribution (ESD) of the oulation eigenvalues is defined by H (t) [t,i,+ ), t R, i= where A denotes the indicator function of set A. It is assumed that H (t) converges to some limit H(t) at all oints of continuity of H. (A4) Su(H), the suort of H, is the union of a finite number of closed intervals, bounded away from zero and infinity. Furthermore, there exists a comact interval in (0, + ) that contains Su(H ) for all N large enough. Let S = N Y T Y. As noted in Section 5, all the results given in this aer still hold in the case of E[y,j ] = µ. A system of eigenvalues of S, sorted in decreasing order, and the 3

5 corresonding eigenvectors are denoted by l = (l,,, l, ) T and (u,..., u ), resectively. The emirical sectral distribution (ESD) of S is defined by F (t) [l,i,+ ), t R. i= Marčenko and Pastur (967) greatly contributed the research on large-dimensional asymtotic theories of eigenvalues of S. This asymtotic theory is called the random matrix theory, and has been generalized by Silverstein (995), Silverstein and Bai (995), Silverstein and Choi (995) among others. These literatures imly that under assumtions (A)-(A4), there exists a distribution function F such that F (x) F (x), x R\{0}, which is called the limiting sectral distribution (LSD). Silverstein and Choi (995) showed that F is everywhere continuous excet at zero, and that the mass of F at zero is given by by F (0) = max{ y, H(0)}. (2.) For a nondecreasing function G on the real line, the stieltjes transform m G of G is defined m G (z) x z dg(x), z C+. The stieltjes transform has the well-known inversion formula G{[a, b]} = π lim η 0 + b a I ( m G (ξ + iη) ) dξ, (2.2) if G is continuous at a and b. Silverstein (995) develoed the most general version of the equation which relates F to H and y, namely, m m F (z) is the unique solution in the set {m C : ( y)/z + ym C + } to the equation m F (z) = t( y yzm F (z)) z dh(t), z C+. (2.3) Silverstein and Choi (995) showed that m F is extended to R, that is, lim z C + λm F (z) m F (λ) for λ R. It is noted from the inversion formula that F (λ) = π I[ m F (λ)] on R. We next consider the limiting sectral distribution (LSD) of the eigenvalues of N Y Y T = N X Σ X T. Denote the LSD by F. The eigenvalues of N Y T Y and N Y Y T only differs by N zero eigenvalues. It then holds: F (x) =( y) [0,+ ) (x) + yf (x), x R, F (x) = y y [0,+ ) (x) + F, x R, y m F (z) = y z m F (z) = y yz + ym F (z), z C +, + y m F (z), z C +. 4

6 2.2 Setu of the roblem Consider the roblem of estimating Σ by estimator Ω, where the estimator is evaluated by the risk function relative to the loss function L (Σ, Ω ) tr (Ω Σ I )(Ω Σ I ) T. (2.4) This loss function is based on the Frobenius norm A F = tr (AA T ), namely, L (Σ, Ω ) = Ω Σ I 2 F. We thus call it the Frobenius loss function. The loss function satisfies that L (Σ, Ω ) 0 and L (Σ, Σ ) = 0. For remarks of other loss functions, see Section 5. Ledoit and Wolf (202, 5) considered a class of rotation-equivariant estimators in estimation of Σ and Σ. An estimator Σ = Σ (Y ) of Σ is called rotation-equivariant when for any -dimensional orthogonal matrix W, Σ (Y W ) = W T Σ (Y )W. For the samle covariance matrix, it is well known that eigenvalues of S diverge more widely than those of Σ. Rotationequivariant estimators have the same eigenvectors as S and relace eigenvalues of S with other values shrunk towards a center of the eigenvalues. In the estimation of the inverse Σ, rotationequivariant estimators Ω = Ω (Y ) satisfy Ω (Y W ) = W T Ω (Y )W for any orthogonal matrix W. Thus, every rotation-equivariant estimator of Σ is of the form Ω (A ) = U A U T, A = diag(a,, a ), (2.5) where U = (u,..., u ), and a i may deend on the eigenvalues l of S. The class of rotation-equivariant estimators include various estimators. The ridge estimator given by Wang, et al. (204) corresonds to the case that a i = α/(l i + γ), and the linear estimator treated by Bodnar, et al. (204) corresonds to a i = α/l i + β for N > and a i = β for N <. Ledoit and Wolf (205) treated the general rotation-equivariant estimator directly. The roblem is how to estimate the arameters a i s, α, γ or β. A reasonable way is the estimation of the otimal a i s which minimize the risk function R (Σ, Ω ) E[L (Σ, Ω )]. Instead of minimizing risk functions, Ledoit and Wolf (202) obtained the otimal a i s which minimize the loss function and estimated the asymtotic otimal a i s using the random matrix theory. This argument is used in the next section to estimate the arameters in the estimation of the recision matrix Σ under the Frobenius loss (2.4). 3 Non-linear Shrinkage Estimation of the Precision Matrix In this aer, we treat two non-linear shrinkage estimators: the Ledoit-Wolf tye estimator and the linear ridge estimator. The estimators of the otimal arameters are rovided based on the random matrix theory. 3. Ledoit-Wolf tye estimator The Ledoit-Wolf tye estimator can be derived based on the otimal a i s which minimize the Frobenius loss (2.4) of the rotation-equivariant estimator Ω (A ) in (2.5). In fact, the loss function is written as L (Σ, Ω (A )) = i= {a2 i ut i Σ2 u i 2a i u T i Σ u i } +, which is minimized at a oracle i = u T i Σ u i /u T i Σ 2 u i, (3.) 5

7 for i =,...,. Since a oracle i s deend on Σ, we need to estimate them. Using the random matrix theory, Ledoit and Péché (20) showed that under (A)-(A4), u T i Σ u i can be aroximated as δ(l i ), where x y yx m F (x) 2 if x > 0, δ(x) = (3.2) if x = 0, y >, (y ) m F (0) 0 otherwise. Concerning the term u T i Σ2 u i, we can use a similar argument as in Ledoit and Péché (20). In fact, the following theorem guarantees that it can be aroximated by ϕ(l i ), where x 2 { y 2 2y 2 xr[ m F (x)] y 2 x 2 m F (x) 2 } ( y yx m F (x)) tdh(t)yx y yx m F (x) 2 if x > 0, ϕ(x) = y y m F (0) ( tdh(t) ) y m F (0) if x = 0, y >, 0 otherwise. (3.3) Theorem Under (A)-(A4), the rotation-equivariant estimator Ω (A ) with a i = a i (l i ) has the almost sure limit given by {a L (Σ, Ω (A )) 2 (x)ϕ(x) 2a(x)δ(x) } df (x) +. (3.4) The roof is given in the Aendix. Hence, one gets the Ledoit-Wolf tye estimator Ω LW =U A U T, A =diag(a,, a ), a i = δ(l i )/ϕ(l i ). (3.5) Theorem imlies that the estimator Ω LW attains the minimum of the limiting loss function, given by [{δ(x)} 2 /ϕ(x)]df (x) +. Since y, F and H are unknown, however, Ω LW is not feasible. The way for obtaining a bona fide estimator in the oracle estimator Ω LW is rovided by Ledoit and Wolf (205). Here, the estimators of m F (x) and m F (0) are denoted by mf (x) and m F (0), (3.6) resectively, where the detail for the derivation is given in the Aendix. Also tdh(t) in equation (3.3) is unknown, but this is the limit of tr [Σ ] and can be estimated by tr [S ]. Relacing m F (x), m F (0) and tdh(t) in equations (3.2) and (3.3) with their estimators, and relacing the limiting concentration ratio y with /N, we have the consistent estimators for δ(x) and ϕ(x). Then, we get the bona fide estimator δ(x) and ϕ(x), (3.7) Ω LW =U Â U T, Â =diag(â,, â ), â i = δ(l i )/ ϕ(l i ). (3.8) 6

8 3.2 Linear ridge estimator In the estimation of the inverse Σ, we here consider a new tye of the estimator Ω (α, β, γ) = α(s + γi ) + βi, (3.9) where α, β and γ are unknown arameters. This estimator not only belongs to the class of rotation-equivariant estimators (2.5) with a i = α/(l i + γ) + β, but also corresonds to the ridge estimator for β = 0 and the linear estimator for γ = 0. We call Ω (α, β, γ) the linear ridge estimator. This estimator is well-conditioned and available when > N. The use of the form β 0 I for constant β 0 has been suggested as an estimator of Σ in the literature in the ultra-high dimensional case. The ridge-tye estimator α 0 (S + γi ) for constant α 0 is reasonable unless is ultra-high dimensional. Thus, the linear ridge estimator (3.9) is interreted as a convex combination of α 0 (S +γi ) and β 0 I, namely, for 0 < w <, ( w)α 0 (S +γi ) +wβ 0 I = α(s + γi ) + βi for α = ( w)α 0 and β = wβ 0. A related exlanation is given in the next subsection from a Bayesian asect. The Frobenius loss function of Ω (α, β, γ) is written as L (Σ, Ω (α, β, γ)) = tr [ {α(s + γi ) Σ + βσ I }{α(s + γi ) Σ + βσ I } T]. Given γ, the otimal α and β are given by α (γ) = tr [(S + γi ) Σ ]tr [Σ 2 ] tr [(S + γi ) Σ 2 ]tr [Σ ] tr [(S + γi ) 2 Σ 2 ]tr [Σ 2 ] {tr [(S + γi ) Σ 2 ]} 2, β (γ) = tr [(S + γi ) 2 Σ 2 ]tr [Σ ] tr [(S + γi ) Σ ]tr [(S + γi ) Σ 2 ] tr [(S + γi ) 2 Σ 2 ]tr [Σ 2 ] {tr [(S + γi ) Σ 2 ]} 2. Substituting these otimal quantities into the loss function, we can get L (γ) =L (Σ, Ω LR (α (γ), β (γ), γ)) = [tr [(S + γi ) 2 Σ 2 ]tr [Σ 2 ] {tr [(S + γi ) Σ 2 ]} 2] [ {tr [(S + γi ) Σ ]} 2 tr [Σ 2 ] tr [(S + γi ) 2 Σ 2 ](tr [Σ ]) 2 ] + 2 tr [(S + γi ) Σ ]tr [(S + γi ) Σ 2 ]tr [Σ ] +. The asymtotic behaviors of α (γ), β (γ) and L (γ) can easily be checked by alying the results of the literature. The results are given in the following theorem, which will be roved in the Aendix. Theorem 2 Let t = tdh(t), t 2 = t 2 dh(t), B(γ) = γ m F ( γ) y( γ m F ( γ)) and B (γ) = d dγ B(γ). 7

9 Under (A)-(A4), α (γ), β (γ) and L (γ) converge to α(γ), β(γ) and L(γ), resectively, where where α(γ) = { } t yγb 2 (γ) + (t 2 + t γ t 2 c(γ) y)b(γ) t 2, β(γ) = { } yγb 3 (γ) + γb 2 (γ) + 2t yγb(γ)b (γ) + t (γ t )B (γ), c(γ) L(γ) = + { 2t yγb 3 (γ) + ( t 2 + t 2 c(γ) y 2t γ)b 2 (γ) + t 2 B(γ) } 2t 2 yγb(γ)b (γ) + t 2 (t γ)b (γ), c(γ) = y 2 γ 2 B 4 (γ) + {t 2 yγ + 2yγ(t y γ)}b 3 (γ) + (t 2 γ t 2 y 2 + 4t yγ γ 2 )B 2 (γ) 2t (yt γ)b(γ) + 2t t 2 γb(γ)b (γ) + t t 2 (γ t )B (γ) t 2. (3.0) To estimate α, β and γ, we need to estimate α(γ), β(γ) and L(γ). As mentioned in Wang, et al. (205), for γ > 0, m F ( γ) can easily be estimated by tr [(S + γi ) ]/, since m F ( γ) is the limit of m F ( γ) = tr [(S + γi ) ]/. Thus, consistent estimators of B(γ) and B (γ) given γ are γtr [(S + γi ) B(γ) ]/ = (/N) + γtr [(S + γi ) ]/N, B (γ) = tr [(S + γi ) ]{(/N) B(γ) } (/N) + γtr [(S + γi ) ]/N. (3.) Since t = tdh(t) is the limit of a = tr [Σ ]/, it is estimated by â = tr [S ]/. Also, note that t 2 = t 2 dh(t) is the limit of a 2 = tr [Σ 2 ]/. Then we can use the consistent estimator â 2 = N { (N )(N 2)tr [ S 2 } N(N 2)(N 3) ] + {tr [ S ]} 2 NQ, where S = (N ) N i= (y,i y )(y,i y ) T and Q = (N ) N i= {(y,i y ) T (y,i y )} 2 for y = N N i= y,i. For the details of the estimator â 2, see Himeno and Yamada (204). Thus, consistent estimators of t and t 2 are given by â and â 2. Relacing B(γ), B (γ), t, t 2 and y with B(γ), B (γ), â, â 2 and /N in (3.0), we have α (γ), β (γ) and L (γ). Let γ be the solution of minimizing L (γ), namely γ = argmin γ>0 L (γ). Then, α and β are estimated by α = α ( γ) and β = β ( γ). Although α, β and γ are the consistent estimators of α, β and γ as (N, ) with /N y, it is not guaranteed that α, β and γ are inside reasonable ranges for finite and N. For ractical use, we need to adjust the values of α, β and γ to be inside aroriate ranges. An aroach to this end is the Bayesian argument given in the next subsection. We shall suggest the linear ridge estimator Ω LR in (3.6) through (3.3), (3.4) and (3.5). 8

10 3.3 Bayesian motivation and the suggested linear ridge estimator We give a Bayesian motivation of the linear ridge estimator (3.9). This tells us about how the arameters α, β and γ should be restricted. Let V = NS. Consider the case of N > and assume that V has a Wishart distribution W (N, Σ ). The density function of V is denoted by f(v Σ ). Let η be a random variable having a Bernoulli distribution Ber(θ) with P (η = ) = θ. As a rior distribution of Σ, consider a sike and slab rior distribution: Given η = 0, Σ Σ has a one-oint distribution P (Σ and η = are denoted by π(σ to satisfy that δ Λ0 (Σ )dσ marginal density functions of (V, Σ f(v, Σ W (k, Λ ), and given η =, = Λ 0 ) =. The density functions of Σ given η = 0 Λ ) and δ Λ0 (Σ ), resectively, where δ Λ0 (Σ ) is assumed = and Σ δ Λ0 (Σ )dσ = Λ 0. Then, the joint and ) and V are written as, resectively, ) =f(v Σ ){( θ)π(σ f(v ) =( θ) Thus, the Bayes estimator, given by Ω Bayes where Ω Bayes = It can be easily seen that Σ f(v, Σ = ( θ) Σ Λ ) + θδ Λ0 (Σ )}, f(v Σ )π(σ Λ )dσ + θf(v Λ 0 ). = E[Σ )dσ /f(v ) ( θ) f(v Σ Σ f(v Σ =( w 0 ) f(v Σ V ], is exressed as f(v Σ )π(σ Λ )dσ )π(σ )π(σ )π(σ + θλ 0 f(v Λ 0 ) Λ )dσ + θf(v Λ 0 ) Λ )dσ Λ )dσ + w 0 Λ 0, θf(v Λ 0 ) w 0 = ( θ) f(v Σ )π(σ Λ )dσ + θf(v Λ 0 ). Σ f(v Σ )π(σ f(v Σ )π(σ Λ )dσ Λ )dσ = (N + k)(v + Λ ) = v 0 (S + N Λ ), for v 0 = (N + k)/n. Thus, we get Ω Bayes = ( w 0 )v 0 (S + N Λ ) + w 0 Λ 0, where v 0 and w 0 satisfy that v 0 > and 0 < w 0 <. Motivated from the Bayes estimator Ω Bayes, we consider the following linear ridge estimator: Let Λ = NγI and Λ 0 = (/l)i for l = i= l i/ = tr [S ]/. Then the linear ridge estimator we treat is linear ridge Ω = v( w)(s + γi ) + w(/l)i, (3.2) where v, w and γ are constants which satisfy the restrictions v >, 0 < w <, and γ > 0. (3.3) 9

11 This estimator corresonds to (3.9) with α = v( w) and β = w/l. On the other hand, from the results below (3.), we have the consistent estimators α, β and γ for α, β and γ. Thus, one gets the equations α = ṽ( w) and β = w/l, which yields that w = l β and ṽ = α/( w) = α/( l β). Since ṽ, w and γ need to satisfy the restriction (3.3), the arameters v, w and γ are estimated by ŵ = 0 w = 0 (l β), v = ṽ = α/( w), γ = γ 0, (3.4) where a b = max(a, b) and a b = min(a, b). This gives α = v( ŵ) and β = ŵ/l, (3.5) and we suggest the use of the linear ridge estimator Ω LR = v( ŵ)(s + γi ) + ŵ(/l)i = α(s + γi ) + βi. (3.6) Note that when ŵ = 0 or β = 0, the linear ridge estimator becomes the ridge-tye estimator α(s + γi ) = v(s + γi ). On the other hand, when ŵ = or α = 0, the linear ridge estimator becomes βi = (/l)i, which is roosed frequently as an estimator of an ultra-high dimensional recision matrix. Generally, in the case that 0 < ŵ <, the linear ridge estimator is a convex combination of the ridge estimator and (/l)i. 3.4 Ridge and linear shrinkage estimators for comarison In the revious subsections, we have derived the Ledoit-Wolf tye estimator (3.8) and the linear ridge estimator (3.6) relative to the Frobenius loss function (2.4). To comare them with the existing other estimators, we here look at two estimators: the ridge and the linear shrinkage estimators. Both estimators are secial cases of the linear ridge estimator (3.9), and we can investigate how the linear ridge estimator is effective in comarison with them. Thus, we derive the otimal values of the arameters in the ridge and the linear shrinkage estimators with resect to the Frobenius loss. [] Ridge estimator. The ridge estimator is of the form Ω ridge = α(s + γi ), (3.7) which has been used and studies in the literature. For examle, see Srivastava and Kubokawa (2007). Although most results were obtained under Gaussian distributions, Wang, et al. (204) derived the estimators of the otimal α and γ using the random matrix theory without any assumtion of underlying distributions, where they treated another quadratic loss function. Given γ, the otimal α relative to the Frobenius loss is α R (γ) = which leads to the reduced loss function L (Σ tr [(S + γi ) Σ ] tr [(S + γi ) Σ 2 (S + γi ) ],, Ω ridge (α R (γ), γ)) = {tr [(S + γi ) Σ ]} 2 tr [(S + γi ) Σ 2 (S + γi ) ]. 0

12 Then, it follows from the above results that α R (γ) and L (Σ, Ω ridge (α R (γ), γ)) converge to α R B(γ) (γ) = yb 2 (γ) + B(γ) + 2yγB(γ)B (γ) + (γ t )B (γ), L R (γ) = B 2 (γ) yb 2 (γ) + B(γ) + 2yγB(γ)B (γ) + (γ t )B (γ), for B(γ) and B (γ) in Theorem 2. Relacing B(γ), B (γ), t and y with B(γ), B (γ), â and /N, we have α R and L R (γ), where B(γ) and B (γ) are given in (3.). Let γ R be the solution of minimizing L R (γ). Then, we get α R = α R ( γ R ), which yields the ridge estimator. [2] Linear shrinkage estimator. The linear shrinkage estimator suggested by Bodnar, et al. (204) is of the form { Ω linear αs + β I = if N > αs + + β I if N <. (3.8) In the case of N >, we can use the results given in the revious subsections and Theorem 3.2 in Bodnar, et al. (204) to show that the Frobenius loss function of the linear shrinkage estimator has the limit lim L (Σ, Ω linear ) N, { = lim α 2 tr [S 2 N, ( = + ( y) 2 + t } Σ 2 ] + 2αβtr [S Σ 2 ] + β 2 tr [Σ 2 ] 2αtr [S Σ ] 2βtr [Σ ] + + m F (0) ) α t y y αβ + t 2β 2 2 y α 2t β, where t and t 2 are given in Theorem 2. Then, the otimal α and β in terms of minimizing the limit of the loss function are given by α L t 2 t 2 = t 2 { + t ( + m F (0))( y)}, β L = t ( αl ). t 2 y Note that m F (0) can be consistently estimated by tr [S ]/ when N >. Relacing t, t 2 and m F (0) with their estimators and relacing y with /N, one gets the estimators α L and β L, which yields the linear estimator. In the case of N <, Bodnar, et al. (204) cannot rovide estimators for general Σ. This is because in their settings one needs the limit of tr [S + Σ ], but this cannot be obtained without assuming a structure such as Σ = σ 2 I for scalar σ 2. Without assuming such a structure, however, we can obtain estimators of the otimal α and β using equation (3.3). In our setting, we can exand the loss function as L (Σ, Ω linear ) = { } α 2 tr [(S + ) 2 Σ 2 ] + 2αβtr [S + Σ 2 ] + β 2 tr [Σ 2 ] 2αtr [S + Σ ] 2βtr [Σ ] +,

13 so that we need the limit of tr [(S + ) 2 Σ 2 ], tr [S + Σ 2 ] and tr [S + Σ ]. Using equation (3.3) and Theorem 3.3 in Bodnar, et al. (204), one gets lim L (Σ, Ω linear N, ) = + α 2 lim N, = + α 2 ϕ(x) x 2 N i= ϕ(l i ) (l i ) df (x) + 2αβ 2 + 2αβ lim N, N i= ϕ(l i ) l i + t 2 β 2 2 y α 2t β ϕ(x) x df (x) + t 2β 2 2 y α 2t β, where ϕ( ) is given in (3.3) and l i, i =,, N are the eigenvalues of S. Then, the otimal α and β in terms of minimizing the limit of the loss function are given by α L t 2 /(y ) t {ϕ(x)/x}df (x) = t 2 {ϕ(x)/x 2 }df (x) [ {ϕ(x)/x}df (x)] 2, β L = t {ϕ(x)/x 2 }df (x) {ϕ(x)/x}df (x)/(y ) t 2 {ϕ(x)/x 2 }df (x) [ {ϕ(x)/x}df (x)] 2. Note that x k ϕ(x)df (x) for integer k can be estimated by N i= lk i ϕ(l i ) for ϕ(x) given in (3.7). Relacing t, t 2, {ϕ(x)/x}df (x) and {ϕ(x)/x 2 }df (x) with their estimators and relacing y with /N, we have the consitent estimators α L and β L, which roduces the linear estimator for > N. 4 Simulation and Real Data Analysis 4. Simulation study In this section, we investigate erformances of the rocedures suggested in the revious sections through Monte Carlo simulation. In our simulation, eigenvalues of a oulation covariance matrix are fixed as follows: Let F (a,b) (x) be a cumulative distribution function of the beta distribution beta(a, b), namely F (a,b) (x) = Γ(a + b) Γ(a)Γ(b) x 0 t a ( t) b dt, x [0, ]. Since the suort of the beta distribution is [0, ], we transform the interval linearly to [, 0], and the oulation eigenvalues are given by This setu is the same as in Ledoit and Wolf (202). ( i + 9F (a,b) ), i =,,. (4.) 2 We first treat the case of (a, b) = (, ), and investigate the numerical erformances of the risk of Ω oracle, Ω LW, Ω LR, Ω ridge and Ω linear, which are given in (3.), (3.8), (3.6), (3.7) and (3.8), resectively. These estimators are denoted by oracle, LW, LR, ridge and linear in the tables given below. For comutation of Ω LW, we need the QuEST function introduced in Ledoit and Wolf (205). Using this ackage, we carried out the simulation 2 and Ω linear

14 exeriments on the Matlab. Random observations y,..., y N are generated as y i = Σ /2 x i for x i = (x i,..., x i ) T where x i,..., x i are mutually indeendently distributed as (Case I) x ij N (0, ), (Case II) x ij (m 2)/mz i,j for z i,j t m, where t m is a t-distribution with m degrees of freedom. (Case II) is an examle of a heavy-tailed distribution and we treat the case of m = 0. The simulation exeriments are carried out under the above setu for N = 50, = 30, 70, 50, 250, 500 and 700, and emirical risks of these estimators are calculated based on, 000 relications. Table rovides values of emirical risks of the estimators in normal and 4/5t 0 distributions with m = 0, where eigenvalues of Σ are generated by (4.) for (a, b) = (, ). Definitely, the linear ridge estimator Ω LR erforms better than the ridge and linear estimators Ω ridge and Ω linear. Comaring Ω LR and Ω LW, we can see that the linear ridge estimator Ω LR is better than the Ledoit-Wolf tye estimator Ω LW excet for the cases of = 70 or 50. Esecially, in the case of = 50, the estimator Ω LW gives a smaller risk than Ω LR in the normal and t 0 distributions, but the imrovement is not significant. When is small or /N is comaratively large, Ω LR erforms better than Ω LW. It may be because in such cases, δ(l i ) and ϕ(l i ) in equation (3.8) give oor aroximations for u T i Σ u i and u T i Σ2 u i in equation (3.), resectively. The resulting estimate δ(l i )/ ϕ(l i ) causes over-shrinkage or under-shrinkage, which leads to the increased risk of Ω LW. For the linear ridge estimator, on the other hand, we only have to calculate tr [(S + γi ) ], namely, we can use all the samle eigenvalues to estimate arameters. This leads to a stability of Ω LR in the cases that N and are small or /N is comaratively large. Comaring Ω LW and Ω LR with the oracle estimator, we observe that as gets larger, the risks of Ω LW and Ω LR become more inflated in comarison with the oracle estimator, namely, the estimation error of â i in (3.8) for estimating a i in the Ledoit-Wolf tye estimator increases for large. The inflation in the risk of Ω LR seems relatively smaller than that of Ω LW for large. Table : Emirical Risks of Ω oracle, Ω LW, Ω LR, Ω ridge and Ω linear with N = 50, (a, b) = (, ) oracle LW LR ridge linear Normal (a,b)=(,) t (a,b)=(,)

15 We next investigate the erformances of the estimators for some variety of shaes of oulation sectral distribution. As the shae arameters (a, b) in the beta distribution, we deal with the cases of (a, b) = (.5,.5), (0.5, 0.5), (5, 5) and (2, 5). For grahical illustration of the beta distribution under these arameters, see Ledoit and Wolf (202). We carry out similar simulation exeriments as described above for N = 50 and = 30, 70, 50, 250 and 500. Poulation eigenvalues are generated by equation (4.). The results are reorted in Table 2. It is revealed that we have similar risk erformances as mentioned in the case of (a, b) = (, ). Thus, the results given in Tables and 2 show that the suggested linear ridge estimator Ω LR gives a relatively good erformance among the estimators comared in this simulation. Table 2: Emirical Risks of Ω oracle Distribution, Ω LW, Ω LR, Ω ridge and Ω linear with N = 50 under Normal (a, b) oracle LW LR ridge linear (.5,.5) (0.5,0.5) (5,5) (2,5) Illustrative examle We now comare the erformance of several estimators of the recision matrix through the quadratic discriminant analysis (QDA). In QDA, when observation x i,j, i =, 2, j =,..., N i are obtained from the oulation grou Π i, i =, 2, a new observation x is classified into Π or Π 2 via the classification rule (x x ) T Ω, (x x ) (x x 2 ) T Ω 2, (x x 2 ) < (res. >) 0 = x Π (res.π 2 ), 4

16 where x i = Ni Ni j= x i,j, i =, 2 and Ω i,, i =, 2 is an estimator of the recision matrix of the oulation Π i. The classification rules given by substituting the estimators Ω, and Ω 2, corresonding to Ω LW, Ω LR, Ω ridge, Ω linear, Ω MP and Ω diag are denoted by LW, LR, ridge, linear, MP and diag, where Ω MP is the diagonal matrix Ω diag is the Moore-Penrose generalized inverse matrix and Ω diag = (diag(s )). It is known in the literature that Ω diag has a good erformance in discriminant analysis. To estimate these estimators, the samle covariance matrix of the oulation Π i is estimated by S i, = N i N i j= (x i,j x i )(x i,j x i ) T. We use the microarray data described in a colon cancer study by Alon, et al. (999) where exression levels for 2000 genes were measured on 40 normal and on 22 colon tumor tissues. The dataset is available at htt://genomics-ubs.rinceton.edu/oncology/affydata. A base 0 logarithmic transformation is alied. The descrition of the above datasets and rerocessing are due to Dettling and Buhlmann (2003) excet standardization is not followed. This dataset was used by Srivastava and Kubokawa (2007), Fisher and Sun (20) and Touloumis (205). Using this dataset, we estimate the covariance matrices of the normal and colon cancer grous of the to genes, where = 00, 250, 500 and 900. The correct classification rates based on the estimators Ω LW, Ω LR, Ω ridge, Ω linear, Ω MP and Ω diag are estimated by a leave-one-out cross validation. The estimated correct classification rates are reorted in Table 3. A relative gain in correct classification rate for our Ω LR is at least 28.7 over Ω LW and is at most 28.7 over Ω ridge. Moreover, Ω LR shows equal or slightly higher correct classification rates than Ω linear and Ω diag. Lastly we can see that the correct classification rate of Ω linear is higher than that of Ω LW and Ω ridge, which seems to contradict the results in the revious simulation study. This is because this data set of the colon cancer has considerable sarsity in the covariance structure. Thus, we may imrove the correct classification rate by assuming the shericity or diagonality for the recision matrix. Table 3: Correct Classification Rates in the Colon Cancer Dataset LW LR ridge linear MP diag Concluding Remarks In this aer, we have addressed the roblem of estimating the high-dimensional recision matrix Σ relative ot the Frobenius loss function L (Σ, Ω ) = tr (Ω Σ I )(Ω Σ I ) T. We have suggested the linear ridge estimator of the recision matrix motivated from the Bayesian asect, and rovided the estimators of the otimal arameters using the random matrix theory. 5

17 We also have derived the Ledoit-Wolf tye estimators and comared it with the suggested linear ridge estimator and the other existing estimators. Some simulation results show that the linear ridge estimator erforms better than the other existing estimators when N and are small, or when /N is large. It is worth mentioning the estimation under other loss functions. A related loss function is the quadratic loss tr [(Ω Σ I ) 2 ]. Under this loss fucntion, we can obtain the ridge, linear and linear ridge estimators. However, we could not derive a Ledoit-Wolf tye nonlinear shrinkage estimator, because we need the asymtotic behavior of u T i Σ u j for i j. In our numerical investigation, these quantities seem almost zero, but not just zero. This is left for future research. Ledoit and Wolf (202) treated the loss functions tr [(Ω Σ ) 2 ] and {tr [Ω Σ] log Ω Σ and derived the nonlinear shrinkage estimators. We can derive the ridge, linear and linear ridge estimators relative to these loss functions, and we shall comare those estimators numerically under the loss function in a future study. To get estimators of the arameters a i, α, β and γ under those loss functions, we have to solve equation (2.3) by resorting to the strong ackage QuEST run on Matlab. An advantage of the Frobenius loss function treated in this aer is that the estimators of α and β in the linear ridge estimator and the ridge estimator are exlicitly exressed for fixed γ, and the resulting linear ridge and ridge estimators are written in closed forms. Since the other estimators can be derived by using the QuEST, we can comare the erformance of the suggested linear ridge estimator with the existing other shrinkage ones. Concerning the mean µ = E[x,i ], we have treated the case of µ = 0 in this aer. In the case of µ 0, however, it is noted that all the results in this aer still hold by relacing y,i and N with y,i y and n = N for y = N N i= y i as long as we consider the asymtotics on N and. That is, the samle covariance matrix is S = N (y n,i y )(y,i y ) T, which is rewritten as S = N n N i= N (y,i µ)(y,i µ) T N n (y µ)(y µ) T. i= Leaving the second term out of consideration does not affect the LSD of S since rank((y µ)(y µ) T ) = and is negligible in our arguments for large N and. It would be interesting to use the suggested estimator for testing the hyothesis H 0 : µ = 0. Reasonable test statistics are functions of y T Ω y. For examle, test statistics of Bai and Saradanasa (996) and Srivastava (2007) are based on y T y /( tr [S ]) and y T S + y, resectively, which are corresond to Ω = (/tr S)I = (/l)i and Ω = S +. Using the linear ridge estimator for Ω, we can suggest a test statistic based on ) y T Ω LR y =y T (( w)α 0 (S + γi ) + w(/l)i y =( w)α 0 y T (S + γi ) y + w(/l)y T y, which is a convex combination of the Bai-Saranadasa statistic and the Hotelling T -statistic relacing S or S + with the ridge estimator. We will study the erformance of a test statistic based y T Ω LR y as a future roject. 6

18 A Aendix A. Proof of Theorem The Frobenius loss function of the rotation-equivariant estimator is L (Σ, Ω (A )) = {a 2 (l i )u T i Σ 2 u i 2a(l i )u T i Σ u i } +. Theorem 4 in Ledoit and Péché (20) shows that a(l i )u T i Σ u i a(x)δ(x)df (x), i= i= where δ(x) is given in (3.2). Thus, we shall evaluate the asymtotic quantity of i= a(l i)u T i Σ2 u i. The same arguments as used in the roof of Theorem 4 in Ledoit and Péché (20) are heavily exloited for the evaluation. Let (2) (x) be the nondecreasing function defined by (2) (x) = u T j Σ 2 u j [lj,+ )(x), x R. j= When all the samle eigenvalues are distinct, it is noted that u T i Σ2 u i (i =,, ) can be recovered from (2) as u T i Σ 2 u i = lim ϵ 0 + (2) (l i + ϵ) (2) (l i ϵ). F (l i + ϵ) F (l i ϵ) Thus, it suffices to examine the asymtotic behavior of (2) (x). Note that the stieltjes transform of (2) (x) is the same with Euation (3) of Ledoit and Péché (20) for g(τ j ) = τj 2. Let Θ(2) (z) be the stieltjes transform of (2) (x). Then, Θ (2) (z) converges a.s. to Θ (2) (z) for all z C +. Moreover, it follows from Lemmas 2 and 3 in Ledoit and Péché (20) that Θ (2) (z) = z + z 2 m F (z) tdh(t) ( y yzm F (z)) 2 + y yzm F (z), z C+. (A.) Using Lemma 6 in Ledoit and Péché (20), we can see that lim (2) (x) (2) (x) exists and is equal to (2) x (x) = lim I[Θ (2) (λ + iη)]dλ. (A.2) η 0 + π for every continuous oint x R of (2) (x). Plugging (A.) into (A.2) yields (2) (x) = lim η 0 + π = lim η 0 + π x + lim η 0 + π x x [ (λ + iη) + (λ + iη) 2 m F (λ + iη) I ( y y(λ + iη)m F (λ + iη)) 2 + [ (λ + iη) + (λ + iη) 2 m F (λ + iη) ] I ( y y(λ + iη)m F (λ + iη)) 2 dλ [ tdh(t) ] I dλ. y y(λ + iη)m F (λ + iη) tdh(t) ] dλ y y(λ + iη)m F (λ + iη) (A.3) 7

19 We begin by evaluating (2) (x) in the case of y (0, ). The first term in RHS of (A.3) is lim η 0 + π x = lim η 0 + π x [ (λ + iη) + (λ + iη) 2 m F (λ + iη) ] I ( y y(λ + iη)m F (λ + iη)) 2 which is rewritten as π x = π = x x dλ [ { } I (λ + iη) + (λ + iη)m F (λ + iη) { ( y yr[(λ + iη)m F (λ + iη)]) 2 y 2 (I[(λ + iη)m F (λ + iη)]) 2 }] + 2iyI[(λ + iη)m F (λ + iη)]( y yr[(λ + iη)m F (λ + iη)]) ( ( y y(λ + iη)m F (λ + iη)) 2 2 ) dλ, λi[λm F (λ)] { ( + y + yr[λm F (λ)])( y yr[λm F (λ)]) y 2 (I[λm F (λ)]) 2} ( y yλm F (λ)) 2 2 dλ I[m F (λ)]λ 2 { y 2 2y 2 λr[m F (λ)] y 2 λ 2 m F (λ) 2 } ( y yλm F (λ)) 2 2 dλ λ 2 { y 2 2y 2 λr[m F (λ)] y 2 λ 2 m F (λ) 2 } ( y yλm F (λ)) 2 2 df (λ). The second term in RHS of (A.3) is x tdh(t) lim I[ η 0 + π y y(λ + iη)m F (λ + iη) ]dλ = tdh(t) x π lim yi[(λ + iη)m F (λ + iη)] η 0 + y y(λ + iη)m F (λ + iη) 2 dλ, which is equal to tdh(t) x π x = tdh(t) Hence, for x > 0, it is concluded that yi[λm F (λ)] y yλm F (λ) 2 dλ yλf (λ) y yλm F (λ) 2 dλ = (2) (x) = x x ϕ(λ)df (λ), for ϕ( ) given in (3.3). This shows Theorem in the case of y (0, ). tdh(t)yλ 2 df (λ). y yλm F (λ) We next consider the case of y (, ). Rewriting Θ (2) (z) based on the relation between m F (z) and m F (z) given by y + yzm F (z) = + zm F (z), we have Θ (2) (z) = zm F (z) = zm F (z) ( y z + y zm F (z) zm F (z) ( + zmf (z) ym F (z) + ) tdh(t) ) tdh(t) = µ(z) z 8

20 where µ(z) = ( + zm F (z) + m F (z) ym F (z) The inversion formula for the stieltjes transform imlies that: lim (ϵ) (2) ( ϵ)) = lim ϵ 0 +( (2) ϵ 0 + lim η 0 + π ϵ ϵ lim ϵ ϵ 0 + η 0 + π ϵ = lim ) tdh(t). I[Θ (2) (λ + iη)]dλ [ I =µ(0) = ( m F (0) ym F (0) + µ(λ + iη) ] dλ λ + iη ) tdh(t), (A.4) The third equality follows from Lemma 9 in Ledoit and Péché (20), since µ is a comlex holomorhic function and µ(0) R. Also from Lemma 8 in Ledoit and Péché (20)], it is noted that F (λ) = ( y ) [0,+ ) (λ) for λ in a neighborhood of zero. For x in a neighborhood of zero, from (A.4), it can be seen that (2) (x) = x m F (0) ( ym F (0) + ) tdh(t) d [0,+ ) (λ)λ. Comaring the two exressions, we can see that for x in a neighborhood of zero, x (2) y ( ) (x) = y m F (0) ym F (0) + tdh(t) df (λ) Therefore, for x in a neighborhood of zero, which roves Theorem. A.2 Proof of Theorem 2 (2) (x) = x ϕ(λ)df (λ), We begin by roviding the roof for the limit of α (γ). Given γ > 0, the otimal α is α (γ) = tr [(S + γi ) Σ ] tr [Σ2 ] tr [(S + γi ) Σ 2 ] tr [Σ ] tr [(S + γi ) 2 Σ 2 ] tr [Σ2 ] { tr [(S + γi ) Σ 2 ]} 2 It follows from Theorem 2 in Wang, et al. (204) that when N and /N y (0, ), tr [(S + γi ) Σ ] γ m F ( γ) y( γ m F ( γ)). Let B(γ) = { γ m F ( γ)}/{ y( γ m F ( γ))}. Since tr [(S + γi ) Σ 2 ] is Θ (2) ( γ) in the roof of Theorem, it can be seen from the equation (A.) that tr [(S + γi ) Σ 2 ] γ + γ 2 m F ( γ) ( y( yγ m F ( γ))) 2 + =( tdh(t) γb(γ)) ( + yb(γ) ). tdh(t) y( yγ m F ( γ)) 9

21 Since tr [(S + γi ) 2 Σ 2 ] = d dγ tr [(S + γi ) Σ 2 ], the limit of tr [(S + γi ) 2 Σ 2 ] is d ( ( ) tdh(t) γb(γ)) + yb(γ) dγ ( = yb 2 (γ) + B(γ) + 2yγB(γ)B (γ) + γ ) tdh(t) B (γ). For β (γ) and L (γ), we can rove their limits using the same arguments, and the roof of Theorem 2 is comlete. A.3 Consistent estimator of the stieltjes transform m F (x) This section exlains briefly the method of Ledoit and Wolf (205) for estimating m F (x). Ledoit and Wolf (205) rovided the nonrandom multivariate function called the Quantized Eigenvalues Samling Transforms or QuEST for short. For any ositive integers N and, the QuEST function denoted by Q N, is defined as Q N, : [0, ) [0, ) (A.5) t (t,, t ) T Q N, (t) (q N,(t),, q N, (t))t. (A.6) For z C +, m m t N, (z) is the unique solution in the set {m C : (N )/Nz+m/N C+ } to the equation m = i= t i { (/N) (/N)zm} z, (A.7) x R, FN,(x) t { max N, lim η 0 + π x i= (ti =0) } I[m t N,(ξ + iη)]dξ if x = 0, otherwise, (A.8) u [0, ], (F t N,) (u) su{x R : F t N,(x) u}, (A.9) and i/ i =,,, qn,(t) i (FN,) t (u)du. (A.0) (i )/ It can be seen that equation (A.7) quantizes equation (2.3), and that equation (A.8) quantizes equations (2.) and (2.2). FN, t is the limiting distribution of samle eigenvalues corresonding to the oulation sectral distribution i= (t i =0). By equation (A.9), (FN, t ) reresents the inverse sectral distribution function, or the quantile function. By equation (A.0), qn, i (t) 20

22 can be interreted as a smoothed version of the (i 0.5)/ quantile of FN, t. Ledoit and Wolf (205) estimates the eigenvalues of the oulation covariance matrix by ˆt = argmin t [0, ) [qn,(t) i l i ] 2, where l = (l,..., l ) T are eigenvalues of the samle covariance matrix, and showed that i= [ˆt i t i ] 2 0. i= Lastly, the estimator of m F (x) is obtained as the unique solution ˆm R C + to the equation m = i= ˆt i { (/N) (/N)xm} x. Acknowledgments. First of all, we are really grateful to Professor Olivier Ledoit for his kindness of giving us his rograms for comutation. Research of the second author was suorted in art by Grant-in- Aid for Scientific Research (5H0943 and ) from Jaan Society for the Promotion of Science. References [] Alon, U., Barkai, N., Motterman, D., Gish, K., Mack, S., and Levine, J. (999). Broad atterns of gene exression revealed by clustering analysis of tumor and normal colon tissues robed by oligonucleotide arrays. Proceeding of the National Academy of Science of the United States of America, 96, [2] Bai, Z., and Saranadasa, H. (996). Effect of high dimension: By an examle of a two samle roblem. Statist. Sinica, 6, [3] Bai, J., and Shi, S. (20). Estimating high dimensional covariance matrices and its alications. Ann. Economics and finance, 2, [4] Bodnar, T., Guta, A.K., and Parolya, N. (205). Direct shrinkage estimation of large dimensional recision matrix. J. Multivariate Analysis, to aear. [5] Dettling, M., and Buhlmann, P. (2002). Boosting for tumor classification with gene exression data. Bioinfomatics, 9, [6] Fisher, T.J., and Sun, X. (20). Imroved Stein-tye shrinkage estimators for the highdimensional multivariate normal covariance matrix. Com. Statist. Data Analysis, 55,

23 [7] Himeno, T. and Yamada, T. (204). Estimations for some functions of covariance matrix in high dimension under non-normality and its alications. J. Multivariate Analysis, 30, [8] Ledoit, O., and Wolf, M. (2004). A well-conditioned estimator for large-dimensional covariance matrices. J. Multivariate Analysis, 88, [9] Ledoit, O., and Péché, S. (20). Eigenvectors of some large samle covariance matrix. Prob. Theory Relat. Fields, 52, [0] Ledoit, O., and Wolf, M. (202). Nonlinear shrinkage estimation of large-dimensional covariance matrices. Ann. Statist., 40, [] Ledoit, O., and Wolf, M. (205). Sectrum estimation: A unified framework for covariance matrix estimation and PCA in large dimensions. J. Multivariate Anal., 39, [2] Marčenko, V.A., and Patur, L.A. (967). Distribution of eigenvalues for some sets of random matrices. Math. USSR-Sb.,, [3] Silverstein, J. (995). Strong convergence of the emirical distribution of eigenvalues of large dimensional random matrices. J. Multivariate Anal., 55, [4] Silverstein, J., and Bai, Z.D. (995). On the emirical distribution of eigenvalues of a class of large dimensional random matrices. J. Multivariate Anal., 54, [5] Silverstein, J., and Choi, S. (995). Analysis of the limiting sectral distribution of large dimensional random matrices. J. Multivariate Anal., 54, [6] Srivastava, M.S. (2007). Multivariate theory for analysing high dimensional data. J. Jaan Statist. Soc., 37, [7] Srivastava, M.S., and Kubokawa, T. (2007). Comarison of discrimination methods for high dimensional data. J. Jaan Statist. Soc., 37, [8] Wang, C., Pan, G., Tong, T., and Zhu, L. (205). Shrinkage estimation of large dimensional recision matrix using random matrix theory. Statistica Sinica, 25, [9] Touloumis, A. (205). Nonarametric Stein-tye shrinkage covariance matrix estimators in high-dimensional setting. Com. Statist. Data Analysis, 83,

Estimation of the large covariance matrix with two-step monotone missing data

Estimation of the large covariance matrix with two-step monotone missing data Estimation of the large covariance matrix with two-ste monotone missing data Masashi Hyodo, Nobumichi Shutoh 2, Takashi Seo, and Tatjana Pavlenko 3 Deartment of Mathematical Information Science, Tokyo

More information

arxiv: v2 [stat.me] 3 Nov 2014

arxiv: v2 [stat.me] 3 Nov 2014 onarametric Stein-tye Shrinkage Covariance Matrix Estimators in High-Dimensional Settings Anestis Touloumis Cancer Research UK Cambridge Institute University of Cambridge Cambridge CB2 0RE, U.K. Anestis.Touloumis@cruk.cam.ac.uk

More information

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE

ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE J Jaan Statist Soc Vol 34 No 2004 9 26 ASYMPTOTIC RESULTS OF A HIGH DIMENSIONAL MANOVA TEST AND POWER COMPARISON WHEN THE DIMENSION IS LARGE COMPARED TO THE SAMPLE SIZE Yasunori Fujikoshi*, Tetsuto Himeno

More information

Estimation of the Global Minimum Variance Portfolio in High Dimensions

Estimation of the Global Minimum Variance Portfolio in High Dimensions Estimation of the Global Minimum Variance Portfolio in High Dimensions Taras Bodnar, Nestor Parolya and Wolfgang Schmid 07.FEBRUARY 2014 1 / 25 Outline Introduction Random Matrix Theory: Preliminary Results

More information

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Journal of Modern Alied Statistical Methods Volume Issue Article 7 --03 A Comarison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Ghadban Khalaf King Khalid University, Saudi

More information

Linear diophantine equations for discrete tomography

Linear diophantine equations for discrete tomography Journal of X-Ray Science and Technology 10 001 59 66 59 IOS Press Linear diohantine euations for discrete tomograhy Yangbo Ye a,gewang b and Jiehua Zhu a a Deartment of Mathematics, The University of Iowa,

More information

Journal of Multivariate Analysis

Journal of Multivariate Analysis Journal of Multivariate Analysis 101 010 55 570 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homeage: www.elsevier.com/locate/jmva A new test for shericity of the

More information

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek Use of Transformations and the Reeated Statement in PROC GLM in SAS Ed Stanek Introduction We describe how the Reeated Statement in PROC GLM in SAS transforms the data to rovide tests of hyotheses of interest.

More information

substantial literature on emirical likelihood indicating that it is widely viewed as a desirable and natural aroach to statistical inference in a vari

substantial literature on emirical likelihood indicating that it is widely viewed as a desirable and natural aroach to statistical inference in a vari Condence tubes for multile quantile lots via emirical likelihood John H.J. Einmahl Eindhoven University of Technology Ian W. McKeague Florida State University May 7, 998 Abstract The nonarametric emirical

More information

Asymptotically Optimal Simulation Allocation under Dependent Sampling

Asymptotically Optimal Simulation Allocation under Dependent Sampling Asymtotically Otimal Simulation Allocation under Deendent Samling Xiaoing Xiong The Robert H. Smith School of Business, University of Maryland, College Park, MD 20742-1815, USA, xiaoingx@yahoo.com Sandee

More information

Maxisets for μ-thresholding rules

Maxisets for μ-thresholding rules Test 008 7: 33 349 DOI 0.007/s749-006-0035-5 ORIGINAL PAPER Maxisets for μ-thresholding rules Florent Autin Received: 3 January 005 / Acceted: 8 June 006 / Published online: March 007 Sociedad de Estadística

More information

Covariance Matrix Estimation for Reinforcement Learning

Covariance Matrix Estimation for Reinforcement Learning Covariance Matrix Estimation for Reinforcement Learning Tomer Lancewicki Deartment of Electrical Engineering and Comuter Science University of Tennessee Knoxville, TN 37996 tlancewi@utk.edu Itamar Arel

More information

An Improved Generalized Estimation Procedure of Current Population Mean in Two-Occasion Successive Sampling

An Improved Generalized Estimation Procedure of Current Population Mean in Two-Occasion Successive Sampling Journal of Modern Alied Statistical Methods Volume 15 Issue Article 14 11-1-016 An Imroved Generalized Estimation Procedure of Current Poulation Mean in Two-Occasion Successive Samling G. N. Singh Indian

More information

ON THE LEAST SIGNIFICANT p ADIC DIGITS OF CERTAIN LUCAS NUMBERS

ON THE LEAST SIGNIFICANT p ADIC DIGITS OF CERTAIN LUCAS NUMBERS #A13 INTEGERS 14 (014) ON THE LEAST SIGNIFICANT ADIC DIGITS OF CERTAIN LUCAS NUMBERS Tamás Lengyel Deartment of Mathematics, Occidental College, Los Angeles, California lengyel@oxy.edu Received: 6/13/13,

More information

4. Score normalization technical details We now discuss the technical details of the score normalization method.

4. Score normalization technical details We now discuss the technical details of the score normalization method. SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules

More information

Research Article An iterative Algorithm for Hemicontractive Mappings in Banach Spaces

Research Article An iterative Algorithm for Hemicontractive Mappings in Banach Spaces Abstract and Alied Analysis Volume 2012, Article ID 264103, 11 ages doi:10.1155/2012/264103 Research Article An iterative Algorithm for Hemicontractive Maings in Banach Saces Youli Yu, 1 Zhitao Wu, 2 and

More information

An Analysis of Reliable Classifiers through ROC Isometrics

An Analysis of Reliable Classifiers through ROC Isometrics An Analysis of Reliable Classifiers through ROC Isometrics Stijn Vanderlooy s.vanderlooy@cs.unimaas.nl Ida G. Srinkhuizen-Kuyer kuyer@cs.unimaas.nl Evgueni N. Smirnov smirnov@cs.unimaas.nl MICC-IKAT, Universiteit

More information

Hotelling s Two- Sample T 2

Hotelling s Two- Sample T 2 Chater 600 Hotelling s Two- Samle T Introduction This module calculates ower for the Hotelling s two-grou, T-squared (T) test statistic. Hotelling s T is an extension of the univariate two-samle t-test

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

Elementary Analysis in Q p

Elementary Analysis in Q p Elementary Analysis in Q Hannah Hutter, May Szedlák, Phili Wirth November 17, 2011 This reort follows very closely the book of Svetlana Katok 1. 1 Sequences and Series In this section we will see some

More information

A well-conditioned estimator for large-dimensional covariance matrices

A well-conditioned estimator for large-dimensional covariance matrices Journal of Multivariate Analysis 88 (004) 365 411 A well-conditioned estimator for large-dimensional covariance matrices Olivier Ledoit a,b and Michael Wolf c,,1 a Anderson Graduate School of Management,

More information

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material Robustness of classifiers to uniform l and Gaussian noise Sulementary material Jean-Yves Franceschi Ecole Normale Suérieure de Lyon LIP UMR 5668 Omar Fawzi Ecole Normale Suérieure de Lyon LIP UMR 5668

More information

CHAPTER 5 STATISTICAL INFERENCE. 1.0 Hypothesis Testing. 2.0 Decision Errors. 3.0 How a Hypothesis is Tested. 4.0 Test for Goodness of Fit

CHAPTER 5 STATISTICAL INFERENCE. 1.0 Hypothesis Testing. 2.0 Decision Errors. 3.0 How a Hypothesis is Tested. 4.0 Test for Goodness of Fit Chater 5 Statistical Inference 69 CHAPTER 5 STATISTICAL INFERENCE.0 Hyothesis Testing.0 Decision Errors 3.0 How a Hyothesis is Tested 4.0 Test for Goodness of Fit 5.0 Inferences about Two Means It ain't

More information

Approximating min-max k-clustering

Approximating min-max k-clustering Aroximating min-max k-clustering Asaf Levin July 24, 2007 Abstract We consider the roblems of set artitioning into k clusters with minimum total cost and minimum of the maximum cost of a cluster. The cost

More information

On split sample and randomized confidence intervals for binomial proportions

On split sample and randomized confidence intervals for binomial proportions On slit samle and randomized confidence intervals for binomial roortions Måns Thulin Deartment of Mathematics, Usala University arxiv:1402.6536v1 [stat.me] 26 Feb 2014 Abstract Slit samle methods have

More information

The spectrum of kernel random matrices

The spectrum of kernel random matrices The sectrum of kernel random matrices Noureddine El Karoui Deartment of Statistics, University of California, Berkeley Abstract We lace ourselves in the setting of high-dimensional statistical inference,

More information

Bayesian Spatially Varying Coefficient Models in the Presence of Collinearity

Bayesian Spatially Varying Coefficient Models in the Presence of Collinearity Bayesian Satially Varying Coefficient Models in the Presence of Collinearity David C. Wheeler 1, Catherine A. Calder 1 he Ohio State University 1 Abstract he belief that relationshis between exlanatory

More information

General Linear Model Introduction, Classes of Linear models and Estimation

General Linear Model Introduction, Classes of Linear models and Estimation Stat 740 General Linear Model Introduction, Classes of Linear models and Estimation An aim of scientific enquiry: To describe or to discover relationshis among events (variables) in the controlled (laboratory)

More information

An Inverse Problem for Two Spectra of Complex Finite Jacobi Matrices

An Inverse Problem for Two Spectra of Complex Finite Jacobi Matrices Coyright 202 Tech Science Press CMES, vol.86, no.4,.30-39, 202 An Inverse Problem for Two Sectra of Comlex Finite Jacobi Matrices Gusein Sh. Guseinov Abstract: This aer deals with the inverse sectral roblem

More information

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK Towards understanding the Lorenz curve using the Uniform distribution Chris J. Stehens Newcastle City Council, Newcastle uon Tyne, UK (For the Gini-Lorenz Conference, University of Siena, Italy, May 2005)

More information

Additive results for the generalized Drazin inverse in a Banach algebra

Additive results for the generalized Drazin inverse in a Banach algebra Additive results for the generalized Drazin inverse in a Banach algebra Dragana S. Cvetković-Ilić Dragan S. Djordjević and Yimin Wei* Abstract In this aer we investigate additive roerties of the generalized

More information

Lower Confidence Bound for Process-Yield Index S pk with Autocorrelated Process Data

Lower Confidence Bound for Process-Yield Index S pk with Autocorrelated Process Data Quality Technology & Quantitative Management Vol. 1, No.,. 51-65, 15 QTQM IAQM 15 Lower onfidence Bound for Process-Yield Index with Autocorrelated Process Data Fu-Kwun Wang * and Yeneneh Tamirat Deartment

More information

Tests for Two Proportions in a Stratified Design (Cochran/Mantel-Haenszel Test)

Tests for Two Proportions in a Stratified Design (Cochran/Mantel-Haenszel Test) Chater 225 Tests for Two Proortions in a Stratified Design (Cochran/Mantel-Haenszel Test) Introduction In a stratified design, the subects are selected from two or more strata which are formed from imortant

More information

arxiv: v2 [stat.me] 27 Apr 2018

arxiv: v2 [stat.me] 27 Apr 2018 Nonasymtotic estimation and suort recovery for high dimensional sarse covariance matrices arxiv:17050679v [statme] 7 Ar 018 1 Introduction Adam B Kashlak and Linglong Kong Deartment of Mathematical and

More information

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO)

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO) Combining Logistic Regression with Kriging for Maing the Risk of Occurrence of Unexloded Ordnance (UXO) H. Saito (), P. Goovaerts (), S. A. McKenna (2) Environmental and Water Resources Engineering, Deartment

More information

Machine Learning: Homework 4

Machine Learning: Homework 4 10-601 Machine Learning: Homework 4 Due 5.m. Monday, February 16, 2015 Instructions Late homework olicy: Homework is worth full credit if submitted before the due date, half credit during the next 48 hours,

More information

t 0 Xt sup X t p c p inf t 0

t 0 Xt sup X t p c p inf t 0 SHARP MAXIMAL L -ESTIMATES FOR MARTINGALES RODRIGO BAÑUELOS AND ADAM OSȨKOWSKI ABSTRACT. Let X be a suermartingale starting from 0 which has only nonnegative jums. For each 0 < < we determine the best

More information

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests 009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 0-, 009 FrB4. System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests James C. Sall Abstract

More information

Sharp gradient estimate and spectral rigidity for p-laplacian

Sharp gradient estimate and spectral rigidity for p-laplacian Shar gradient estimate and sectral rigidity for -Lalacian Chiung-Jue Anna Sung and Jiaing Wang To aear in ath. Research Letters. Abstract We derive a shar gradient estimate for ositive eigenfunctions of

More information

Asymptotic non-null distributions of test statistics for redundancy in high-dimensional canonical correlation analysis

Asymptotic non-null distributions of test statistics for redundancy in high-dimensional canonical correlation analysis Asymtotic non-null distributions of test statistics for redundancy in high-dimensional canonical correlation analysis Ryoya Oda, Hirokazu Yanagihara, and Yasunori Fujikoshi, Deartment of Mathematics, Graduate

More information

Eigenvectors of some large sample covariance matrix ensembles

Eigenvectors of some large sample covariance matrix ensembles Probab. Theory Relat. Fields DOI 0.007/s00440-00-0298-3 Eigenvectors of some large sample covariance matrix ensembles Olivier Ledoit Sandrine Péché Received: 24 March 2009 / Revised: 0 March 200 Springer-Verlag

More information

MATH 2710: NOTES FOR ANALYSIS

MATH 2710: NOTES FOR ANALYSIS MATH 270: NOTES FOR ANALYSIS The main ideas we will learn from analysis center around the idea of a limit. Limits occurs in several settings. We will start with finite limits of sequences, then cover infinite

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

A New Asymmetric Interaction Ridge (AIR) Regression Method

A New Asymmetric Interaction Ridge (AIR) Regression Method A New Asymmetric Interaction Ridge (AIR) Regression Method by Kristofer Månsson, Ghazi Shukur, and Pär Sölander The Swedish Retail Institute, HUI Research, Stockholm, Sweden. Deartment of Economics and

More information

POINTS ON CONICS MODULO p

POINTS ON CONICS MODULO p POINTS ON CONICS MODULO TEAM 2: JONGMIN BAEK, ANAND DEOPURKAR, AND KATHERINE REDFIELD Abstract. We comute the number of integer oints on conics modulo, where is an odd rime. We extend our results to conics

More information

Comparison of Discrimination Methods for High Dimensional Data

Comparison of Discrimination Methods for High Dimensional Data Comparison of Discrimination Methods for High Dimensional Data M.S.Srivastava 1 and T.Kubokawa University of Toronto and University of Tokyo Abstract In microarray experiments, the dimension p of the data

More information

CERIAS Tech Report The period of the Bell numbers modulo a prime by Peter Montgomery, Sangil Nahm, Samuel Wagstaff Jr Center for Education

CERIAS Tech Report The period of the Bell numbers modulo a prime by Peter Montgomery, Sangil Nahm, Samuel Wagstaff Jr Center for Education CERIAS Tech Reort 2010-01 The eriod of the Bell numbers modulo a rime by Peter Montgomery, Sangil Nahm, Samuel Wagstaff Jr Center for Education and Research Information Assurance and Security Purdue University,

More information

Stationary Monetary Equilibria with Strictly Increasing Value Functions and Non-Discrete Money Holdings Distributions: An Indeterminacy Result

Stationary Monetary Equilibria with Strictly Increasing Value Functions and Non-Discrete Money Holdings Distributions: An Indeterminacy Result CIRJE-F-615 Stationary Monetary Equilibria with Strictly Increasing Value Functions and Non-Discrete Money Holdings Distributions: An Indeterminacy Result Kazuya Kamiya University of Toyo Taashi Shimizu

More information

Multiplicative group law on the folium of Descartes

Multiplicative group law on the folium of Descartes Multilicative grou law on the folium of Descartes Steluţa Pricoie and Constantin Udrişte Abstract. The folium of Descartes is still studied and understood today. Not only did it rovide for the roof of

More information

VIBRATION ANALYSIS OF BEAMS WITH MULTIPLE CONSTRAINED LAYER DAMPING PATCHES

VIBRATION ANALYSIS OF BEAMS WITH MULTIPLE CONSTRAINED LAYER DAMPING PATCHES Journal of Sound and Vibration (998) 22(5), 78 85 VIBRATION ANALYSIS OF BEAMS WITH MULTIPLE CONSTRAINED LAYER DAMPING PATCHES Acoustics and Dynamics Laboratory, Deartment of Mechanical Engineering, The

More information

Sums of independent random variables

Sums of independent random variables 3 Sums of indeendent random variables This lecture collects a number of estimates for sums of indeendent random variables with values in a Banach sace E. We concentrate on sums of the form N γ nx n, where

More information

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0 Advanced Finite Elements MA5337 - WS7/8 Solution sheet This exercise sheets deals with B-slines and NURBS, which are the basis of isogeometric analysis as they will later relace the olynomial ansatz-functions

More information

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)]

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)] LECTURE 7 NOTES 1. Convergence of random variables. Before delving into the large samle roerties of the MLE, we review some concets from large samle theory. 1. Convergence in robability: x n x if, for

More information

Approximation of the Euclidean Distance by Chamfer Distances

Approximation of the Euclidean Distance by Chamfer Distances Acta Cybernetica 0 (0 399 47. Aroximation of the Euclidean Distance by Chamfer Distances András Hajdu, Lajos Hajdu, and Robert Tijdeman Abstract Chamfer distances lay an imortant role in the theory of

More information

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning TNN-2009-P-1186.R2 1 Uncorrelated Multilinear Princial Comonent Analysis for Unsuervised Multilinear Subsace Learning Haiing Lu, K. N. Plataniotis and A. N. Venetsanooulos The Edward S. Rogers Sr. Deartment

More information

ON THE SET a x + b g x (mod p) 1 Introduction

ON THE SET a x + b g x (mod p) 1 Introduction PORTUGALIAE MATHEMATICA Vol 59 Fasc 00 Nova Série ON THE SET a x + b g x (mod ) Cristian Cobeli, Marian Vâjâitu and Alexandru Zaharescu Abstract: Given nonzero integers a, b we rove an asymtotic result

More information

Estimating function analysis for a class of Tweedie regression models

Estimating function analysis for a class of Tweedie regression models Title Estimating function analysis for a class of Tweedie regression models Author Wagner Hugo Bonat Deartamento de Estatística - DEST, Laboratório de Estatística e Geoinformação - LEG, Universidade Federal

More information

On the sphericity test with large-dimensional observations. Citation Electronic Journal of Statistics, 2013, v. 7, p

On the sphericity test with large-dimensional observations. Citation Electronic Journal of Statistics, 2013, v. 7, p Title On the shericity test with large-dimensional observations Authors Wang, Q; Yao, J Citation Electronic Journal of Statistics, 203, v. 7,. 264 292 Issued Date 203 URL htt://hdl.handle.net/0722/89449

More information

Chater Matrix Norms and Singular Value Decomosition Introduction In this lecture, we introduce the notion of a norm for matrices The singular value de

Chater Matrix Norms and Singular Value Decomosition Introduction In this lecture, we introduce the notion of a norm for matrices The singular value de Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A Dahleh George Verghese Deartment of Electrical Engineering and Comuter Science Massachuasetts Institute of Technology c Chater Matrix Norms

More information

Non white sample covariance matrices.

Non white sample covariance matrices. Non white sample covariance matrices. S. Péché, Université Grenoble 1, joint work with O. Ledoit, Uni. Zurich 17-21/05/2010, Université Marne la Vallée Workshop Probability and Geometry in High Dimensions

More information

Applications to stochastic PDE

Applications to stochastic PDE 15 Alications to stochastic PE In this final lecture we resent some alications of the theory develoed in this course to stochastic artial differential equations. We concentrate on two secific examles:

More information

Probability Estimates for Multi-class Classification by Pairwise Coupling

Probability Estimates for Multi-class Classification by Pairwise Coupling Probability Estimates for Multi-class Classification by Pairwise Couling Ting-Fan Wu Chih-Jen Lin Deartment of Comuter Science National Taiwan University Taiei 06, Taiwan Ruby C. Weng Deartment of Statistics

More information

A multiple testing approach to the regularisation of large sample correlation matrices

A multiple testing approach to the regularisation of large sample correlation matrices A multile testing aroach to the regularisation of large samle correlation matrices Natalia Bailey Queen Mary, University of London M. Hashem Pesaran University of Southern California, USA, and rinity College,

More information

Paper C Exact Volume Balance Versus Exact Mass Balance in Compositional Reservoir Simulation

Paper C Exact Volume Balance Versus Exact Mass Balance in Compositional Reservoir Simulation Paer C Exact Volume Balance Versus Exact Mass Balance in Comositional Reservoir Simulation Submitted to Comutational Geosciences, December 2005. Exact Volume Balance Versus Exact Mass Balance in Comositional

More information

Collaborative Place Models Supplement 1

Collaborative Place Models Supplement 1 Collaborative Place Models Sulement Ber Kaicioglu Foursquare Labs ber.aicioglu@gmail.com Robert E. Schaire Princeton University schaire@cs.rinceton.edu David S. Rosenberg P Mobile Labs david.davidr@gmail.com

More information

On the Chvatál-Complexity of Knapsack Problems

On the Chvatál-Complexity of Knapsack Problems R u t c o r Research R e o r t On the Chvatál-Comlexity of Knasack Problems Gergely Kovács a Béla Vizvári b RRR 5-08, October 008 RUTCOR Rutgers Center for Oerations Research Rutgers University 640 Bartholomew

More information

HENSEL S LEMMA KEITH CONRAD

HENSEL S LEMMA KEITH CONRAD HENSEL S LEMMA KEITH CONRAD 1. Introduction In the -adic integers, congruences are aroximations: for a and b in Z, a b mod n is the same as a b 1/ n. Turning information modulo one ower of into similar

More information

The Binomial Approach for Probability of Detection

The Binomial Approach for Probability of Detection Vol. No. (Mar 5) - The e-journal of Nondestructive Testing - ISSN 45-494 www.ndt.net/?id=7498 The Binomial Aroach for of Detection Carlos Correia Gruo Endalloy C.A. - Caracas - Venezuela www.endalloy.net

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell February 10, 2010 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

DEPARTMENT OF ECONOMICS ISSN DISCUSSION PAPER 20/07 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES

DEPARTMENT OF ECONOMICS ISSN DISCUSSION PAPER 20/07 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES DEPARTMENT OF ECONOMICS ISSN 1441-549 DISCUSSION PAPER /7 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES ZuXiang Wang * & Russell Smyth ABSTRACT We resent two new Lorenz curve families by using the basic

More information

State Estimation with ARMarkov Models

State Estimation with ARMarkov Models Deartment of Mechanical and Aerosace Engineering Technical Reort No. 3046, October 1998. Princeton University, Princeton, NJ. State Estimation with ARMarkov Models Ryoung K. Lim 1 Columbia University,

More information

A Note on the Positive Nonoscillatory Solutions of the Difference Equation

A Note on the Positive Nonoscillatory Solutions of the Difference Equation Int. Journal of Math. Analysis, Vol. 4, 1, no. 36, 1787-1798 A Note on the Positive Nonoscillatory Solutions of the Difference Equation x n+1 = α c ix n i + x n k c ix n i ) Vu Van Khuong 1 and Mai Nam

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Numerous alications in statistics, articularly in the fitting of linear models. Notation and conventions: Elements of a matrix A are denoted by a ij, where i indexes the rows and

More information

Journal of Mathematical Analysis and Applications

Journal of Mathematical Analysis and Applications J. Math. Anal. Al. 44 (3) 3 38 Contents lists available at SciVerse ScienceDirect Journal of Mathematical Analysis and Alications journal homeage: www.elsevier.com/locate/jmaa Maximal surface area of a

More information

Journal of Inequalities in Pure and Applied Mathematics

Journal of Inequalities in Pure and Applied Mathematics Journal of Inequalities in Pure and Alied Mathematics htt://jiam.vu.edu.au/ Volume 3, Issue 5, Article 8, 22 REVERSE CONVOLUTION INEQUALITIES AND APPLICATIONS TO INVERSE HEAT SOURCE PROBLEMS SABUROU SAITOH,

More information

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM JOHN BINDER Abstract. In this aer, we rove Dirichlet s theorem that, given any air h, k with h, k) =, there are infinitely many rime numbers congruent to

More information

On-Line Appendix. Matching on the Estimated Propensity Score (Abadie and Imbens, 2015)

On-Line Appendix. Matching on the Estimated Propensity Score (Abadie and Imbens, 2015) On-Line Aendix Matching on the Estimated Proensity Score Abadie and Imbens, 205 Alberto Abadie and Guido W. Imbens Current version: August 0, 205 The first art of this aendix contains additional roofs.

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell October 25, 2009 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

Asymptotic Properties of the Markov Chain Model method of finding Markov chains Generators of..

Asymptotic Properties of the Markov Chain Model method of finding Markov chains Generators of.. IOSR Journal of Mathematics (IOSR-JM) e-issn: 78-578, -ISSN: 319-765X. Volume 1, Issue 4 Ver. III (Jul. - Aug.016), PP 53-60 www.iosrournals.org Asymtotic Proerties of the Markov Chain Model method of

More information

CONVOLVED SUBSAMPLING ESTIMATION WITH APPLICATIONS TO BLOCK BOOTSTRAP

CONVOLVED SUBSAMPLING ESTIMATION WITH APPLICATIONS TO BLOCK BOOTSTRAP Submitted to the Annals of Statistics arxiv: arxiv:1706.07237 CONVOLVED SUBSAMPLING ESTIMATION WITH APPLICATIONS TO BLOCK BOOTSTRAP By Johannes Tewes, Dimitris N. Politis and Daniel J. Nordman Ruhr-Universität

More information

Convex Optimization methods for Computing Channel Capacity

Convex Optimization methods for Computing Channel Capacity Convex Otimization methods for Comuting Channel Caacity Abhishek Sinha Laboratory for Information and Decision Systems (LIDS), MIT sinhaa@mit.edu May 15, 2014 We consider a classical comutational roblem

More information

arxiv: v3 [physics.data-an] 23 May 2011

arxiv: v3 [physics.data-an] 23 May 2011 Date: October, 8 arxiv:.7v [hysics.data-an] May -values for Model Evaluation F. Beaujean, A. Caldwell, D. Kollár, K. Kröninger Max-Planck-Institut für Physik, München, Germany CERN, Geneva, Switzerland

More information

2 K. ENTACHER 2 Generalized Haar function systems In the following we x an arbitrary integer base b 2. For the notations and denitions of generalized

2 K. ENTACHER 2 Generalized Haar function systems In the following we x an arbitrary integer base b 2. For the notations and denitions of generalized BIT 38 :2 (998), 283{292. QUASI-MONTE CARLO METHODS FOR NUMERICAL INTEGRATION OF MULTIVARIATE HAAR SERIES II KARL ENTACHER y Deartment of Mathematics, University of Salzburg, Hellbrunnerstr. 34 A-52 Salzburg,

More information

ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION

ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION STATISTICA, anno LXVIII, n., 008 ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION 1. INTRODUCTION The Inverse Gaussian distribution was first introduced

More information

7.2 Inference for comparing means of two populations where the samples are independent

7.2 Inference for comparing means of two populations where the samples are independent Objectives 7.2 Inference for comaring means of two oulations where the samles are indeendent Two-samle t significance test (we give three examles) Two-samle t confidence interval htt://onlinestatbook.com/2/tests_of_means/difference_means.ht

More information

SAS for Bayesian Mediation Analysis

SAS for Bayesian Mediation Analysis Paer 1569-2014 SAS for Bayesian Mediation Analysis Miočević Milica, Arizona State University; David P. MacKinnon, Arizona State University ABSTRACT Recent statistical mediation analysis research focuses

More information

Preconditioning techniques for Newton s method for the incompressible Navier Stokes equations

Preconditioning techniques for Newton s method for the incompressible Navier Stokes equations Preconditioning techniques for Newton s method for the incomressible Navier Stokes equations H. C. ELMAN 1, D. LOGHIN 2 and A. J. WATHEN 3 1 Deartment of Comuter Science, University of Maryland, College

More information

A MIXED CONTROL CHART ADAPTED TO THE TRUNCATED LIFE TEST BASED ON THE WEIBULL DISTRIBUTION

A MIXED CONTROL CHART ADAPTED TO THE TRUNCATED LIFE TEST BASED ON THE WEIBULL DISTRIBUTION O P E R A T I O N S R E S E A R C H A N D D E C I S I O N S No. 27 DOI:.5277/ord73 Nasrullah KHAN Muhammad ASLAM 2 Kyung-Jun KIM 3 Chi-Hyuck JUN 4 A MIXED CONTROL CHART ADAPTED TO THE TRUNCATED LIFE TEST

More information

Deriving Indicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V.

Deriving Indicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V. Deriving ndicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V. Deutsch Centre for Comutational Geostatistics Deartment of Civil &

More information

Applied Mathematics and Computation

Applied Mathematics and Computation Alied Mathematics and Comutation 217 (2010) 1887 1895 Contents lists available at ScienceDirect Alied Mathematics and Comutation journal homeage: www.elsevier.com/locate/amc Derivative free two-oint methods

More information

Estimation of Separable Representations in Psychophysical Experiments

Estimation of Separable Representations in Psychophysical Experiments Estimation of Searable Reresentations in Psychohysical Exeriments Michele Bernasconi (mbernasconi@eco.uninsubria.it) Christine Choirat (cchoirat@eco.uninsubria.it) Raffaello Seri (rseri@eco.uninsubria.it)

More information

Bayesian Model Averaging Kriging Jize Zhang and Alexandros Taflanidis

Bayesian Model Averaging Kriging Jize Zhang and Alexandros Taflanidis HIPAD LAB: HIGH PERFORMANCE SYSTEMS LABORATORY DEPARTMENT OF CIVIL AND ENVIRONMENTAL ENGINEERING AND EARTH SCIENCES Bayesian Model Averaging Kriging Jize Zhang and Alexandros Taflanidis Why use metamodeling

More information

Uniform Law on the Unit Sphere of a Banach Space

Uniform Law on the Unit Sphere of a Banach Space Uniform Law on the Unit Shere of a Banach Sace by Bernard Beauzamy Société de Calcul Mathématique SA Faubourg Saint Honoré 75008 Paris France Setember 008 Abstract We investigate the construction of a

More information

On Wrapping of Exponentiated Inverted Weibull Distribution

On Wrapping of Exponentiated Inverted Weibull Distribution IJIRST International Journal for Innovative Research in Science & Technology Volume 3 Issue 11 Aril 217 ISSN (online): 2349-61 On Wraing of Exonentiated Inverted Weibull Distribution P.Srinivasa Subrahmanyam

More information

Sign Tests for Weak Principal Directions

Sign Tests for Weak Principal Directions Sign Tests for Weak Princial Directions DAVY PAINDAVEINE, JULIEN REMY and THOMAS VERDEBOUT arxiv:1812.09367v1 [math.st] 21 Dec 2018 Université libre de Bruxelles ECARES and Déartement de Mathématique Avenue

More information

RANDOM MATRIX-IMPROVED KERNELS FOR LARGE DIMENSIONAL SPECTRAL CLUSTERING. Hafiz Tiomoko Ali, Abla Kammoun, Romain Couillet

RANDOM MATRIX-IMPROVED KERNELS FOR LARGE DIMENSIONAL SPECTRAL CLUSTERING. Hafiz Tiomoko Ali, Abla Kammoun, Romain Couillet RANDOM MATRIX-IMPROVED KERNELS FOR LARGE DIMENSIONAL SPECTRAL CLUSTERING Hafiz Tiomoko Ali, Abla Kammoun, Romain Couillet CentraleSuélec, Université Paris-Saclay, Gif-sur-Yvette, France King Abdullah University

More information

Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Application

Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Application BULGARIA ACADEMY OF SCIECES CYBEREICS AD IFORMAIO ECHOLOGIES Volume 9 o 3 Sofia 009 Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Alication Svetoslav Savov Institute of Information

More information

Statics and dynamics: some elementary concepts

Statics and dynamics: some elementary concepts 1 Statics and dynamics: some elementary concets Dynamics is the study of the movement through time of variables such as heartbeat, temerature, secies oulation, voltage, roduction, emloyment, rices and

More information

Positivity, local smoothing and Harnack inequalities for very fast diffusion equations

Positivity, local smoothing and Harnack inequalities for very fast diffusion equations Positivity, local smoothing and Harnack inequalities for very fast diffusion equations Dedicated to Luis Caffarelli for his ucoming 60 th birthday Matteo Bonforte a, b and Juan Luis Vázquez a, c Abstract

More information

Notes on Instrumental Variables Methods

Notes on Instrumental Variables Methods Notes on Instrumental Variables Methods Michele Pellizzari IGIER-Bocconi, IZA and frdb 1 The Instrumental Variable Estimator Instrumental variable estimation is the classical solution to the roblem of

More information