Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix

Similar documents
Measures of uncertainty in market network analysis

Testing Equality of Natural Parameters for Generalized Riesz Distributions

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Testing Statistical Hypotheses

TEST FOR INDEPENDENCE OF THE VARIABLES WITH MISSING ELEMENTS IN ONE AND THE SAME COLUMN OF THE EMPIRICAL CORRELATION MATRIX.

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques

Completeness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all.

A Statistical Analysis of Fukunaga Koontz Transform

Testing Statistical Hypotheses

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

ONE DIMENSIONAL MARGINALS OF OPERATOR STABLE LAWS AND THEIR DOMAINS OF ATTRACTION

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Finite Sample Analysis of Weighted Realized Covariance with Noisy Asynchronous Observations

Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

arxiv: v2 [q-fin.st] 24 Aug 2018

On Reduction and Q-conditional (Nonclassical) Symmetry

Some History of Optimality

An Introduction to Multivariate Statistical Analysis

Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data

Choosing the best set of variables in regression analysis using integer programming

ALEXEY V. GAVRILOV. I p (v) : T p M T Expp vm the operator of parallel transport along the geodesic γ v. Let E p be the map

On the GLR and UMP tests in the family with support dependent on the parameter

SOME TECHNIQUES FOR SIMPLE CLASSIFICATION

Inference on reliability in two-parameter exponential stress strength model

A Note on Some Wishart Expectations

A method for construction of Lie group invariants

Some New Properties of Wishart Distribution

GENERAL PROBLEMS OF METROLOGY AND MEASUREMENT TECHNIQUE

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Advanced Econometrics I

On the Trivariate Non-Central Chi-Squared Distribution

Statistical Signal Processing Detection, Estimation, and Time Series Analysis

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Stochastic Design Criteria in Linear Models

On the Midpoint Method for Solving Generalized Equations

Effective Sample Size in Spatial Modeling

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Statistical models and likelihood functions

Probabilities & Statistics Revision

Some General Types of Tests

Lecture 19. The Kadison-Singer problem

Lecture 12 November 3

Mean. Pranab K. Mitra and Bimal K. Sinha. Department of Mathematics and Statistics, University Of Maryland, Baltimore County

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

Information Theory and Hypothesis Testing

CS 6375 Machine Learning

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

CS Lecture 19. Exponential Families & Expectation Propagation

Mean-field equations for higher-order quantum statistical models : an information geometric approach

Invariant HPD credible sets and MAP estimators

Lecture 11. Multivariate Normal theory

STAT Financial Time Series

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Robustness of the Quadratic Discriminant Function to correlated and uncorrelated normal training samples

Multivariate Distributions

Calculation of Bayes Premium for Conditional Elliptical Risks

MATH5745 Multivariate Methods Lecture 07

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

The regression model with one fixed regressor cont d

An Algebraic and Geometric Perspective on Exponential Families

Chapter 4. Theory of Tests. 4.1 Introduction

Volatility. Gerald P. Dwyer. February Clemson University

Application of Parametric Homogeneity of Variances Tests under Violation of Classical Assumption

STAT 4385 Topic 01: Introduction & Review

10708 Graphical Models: Homework 2

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Admissible Significance Tests in Simultaneous Equation Models

Application of Homogeneity Tests: Problems and Solution

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics


POSITIVE DEFINITE FUNCTIONS AND MULTIDIMENSIONAL VERSIONS OF RANDOM VARIABLES

Multiple orthogonal polynomials. Bessel weights

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

The Wiener Itô Chaos Expansion

Mi-Hwa Ko. t=1 Z t is true. j=0

c 2005 Society for Industrial and Applied Mathematics

Statistical Inference On the High-dimensional Gaussian Covarianc

An introduction to Bayesian inference and model comparison J. Daunizeau

Support Vector Method for Multivariate Density Estimation

5.1 Uniformly Most Accurate Families of Confidence

Combinatorial Dimension in Fractional Cartesian Products

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

topics about f-divergence

FDR-CONTROLLING STEPWISE PROCEDURES AND THEIR FALSE NEGATIVES RATES

ARANDOM-MATRIX-THEORY-BASEDANALYSISOF STOCKS OF MARKETS FROM DIFFERENT COUNTRIES

Multivariate Normal-Laplace Distribution and Processes

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

Gaussian representation of a class of Riesz probability distributions

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood

Theory and Methods of Statistical Inference. PART I Frequentist likelihood methods

Ch. 5 Hypothesis Testing

Hypothesis testing (cont d)

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity

Using Hankel structured low-rank approximation for sparse signal recovery

Stein s Method and Characteristic Functions

Statistical Inference

Transcription:

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix Alexander P. Koldanov and Petr A. Koldanov Abstract A multiple decision statistical problem for the elements of inverse covariance matrix is considered. Associated optimal unbiased multiple decision statistical procedure is given. This procedure is constructed using the Lehmann theory of multiple decision statistical procedures and the conditional tests of the Neyman structure. The equations for thresholds calculation for the tests of the Neyman structure are analyzed. Keywords Inverse covariance matrix Tests of the Neyman structure Multiple decision statistical procedure Generating hypothesis 1 Introduction A market network is constructed by means of some similarity measure between every pairs of stocks. The most popular measure of similarity between stocks of a market is the correlation between them [1 4]. The analysis of methods of market graph construction [2] from the statistical point of view was started in [5]. In [5] multiple decision statistical procedure for market graph construction based on the Pearson test is suggested. The authors of [5] note that a procedure of this type can be made optimal in the class of unbiased multiple decision statistical procedures if one uses the tests of the Neyman structure for generating hypothesis. In the present paper we use the partial correlations as a measure of similarity between stocks. In this case the elements of inverse covariance matrix are the weights of links between stocks in the market network [6]. A.P. Koldanov P.A. Koldanov ( ) National Research University, Higher School of Economics, Bolshaya Pecherskaya 25/12, Nizhny Novgorod, 63155, Russia e-mail: akoldanov@hse.ru; pkoldanov@hse.ru V.F. Demyanov et al. (eds.), Constructive Nonsmooth Analysis and Related Topics, Springer Optimization and Its Applications 87, DOI 1.17/978-1-4614-8615-2 13, Springer Science+Business Media New York 214 25

26 A.P. Koldanov and P.A. Koldanov The main goal of the paper is investigation of the problem of identification of inverse covariance matrix as a multiple decision statistical problem. As a result an optimal unbiased multiple decision statistical procedure for identification of inverse covariance matrix is given. This procedure is constructed using the Lehmann theory of multiple decision statistical procedures and the conditional tests of the Neyman structure. In addition the equations for thresholds calculation for the tests of the Neyman structure are analyzed. The paper is organized as follows. In Sect. 2 we briefly recall the Lehmann theory of multiple decision statistical procedures. In Sect. 3 we describe the tests of the Neyman structure. In Sect. 4 we formulate the multiple decision problem for identification of inverse covariance matrix. In Sect. 5 we construct and study the optimal tests for testing of generating hypothesis for the elements of an inverse covariance matrix. In Sect. 6 we construct the multiple statistical procedure and consider some particular cases. In Sect. 7 we summarize the main results of the paper. 2 Lehmann Multiple Decision Theory In this section we recall for the sake of completeness the basic idea of the Lehmann theory following the paper [5]. Suppose that the distribution of a random vector R is taken from the family f (r, θ) : θ Ω, where θ is a parameter, Ω is the parametric space, r is an observation of R. We need to construct a statistical procedure for the selection of one from the set of L hypotheses, which in the general case can be stated as: H i : θ Ω i,i = 1,...,L, Ω i Ω = /, i, L i=1 Ω i = Ω. (1) The most general theory of a multiple decision procedure is the Lehmann theory [7]. The Lehmann theory is based on three concepts: generating hypothesis, generating hypothesis testing and compatibility conditions, and additivity condition for the loss function. The multiple decision problem (1) of selecting one from the set of L hypothesesh i : θ Ω i ;i = 1,...,L is equivalent to a family of M two decision problems: with H : θ ω vs K : θ ω 1, = 1,...,M, (2) L i=1 Ω i = ω ω 1 = Ω.

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix 27 This equivalence is given by the relations: where Ω i = M =1 ω χ i,, ω = {i:χ i, =1} Ω i, { 1, Ωi ω χ i, = /, 1, Ω i ω = /. (3) Hypotheses H : ( = 1,...,M) are called generating hypotheses for the problem (1). The equivalence between problem (1) and the family of problems (2) reduces the multiple decision problem to the testing of generating hypothesis. Any statistical procedure δ for hypothesis testing of H can be written in the following form {, r X, δ (r)=, r X 1 1, where is the decision of acceptance of H 1 and of K, X is the acceptance region of H and X 1 (4) is the decision of acceptance is the acceptance region of K (reection region of H ) in the sample space. One has X X 1 = /, X X 1 = X, X being the sample space. Define the acceptance region for H i by D i = M =1 X χ i,, (5) where χ i, are defined by (3), and put X 1 = X. Note that L i=1 D i X, but it is possible that L i=1 D i X. Therefore, if D 1,D 2,...,D L is a partition of the sample space X, then one can define the statistical procedure δ(r) by d 1, r D 1, d δ(r)= 2, r D 2,......... d L, r D L. According to [8,9] we define the conditional risk for multiple decision statistical procedure by risk(θ,δ)=e θ w(θ,δ(r)) = L k=1 (6) w(θ,d k )P θ (δ(r)=d k ), (7)

28 A.P. Koldanov and P.A. Koldanov where E θ is the expectation according to the density function f (x,θ) and w(θ,d k ) is the loss from decision d k under the condition that θ is true, θ Ω. Under the additivity condition of the loss function (see [5] for more details) the conditional risk can be written as risk(θ,δ)= M =1 risk(θ,δ ), θ Ω. (8) We call statistical procedure optimal in a class of statistical procedures if it has minimal conditional risk for all θ Ω in this class. The main result of the Lehmann theory (see [7]) states that if Eq. (8) is satisfied and statistical procedures (4) are optimal in the class of unbiased statistical tests, then the associated multiple decision statistical procedure (6) is optimal in the class of unbiased multiple decision procedures. 3 Unbiasedness and Tests of the Neyman Structure The class of unbiased multiple decision statistical procedures according to Lehmann [9, 1] is defined by: E θ w(θ,δ(r)) E θ w(θ,δ(r)) f or any θ,θ Ω. Let f (r;θ) be the density of the exponential family: f (r;θ)=c(θ)exp( M =1 θ T (r))h(r), (9) where c(θ) is a function defined in the parameters space, h(r), T (r) are functions defined in the sample space, and T (R) are the sufficient statistics for θ, = 1,...,M. Suppose that generating hypotheses (2) has the form: H : θ = θ vs K : θ θ, = 1,2,...,M, (1) where θ are fixed. For a fixed the parameter θ is called information or structural parameter and θ k, k are called nuisance parameters. According to [9] the optimal unbiased tests for generating hypotheses (1) are: δ = {, c 1 (t 1,...,t 1,t +1,...,t M ) < t < c 2 (t 1,...,t 1,t +1,...,t M ), 1, otherwise, (11)

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix 29 where t i = T i (r),i = 1,...,M and constants c 1 (t 1,...,t 1,t +1,...,t M ), c 2 (t 1,..., t 1,t +1,...,t M ) are defined from the equations c 2 c 1 f (t,θ T i = t i,i = 1,...,M;i )dt = 1 α, (12) and c 1 + + c 2 t f (t,θ T i = t i,i = 1,...,M;i )dt t f (t,θ T i = t i,i = 1,...,M;i )dt + = α t f (t,θ T i = t i,i = 1,...,M;i )dt, (13) where f (t,θ T i = t i,i = 1,...,M;i ) is the density of conditional distribution of statistics T and α is the level of significance of the test. A test satisfying (12) is said to have Neyman structure. This test is characterized by the fact that the conditional probability of reection of H (under the assumption that H is true) is equal to α on each of the surfaces k (x) =t k ). Therefore k (T the multiple decision statistical procedure associated with the tests of the Neyman structure (11), (12), and (13) is optimal in the class of unbiased multiple decision procedures. 4 Problem of Identification of Inverse Covariance Matrix In this section we formulate the multiple decision problem for elements of inverse covariance matrix. Let N be the number of stocks on a financial market, and let n be the number of observations. Denote by r i (t) the daily return of the stock i for the day t (i = 1,...,N;t = 1,...,n). We suppose r i (t) to be an observation of the random variable R i (t). We use the standard assumptions: the random variables R i (t), t = 1,...,n are independent and have all the same distribution as a random variable R i (i = 1,...,N). The random vector (R 1,R 2,...,R N ) describes the oint behavior of the stocks. We assume that the vector (R 1,R 2,...,R N ) has a multivariate normal distribution with covariance matrix σ i where σ i = cov(r i,r )=E(R i E(R i ))(R E(R )),ρ i =(σ i )/( σ ii σ ),i, = 1,...,N, E(R i ) is the expectation of the random variable R i. We define a sample space as R N n with the elements (r i (t)). Statistical estimation of σ i is s i = Σt=1 n (r i(t) r i )(r (t) r ) where r i =(1/n) t=1 n r i(t).the sample correlation between the stocks i and is defined by r i =(s i )/( s ii s ).

21 A.P. Koldanov and P.A. Koldanov It is known [11] that for a multivariate normal vector the statistics (r 1,r 2,...,r N ) and s i (matrix of sample covariances) are sufficient. Let σ i be the elements of inverse covariance matrix σ i. Then the problem of identification of inverse covariance matrix can be formulated as a multiple decision problem of the selection of one from the set of hypotheses: H 1 : σ i = σ i, i, = 1,...,N, i <, H 2 : σ 12 σ 12,σi = σ i,i, = 1,...,N,(i, ) (1,2), i <, H 3 : σ 13 σ 13,σi = σ i,i, = 1,...,N,(i, ) (1,3), i <, H 4 : σ 12 σ 12,σ13 σ 13,σi = σ i,i, = 1,...,N,(i, ) (1,2), (i, ) (1,3),... H L : σ i σ i,i, = 1,...,N, i <, (14) where L = 2 M with M = N(N 1)/2. Multiple decision problem (14) is a particular case of the problem (1). The parameter space Ω is the space of positive semi-definite matrices σ i, Ω k is a domain in the parameters space associated with the hypothesis H k from the set (14) k = 1,...,L. For the multiple decision problem (14) we introduce the following set of generating hypotheses: h i, : σ i = σ i vs k i, : σ i σ i, i, = 1,2,...,N, i <, (15) We use the following notations: i, is the decision of acceptance of the hypothesis h i, and 1 i, is the decision of reection of h i,. 5 Tests of the Neyman Structure for Testing of Generating Hypothesis Now we construct the optimal test in the class of unbiased tests for generating hypothesis (15). To construct these tests we use the sufficient statistics s i with the following Wishart density function [11]: f ({s k,l })= [det(σ kl )] n/2 [det(s kl )] (n N 2)/2 exp[ (1/2) k l s k,l σ kl ] 2 (Nn/2) π N(N 1)/4 Γ (n/2)γ ((n 1)/2) Γ ((n N + 1)/2) if the matrix (s kl ) is positive definite, and f ({s kl })= otherwise. One has for a fixed i < : f ({s kl })=C({σ kl }) exp[ σ i s i 1 2 (k,l) (i, );(k,l) (,i) s kl σ kl ] h({s kl }),

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix 211 where C({σ kl })= 1 q det(σ kl )] n/2, q = 2 (Nn/2) π N(N 1)/4 Γ (n/2)γ ((n 1)/2) Γ ((n N + 1)/2), h({s kl })=[det(s kl )] (n N 2)/2. Therefore, this distribution belongs to the class of exponential distributions with parameters σ kl. The optimal tests of the Neyman structure (11), (12), and (13) for generating hypothesis (15) take the form: δ i, ({s kl })= { i,, c 1 i, ({s kl}) < s i < c 2 i, ({s kl}), (k,l) (i, ), 1 i,, s i c 1 i, ({s kl}) or s i c 2 i, ({s kl}), (k,l) (i, ), (16) where the critical values are defined from the equations I [c 1i, ;c2i, I i ] exp[ σ s i][det(s kl )] (n N 2)/2 ds i i exp[ σ s = 1 α i,, (17) i][det(s kl )] (n N 2)/2 ds i s i exp[ σ i I [ ;c 1 i, ] s i][det(s kl )] (n N 2)/2 ds i + s i exp[ σ i I [c 2 i, ;+ ] s i][det(s kl )] (n N 2)/2 ds i = α i, s i exp[ σ i s i][det(s kl )] (n N 2)/2 ds i, I where I is the interval of values of s i such that the matrix (s kl ) is positive definite and α i is the level of significance of the tests. Consider Eqs. (17) and (18). Note that det(s kl ) is a quadratic polynomial of s i. Let det(s k,l )= C 1 s 2 i, +C 2s i, +C 3 = C 1 ( s 2 i, +As i, +B), C 1 > then the positive definiteness of the matrix (s k,l ) for a fixed s k,l, (k,l) (i, ) gives the following interval for the value of s i, : (18) I = {x : A A 2 2 4 + B < x < A A 2 + 2 + B}. (19) 4 Now we define the functions x Ψ 1 (x)= ( t 2 + At + B) K exp( σ i t)dt, (2) Ψ 2 (x)= x t( t 2 + At + B) K exp( σ i t)dt, (21)

212 A.P. Koldanov and P.A. Koldanov where K =(n N 2)/2. One can calculate the critical values of c 1 i,,c2 i, from the equations: Ψ 1 (c 2 i) Ψ 1 (c 1 i)=(1 α i )(Ψ 1 ( A A 2 + 2 4 + B) Ψ 1( A A 2 2 + B)), (22) 4 Ψ 2 (c 2 i) Ψ 2 (c 1 i)=(1 α i )(Ψ 2 ( A A 2 + 2 4 + B) Ψ 2( A A 2 2 + B)). (23) 4 The test (16) can be written in terms of sample correlations. First note that det(s kl )=det(r kl )s 11 s 22...s NN where r kl are the sample correlations. One has J [e 1i, ;e2i, i ] exp[ σ r i s ii s ][det(r kl )] (n N 2)/2 dr i i J exp[ σ r = 1 α i,, (24) i s ii s ][det(r kl )] (n N 2)/2 dr i r i exp[ σ i J [ ;e 1 i, ] r i sii s ][det(r kl )] (n N 2)/2 dr i + r i exp[ σ i J [e 2 i, ;+ ] r i sii s ][det(r kl )] (n N 2)/2 dr i = α i, r i exp[ σ i r i sii s ][det(r kl )] (n N 2)/2 dr i, J (25) where I = J s ii s, c k i, = ek i, sii s ;k = 1,2. Therefore, the tests (16) take the form δ i, ({r kl })= i,, e 1 i, (s ii,s,{r kl }) < r i < e 2 i, (s ii,s,{r kl }), (k,l) (i, ), 1 i,, r i e 1 i, (s ii,s {r kl }) or r i e 2 i, (s ii,s {r kl }), (k,l) (i, ). It means that the tests of the Neyman structure for generating hypothesis (15) do not depend on s k,k, k i, k. In particular for N = 3, (i, )=(1,2) one has { 1,2, e 1 12 δ 1,2 ({r k,l })= (s 11,s 22,r 13,r 23 ) < r 12 < e 2 12 (s 11,s 22,r 13,r 23 ), 1,2 1, r 12 e 1 12 (s 11,s 22,r 13,r 23 ) or r 12 e 2 12 (s 11,s 22,r 13,r 23 ). (27) To emphasize the peculiarity of the constructed test we consider some interesting particular cases. n N 2 =, σ i. In this case expressions (2) and (21) can be simplified. Indeed one has in this case x Ψ 1 (x)= exp( σ i t)dt = 1 σ i (1 exp( σ i x)), (26)

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix 213 x Ψ 2 (x)= t exp( σ i t)dt = x σ i exp( σ i x) 1 (σ i )2 exp( σ i x)+ 1 (σ i )2. Finally one has the system of two equations for defining constants c 1 i,c2 i : exp( σ i c1 i) exp( σ i c2 i) =(1 α i ){exp( σ i A (A 2 2 i A + B)) exp( σ 4 (A 2 + 2 (28) 4 + B))}, c 1 iexp( σ i c1 i) c 2 iexp( σ i c2 i) =(1 α i ){( A 2 A 2 ( A 2 + A 2 4 4 + B)exp( σ i (A 2 A 2 4 + B)) + B)exp( σ i (A 2 + A 2 4 + B))}. (29) σ =. In this case the critical values are defined by the system of algebraic equations (22) and (23) where the functions Ψ 1, Ψ 2 are defined by x Ψ 1 (x)= ( t 2 + At + B) K dt, Ψ 2 (x)= x t( t 2 + At + B) K dt, In this case the tests of the Neyman structure have the form { i,, e 1 i, δ i, ({r k,l })= ({r kl} < r i < e 2 i, ({r kl}, (k,l) (i, ), i, 1, r i e 1 i, ({r kl} or r i e 1 i, ({r kl}, (k,l) (i, ). (3) n N 2 =, σ =. In this case one has c 1 i, = A 2 (1 α A i) 2 4 + B, (31) c 2 i, = A 2 +(1 α A i) 2 4 + B. 6 Multiple Statistical Procedure Based on the Tests of the Neyman Structure Now it is possible to construct the multiple decision statistical procedure for problem (14) based on the tests of Neyman structure. Then the multiple decision statistical procedure (6) takes the form:

214 A.P. Koldanov and P.A. Koldanov d 1, c 1 i ({s kl}) < r i < c 2 i ({s kl}), i <,i,,k,l = 1,...,N,(k,l) (i, ), d 2, r 12 c 1 12 ({s kl}) or r 12 c 2 12 ({s kl}),c 1 i ({s kl}) < r i < c 2 i ({s kl}), δ = i <,i,,k,l = 1,...,N,(k,l) (i, ),...... d L, r i c 1 i ({s kl}) or r i c 2 i ({s kl}), i <,i,,k,l = 1,...,N,(k,l) (i, ), (32) where c i ({s kl }) are defined from the Eqs. (17) and (18). One has D k = {r R N n : δ(r)=d k }, k = 1,2,...,L. It is clear that L D k = R N n. k=1 Then D 1,D 2,...,D L is a partition of the sample space R N n. The tests of the Neyman structure for generating hypothesis (15) are optimal in the class of unbiased tests. Therefore if the condition of the additivity (8) of the loss function is satisfied, then the associated multiple decision statistical procedure is optimal. For discussion of additivity of the loss function see [5]. We illustrate statistical procedure (32) with an example. Let N = 3. In this case problem (14) is the problem of the selection of one from eight hypotheses: H 1 : σ 12 = σ 12, σ 13 = σ 13, σ 23 = σ 23, H 2 : σ 12 σ 12, σ 13 = σ 13, σ 23 = σ 23, H 3 : σ 12 = σ 12, σ 13 σ 13, σ 23 = σ 23, H 4 : σ 12 = σ 12, σ 13 = σ 13, σ 23 σ 23, H 5 : σ 12 σ 12, σ 13 σ 13, σ 23 = σ 23, H 6 : σ 12 = σ 12, σ 13 σ 13, σ 23 σ 23, H 7 : σ 12 σ 12, σ 13 = σ 13, σ 23 σ 23, H 8 : σ 12 σ 12, σ 13 σ 13, σ 23 σ 23. (33) Generating hypotheses are: h 1,2 : σ 12 = σ 12 vs k 1,2 : σ 12 σ 12, σ 13,σ 23 are the nuisance parameters. h 1,3 : σ 13 = σ 13 vs k 1,3 : σ 13 σ 13, σ 12,σ 23 are the nuisance parameters. h 2,3 : σ 23 = σ 23 vs k 2,3 : σ 23 σ 23, σ 12,σ 13 are the nuisance parameters.

Optimal Multiple Decision Statistical Procedure for Inverse Covariance Matrix 215 In this case multiple statistical procedure for problem (33) (ifσ ) is: d 1, c 1 12 < r 12 < c 2 12, c1 13 < r 13 < c 2 13, c1 23 < r 23 < c 2 23, d 2, r 12 c 1 12 or r 12 c 2 12, c1 13 < r 13 < c 2 13, c1 23 < r 23 < c 2 23, d 3, c 1 12 < r 12 < c 2 12, r 13 c 1 13 or r 13 c 2 13, c1 23 < r 23 < c 2 23, d 4, c 1 12 < r 12 < c 2 12, c1 13 < r 13 < c 2 13, r 23 c 1 23 or r 23 c 2 23, δ = d 5, r 12 c 1 12 or r 12 c 2 12, r 13 c 1 13 or r 13 c 2 13, c1 23 < r 23 < c 2 23, d 6, c 1 12 < r 12 < c 2 12, r 13 c 1 13 or r 13 c 2 13, r 23 c 1 23 or r 23 c 2 23, d 7, r 12 c 1 12 or r 12 c 2 12, c1 13 < r 13 < c 2 13, r 23 c 1 23 or r 23 c 2 23, d 8, r 12 c 1 12 or r 12 c 2 12, r 13 c 1 13 or r 13 c 2 13, r 23 c 1 23 or r 23 c 2 23. (34) The critical values c k 12 = ck 12 (r 13,r 23,s 11 s 22 ), c k 13 = ck 13 (r 12,r 23,s 11 s 33 ), c k 23 = c k 23 (r 12,r 13,s 22 s 33 ); k = 1,2 are defined from Eqs. (24) and (25). If n = 5; σ i ; i, = 1,2,3, then the critical values c k i ; k = 1,2 are defined from (28) and (29). If σ i =, i, and n = 5, then tests (3) for generating hypothesis depend on the sample correlation only. Therefore the corresponding multiple statistical procedure with L decisions depends only on the sample correlation too. This procedure is (34) where constants c k 12 = ck 12 (r 13,r 23 ),c k 13 = ck 13 (r 12,r 23 ),c k 23 = ck 23 (r 12,r 13 );k = 1,2. In this case I 1,2 =(r 13 r 23 G 1,2 ;r 13 r 23 + G 1,2 ), I 1,3 =(r 12 r 23 G 1,3 ;r 12 r 23 + G 1,3 ), I 2,3 =(r 12 r 13 G 2,3 ;r 12 r 13 + G 2,3 ), where G 1,2 = (1 r13 2 )(1 r2 23 ), G 1,3 = (1 r12 2 )(1 r2 23 ), G 2,3 = (1 r12 2 )(1 r2 13 ), and the critical values are c 1 12 = r 13 r 23 (1 α 12 )G 1,2, c 2 12 = r 13 r 23 +(1 α 12 )G 1,2, c 1 13 = r 12 r 23 (1 α 13 )G 1,3, c 2 13 = r 12 r 23 +(1 α 13 )G 1,3,

216 A.P. Koldanov and P.A. Koldanov c 1 23 = r 12 r 13 (1 α 23 )G 2,3, c 2 23 = r 12 r 13 +(1 α 23 )G 2,3. Note that in this case test (34) has a very simple form. 7 Concluding Remarks Statistical problem of identification of elements of inverse covariance matrix is investigated as multiple decision problem. Solution of this problem is developed on the base of the Lehmann theory of multiple decision procedures and theory of tests of the Neyman structure. It is shown that this solution is optimal in the class of unbiased multiple decision statistical procedures. Obtained results can be applied to market network analysis with partial correlations as a measure of similarity between stocks returns. Acknowledgements The authors are partly supported by National Research University, Higher School of Economics, Russian Federation Government grant, N. 11.G34.31.57 References [1] Mantegna, R.N.: Hierarchical structure in financial market. Eur. Phys. J. series B 11, 193 197 (1999) [2] Boginsky V., Butenko S., Pardalos P.M.: On structural properties of the market graph. In: Nagurney, A. (ed.) Innovations in Financial and Economic Networks. pp. 29 45. Edward Elgar Publishing Inc., Northampton (23) [3] Boginsky V., Butenko S., Pardalos P.M.: Statistical analysis of financial networks. Comput. Stat. Data. Anal. 48, 431 443 (25) [4] M. Tumminello, T. Aste, T. Di Matteo, R.N. Mantegna, H.A.: Tool for Filtering Information in Complex Systems. Proc. Natl. Acad. Sci. Uni. States Am. 12 3, 1421 1426 (25) [5] Koldanov, A.P., Koldanov, P.A., Kalyagin, V.A., Pardalos, P.M.: Statistical Procedures for the Market Graph Construction. Comput. Stat. Data Anal. 68, 17 29, DOI: 1.116/.csda.213.6.5. (213) [6] Hero, A., Raaratnam, B.: Hub discovery in partial correlation graphical models. IEEE Trans. Inform. Theor. 58 (9), 664 678 (212) [7] Lehmann E.L.: A theory of some multiple decision procedures 1. Ann. Math. Stat. 28, 1 25 (1957) [8] Wald, A.: Statistical Decision Function. Wiley, New York (195) [9] Lehmann E.L., Romano J.P.: Testing Statistical Hypothesis. Springer, New York (25) [1] Lehmann E.L.: A general concept of unbiasedness. Ann. Math. Stat. 22, 587 597 (1951) [11] Anderson T.W.: An Introduction to Multivariate Statistical Analysis. 3 edn. Wiley- Interscience, New York (23)