ACTA UNIVERSITATIS LODZIENSIS. Andrzej Grzybowski*

Similar documents
A L A BA M A L A W R E V IE W

A C T A U N I V E R S I T A T I S L O D Z I E N S I S. Czeslaw D om ański*, Dariusz Parys** THE CLOSED BOOTSTRAP MULTIPLE TEST PROCEDURE

Quantum Minimax Theorem

Chapter 5 Matrix Approach to Simple Linear Regression

OH BOY! Story. N a r r a t iv e a n d o bj e c t s th ea t e r Fo r a l l a g e s, fr o m th e a ge of 9

ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS. Abstract. We introduce a wide class of asymmetric loss functions and show how to obtain

. a m1 a mn. a 1 a 2 a = a n

114 A^VÇÚO 1n ò where y is an n 1 random vector of observations, X is a known n p matrix of full column rank, ε is an n 1 unobservable random vector,

NORMS ON SPACE OF MATRICES

c. What is the average rate of change of f on the interval [, ]? Answer: d. What is a local minimum value of f? Answer: 5 e. On what interval(s) is f

Type Dimensions LIPI Type Sign FU0024D0

ACTA UNIVERSITATIS LODZIENSIS. Grzegorz Kończak*

304 A^VÇÚO 1n ò while the commonly employed loss function for the precision of estimation is squared error loss function ( β β) ( β β) (1.3) or weight

Table of C on t en t s Global Campus 21 in N umbe r s R e g ional Capac it y D e v e lopme nt in E-L e ar ning Structure a n d C o m p o n en ts R ea

USI TYPE PACKINGS FOR BOTH PISTON AND ROD SEALS IRON RUBBER (AU) Material. Please designate NOK Part number and type & size on your order.

Stochastic Design Criteria in Linear Models

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

On a theorem of Weitzenbiick in invariant theory

Summer School: Semidefinite Optimization

Software Process Models there are many process model s in th e li t e ra t u re, s om e a r e prescriptions and some are descriptions you need to mode

Knowledge Discovery and Data Mining 1 (VO) ( )

EKOLOGIE EN SYSTEMATIEK. T h is p a p e r n o t to be c i t e d w ith o u t p r i o r r e f e r e n c e to th e a u th o r. PRIMARY PRODUCTIVITY.

On the Efficiencies of Several Generalized Least Squares Estimators in a Seemingly Unrelated Regression Model and a Heteroscedastic Model

1 Last time: least-squares problems

NUMERICAL SIMULATION OF MHD-PROBLEMS ON THE BASIS OF VARIATIONAL APPROACH

I M P O R T A N T S A F E T Y I N S T R U C T I O N S W h e n u s i n g t h i s e l e c t r o n i c d e v i c e, b a s i c p r e c a u t i o n s s h o

A NOTE ON A SINGLE VEHICLE AND ONE DESTINATION ROUTING PROBLEM AND ITS GAME-THEORETIC MODELS

Covariance function estimation in Gaussian process regression

S ca le M o d e l o f th e S o la r Sy ste m

A characterization of consistency of model weights given partial information in normal linear models

Optimal Control of Linear Systems with Stochastic Parameters for Variance Suppression: The Finite Time Case

ORTHOGONALLY INVARIANT ESTIMATION OF THE SKEW-SYMMETRIC NORMAL MEAN MATRIX SATOSHI KURIKI

Non-absolutely monotonic functions which preserve non-negative definiteness

BOUNDS OF MODULUS OF EIGENVALUES BASED ON STEIN EQUATION

Bayesian Estimation of Regression Coefficients Under Extended Balanced Loss Function

T h e C S E T I P r o j e c t

On the convergence of solutions of the non-linear differential equation

Real Gas Equation of State for Methane

Course Summary Math 211

ELEMENTARY LINEAR ALGEBRA

2. Matrix Algebra and Random Vectors

ADMISSIBILITY OF UNBIASED TESTS FOR A COMPOSITE HYPOTHESIS WITH A RESTRICTED ALTERNATIVE

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Alles Taylor & Duke, LLC Bob Wright, PE RECORD DRAWINGS. CPOW Mini-Ed Conf er ence Mar ch 27, 2015

ITERATION OF ANALYTIC NORMAL FUNCTIONS OF MATRICES

THE NUMBER OF LOCALLY RESTRICTED DIRECTED GRAPHS1

Hausdorff operators in H p spaces, 0 < p < 1

Introd uction to Num erical Analysis for Eng ineers

C o r p o r a t e l i f e i n A n c i e n t I n d i a e x p r e s s e d i t s e l f

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

9.520: Class 20. Bayesian Interpretations. Tomaso Poggio and Sayan Mukherjee

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij

Form and content. Iowa Research Online. University of Iowa. Ann A Rahim Khan University of Iowa. Theses and Dissertations

I N A C O M P L E X W O R L D

P a g e 5 1 of R e p o r t P B 4 / 0 9

A matrix over a field F is a rectangular array of elements from F. The symbol

Sociedad de Estadística e Investigación Operativa

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Product distance matrix of a tree with matrix weights

Two remarks on normality preserving Borel automorphisms of R n

Distances, volumes, and integration

Density estimators for the convolution of discrete and continuous random variables

H STO RY OF TH E SA NT

SYMMETRICOMPLETIONS AND PRODUCTS OF SYMMETRIC MATRICES

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Statistical Machine Learning for Structured and High Dimensional Data

A Tropical Extremal Problem with Nonlinear Objective Function and Linear Inequality Constraints

Regression. Oscar García

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Unbiased prediction in linear regression models with equi-correlated responses

COM Optimization for Communications 8. Semidefinite Programming

176 5 t h Fl oo r. 337 P o ly me r Ma te ri al s

CATAVASII LA NAȘTEREA DOMNULUI DUMNEZEU ȘI MÂNTUITORULUI NOSTRU, IISUS HRISTOS. CÂNTAREA I-A. Ήχος Πα. to os se e e na aș te e e slă ă ă vi i i i i

ESTIMATORS FOR GAUSSIAN MODELS HAVING A BLOCK-WISE STRUCTURE

The Ability C ongress held at the Shoreham Hotel Decem ber 29 to 31, was a reco rd breaker for winter C ongresses.

(2009) Journal of Rem ote Sensing (, 2006) 2. 1 (, 1999), : ( : 2007CB714402) ;

Statistics & Decisions 7, (1989) R. Oldenbourg Verlag, München /89 $

MIVQUE and Maximum Likelihood Estimation for Multivariate Linear Models with Incomplete Observations

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Basic Concepts in Matrix Algebra

Uniqueness of the Solutions of Some Completion Problems

С-4. Simulation of Smoke Particles Coagulation in the Exhaust System of Piston Engine

Beechwood Music Department Staff

A Note on Nonconvex Minimax Theorem with Separable Homogeneous Polynomials

Efficient robust optimization for robust control with constraints Paul Goulart, Eric Kerrigan and Danny Ralph

What are S M U s? SMU = Software Maintenance Upgrade Software patch del iv ery u nit wh ich once ins tal l ed and activ ated prov ides a point-fix for

Factorization of Seperable and Patterned Covariance Matrices for Gibbs Sampling

Designs for weighted least squares regression, with estimated weights

A Robust von Neumann Minimax Theorem for Zero-Sum Games under Bounded Payoff Uncertainty

SUPPLEMENTARY MATERIAL COBRA: A Combined Regression Strategy by G. Biau, A. Fischer, B. Guedj and J. D. Malley

QUASI-ORTHOGONAL ARRAYS AND OPTIMAL FRACTIONAL FACTORIAL PLANS

Measures and Jacobians of Singular Random Matrices. José A. Díaz-Garcia. Comunicación de CIMAT No. I-07-12/ (PE/CIMAT)

Lecture notes on statistical decision theory Econ 2110, fall 2013

FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES. Danlei Chu, Tongwen Chen, Horacio J. Marquez

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics

LSU Historical Dissertations and Theses

Computational and Statistical Aspects of Statistical Machine Learning. John Lafferty Department of Statistics Retreat Gleacher Center

ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters

EXISTENCE OF PURE EQUILIBRIA IN GAMES WITH NONATOMIC SPACE OF PLAYERS. Agnieszka Wiszniewska Matyszkiel. 1. Introduction

Transcription:

ACTA UNIVERSITATIS LODZIENSIS FOLIA OECONOMICA 12, 2000 Andrzej Grzybowski* O N UNCERTAIN TY CLA SSES AND M IN IM AX E STIM A T IO N IN T H E L IN EA R R E G R E SSIO N M O D E L S W IT H H E T E R O S C E D A S T IC IT Y AND CO R RELA TED E R R O R S1 Abstract. The problem of minimax estimation in the linear regression model is considered under the assumption that a prior information about the regression parameter and the covariance matrix of random component (error) is available for the decision-maker. Two models of the uncertainty of the prior knowledge (so called uncertainly classes) are proposed. The first one may represent the problem of estimation for heteroscedastic model, the other may reflect the uncertainty connected with the presence of the correlation among errors. Minimax estimators for considered classes are obtained. Some numerical examples are discussed as well. L INTRODUCTION Let us consider the ordinary linear regression m odel Y = X/f + Z (1) where Y is an n-dimensional vector of observations o f the dependent variable, X is a given nonstochastic (n x к) m atrix with the rank k, ß is a /с-dimensional vector of unknow n regression coefficients, Z is an n- -dimensional vector of random errors (random com ponents o f the m odel). We assum e E(Y) = X ß and cov (Y) = E. Various papers deal with the problem of the regression estim ation in the presence of prior knowledge about the param eter ß. Some o f them study the problem s where the prior inform ation is of the form o f a restricted * Technical University of Częstochowa, Chair of Applied Mathematics. 1 Support by KBN. Grant No. 1 H02B 013 1 is gratefully acknowledged.

param eter space (see e.g.: D r y gas, 1996; Drygas, Pilz, 1996; G i r к о, 1996; Hoffman, 1996; Pilz, 1996). O ther papers arc devoted to the problems where the prior inform ation is expressed in terms of the probability distribution of the param eter ß. In such a case the distribution o f the param eter is often assumed to belong to a given class of distributions, (see: Berger, 1982; Berger, Chen, 1987; Berger, 1990; Grzybowski, 1997; Verdu, Poor, 1984). This class models the prior knowledge as well as its uncertainty, so we call it an uncertainty class, see Verdu, Poor (1984). Sometimes this approach leads to game theoretic form ulation of the original decision problem (see: Verdu, Poor, 1984; Grzybowski, 1997). In this paper we adopt the latter approach. We introduce two uncertainty classes reflecting the uncertainty in two com m on situations. Section 3 is concerned with the regression estim ation in heteroscedastic m odels. Section 4 deals with the problems associated with the presence of correlation between errors. In each case we solve the game connected with the introduced uncertainty class, i.e. we find m inim ax estim ators and the least favourable states of nature. Some numerical examples are also presented to illustrate im po rtan t features o f the obtained solutions. 2. PRELIMINARY DEFINITIONS AND NOTATION Let L (.,.) be a quadratic loss function, i.e. L(ß, a) = (ß a)th(ß a), for a given nonnegative definite (k x k) m atrix II. F o r a given estim ator d the risk function R(.,.) is defined by R (ß, d) = EpL(ß, d(y)). Let the param eter ß have a distribution n. The Bayes risk r(.,.) connected with the estim ator d is defined as usual: r(n, d) = EnR(Jl, d). Let D denote the set o f all allowable for the decision m aker estim ators. We assum e that all o f them have the Bayes risk finite. The estim ator d* e l ) which satisfies the condition r(n, d#) = inf r(n, d) ded is called a Bayes estimator (with respect to n). Let us consider a game (Г, D, r ), where Г is a given class o f distributions o f the param eter ß. Any distribution n* e I \ satisfying the condition inf г(я*, d) = sup inf r (it, d) del) лег ded is called the least favourable distribution (state of nature). T he minimax estimator is defined as the estim ator d * e l) which satisfies the following condition:

sup r(n, (1*) = inf sup r(n, d). ITe r del) пег Sometimes the robustness of estim ators is described in term s o f the supremum of the Bayes risks, see e.g. Berger (1982, 1990), Verdu, Poor (1984). 'I hen the above estim ator is called minimax-robust. Various problems o f m inim ax-robust regression were discussed e.g. in Berger (1982, 1990), Berger, Chen (1987), Chaturvedi, Srivastawa (1992), Grzybowski (1997), Pinelis (1991). The relations between Bayesian analysis and m inim ax estim ation were exam ined in Berger, Chen (1987), D r y gas, Pilz (1996), Hoffman (1996), Pilz (1986). Review o f recent results on robust Bayesian analysis and interesting references can be found in Męczarski (1998). Various classes o f stable (robust, m inim ax and other) estim ators are also discussed in Milo (199). Let us consider the case where our inform ation about the param eter ß is described with the help o f the following class r SiK of distributions л: r SiK = {n: EJi = 9, cov (//) = A e K c M*}, where К is a given subset of a space M k of all positive definite (k x k) m atrices. I he point 9, fixed throughout this paper, can be thought o f as a prior guess for //, while the set К reflects our uncertainty connected with the guess. Let us assume the covariance m atrix E o f the random disturbance Z belongs to a given subset Q o f the space M. Problem s of m inimax regression estim ation in the presence of such a prior inform ation about E were considered in Grzybowski (1997), Hoffman (1996), Pinelis (1991). The set G = K x Í2 is the uncertainty class in our problem. Let us consider the situation when the set o f allowable estim ators L consists of all affine linear estim ators d, i.e. estim ators having the form d(y) = AY + B, with Л and В being a matrix and a vector of the appropriate dim ensions. T he original problem o f estim ation of the param eter ß now can be treated as a game <G, L, r). The solution of the game was found in Grzybowski (1997). It was proved that if the uncertainty class G is convex then any aftinc linear estim ator d* which is Bayes with respect to the least favourable pair oj matrices is m inim ax-robust (note that the Bayes risk for affine linear estim ators is determ ined by the first two m om ents of the prior distribution). The least favourable m atrices A* and E* satisfy the condition: where C(A, E) = (X'E~LX -f A-*) following fam iliar form ula: tr(c(a*, E*)H = sup tr(c(a, E)H ) (2) (A, L)eG while the estim ator d* is given by the

d*(a*, *) = C(A*, Z*)Xr (E*)_ ły + С(АФ, (3) Last year, during M SA 97, we examined an uncertainty class with the sets К and Q defined as follows: K = {A: A = dlk, d e (0, d0]}, Í2 = {E: E = sln, s e (0, s0]}, with given real values d0, s0 and I*, I being the identity m atrices in the spaces M k and M n, respectively, sec Grzybowski (1997), Hoffman (1996), Pinelis (1991). Practically, the class m ay represent the case where the uncertainties connected with each coefficient ßt arc independent and the same while the regression model satisfies two assumptions: homoscedasticity and independence o f errors Z,. In the sequel o f the paper we propose uncertainty classes representing the situation when the uncertainties connected with each coefficient of regression param eter ß m ay be different and the above two assum ptions ab o u t the regression m odel m ay not be satisfied. F o r convenience we adopt the following notation. F o r any n-dimcnsional vectors a, b we write a > b if ai ^ h i, í = 1, n. We write a > 0 if all com ponents of the vector a are positive. For any m atrix A we write A > 0 (A > 0) if the m atrix is positive (nonnegative) definite. F o r any vector a the symbol diag(a) stands for a diagonal m atrix with the com ponents of a on the m ain diagonal. 3. MINIMAX ESTIMATION IN HETEROSCEDASTIC MODELS Let S e R K and o er " be given vectors. Let G(ô, a) = K x Í2 denote an uncertainty class where the sets К and Ü are defined as follows: К = {A e M k: A = diag(d), 0<d^<}, Q = { Z e M n: E = diag(s), 0<s<cr}. T he class m ay represent the case where the uncertainty connected with coefficients ß. of the regression param eter are different and the random errors Z i are independent but they m ay have different standard deviation. T he following proposition provides the m inim ax estim ators for such problem s. Proposition 1. Let As = diag(c) and = diag(<x). The estim ator d*(ai; given by (3) is the m inim ax estim ator for the game <G(<, a), L, r. In view of the above m entioned results, in order to prove the Proposition it is sufficient to show that the tw o m atrices (A6, E J satisfy the condition (2). F o r this purpose we need the following lemmas. Lemma 1. Let A = [fly]* * > 0 and H = [hu]k * > 0. Let A and H n be the subm atrices of A and H, respectively, obtained by deleting the first row and the first colum n. T hen P ro o f o f Lem m a 1. t r í A ^ H J - t r í A f ^ H n ^ O (4)

Let us w rite dow n the m atrix A in the following form: O ne can verify th at where tr(a - 41) - t r í A í / H n ) = tr(m H ), with с = (flu wta 111w)- 1. c- T._, det(a) Since a u w А ц w = \ we see that c>0. det(a u ) It appears th at M > 0. Indeed, for an arbitrary ^-dim ensional vector x T = (xl,x l), * 1 e R, x 2e R * -1 we have x r M x = c(x x - b)2, with b = wta ii1x 2. Now, let e 7 = (1, 1, 1) and let M * H denote the H adam ard product o f the m atrices M, H. Then tr(m H ) = e r (M * H)e, see R a o (1973). On the other hand, from the Schur lemma we know that (M*H)>0 for any m atrices M > 0 and H ^ 0. It follows that er (M * H )c ^ 0 and the p ro o f is com pleted. Lemma 2. Let A>0, H^O be given k x к m atrices. Let for i= 1,..., к and x > 0 functions f t be defined as follows: /,(x ) = tr(a " H ), where a ri. i, \ au + - A* = [ M tx t with bjt = i x for j = l = i ajt otherwise. Then for each i = 1,..., к the fu n c tio n /, is non-decreasing. P ro o f of Lem m a 2. W ithout loss o f generality we m ay consider i = 1. A little calculation shows that for each x > 0 the derivative of f l does exist and = det(a n )det(a )[tr(a - lh) - t ^ A ^ H n ) ] [det(au ) + x d e t(a )]2

In view of Lemma I and our assum ptions about the m atrices Л and B, the derivative is nonnegativc, which com pletes the proof. P roof of the Proposition 1. F or any A = d ia g ^,, d2,..., d*) and L = diag(.st, s2,..., s ), A>0, E > 0, let the function g(dlt..., dk, s t, s j be defined as follows: g(dt, dk, su..., s ) = T(A, I ) = tr{[xr d ia g (, -) X + diag (J, - ) ] ' 1 H () Sj Sn 1 & к It is easy to check that the function g can also be expressed in the following way: g(dx, dk, su..., sn) = tr[diag(dp dk)h\ - tr[xdiag(</j, dt)x r + d ia g (s1, s )]_1N (6) where M = X d iag (d 1, dk) H diag(ť/j, dk) \ '. N ote that M>0. It can be seen from () that for each i = 1, к the function g as a function of dt is of the same form as the functions /, from Lem m a 2. T he relation (6) shows th at g as a function o f s, has got the form: const - /jf1). T hus, in view of (), (6) and Lem m a 2, the function g is \ s ij non-decreasing w.r.t. each variable dlt dk, sit..., s. It results that on the set {(d, s )e R x R ': 0<d^<, 0<s^o-} it achieves its m axim um at the point (ô, и). Since the set is convex the condition (2) yields the desired result. One m ay notice that in the considered case the least favourable states o f nature (and associated with them m inim ax estim ators) are intuitive - the m atrices As, are connected with greatest values o f variances o f the regression param eter and error, respectively. In the next section we discuss a problem where there is no such predictable value of the least favourable state o f nature. 4. MINIMAX ESTIMATION IN SOME PROBLEMS CONNECTED WITH CORRELATION BETWEEN ERRORS рн-л Let P(p) denote a m atrix with elements pu = y 2, /j < 1. Such m atrices appear in a natural way in the case where the dependence between errors can be described by the following first order au toco rrelation process:

where V = (Kj, V2, Vn) is a random vector with E(V) = 0 and D 2(V) = a>i, 0<co<co. It is well-known that then Cov(Z) = a)p(p). I'or given constants < u > 0, \<.p l <.p2 <. \ let us consider an uncertainty class G(S, о), p u p2) = K x 2 with the set К defined as previously while Í2 = { e M, : I = E(w, a) = w fa P ^ j) + (1 - a)p(/?2)], 0 < w < w, 0 < ос < 1}. The following proposition provides the m inim ax estim ators when the uncertainty class is G(<, со, p,, p2). Proposition 2. There exists a num ber a og[0> 1] such that the pair of matrices As, E(ro, a 0) is the least favourable state of nature and the estim ator d*[a)> a o)] *s the m inim ax estim ator in the game <G(á, со, p u p 2)t L,r). I he num ber a o e[0, 1] depends on the m atrices X and H. Proposition 2 states that, as in the previous case, the least favourable m atrix A is associated with the greatest variances of /У,. On the other hand the proposition asserts that the least favourable value o f the param eter a depends on the m atrices X, H and, in that sense, the least favourable covariancc m atrix o f the vector Z is unpredictable. I he proof oi Proposition 2 is based on Lemma 2 and will be om itted. In the sequel we present some numerical examples to show how a 0 depends on the m atrix II determ ining the loss. The dependence yields that our solution does not have the feature: m inim ax prediction equals prediction based on m inim ax estim ate o f the regression param eter. T he solution of the problem s considered in the previous Section has got such a property. In our examples we consider the m odel (1) with the following fixed values o f its characteristics: к = 3, n ш 7, X = - - 1 -. 1 1-1 1 1-1 - In all examples we consider the class G(<, w, 0, 0.) with fixed values < = (,..., ) and со =. This is because we already know how the estim ators depend upon these values. T o simplify the notation we write Tr(ot) instead o f tr{c[ai; E(w, a)] H}. Example 1. The classical problem of estim ation of the regression param eter. In this example we consider the case where H = I, i.e. the classical problem of estim ation of the param eter /?. Figure 1 shows the graph of the function Ir. We can see th at it has got one m axim um.

Fig. 1. The function Tr for the problem of estimation of [i. It can be num erically verified that the m axim um is taken on for <*max = 0,728304. Exam ple 2. The prediction o f the dependent variable. Now let us consider a problem of prediction of the value o f the dependent variable Y when the independent variables take on the following values: = 1, x 2 =, x 3 = 9. The corresponding m atrix determ ining the 'l 9' loss function is of the following form: H = 2 4. T he graph o f the 9 4 81 function T r in this case is presented in Fig. 2. Num erical calculation shows th a t the only m axim um o f T r is achieved for a mttx= 0,81236. Fig. 2. The function Tr for the problem of prediction for x, = 1, x 2 -, x 3 = 9

The two above examples show that the least favourable value of the param eter a (and the associated covariance m atrix ) can hardly be considered as intuitive. T he value changes for different m atrices H. It seems that in such situations we have different m inimax estim ators for various purposes (such as the estim ation of regression param eter, prediction for different values o f independent variables etc.) even in the same m odel. So, in the case of correlated errors one should be partiularly aw are o f the purpose o f the m inim ax estim ation. REFERENCES Berger J. (1982), Bayesian Robustness and the Stein Effect, J. Amer. Stat. Assoc. 38-368. Berger J., Chen S. (1987), Minimaxity o f Empirical Bayes Estimators Derived from Subjective Hyperpriors, [in:] Advances of the Multivariate Statistical Analysis, ed. A. K. Gupta, 1-12. Berger J. (1990), Robust Bayesian Analysis: Sensitivity to the Prior, J. Statist. Plann Inference, 2, 303-328. Chaturvedi A., Srivastawa A. K. (1992), Families o f Minimax Estimators in Linear Regression Model, Sankhya, 4, Ser. B, 278-288. D ry g a s H. (1996), Spectral Methods in Linear Minimax Estimation, Acta Appl. Math 43 17-42. Drygas H., Pilz J. (1996), On the Equivalence o f the Spectral Theory and Bayesian Analysis in Minimax Linear Estimation, Acta Appl. Math., 43, 43-7. Girko V. L. (1996), Spectral Theory o f Minimax Estimation, Acta Appl. Math., 43, 9-69. Grzybowski A. (1997), Minimaksowo-odporna estymacja parametrów regresji, Przegląd Statystyczny", 3, 427-43. Grzybowski A. (1997), On Some Model o f Uncertainty in the Minimax-Robust Multivariate Linear Regression, Proceedings of Iď International Conference on Multivariate Statistical Analysis MSA 97, 203-212. Hoffman K. (1996), A Subclass o f Bayes Linear Estimators that are Minimax, Acta Appl Math., 43, 87-9. Męczarski M. (1998), Problemy odporności w bayesowskiej analizie statystycznej, ser. Monografie i Opracowania, 446, SGH, Warszawa. Milo W. (199), Stabilność i wrażliwość metod ekonometrycznych, Wydawnictwo Uniwersytetu Łódzkiego, Łódź. Pilz J. (1986), Minimax Linear Regression Estimation with Symmetric Parameter Restrictions, J. Statist. Plann. Inference, 13, 297-318. Pinelis I. F. (1991), O minimaksnom ocenivanii regressii, Teorija Verojatnostej i ее primenenija 3 (3), s. 494-0. R a o C. R. (1973), Linear Statistical Inference and its Applications, John Wiley and Sons, New York. (Polish translation: Modele liniowe statystyki matematycznej, PWN, Warszawa 1984). V er du S., Poor H. V. (1984), On Minimax Robustness: A General Approach and Applications, IEEE Trans. Inform. Theory IT-30, 328-340.