Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Similar documents
Bayesian Nonparametric Point Estimation Under a Conjugate Prior

Asymptotic efficiency of simple decisions for the compound decision problem

Mark Gordon Low

Equivalence Theory for Density Estimation, Poisson Processes and Gaussian White Noise With Drift

L. Brown. Statistics Department, Wharton School University of Pennsylvania

Wavelet Shrinkage for Nonequispaced Samples

Optimal global rates of convergence for interpolation problems with random design

arxiv:math/ v1 [math.st] 29 Mar 2005

Estimation of a quadratic regression functional using the sinc kernel

Chi-square lower bounds

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin

Optimal Estimation of a Nonsmooth Functional

RANDOM WALKS WHOSE CONCAVE MAJORANTS OFTEN HAVE FEW FACES

Lecture 7 Introduction to Statistical Decision Theory

D I S C U S S I O N P A P E R

Inference for High Dimensional Robust Regression

STATISTICS SYLLABUS UNIT I

Lecture 8: Information Theory and Statistics

Distance between multinomial and multivariate normal models

Studentization and Prediction in a Multivariate Normal Setting

A COUNTEREXAMPLE TO AN ENDPOINT BILINEAR STRICHARTZ INEQUALITY TERENCE TAO. t L x (R R2 ) f L 2 x (R2 )

Root-Unroot Methods for Nonparametric Density Estimation and Poisson Random-Effects Models

Why experimenters should not randomize, and what they should do instead

Can we do statistical inference in a non-asymptotic way? 1

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

Minimax Estimation of Kernel Mean Embeddings

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES

Asymptotically sufficient statistics in nonparametric regression experiments with correlated noise

Regularity of the density for the stochastic heat equation

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

1 Exercises for lecture 1

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

COMPLETE QTH MOMENT CONVERGENCE OF WEIGHTED SUMS FOR ARRAYS OF ROW-WISE EXTENDED NEGATIVELY DEPENDENT RANDOM VARIABLES

LAN property for ergodic jump-diffusion processes with discrete observations

Goodness-of-fit tests for the cure rate in a mixture cure model

Asymptotic Statistics-III. Changliang Zou

Max stable Processes & Random Fields: Representations, Models, and Prediction

A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis

Statistical Inference

Variance Function Estimation in Multivariate Nonparametric Regression

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

A NOTE ON THE COMPLETE MOMENT CONVERGENCE FOR ARRAYS OF B-VALUED RANDOM VARIABLES

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

SPATIAL STRUCTURE IN LOW DIMENSIONS FOR DIFFUSION LIMITED TWO-PARTICLE REACTIONS

The Azéma-Yor Embedding in Non-Singular Diffusions

Regular Variation and Extreme Events for Stochastic Processes

Bayesian Inference and the Symbolic Dynamics of Deterministic Chaos. Christopher C. Strelioff 1,2 Dr. James P. Crutchfield 2

Efficient Robbins-Monro Procedure for Binary Data

Anomalous transport of particles in Plasma physics

Evgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields

Quickest Detection With Post-Change Distribution Uncertainty

DA Freedman Notes on the MLE Fall 2003

Minimax Risk: Pinsker Bound

Robust Backtesting Tests for Value-at-Risk Models

ECE531 Lecture 10b: Maximum Likelihood Estimation

1.1 Basis of Statistical Decision Theory

8 Basics of Hypothesis Testing

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Sub-Gaussian estimators under heavy tails

Discussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan

The Codimension of the Zeros of a Stable Process in Random Scenery

Limiting Distributions

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Qualifying Exam in Probability and Statistics.

Ensemble Minimax Estimation for Multivariate Normal Means

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Limiting Distributions

A CLT FOR MULTI-DIMENSIONAL MARTINGALE DIFFERENCES IN A LEXICOGRAPHIC ORDER GUY COHEN. Dedicated to the memory of Mikhail Gordin

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0}

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Bickel Rosenblatt test

Submitted to the Brazilian Journal of Probability and Statistics

Inverse Statistical Learning

Minimum Message Length Analysis of the Behrens Fisher Problem

Covariance function estimation in Gaussian process regression

04. Random Variables: Concepts

Mi-Hwa Ko. t=1 Z t is true. j=0

On Extracting Common Random Bits From Correlated Sources on Large Alphabets

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh

Lecture 4: September Reminder: convergence of sequences

Negative Association, Ordering and Convergence of Resampling Methods

OPTIMAL POINTWISE ADAPTIVE METHODS IN NONPARAMETRIC ESTIMATION 1

Asymptotic normality of the L k -error of the Grenander estimator

Optimal Sequential Selection of a Unimodal Subsequence of a Random Sequence

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA

Exact Simulation of Diffusions and Jump Diffusions

Uniform Correlation Mixture of Bivariate Normal Distributions and. Hypercubically-contoured Densities That Are Marginally Normal

Statistical Measures of Uncertainty in Inverse Problems

Spring 2012 Math 541B Exam 1

Empirical Risk Minimization is an incomplete inductive principle Thomas P. Minka

Semiparametric posterior limits

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Transcription:

University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University of Pennsylvania Cun-Hui Zhang Follow this and additional works at: http://repository.upenn.edu/statistics_papers Part of the Statistics and Probability Commons Recommended Citation Brown, L. D., & Zhang, C. (1998). Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½. The Annals of Statistics, 26 (1), 279-287. http://dx.doi.org/10.1214/aos/1030563986 This paper is posted at ScholarlyCommons. http://repository.upenn.edu/statistics_papers/314 For more information, please contact repository@pobox.upenn.edu.

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Abstract An example is provided to show that the natural asymptotic equivalence does not hold between any pairs of three nonparametric experiments: density problem, white noise with drift and nonparametric regression, when the smoothness index of the unknown nonparametric function class is ½ Keywords risk equivalence, nonparametric regression, density estimation, white noise Disciplines Statistics and Probability This journal article is available at ScholarlyCommons: http://repository.upenn.edu/statistics_papers/314

The Annals of Statistics 1998, Vol. 26, No. 1, 279 287 ASYMPTOTIC NONEQUIVALENCE OF NONPARAMETRIC EXPERIMENTS WHEN THE SMOOTHNESS INDEX IS 1/2 By Lawrence D. Brown 1 and Cun-Hui Zhang 2 University of Pennsylvania and Rutgers University An example is provided to show that the natural asymptotic equivalence does not hold between any pairs of three nonparametric experiments: density problem, white noise with drift and nonparametric regression, when the smoothness index of the unknown nonparametric function class is 1/2. 1. Introduction. There have recently been several papers demonstrating the global asymptotic equivalence of certain nonparametric problems. See especially Brown and Low (1996), who established global asymptotic equivalence of the usual white-noise-with-drift problem to the nonparametric regression problem, and Nussbaum (1996), who established global asymptotic equivalence to the nonparametric density problem. In both these instances the results were established under a smoothness assumption on the unknown nonparametric drift, regression or density function. In both cases such functions were assumed to have smoothness coefficient α>1/2, for example, to satisfy 1 1 f x f y M x y α for all x y in their domain of definition. This note contains an example which shows that such a condition is necessary, in the sense that global asymptotic equivalence may fail between any pairs of the above three nonparametric experiments when (1.1) fails in a manner that the nonparametric family of unknown functions contains functions satisfying (1.1) with α = 1/2 but not with any α>1/2. Efromovich and Samarov (1996) have already shown that asymptotic equivalence of nonparametric regression and white noise may fail when α<1/4 in (1.1). The present counterexample to equivalence is somewhat different from theirs and carries the boundary value α = 1/2. Section 2 contains a brief formal description of the nonparametric problems and of global asymptotic equivalence. Section 3 describes the example. Received July 1996; revised May 1997. 1 Supported in part by NSF Grant DMS-91-07842. 2 Supported in part by the National Security Agency. AMS 1991 subject classifications. Primary 62G07; secondary 62G20, 62M05. Key words and phrases. Risk equivalence, nonparametric regression, density estimation, white noise. 279

280 L. D. BROWN AND C.-H. ZHANG 2. The problem. Density problem ξ 0 n. In the nonparametric density problem one observes i.i.d. variables X 1 X n having some density g = f 2 /4. Assume the support of f is 0 1. About f it is assumed only that f f f 0 1 0 f2 x dx = 4, where is some (large) class of functions in L 2 0 1. In general, the goal is to use the observations to make some sort of inference about f. White noise ξ 1 n. In the white-noise problem one observes a Gaussian process Z n t 0 t 1 which can be symbolically written as dz n t =f t dt + 1 db t n where B t denotes the standard Brownian motion on 0 1. Again f. Nonparametric regression ξ 2 n. Here one observes Y i X i 1 i n. Given X i 1 i n, the Y i are independent normal variables with mean f X i and unit variance, f. For deterministic X i [e.g., X i = i/ n + 1 ] and α = 1/2, the asymptotic nonequivalence of nonparametric regression and white noise has already been established in Brown and Low [(1996), Remark 4.6]. Hence, in the case of current interest the X i are i.i.d. uniform random variables on 0 1. Asymptotic equivalence. The assertion that two of these formulations say, nonparametric regression and white noise are globally asymptotically equivalent is equivalent to the following assertion: for each n, let A n be an action space and L n be a loss function with L n 1, and let δ n be a procedure in one of the problems. Then there exists a corresponding procedure δ n in the other problem such that the sequences δ n and δ n are asymptotically equivalent, which means ( 2 1 lim Ln f δ n ) ( E n Ln f δ n ) = 0 sup n f E n f [The expectations in (2.1) are computed under f for the nth form of the first and second problems, respectively. The notation should be interpreted to allow for randomized decision rules. The convergence in (2.1) is uniform in the loss functions and decision rules as they are allowed to change with n.] The asymptotic equivalence established in Brown and Low (1996) is a little different from the above statement. The equivalence expressed above is established in Brown and Zhang (1996). The equivalence assertion involving the density problem was established by Nussbaum (1996) under (1.1) and the additional assumption that min 0 x 1 f x is uniformly bounded away from 0 over : it is that the density problem with unknown density g = f 2 /4 is asymptotically equivalent to the f

sup n f ASYMPTOTIC NONEQUIVALENCE 281 white noise with drift f. The statement of (2.1) thus becomes ( lim Ln f δ n ) ( E n f Ln f δ n ) = 0 E n f 2 /4 where the first expectation refers to the density problem and the second to either the white-noise or the nonparametric regression problem. Hence, in order to prove two formulations nonequivalent in this sense, it suffices to find some action spaces A n, uniformly bounded loss functions L n and priors on for which the Bayes risks converge to different values in different problems. 3. The example. Let α M be the class of all functions f with support 0 1 such that (1.1) holds for all 0 x<y 1. Here we shall give an example to show that the white-noise and nonparametric regression experiments are not asymptotically equivalent for = 1/2 M, and also that under the stronger restriction = 1/2 M f f>ε 0 f 2 2 = 4, the density problem is not asymptotically equivalent to either the white noise or the nonparametric regression for every 0 ɛ 0 < 2. Let ψ x be a function in 1 M such that 1 0 ψ x dx = 1 0 ψ3 x dx = 0 and ψ x =0 for x 0 1. Form = m n, define 3 1 φ j x =m 1/2 ψ m x j 1 /m 1 j m For θ = θ 1 θ m with θ j 1, define 3 2 φ θ x = m θ j φ j x where c m = c m θ is chosen so that f θ /2 2 = 1. f θ x =2 + φ θ x c m Motivation. Examples of asymptotic nonequivalence can be constructed by finding Bayes problems for which the Bayes risks converge to different limits for different sequences of experiments. Consider prior distributions on the subspaces m 1/2 α φ θ of α M with φ θ in (3.2) and m n. Due to the normality of the errors, the nonparametric regression ξ 2 n is characterized by the Fisher information σα 2 j = m1 2α i φ 2 j X i given X i,1 j m, so that its Bayes risk is of the form Er n σ α 1 σ α n, with r n being the conditional Bayes risk. It is easily seen that the corresponding Bayes risk for the white noise ξ 1 n is r n Eσα 2 1 Eσα 2 n. For tractable Bayes problems, the limit Bayes risk is often found via Taylor expansions of certain components of r n and applications of limiting theorems to j σα k j for ξ 2 n. Since σ α j is proportional to m α, higher-order terms are needed in Taylor expansions only for small α, whereas lower-order terms are more likely to be crucial for large α. Since we are interested in the largest possible α = 1/2 for nonequivalence examples, we shall look for Bayes problems in which j σ α j plays an important role. This leads to the following variation of the compound hypothesis testing problem of Robbins (1951).

282 L. D. BROWN AND C.-H. ZHANG Let G be the prior such that θ j,1 j m, are i.i.d. Bernoulli variables with 3 3 P G θ j = 1 =P G θ j = 1 =1/2 Consider the estimation of θ = θ 1 θ m with the 0 1 loss for picking more than half of the θ j wrong, m 3 4 L θ a =I I θ j a j >m/2 where a = a 1 a m is the action. We shall show that the Bayes risks for the three sequences of experiments converge to different limits as n and n/m λ, 3 5 R G d G ξ j n τ j λ ψ where ξ j n j = 0 1 2, are, respectively, the density problem, white noise and nonparametric regression, d G = d G ξ j n is the Bayes rule with experiment ξ j n, is the standard normal distribution function and τ j is defined below. Let N be a Poisson variable with EN = λ and let U i i 1, be i.i.d. uniform 0 1 variables independent of N. The functions τ j λ ψ in (3.5) are analytically different and are given by N τ 0 λ ψ =E ψ U i 3 6 2λ τ 1 λ ψ = π ψ 2 τ 2 λ ψ =E 2 N ψ π 2 U i By the Schwarz inequality, τ 2 λ ψ <τ 1 λ ψ. For small λ, τ 1 λ ψ >τ 0 λ ψ = ψ 1 + o 1 λ >τ 2 λ ψ = 2/π ψ 1 + o 1 λ [By the moment convergence in the strong law of large numbers and the central limit theorem, τ i λ ψ /τ j λ ψ 1asλ for all 0 i<j 2.] For = 1/2 M, the asymptotic nonequivalence between the white noise and the nonparametric regression follows from (3.5) and (3.6) as ψ 1 M implies f θ 1/2 M in view of (3.1) and (3.2). Since φ j x φ k x =0 for j k, 3 7 φ θ 2 2 = m φ j 2 2 = m 1 ψ 2 2 φ θ = ψ / m Since g θ = f θ /2 2 is a density, by (3.7) c m do not depend on θ under (3.3) and 3 8 c m = 2 1 1 φ θ 22 /4 = m 1 ψ 2 2 /4 + O m 2 The asymptotic nonequivalence between the density problem and either the white noise or the nonparametric regression also follows from (3.5) and (3.6)

ASYMPTOTIC NONEQUIVALENCE 283 as the further restrictions f θ 2 2 = 4 and f θ >ɛ 0 hold for large m and fixed 0 ɛ 0 < 2 by (3.2), (3.7), (3.8) and the fact that ψ M. Motivation (continued). The decision problem (3.4) is closely related to compound hypothesis testing and estimation with L 1 loss. In the compound hypothesis testing problem, the loss function is m 1 m I θ j a j, the Bayes rules for the prior (3.3) are the same as with the loss (3.4), and the Bayes risks are 1/2 λ/ 4n 1/2 τ j λ ψ +o n 1/2 for ξ j n. In the estimation problem with the L 1 loss 1 0 a t f t dt, the Bayes rules are fg = m d G j φ j with φ j in (3.1) and d G = d G 1 dg m in (3.5), and the Bayes risks are [ λ/n ψ 1 1/2 λ/ 4n 1/2 τ j λ ψ ] + o n 1 Thus, the difference among the nonparametric experiments is recovered in the second-order asymptotics in the above two decision problems. The proofs of the above statements are omitted as they are similar to and simpler than the calculation of the Bayes risks (3.5) and (3.6). Remark 1. In nonparametric regression with deterministic X i = i/ n+1, Y i contain no information about f in the subspace (3.2) with f = f θ and m = n + 1 as in Brown and Low (1996), so that the Bayes risk for the decision problem (3.4) is 1/2 under the prior (3.3). Since τ j 1 ψ > 0 for ψ 2 > 0 in (3.5) for each j = 0 1 2, the nonparametric regression with deterministic X i is asymptotically nonequivalent to ξ j n for α = 1/2. Remark 2. Many applications of the nonparametric experiments discussed here involve their d-dimensional versions with f being an unknown function of d real variables (e.g., E Y i X i =f X i with X i being uniform 0 1 d in the case of nonparametric regression). The above example can be easily modified to show the asymptotic nonequivalence of these nonparametric experiments when the smoothness index is d/2. 4. Calculation of Bayes risks. In this section we calculate the limit of the Bayes risks given in (3.5) and (3.6). Nonparametric regression ξ 2 n. Let Pn be the conditional probability given X 1 X n and under P G. Set n n S j = φ j X i Y i 2 + c m σj 2 = φ 2 j X i Since φ j have disjoint support sets, by (3.2) S j are sufficient for θ j.in addition, S j θ j 1 j m, are independent random vectors under P n with P n S j t θ j = ( t θ j σ 2 j /σ j) P n θ j =±1 =1/2

284 L. D. BROWN AND C.-H. ZHANG Since the loss function in (3.4) is increasing in I θ j a j, the Bayes rule d G = d G 1 dg m is given by d G j = I S j > 0 I S j 0 and I θ j d G j are independent Bernoulli random variables under P n with 4 1 p j = P n θj d G j = P n θj S j < 0 θ j = σj Since σj 2 n φ j 2 = n/m ψ 2 = O 1, p j 1 p j are uniformly bounded away from zero, so that ( m ) 1/2 m m 4 2 p j 1 p j I θ j d G j p j N 0 1 uniformly under Pn. Let N j = # i j 1 /m<x i j/m. Since X i are i.i.d. uniform, N 1 N m is a multinomial vector with EN j = n/m λ. It follows that N1 meσ1 = E and meσ 2 1 = mn φ 1 2 2 = n m ψ 2 2 N1 meσ 1 σ 2 = E ψ 2 U i 1/2 π ψ 2 U i τ 2 λ ψ 2 N 2 i=n 1 +1 1/2 ψ 2 U i τ2 2 λ ψ π 2 as n and n/m λ, where U i are i.i.d. uniform 0 1 variables independent of X i. These lead to meσ1 2 = O 1 and ( m ) Var σ j meσ1 2 + m m 1 Eσ 1 σ 2 Eσ 1 2 = o m By (4.1) and the boundedness and Taylor expansion of, p j 1/2 = σ j 1/2 = σ j / 2π + O σj 2 and p j 1 p j =1/4 p j 1/2 2 = 1/4 + O σj 2 with O 1 uniformly bounded by a universal constant, so that by the Chebyshev inequality, m p j 1/2 m p j 1 p j τ 2 λ ψ and 1 m/4 m 4 in probability. This and (4.2) imply (3.5) and (3.6) for ξ 2 n,as m P I (( θ j d G m ) 1/2 m ( ) ) j >m/2 = E p j 1 p j pj 1 + o 1 2 = τ 2 λ ψ + o 1

ASYMPTOTIC NONEQUIVALENCE 285 White noise ξ 1 n. The calculation is similar but much simpler compared with nonparametric regression. The sufficient statistics are Z j = n φ j t dz n t N θ j n φ j 2 2 n φ j 2 2 so that the Bayes rule d G = d G 1 dg m is given by dg j = I Z j > 0 I Z j 0. Consequently, I θ j d G j, 1 j m, are i.i.d. Bernoulli variables with P θ j d G j = P θ j d G j θ j = n φ j 2. Since n φ j 2 = n ψ 2 /m = λ + o 1 ψ 2 / m and t 1/2 + t/ 2π for small t, m P I θ j d G j >m/2 ψ 2 2λ/π Density problem ξ 0 n. For 1 j m, define [ ( n 1 ± φj X i j ±1 =exp 2 log c ) m j 1 I 2 2 m <X i j ] m Since the observations X 1 X n are i.i.d. from g θ = f θ /2 2, by (3.2) the likelihood is n m g θ X i = j θ j so that θ j 1 j m, are independent given X 1 X n and j ±1 are sufficient for θ j. Consequently, the Bayes rule d G = d G 1 dg m is given by d G j = I j 1 > j 1 I j 1 j 1 and I θ j d G j 1 j m, are independent variables given X 1 X n with 4 3 p j = Pn θ j d G j =P n θ j d G j θ j = min j 1 j 1 j 1 + j 1 where Pn is the conditional probability given X 1 X n (the posterior probability measure with respect to θ). Taking Taylor expansions, we find by (3.7) and (3.8), 2 log 1 ± φ j X i /2 c m /2 =±φ j X i c m φ 2 j X i /4 + O m 3/2 with uniform O 1, so that 4 4 log j 1 / j 1 = 2 S j + O N j m 3/2 where S j = n φ j X i. Since 1 0 ψ x dx = 1 0 ψ3 x dx = 0, E θ φ j X i = φ j x 1 + θ j φ j x /2 c m /2 2 dx = 1 c m /2 θ j φ j 2 2 = 1 c m /2 θ j ψ 2 2 /m2

286 L. D. BROWN AND C.-H. ZHANG and by (3.7) and (3.8), E θ φ 2 j X i = 1 c m /2 2 φ j 2 2 + φ j 4 4 /4 1 ψ 2 2 / 4m ψ 2 2 /m2 + ψ 4 4 / 4m3 ψ 2 /m2 Since φ j x = ψ / m, by the Bernstein inequality there exists a constant C such that C log m + 1 P θ S j > 1 m m 2 uniformly in θ, so that P C log m + 1 m 1 m 0 max S j > 1 j m This and (4.4) allow us to take the Taylor expansion of p j in (4.3), 1 2p j = j 1 / j 1 1 j 1 / j 1 +1 4 5 = 2 S j + 2 S 2 j 2 + 2 S j + O m 3/2 N j + log m 3 = S j +O m 3/2 N j + log m 3 Since m N j = n = O m, max 1 j m 1 2 p j 0in probability, so that (4.2) holds under the current Pn over some events C n σ X 1 X n with P C n 1. Thus (( m ) 4 6 R G d G 1/2 m ) ξ 0 n =E p j 1 p j p j 1/2 + o 1 In addition, N 1 meθ S 1 =E θ ψ Ũ i1 N 1 N 2 me θ S 1 S 2 =E θ ψ Ũ i1 ψ Ũ i2 where Ũ ij i 1, are i.i.d. with density 1 + θ j m 1/2 ψ x /2 c m /2 2 under P θ. Since these density functions converge uniformly in x θ j to the uniform 0 1 density and N 1 N m is a multinomial vector with EN j = E θ N j = n/m, N me S 1 E ψ U i = τ 0 λ ψ me S 1 S 2 τ0 2 λ ψ

ASYMPTOTIC NONEQUIVALENCE 287 and E S 2 1 neφ2 j X 1 n ψ 2 /m2, so that E 1 m 2 S j τ 0 λ ψ m 0 m E S j 2 = O 1 Hence, by (4.5) and (4.6), R G d G ξ 0 n =E τ 0 λ ψ + o 1. REFERENCES Brown, L. D. and Low, M. G. (1996). Asymptotic equivalence of nonparametric regression and white noise. Ann. Statist. 24 2384 2398. Brown, L. D. and Zhang, C.-H. (1996). Coupling inequalities for some random design matrices and asymptotic equivalence of nonparametric regression and white noise. Preprint. Efromovich, S. and Samarov, A. (1996). Asymptotic equivalence of nonparametric regression and white noise has its limits. Statist. Probab. Lett. 28 143 145. Nussbaum, M. (1996). Asymptotic equivalence of density estimation and white noise. Ann. Statist. 24 2399 2430. Robbins, H. (1951). Asymptotically subminimax solutions of compound statistical decision problems. Proc. Second Berkeley Symp. Math. Statist. Probab. 1 131 148. Univ. California Press, Berkeley. Department of Statistics The Wharton School University of Pennsylvania Philadelphia, Pennsylvania 19104-6302 E-mail: lbrown@stat.wharton.upenn.edu Department of Statistics Rutgers University Busch Campus Piscataway, New Jersey 08855 E-mail: czhang@stat.rutgers.edu