New Entropy Estimator with an Application to Test of Normality

Size: px
Start display at page:

Download "New Entropy Estimator with an Application to Test of Normality"

Transcription

1 New Entropy Estimator with an Application to Test of Normality arxiv: v1 [math.st] 15 Oct 211 Salim BOUZEBDA 1, Issam ELHATTAB 1, Amor KEZIOU 1 and Tewfik LOUNIS 2 L.S.T.A., Université Pierre et Marie Curie 4 place Jussieu Paris Cedex 5 Laboratoire de Mathématiques Nicolas Oresme CNRS UMR 6139 Université de Caen 2 Abstract In the present paper we propose a new estimator of entropy based on smooth estimators of quantile density. The consistency and asymptotic distribution of the proposed estimates are obtained. As a consequence, a new test of normality is proposed. A small power comparison is provided. A simulation study for the comparison, in terms of mean squared error, of all estimators under study is performed. AMS Subject Classification: : 62F12 ; 62F3 ; 62G3 ; 6F17 ; 62E2. Keywords: Entropy; Quantile density; Kernel estimation; Vasicek s estimator; Spacingbased estimators; Test of nomality; Entropy test. 1 Introduction and estimation Let X be a random variable [r.v.] with cumulative distribution function [cdf] F(x := P(X x for x R and a density function f( with respect to Lebesgue measure on R. Then its differential (or Shannon entropy is defined by H(X := f(xlogf(xdx, (1.1 R salim.bouzebda@upmc.fr (corresponding author issam.elhattab@upmc.fr amor.keziou@upmc.fr tewfik.lounis@gmail.com 1

2 where dx denotes Lebesgue measure on R. We assume that H(X is properly defined by the integral (1.1, in the sense that H(X <. (1.2 The concept of differential entropy was introduced in Shannon s original paper [Shannon (1948]. Since this early epoch, the notion of entropy has been the subject of great theoretical and applied interest. Entropy concepts and principles play a fundamental role in many applications, such as statistical communication theory [Gallager (1968], quantization theory [Rényi (1959], statistical decision theory [Kullback (1959], and contingency table analysis [Gokhale and Kullback (1978]. Csiszár (1962 introduced the concept of convergence in entropy and showed that the latter convergence concept implies convergence in L 1. This property indicates that entropy is a useful concept to measure closeness in distribution, and also justifies heuristically the usage of sample entropy as test statistics when designing entropy-based tests of goodness-of-fit. This line of research has been pursued by Vasicek (1976, Prescott (1976, Dudewicz and van der Meulen (1981, Gokhale (1983, Ebrahimi et al. (1992 and Esteban et al. (21 [including the references therein]. The idea here is that many families of distributions are characterized by maximization of entropy subject to constraints (see, e.g., Jaynes (1957 and Lazo and Rathie (1978. There is a huge literature on the Shannon s entropy and its applications. It is not the purpose of this paper to survey this extensive literature. We can refer to Cover and Thomas (26 (see their Chapter 8, for a comprehensive overview of differential entropy and their mathematical properties. In the literature, Several proposals have been made to estimate entropy. Dmitriev and Tarasenko (1973 and Ahmad and Lin (1976 proposed estimators of the entropy using kernel-type estimators of the density f(. Vasicek (1976 proposed an entropy estimator based on spacings. Inspired by the work of Vasicek (1976, some authors [van Es (1992, Correa (1995 and Wieczorkowski and Grzegorzewski (1999] proposed modified entropy estimators, improving in some respects the properties of Vasicek s estimator (see Section 4 below. The reader finds in Beirlant et al. (1997 detailed accounts of the theory as well as surveys for entropy estimators. This paper aims to introduce a new entropy estimator and obtains its asymptotic properties. Comparison Simulations indicate that our estimator produces smaller mean squared error than the other well-known competitors considered in this work. As a consequence, we propose a new test of normality. First, we introduce some definitions and notations. For each distribution function F(, we define the quantile function by Let Q(t := inf{x : F(x t}, < t < 1. x F := sup{x : F(x = } and x F := inf{x : F(x = 1}, x F < x F. 2

3 We assume that the distribution function F( has a density f( (with respect to Lebesgue measure on R, and that f(x > for all x (x F,x F. Let q(x := dq(x/dx = 1/f(Q(x, < x < 1, be the quantile density function [qdf]. The entropy H(X, defined by (1.1, can be expressed in the form of quantile-density functional as ( d H(X = log dx Q(x dx = log ( q(x dx. (1.3 [,1] [,1] The Vasicek s estimator was constructed by replacing the quantile function Q( in (1.3 by the empirical quantile function and using a difference operator instead of the differential one. The derivative (d/dxq(x is then estimated by a function of spacings. In this work, we construct our estimator of entropy by replacing q(, in (1.3, by an appropriate estimator q n ( ofq(. We shall consider the kernel-type estimator ofq( introduced by Falk (1986 and studied by Cheng and Parzen (1997. Our choice is motivated by the well asymptotic behavior properties of this estimator. Cheng and Parzen (1997 were established the asymptotic properties of q n ( on all compact U ],1[, which avoids the boundary problems. Since the entropy is definite as an integral on ],1[ of a functional of q(, it is not suitable to substitute directly q n ( in (1.3 to estimate H(X. To circumvent the boundary effects, we will proceed as follows. We set for small ε ],1/2[, H ε (X := εlog ( q(ε +εlog ( q(1 ε + In view (1.3, we can see that 1 ε ε log ( q(x dx. H(X H ε (X = o ( η(ε, (1.4 where η(ε, as ε. The choice of ε close to zero guaranteed the closeness of H ε (X to H(X, then the problem of the estimation of H(X is reduced to estimate H ε (X. Given an independent and identically distributed random [i.i.d.] sample X 1,...,X n, and let ε ],1/2[, an estimator of H ε (X can be defined as Ĥ ε;n (X = εlog( q n (ε+εlog( q n (1 ε+ 1 ε ε log ( q n (x dx, (1.5 where the estimator q n ( of the qdf q( is defined as follows. Let X 1;n X n;n denote the order statistics of X 1,...,X n. The empirical quantile function Q n ( based upon these random variables is given by Q n (t := X k;n, (k 1/n < t k/n, k = 1,...,n. (1.6 3

4 Let {K n (t,x,(t,x [,1] ],1[} n 1 be a sequence of kernels and {µ n ( } n 1 a sequence of σ-finite measures on [,1]. A smoothed version of Q n ( (see, e.g., Cheng and Parzen (1997 can be defined as Q n (t := 1 Finally, we estimate q( by q n (t := d dt Q n (t = d dt Q n (xk n (t,xdµ n (x, t ],1[. 1 Q n (xk n (t,xdµ n (x, t ],1[. (1.7 Clearly, in order to obtain a meaningful qdf estimator in this way, the sequence of kernels K n (, must satisfy certain differentiability conditions, and together with the sequence of measures µ n (, must satisfy certain variational conditions. These conditions will be detailed in the next section. A familiar example is the convolution-kernel estimator q n (t := d dt 1 h 1 n Q n(xk ( t x h n dx, t ],1[, where K( denotes a kernel function, namely a measurable function integrating to 1 on R, and has bounded derivative, and {h n } n 1 is a sequence of positive reals fulfills h n and nh n as n. In this case, Csörgő and Révész (1984 and Csörgő et al. (1991 define n/(n+1 ( t x q n (t := h 1 n K dq n (x = h 1 n 1/(n+1 n 1 ( t i/n K i=1 h n h n (X i+1;n X i;n, t [,1]. Calculations using summation by parts show that, for all t ],1[, [ ( ( ] t 1 t q n (t = q n (t+h 1 n K X n;n K X 1;n. h n In the sequel, we shall consider the general family of qdf estimators defined in (1.7. h n The remainder of the present article is organized as follows. The consistency and normality of our estimator are discussed in the next section. Our arguments, used to establish the asymptotic normality of our estimator, make use of an original application of the invariance principle for the empirical quantile process. In Section 3.1, we discuss briefly the smoothed version of Parzen s entropy estimator. In Section 4, we investigate the finitesample performance of the newly proposed estimator and compare the latter with the performances of existing estimators. In section 5, a new test of normality is proposed and compared with other competitor. Some concluding remarks and possible future developments are mentioned in Section 6. To avoid interrupting the flow of the presentation, all mathematical developments are relegated to Section 7. 4

5 2 Main results Throughout U(ε := [ε,1 ε], for an ε ],1/2[ is arbitrarily fixed (free of the sample size n. The following conditions are used to establish the main results. We recall the notations of Section 1. (Q.1 The quantile density function q( is twice differentiable in ],1[. Moreover, < min{q(,q(1} ; (Q.2 There exists a constant ς > such that { } t(1 t J(t ς, sup t ],1[ where J(t := dlog { q(t } /dt is the score function; (Q.3 Either q( < or q( is nonincreasing in some interval (,t, and either q(1 < or q( is nondecreasing in some interval (t,1, where < t < t < 1. We will make the following assumptions on the sequence of kernels K n (,. (K.1 For each n 1 and each (t,x U(ε ],1[, K n (t,x, and for each t U(ε, 1 K n(t,xdµ n (x = 1; (K.2 There is a sequence δ n such that [ t+δn ] R n := sup 1 K n (t,xdµ n (x, as n ; t U(ε t δ n (K.3 For any function g(, that is at least three times differentiable in ],1[, ĝ n (t := 1 g(xk n (t,xdµ n (x, is differentiable in t on U(ε, and 1 sup g(t g(xk n (t,xdµ n (x = O( n α, α > ; t U(ε sup t U(ε g (t d dt 1 g(xk n (t,xdµ n (x = O( n β, β >. Note that the conditions (Q and (K.1-2-3, which we will use to establish the convergence in probability of Ĥ ε;n (X, match those used by Cheng and Parzen (1997 to establish the asymptotic properties of q n (. Our result concerning consistency of Ĥε;n(X defined in (1.5, where the qdf estimator is given in (1.7, is as follows. 5

6 Theorem 2.1 Let X 1,...,X n be i.i.d. r.v. s with a quantile density function q( fulfilling (Q Let K n (, fulfill (K Then, we have, Ĥε;n(X H(X = O P (n 1/2 M(q;K n +n β +η(ε, (2.1 where M(q;K n := M q Mn (1 δ n logδn 1 +M q M + n (q 2 R n (1+n 1/2 A γ (nm q Mn (1, M n (g := sup u U(ε 1 R n (g := sup u U(ε M g := sup g(u, u U(ε g(xk n (u,x dµ n (x, [,1]\U(ε+δ n and δ n is given in condition (K.2. g(xk n (u,x dµ n (x, The proof of Theorem 2.1 is postponed to the Section 7. To establish the asymptotic normality of Ĥ ε;n (X, we will use the following additional condition. ( (Q.4 E {log 2 q ( F(X } <. Let, for all ε ],1/2[, ( E {log 2 q ( F(X } { εlog 2( q(ε +εlog 2( q(1 ε + 1 ε ε log 2( q(x } ( dx = o ϑ(ε, where ϑ(ε, as ε. The main result, concerning the normality of Ĥ ε;n (X, to be proved here may now be stated precisely as follows. Theorem 2.2 Assume that the conditions (Q and (K hold with α > 1/2 and β > 1/2 in (K.3. Then, we have, n( Ĥ ε;n (X H ε (X ψ n (ε = O P ( { 2εlogε 1 } 1/2 +o P (1, (2.2 where ψ n (ε := U(ε (q (x/q(xb n (xdx, is a centered Gaussian random variable with variance equal to Var ( ψ n (ε { ( = Var log q ( F(X } ( +o ϑ(ε+η(ε. 6

7 The proof of Theorem 2.2 is postponed to the Section 7. Remark 2.3 Condition (Q.4 is extremely weak and is satisfied by all commonly encountered distributions including many important heavy tailed distributions for which the moments do not exists, the interested reader may refer to Song (2 and the references therein. Remark 2.4 We mention that the condition (K.3 is satisfied, when α,β > 1/2, for the difference kernels K n (t,xdµ n (x = h 1 n k((t x/h ndx with h n = O(n ν where 1/4 < ν < 1. A typical example for differences kernels satisfying theses conditions is the gaussian kernel. 3 The smoothed Parzen estimator of entropy We mention that the notations of this section are similar to that used in Parzen (1979 and changes have been made in order to adopt it to our setting. Given a random sample X 1,...,X n of a continuous random variable X with distribution function F(. In this section we will work under the following hypothesis, for all x R, H : F(x = F ( x µ σ, where µ and σ are parameters to be estimated, it is convenient to transform X i to ( Xi µ n Û i = F, σ n where µ n and σ n are efficient estimators of µ and σ. It is more convenient to base tests of the hypothesis on the deviations from identity function t of the empirical distribution function of Û1,...,Ûn. Let denote by D n, (t, t 1 the empirical quantile function of Û1,...,Ûn; it can be expressed in terms of the sample quantile function Q n ( of the original data X 1,...,X n by where D n, (t := F ( Qn (t µ n σ n, ( ( i Q n (t = n n t X i 1;n +n t i 1 X i;n, for i 1 n n t i,i = 1,...,n. n In our case, we may consider the smoothed version of the quantile density given in Parzen (1979 and defined by ( d n, (t := d dt D Qn (t µ n d n, (t = f σ n dt Q n (t 1 σ. n 7

8 Consequently, define where d n (t := f (Q (t d dt Q n (t 1 σ,n, σ,n := 1 f (Q (t d dt Q n (tdt. This is an alternative approach to estimating q( when one desires a goodness of fit test of a location scale parametric model, we refer to Parzen (1991 for more details. We mention that in (Parzen, 1979, Section 7., the smoothed version of d n (t is defined, for the convolution-kernel, by 1 d n (t := d n (u 1 ( t u K du. h n This suggests the following estimator of entropy H n (X := 1 h n log( d n (tdt. (3.1 Under similar conditions to that of Theorem 2.1, we could derive, under H, that, as n H n (X H(X = o P (1. (3.2 For more details on goodness of fit test in connection with Shannon s entropy, the interested reader may refer to Parzen (1991. Note that the normality is an example of location-scale parametric model. Remark 3.1 Recall that there exists several ways to obtain smoothed versions of d n (. Indeed, we can choose the following smoothing method. Keep in mind the following definition q n (t = d dt Q n (t = d dt 1 Then, we can use the following estimator Finally, we estimate the entropy under H H n(x := 1 Q n (xk n (t,xdµ n (x, t ],1[. d n (t := f (Q (t q n (t 1 σ,n. log( d n(tdt. (3.3 Under conditions of Theorem 2.1, we could derive, under H, that, as n H n (X H(X = o P(1. (3.4 8

9 It will be interesting to provide a complete investigation of H n (X and Hn (X and study the test of normality of statistics based on them. This would go well beyond the scope of the present paper. Remark 3.2 The nonparametric approach, that we use to estimate the Shannon s entropy, treats the density parameter-free and thus offers the greatest generality. This is in contrast with the parametric approach where one assumes that the underlying distribution follows a certain parametric model, and the problem reduces to estimating a finite number of parameters describing the model. 4 A comparison study by simulations Assuming that X 1,...,X n, is the sample, the estimator proposed by Vasicek (1976 is given by H m,n (V := 1 n ( n log n 2m {X i+m;n X i m;n }, i=1 where m is a positive integer fulfilling m n/2. Vasicek proved that his estimator is consistent, i.e., H (V m,n P H(X, as n, m and m/n. Vasicek s estimator is also asymptotically normal under appropriate conditions on m and n. van Es (1992 proposed a new estimator of entropy given by H (Van m,n := 1 n m n + k=m n m i=1 log ( n+1 m {X i+m;n X i;n } 1 k +log(m log(n+1. van Es (1992 established, under suitable conditions, the consistency and asymptotic normality of this estimator. Correa (1995 suggested a modification of Vasicek s estimator. In order to estimate the density f( of F( in the interval (X i m;n,x i+m;n he used a local linear model based on 2m+1 points: F(X j;n = α+βx j;n,j = m i,...,m+i. This produces a smaller mean squared error estimator. The Correa s estimator is defined by H m,n (Cor := 1 n m ( i+m } j=i m{ Xj;n X (i (j i log n n i+m { }, j=i m Xj;n X (i i=1 9

10 where X (i := 1 2m+1 i+m j=i m X j;n. Here, m n/2 is a positive integer, X i;n = X 1;n for i < 1 and X i;n = X n;n for i > n. By simulations, Correa showed that, for some selected distributions, his estimator produces smaller mean squared error than Vasicek s estimator and van Es estimator. Wieczorkowski and Grzegorzewski (1999 modified the Vasicek s estimator by adding a bias correction. Their estimator is given by H (WG m,n := H (V m,n logn+log(2m +Ψ(n+1 2 n ( 1 2m n m Ψ(i+m 1, i=1 Ψ(2m where Ψ(x is the digamma function defined by (see e.g., Gradshteyn and Ryzhik (27 Ψ(x := dlogγ(x dx = Γ (x Γ ( x. For integer arguments, we have Ψ(k = k 1 i=1 1 i γ, where γ = is the Euler s constant. A series of experiments were conducted in order to compare the performance of our estimator, in terms of efficiency and robustness, with the following entropy estimators : Vasicek s estimator, van Es estimator, Correa s estimator and Wieczorkowski-Grzegorewski s estimator. We provide numerical illustrations regarding the mean squared error of each of the above estimators of the entropy H(X. The computing program codes were implemented in R. We have considered the following distributions. For each distribution, we give the value of H(X. Standard normal distribution N(,1: H(X = log(σ 2πe = log( 2πe. Uniform distribution: H(X =. Weibull distribution with the shape parameter equal to 2 and the scale parameter equal to.5: H(X =

11 Recall that the entropy for Weibull distribution, with the shape parameter k > and the scale parameter λ >, is given by ( H(X = γ 1 1 ( λ +log +1. k k Exponential distribution with parameter λ = 1: H(X = 1 log(λ = 1. Student s t-distribution with the number of degrees of freedom equal to 1, 3 and 5: H(X = , (degrees of freedom equal to 1 (Cauchy distribution, H(X = , (degrees of freedom equal to 3, H(X = , (degrees of freedom equal to 5. Keeping in mind that the entropy for Student s t-distribution, with the number of degrees of freedom ν, is given by H(X = ν +1 ( ( 1+ν ( ν ( ( νbeta ν Ψ Ψ +log , 1, 2 where Ψ( is digamma function. For sample sizes n = 1, n = 2 and n = 5; 5 samples were generated. All the considered spacing-based estimators depend on the parameter m n/2; the optimal choice of m given to the sample size n is still an open problem. Here, we have chosen the parameter m according to Correa (1995, where m = 3 for n = 1 and n = 2, and m = 4 for n = 5. For our estimator Ĥǫ,n(X, we have used the standard gaussian kernel, and the choice of the bandwidth is done by numerical optimization of the MSE of Ĥ ε;n (X with respect to h n. In Tables 1, 2 and 3 we have considered normal distribution N(,1 with sample sizes n = 1, n = 2 and n = 5. In Tables 4-9, we have considered the uniform distribution, Weibull distribution, exponential distribution with parameter 1, Student t-distribution with parameter 3 and 5 and Cauchy distribution, respectively, and sample size n = 5. In all cases, and for each considered estimator, we compute the bias, variance and MSE by Monte-Carlo through the 5 replications. From Tables 1-9, we can see that our estimator works better than all the others, in the sense that the MSE is smaller. We think that the simulation results may be substantially ameliorated if we choose for example the Beta or Gamma kernel, which have the advantage to take into account the boundary effects. It appears that our estimator, based on the smoothed quantile density estimate, behaves better than the traditional ones in term of 11

12 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 1: Results for n = 1, m = 3, h n =.157, Normal distribution N(,1 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 2: Results for n = 2, m = 3, h n =.81, Normal distribution N(,1 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 3: Results for n = 5, m = 4, h n =.333, Normal distribution N(,1 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 4: Results for n = 5, m = 4, h n =.522, Uniform distribution efficiency. We turn now to compare robustness property of the above estimators of the entropy H(X. The robustness here is to be stand versus contamination and not in terms of influence function or break-down points which make sense only under parametric or semiparametric 12

13 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 5: Results for n = 5, m = 4, h n =.614, Weibull distribution with the shape parameter equal to 2 and the scale parameter equal to.5 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 6: Results for n = 5, m = 4, h n =.712, Exponential distribution with parameter 1 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 7: Results forn = 5, m = 4, h n =,336, Student s t-distribution with the number of degrees of freedom equal to 3 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 8: Results for n = 5, m = 4, h n =.344, Student s t-distribution with the number of degrees of freedom equal to 5 13

14 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 9: Results for n = 5, m = 4, h n =.235, Cauchy distribution settings. According to Huber and Ronchetti (29: Most approaches to robustness are based on the following intuitive requirement: A discordant small minority should never be able to override the evidence of the majority of the observations. In the same reference, it is mentioned that resistant statistical procedure, i.e., if the value of the estimate is insensitive to small changes in the underlying sample, is equivalent to robustness for practical purposes in view of Hampel s theorem, refer to (Huber and Ronchetti, 29, Section and Section 2.6. Typical examples for the notion of robustness in the nonparametric setting are the sample mean and the sample median which are the nonparametric estimates of the population mean and median, respectively. Although nonparametric, the sample mean is highly sensitive to outliers and therefore for symmetric distribution and contaminated data the sample median is more appropriate to estimate the population mean or median, refer to Huber and Ronchetti (29 for more details. In our simulation, we will consider data generated from N(, 1 distribution where a small proportion ǫ of observations were replaced by atypical ones generated from a contaminating distribution F (. We consider two cases with ǫ = 4% and ǫ = 1%, and we choose the contaminating distribution F ( to be the uniform distribution on [,1], the results are presented in Table 1 with ǫ = 4% and Table 11 with ǫ = 1%. Let f( denote the density function of N(,1 and f ( the density of the contaminating distribution F (. The contaminated sample, X 1,...,X n, can be seen as if it has been generated from the density f ǫ ( := (1 ǫf( +ǫf (. Since the sample is contaminated, all the above estimators tend to H(f ǫ and not to H(f. The objective here is to obtain the best (in the sense that the corresponding MSE is the smallest estimate of the entropy H(f from the contaminated data X 1,...,X n. We will consider the five estimates as above, and we compute their bias, variance and MSE by Monte-Carlo using 5 replications. From Tables 1-11, we can see that our estimator is the best one. Remark 4.1 To understand well the behavior of the proposed estimator, it will be interesting to consider several family of distributions: 14

15 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 1: Results for n = 5, m = 4, h n =.333, ǫ = 4%, Normal distribution N(,1 Estimate bias Variance MSE H m,n (V H m,n (Van H m,n (Cor H m,n (WG Ĥ ε;n (X Table 11: Results for n = 5, m = 4, h n =.333, ǫ = 1%, Normal distribution N(,1 1. Distribution with support (, and symmetric. 2. Distribution with support (, and asymmetric. 3. Distribution with support (,. 4. Distribution with support ],1[. Extensive simulations for these families will be undertaken elsewhere. In the following remark, we give a way how the choose the smoothing parameter in practice. Remark 4.2 The choice of the smoothing parameter plays an instrumental role in implementation of practical estimation. We recall that the smoothing parameter h n, in our simulations, has been chosen to minimize the MSE of the estimator, assuming that the underlying distribution is known. In more general case, without assuming any knowledge on the underlying distribution function, one can use, among others, the selection procedure proposed in Jones (1992. Jones (1992 derived that the asymptotic MSE of q n (, in the case of convolution-kernel estimator, which is given by { } 2 AMSE( q n (t = h4 n 4 q 2 (t x 2 K(xdx + 1 nh n q 2 (t R R K 2 (xdx. 15

16 Minimizing the last equation with respect to h n, we find the asymptotically optimal bandwidth for q n ( as { h opt q(t 2 } 1/5 R n = K2 (xdx n(q 2 (t { R x2 K(xdx } 2. Note that h opt n depends on the unknown functions and q(t = 1 f(q(t, q (t = f (Q(tf(Q(t 3[f (Q(t] 2. f 5 (Q(t These functions may be estimated in the classical way, refer to Jones (1992, Silverman (1986 and Cheng and Sun (26 for more details on the subject and the references therein. Another way to estimate the optimal value of h n is to use a cross-validation type method. Remark 4.3 The main problem in using entropy estimates such as in (1.5 is to choose properly the smoothing parameter h n. With a lot more effort, we could derive analog results here for Ĥε;n(X using the methods in Bouzebda and Elhattab (29, 21, 211, as well as the modern empirical process tools developed in Einmahl and Mason (25 in their work on uniform in bandwidth consistency of kernel-type estimators. 5 Test for normality A well-known theorem of information theory [see, e.g., p. 55, Shannon (1948] states that among all distributions that possess a density function f( and have a given variance σ 2, the entropy H(X is maximized by the normal distribution. The entropy of the normal distribution with variance σ 2 is log ( σ 2πe. As pointed out by Vasicek (1976, this property can be used for tests of normality. Towards this aim, one can use the estimate Ĥ ε;n (X of H(X, as follows. Let X 1,...,X n be independent random replicæ of a random variable X with quantile density function q(. Let Ĥε;n(X the estimator of H(X as in (1.5. Let T n denote the statistic T n := log( 2πσ 2 n +.5 Ĥε;n(X, (5.1 where σ 2 n is the sample standard deviation based on X 1,...,X n, defined by σ 2 n := 1 n n ( 2, Xi X n i=1 16

17 where n X n := n 1 X i. The normality hypothesis will be rejected whenever the observed value of T n will be significantly less than, in the sense we precise below. The exact critical values T α of T n at the significance levels α ],1[ are defined by the equation P(T n T α = α. The distribution of T n under the null hypothesis cannot readily be obtained analytically. To evaluate the critical values T α, we have used a Monte Carlo method, for sample sizes 1 n 5 and the significance value given by α =.5. Namely, for each n 5, we simulate 2 samples of size n from the standard normal distribution. Since α =.5 = 1/2, we determine the 1-th order statistic t 1,2 and obtain the critical value T n,.5 through the equation T n,.5 = t 1,2. Our results are presented in Table 12. i=1 Sample size Percentage level n Table 12: Critical Values of T n. Park and Park (23 establish the entropy-based goodness of fit test statistics based on the nonparametric distribution functions of the sample entropy and modified sample entropy Ebrahimi et al. (1994, and compare their performances for the exponential and normal distributions. The authors consider where H Ebr m,n := 1 n n log i=1 ( n c i m {X i+m;n X i m;n } 1+ i 1 if 1 i m, m c i := 2 if m+1 i n m, 1+ n i if n m i n. n Yousefzadeh and Arghami (28 use a new cdf estimator to obtain a nonparametric entropy estimate and use it for testing exponentially and normality, they introduce the, 17

18 following estimator ( n Hm,n You X i+m;n X i m;n := log F n (X i+m;n F n (X i m;n where i=1 F n (X i+m;n F n (X i m;n n i=1 ( F n (X i+m;n F n (X i m;n, F n (x := n 1 n(n+1 ( n n 1 + x X ;n X 2;n X ;n + x X 1;n X 3;n X 1;n ( n 1 i+ 1 + x X i 1;n n(n+1 n 1 X i+1;n X i 1;n + x X i;n X i+2;n X i;n ( n 1 n x X n 2;n n(n+1 n 1 X n;n X n 2;n + x X n 1;n X n+1;n X n 1;n for X 1;n x X 2;n, for X i;n x X i+1;n, i = 2,...,n 2, for X n 1;n x X n,n, and X ;n := X 1;n n n 1 (X 2;n X 1;n, X n+1;n := X n;n n n 1 (X n;n X n 1;n. To compare the power of the proposed test a simulation was conducted. We used tests based on the following entropy estimators: H m,n (WG, Hm,n Ebr, HYou m,n and our test (5.1. In the power comparison, we consider the following alternatives. The Weibull distribution with density function ( f(x;λ,k = k ( ( x k x exp 1{x > }, λ λ λ where k > is the shape parameter and λ > is the scale parameter of the distribution and where 1{ } stands for the indicator function of the set { }. The uniform distribution with density function f(x = 1, < x < 1. The Student t-distribution with density function f(x;ν = Γ((ν +1/2 Γ(ν/2 1 1 νπ (1+x 2 /ν(ν+1/2, ν > 2, < x <. Remark 5.1 We mention the value taken in Table 13 for the statistics based on Hm,n Ebr and Hm,n You are the same to that calculated in Table 6, p of Yousefzadeh and Arghami (28, for the same alternatives that we consider in our comparison. 18

19 Statistics based on Alternatives Ĥ ε;n (X H (WG 4,n H (WG 8,n Hm,n Ebr Hm,n You Uniform Weibull( Student t Student t Table 13: Power estimate of.5 tests against alternatives of the normal distribution based on 2 replications for sample size n=5. Alternatives h n Uniform h n =.5297 Weibull(2 h n =.6555 Student t 5 h n =.31 Student t 3 h n =.189 Table 14: Choice of smoothing parameter for Ĥε;n(X. In Table 13 we have reported the results of power comparison for the sample size is 5. We made 2 Monte-Carlo simulations to compare the powers of the proposed test against 4 alternatives. From Table 13, we can see that the proposed test T n shows better power than all other statistics for the alternative that we consider. It is natural that our test has good performances for unbounded support since the kernel-type estimators behave well in this situation. 6 Concluding remarks and future works We have proposed a new estimator of entropy based on the kernel-quantile density estimator. Simulations show that this estimator behaves better than the other competitors both under contamination or not. More precisely, the MSE is consistently smaller than the MSE of the spacing-based estimators considered in our study. It will be interesting to compare theoretically the power and the robustness of the test of normality, based on the proposed estimator of the present paper, with those considered in Esteban et al. (21. The study of entropy in presence of censored data is possible using a similar methodology as presented here. It would be interesting to provide a complete investigation of the choice of the parameter h n for kernel difference-type estimator which requires nontrivial mathematics, this would go well beyond the scope of the present paper. The problems and the methods described here all are inherently univariate. A natural and useful multivariate extension is the use of copula function. We propose to extend the 19

20 results of this paper to the multivariate case in the following way. Consider a R d -valued random vector X = (X 1,...,X d with joint cdf and marginal cdf s F(x := F(x 1,...,x d := P(X 1 x 1,...,X d x d F j (x j := P(X j x j for j = 1,...,d. If the marginal distribution functions F 1 (,...,F d ( are continuous, then, according to Sklar s theorem [Sklar (1959], the copula function C(, pertaining to F(, is unique and C(u := C(u 1,...,u d := F(Q 1 (u 1,...,Q d (u d, for u [,1] d, where, forj = 1,...,d,Q j (u := inf{x : F j (x u} withu (,1],Q j ( := lim t Q j (t := Q j ( +, is the quantile function of F j (. The differential entropy may represented via density copula c( and quantile densities q i ( as follows H(X = d H(X i +H(c, (6.1 i=1 where H(X i = [,1] log ( q i (u du, (6.2 q i (u = dq i (u/du, for i = 1,...,d and H(c is the copula entropy defined by H(c = c(f 1 (x 1,...,F d (x d log ( c(f 1 (x 1,...,F d (x d dx R d = c(u 1,...,u d log ( c(u 1,...,u d du [,1] d = c(ulog ( c(u du, (6.3 [,1] d where is the copula density. c(u 1,...,u d = d u 1... u d C(u 1,...,u d 7 Proof. This section is devoted to the proofs of our results. The previously defined notation continues to be used below. Proof of Theorem 2.1. Recall that we set for all ε ],1/2[, U(ε = [ε,1 ε], and H ε (X = εlog ( q(ε + log ( q(x dx+εlog ( q(1 ε. U(ε 2

21 Therefore, by (1.4, we have the following Ĥε;n(X H(X Ĥε;n(X H ε (X + H ε (X H(X = Ĥε;n(X H ε (X +o ( η(ε. (7.1 We evaluate the first term of the right hand of (7.1. By the triangular inequality, we have Ĥε;n(X H ε (X = ε( log ( q n (ε log ( q(ε + (log ( q n (x log ( q(x dx We note that for z >, Therefore, we have U(ε +ε ( log ( q n (1 ε log ( q(1 ε ( ε log ( qn (ε log ( q(ε + (log ( q n (x log ( q(x dx U(ε + ε ( log ( qn (1 ε log ( q(1 ε. logz z 1 + 1/z 1. log( qn (x log ( q(x ( qn (x = log q(x q n(x q(x q(x + q n(x q(x. q n (x Under conditions (Q and (K.1-2-3, we may apply Theorem 2.2 in Cheng and Parzen (1997, for all fixed ε ],1/2[ and recall M(q;K n defined in Theorem 2.1, we have, as n, ( sup q n (x q(x = O P n 1/2 M(q;K n +n, β (7.2 x U(ε we infer that we have, uniformly over x U(ε, q n (x (1/2q(x, for all n enough large. This fact implies Ĥε;n(X H ε (X 4 sup q n (x q(x, x U(ε in probability, which implies that Ĥε;n(X H ε (X = O P (n 1/2 M(q;K n +n β +η(ε. Thus the proof is complete. Proof of Theorem 2.2. Throughout, we will make use of the fact (see e.g. Csörgő and Révész 21

22 (1978, 1981 that we can definex 1,...,X n on a probability space which carries a sequence of Brownian bridges {B n (t : t [,1],n 1}, such that, n(qn (t Q(t q(tb n (t = q(te n (t (7.3 and [ ] n 1/2 lim sup q(te n (t C, (7.4 n A γ (n with probability one, where C is a universal constant and { logn, max{q(,q(1} < or γ 2, A γ (n := (loglogn γ (logn (1+ν(1 γ γ > 2, with γ in (Q.2 and an arbitrary positive ν. Using Taylor expansion we get, for λ ],1[, n( Ĥ ε;n (X H ε (X = n(log ( qn (x log ( q(x dx = U(ε + nε(log ( q n (ε log ( q(ε + nε(log ( q n (1 ε log ( q(1 ε ( q n (x q(x n dx+l n (ε λ q n (x+(1 λq(x U(ε =: T n (ε+l n (ε. (7.5 We have, under our assumptions, q n (x q(x in probability, which implies that λ q n (x+ (1 λq(x q(x in probability. Thus, we have, by Slutsky s theorem, as n tends to the infinity, T n (ε has the same limiting law of T n (ε := (1/q(x n( q n (x q(xdx. (7.6 U(ε We have the following decomposition T n (ε = (1/q(x n d ( 1 Q n (vk n (x,vdµ n (v Q(x dx U(ε dx = (1/q(x d ( 1 n(qn (v Q(vK n (x,vdµ n (v dx U(ε dx (1/q(x n d ( 1 Q(vK n (x,vdµ n (v Q(x dx dx U(ε 22

23 Using condition (K.3 and we have the following (1/q(x n d ( 1 dx U(ε U(ε (1/q(xdx 1, Q(vK n (x,vdµ n (v Q(x dx = O(n1/2 β. This, in turn, implies, in connection with (7.3, that T n (ε = (1/q(x d ( 1 (q(vb n (v+q(ve n (vk n (x,vdµ n (v dx dx U(ε +O(n 1/2 β. Recalling the arguments of the proof of Lemma 3.2. of Cheng (1995, for I n (u = [u δ n,u+δ n ] and I c n (u = [,1]\I n(u, we can show that (1/q(x d ( 1 q(ve n (vk n (x,vdµ n (v dx U(ε dx ( (1/q(xdx sup d 1 q(ve n (vk n (x,vdµ n (v. dx U(ε x U(ε Then, it follows (1/q(x d ( 1 q(ve n (vk n (x,vdµ n (v dx U(ε dx ( 1 sup q(v e n(v d x U(ε dx K n(x,v dµ n(v ( 1 sup e n (v q(v d x U(ε dx K n(x,v dµ n(v ( 1 Cn 1/2 A γ (n q(v d dx K n(x,v dµ n(v ( Cn 1/2 A γ (n sup q(x sup d x U(ε δ n x U(ε I n(x dx K n(x,v dµ n(v +Cn 1/2 A γ (n sup d x U(ε dx q(vk n(x,v dµ n(v =: R (1 n (ε+r (2 n (ε. I c n (x 23

24 Since q( is continuous on ],1[, then it is bounded on compact interval U(ε δ n ],1[, which gives, by condition (K.2 R (1 n (ε Cn 1/2 A γ (n sup q(x x U(ε δ n = Cn 1/2 A γ (no(1 = o(1 = o P (1. Let s = 1 + δ with δ > arbitrary, and let r = 1 1/s. Then Hölder s inequality implies { [ [q(v] r d ] 1/r [,1] dx K n(x,vdµ n (v R (2 n (ε Cn 1/2 A γ (n sup x U(ε [ 1 s In(x (v d c [,1] dx K n(x,vdµ n (v Under condition (K.3, we can see that [ sup x U(ε sup [q(v] r d [,1] x U(ε = O(n β. q(x ] 1/r dx K n(x,vdµ n (v [ [q(v] r d [,1] dx K n(x,vdµ n (v This in connection with condition (K.2, implies ( R (2 n (ε = Cn 1/2 A γ (no(1o = o P (1. sup Rn 1/(1/δ x U(ε ] 1/s }. ] 1/r = o(1 Hence U(ε (1/q(x d ( 1 q(ve n (vk n (x,vdµ n (v dx Then, we conclude that T n (ε = (1/q(x d ( 1 dx U(ε +o P (1. dx = o P(1. q(vb n (vk n (x,vdµ n (v dx We have by Cheng and Parzen (1997, p. 297, (we can refer also to Stadtmüller (1988, 1986 and Xiang (1994 for related results on smoothed Wiener process and Brownian 24

25 bridge sup x U(ε 1 sup x U(ε + sup q(vb n (vk n (x,vdµ n (v q(xb n (x 1 x U(ε q(v B n (v B n (x K n (x,vdµ n (v B n(x 1 q(v q(x K n (x,vdµ n (v +OP (n β ( = O P ((2δ n logδn 1 1/2 +O P R 1/(1+δ n = o P (1. This gives, under condition (Q.1, (1/q(x d ( 1 q(vb n (vk n (x,vdµ n (v dx U(ε dx ( 1 1 ε = (1/q(x q(vb n (vk n (x,vdµ n (v ε ( ( d 1 U(ε dx (1/q(x q(vb n (vk n (x,vdµ n (v dx ( = (1/q(xq(xB n (x 1 ε d ε U(ε dx (1/q(x q(xb n (xdx+o P (1 = (q(x /q(xb n (xdx+b n (1 ε B n (ε+o P (1. U(ε Making use of (Csörgő and Révész, 1981, Theorem 1.4.1, for ε sufficiently small, and B n (1 ε B(1 = o P (1 B n (ε B( = o P (1. Which implies that T n (ε = U(ε (q (x/q(xb n (xdx+o P (1. By (Cheng and Parzen, 1997, Theorem 2.2., we have, for β > 1/2 (β in condition (K.3, sup n(log ( qn (u log(q(u = o P (1, (7.7 u U(ε which implies, by using Taylor expansion, that L n (ε = o P (1, (7.8 25

26 where L n (ε is given in (7.5. This gives T n (ε+l n (ε = (q(x /q(xb n (xdx+o P (1. U(ε It follows, n( Ĥ ε;n (X H ε (X = d U(ε (q (x/q(xb n (xdx+o P (1. Here and elsewhere we denote by = d the equality in distribution. Note that (q (x/q(xb n (xdx U(ε is a gaussian random variable with mean ( E (q (u/q(ub n (udu = U(ε =, U(ε (q (u/q(ue(b n (udu and variance ( E (q (u/q(ub n (udu (q (v/q(vb n (vdv U(ε ( U(ε = E (q (u/q(u(q (v/q(vb n (ub n (vdudv U(ε U(ε = (q (u/q(u(q (v/q(ve(b n (ub n (vdudv U(ε U(ε = (q (u/q(u(q (v/q(v(min(u,v uvdudv U(ε U(ε ( v 1 = (q (v/q(v (1 v (q (u/q(uudu+v (q (u/q(u(1 udu dv U(ε v = εlog(q(εlog(q(1 ε+εlog 2 (q(ε log(q(1 ε log(q(xdx = = ( +H ε (X log(q(1 ε H ε (X log 2 (q(xdx+εlog 2 (q(ε+εlog 2 (q(1 ε A 2 (ε U(ε ( = Var log 2 (q(xdx+o [,1] { log ( ϑ(ε ( q ( F(X } ( +o ϑ(ε+η 2 (ε. ( ( 2 H(X+o η(ε U(ε Thus the proof is complete. 26

27 References Ahmad, I. A. and Lin, P. E. (1976. A nonparametric estimation of the entropy for absolutely continuous distributions. IEEE Trans. Information Theory, IT-22(3, Beirlant, J., Dudewicz, E. J., Györfi, L., and van der Meulen, E. C. (1997. Nonparametric entropy estimation: an overview. Int. J. Math. Stat. Sci., 6(1, Bouzebda, S. and Elhattab, I. (29. A strong consistency of a nonparametric estimate of entropy under random censorship. C. R. Math. Acad. Sci. Paris, 347(13-14, Bouzebda, S. and Elhattab, I. (21. Uniform in bandwidth consistency of the kernel-type estimator of the shannon s entropy. C. R. Math. Acad. Sci. Paris, 348(5-6, Bouzebda, S. and Elhattab, I. (211. Uniform-in-bandwidth consistency for kernel-type estimators of Shannon s entropy. Electron. J. Stat., 5, Cheng, C. (1995. Uniform consistency of generalized kernel estimators of quantile density. Ann. Statist., 23(6, Cheng, C. and Parzen, E. (1997. Unified estimators of smooth quantile and quantile density functions. J. Statist. Plann. Inference, 59(2, Cheng, M.-Y. and Sun, S. (26. Bandwidth selection for kernel quantile estimation. J. Chin. Statist. Assoc., 44, Correa, J. C. (1995. A new estimator of entropy. Comm. Statist. Theory Methods, 24(1, Cover, T. M. and Thomas, J. A. (26. Elements of information theory. Wiley- Interscience [John Wiley & Sons], Hoboken, NJ, second edition. Csörgő, M. and Révész, P. (1978. Strong approximations of the quantile process. Ann. Statist., 6(4, Csörgő, M. and Révész, P. (1981. Strong approximations in probability and statistics. Probability and Mathematical Statistics. Academic Press Inc. [Harcourt Brace Jovanovich Publishers], New York. Csörgő, M. and Révész, P. (1984. Two approaches to constructing simultaneous confidence bounds for quantiles. Probab. Math. Statist., 4(2,

28 Csörgő, M., Horváth, L., and Deheuvels, P. (1991. Estimating the quantile-density function. In Nonparametric functional estimation and related topics (Spetses, 199, volume 335 of NATO Adv. Sci. Inst. Ser. C Math. Phys. Sci., pages Kluwer Acad. Publ., Dordrecht. Csiszár, I. (1962. Informationstheoretische Konvergenzbegriffe im Raum der Wahrscheinlichkeitsverteilungen. Magyar Tud. Akad. Mat. Kutató Int. Közl., 7, Dmitriev, J. G. and Tarasenko, F. P. (1973. The estimation of functionals of a probability density and its derivatives. Teor. Verojatnost. i Primenen., 18, Dudewicz, E. J. and van der Meulen, E. C. (1981. Entropy-based tests of uniformity. J. Amer. Statist. Assoc., 76(376, Ebrahimi, N., Habibullah, M., and Soofi, E. (1992. Testing exponentiality based on Kullback-Leibler information. J. Roy. Statist. Soc. Ser. B, 54(3, Ebrahimi, N., Pflughoeft, K., and Soofi, E. S. (1994. Two measures of sample entropy. Statist. Probab. Lett., 2(3, Einmahl, U. and Mason, D. M. (25. Uniform in bandwidth consistency of kernel-type function estimators. Ann. Statist., 33(3, Esteban, M. D., Castellanos, M. E., Morales, D., and Vajda, I. (21. Monte Carlo comparison of four normality tests using different entropy estimates. Comm. Statist. Simulation Comput., 3(4, Gallager, R. (1968. Information theory and reliable communication. New York-London- Sydney-Toronto: John Wiley & Sons, Inc. XVI, 588 p. Gokhale, D. (1983. On entropy-based goodness-of-fit tests. Comput. Stat. Data Anal., 1, Gokhale, D. V. and Kullback, S. (1978. The information in contingency tables, volume 23 of Statistics: Textbooks and Monographs. Marcel Dekker Inc., New York. Gradshteyn, I. S. and Ryzhik, I. M. (27. Table of integrals, series, and products. Elsevier/Academic Press, Amsterdam, seventh edition. Falk, M. (1986. On the estimation of the quantile density function. Statist. Probab. Lett., 4(2, Huber, P. J. and Ronchetti, E. M. (29. Robust statistics. Wiley Series in Probability and Statistics. John Wiley & Sons Inc., Hoboken, NJ, second edition. 28

29 Jaynes, E. T. (1957. Information theory and statistical mechanics. Phys. Rev. (2, 16, Jones, M. (1992. Estimating densities, quantiles, quantile densities and density quantiles. Ann. Inst. Stat. Math., 44(4, Kullback, S. (1959. Information theory and statistics. John Wiley and Sons, Inc., New York. Lazo, A. C. and Rathie, P. N. (1978. On the entropy of continuous probability distributions. IEEE Trans. Inf. Theory, 24, Park, S. and Park, D. (23. Correcting moments for goodness of fit tests based on two entropy estimates. J. Stat. Comput. Simul., 73(9, Parzen, E. (1979. Nonparametric statistical data modeling. J. Amer. Statist. Assoc., 74(365, With comments by John W. Tukey, Roy E. Welsch, William F. Eddy, D. V. Lindley, Michael E. Tarter and Edwin L. Crow, and a rejoinder by the author. Parzen, E. (1991. Goodness of fit tests and entropy. J. Combin. Inform. System Sci., 16(2-3, Prescott, P. (1976. On a test for normality based on sample entropy. J. R. Stat. Soc., Ser. B, 38, Rényi, A. (1959. On the dimension and entropy of probability distributions. Acta Math. Acad. Sci. Hungar., 1, (unbound insert. Sklar, A. (1959. Fonctions de répartition à n dimensions et leurs marges. Publ. Inst. Statist. Univ. Paris, 8, Shannon, C. E. (1948. A mathematical theory of communication. Bell System Tech. J., 27, , Silverman, B. W. (1986. Density estimation for statistics and data analysis. Monographs on Statistics and Applied Probability. Chapman & Hall, London. Song, K.-S. (2. Limit theorems for nonparametric sample entropy estimators. Statist. Probab. Lett., 49(1, Stadtmüller, U. (1986. Asymptotic properties of nonparametric curve estimates. Period. Math. Hungar., 17(2, Stadtmüller, U. (1988. Kernel approximations of a Wiener process. Period. Math. Hungar., 19(1,

30 van Es, B. (1992. Estimating functionals related to a density by a class of statistics based on spacings. Scand. J. Statist., 19(1, Vasicek, O. (1976. A test for normality based on sample entropy. J. Roy. Statist. Soc. Ser. B, 38(1, Wieczorkowski, R. and Grzegorzewski, P. (1999. Entropy estimators improvements and comparisons. Comm. Statist. Simulation Comput., 28(2, Xiang, X. (1994. A law of the logarithm for kernel quantile density estimators. Ann. Probab., 22(2, Yousefzadeh, F. and Arghami, N. R. (28. Testing exponentiality based on type II censored data and a new cdf estimator. Comm. Statist. Simulation Comput., 37(8-1,

Department of Statistics, School of Mathematical Sciences, Ferdowsi University of Mashhad, Iran.

Department of Statistics, School of Mathematical Sciences, Ferdowsi University of Mashhad, Iran. JIRSS (2012) Vol. 11, No. 2, pp 191-202 A Goodness of Fit Test For Exponentiality Based on Lin-Wong Information M. Abbasnejad, N. R. Arghami, M. Tavakoli Department of Statistics, School of Mathematical

More information

Testing Goodness-of-Fit for Exponential Distribution Based on Cumulative Residual Entropy

Testing Goodness-of-Fit for Exponential Distribution Based on Cumulative Residual Entropy This article was downloaded by: [Ferdowsi University] On: 16 April 212, At: 4:53 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 172954 Registered office: Mortimer

More information

ENTROPY-BASED GOODNESS OF FIT TEST FOR A COMPOSITE HYPOTHESIS

ENTROPY-BASED GOODNESS OF FIT TEST FOR A COMPOSITE HYPOTHESIS Bull. Korean Math. Soc. 53 (2016), No. 2, pp. 351 363 http://dx.doi.org/10.4134/bkms.2016.53.2.351 ENTROPY-BASED GOODNESS OF FIT TEST FOR A COMPOSITE HYPOTHESIS Sangyeol Lee Abstract. In this paper, we

More information

Goodness of Fit Test and Test of Independence by Entropy

Goodness of Fit Test and Test of Independence by Entropy Journal of Mathematical Extension Vol. 3, No. 2 (2009), 43-59 Goodness of Fit Test and Test of Independence by Entropy M. Sharifdoost Islamic Azad University Science & Research Branch, Tehran N. Nematollahi

More information

Kullback-Leibler Designs

Kullback-Leibler Designs Kullback-Leibler Designs Astrid JOURDAN Jessica FRANCO Contents Contents Introduction Kullback-Leibler divergence Estimation by a Monte-Carlo method Design comparison Conclusion 2 Introduction Computer

More information

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels

Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Chanseok Park 1 Department of Mathematical Sciences, Clemson University, Clemson, SC 29634 Short Title: Smooth

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,

More information

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION (REFEREED RESEARCH) COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION Hakan S. Sazak 1, *, Hülya Yılmaz 2 1 Ege University, Department

More information

ISSN Asymptotic Confidence Bands for Density and Regression Functions in the Gaussian Case

ISSN Asymptotic Confidence Bands for Density and Regression Functions in the Gaussian Case Journal Afrika Statistika Journal Afrika Statistika Vol 5, N,, page 79 87 ISSN 85-35 Asymptotic Confidence Bs for Density egression Functions in the Gaussian Case Nahima Nemouchi Zaher Mohdeb Department

More information

PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS

PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS PARAMETER ESTIMATION FOR THE LOG-LOGISTIC DISTRIBUTION BASED ON ORDER STATISTICS Authors: Mohammad Ahsanullah Department of Management Sciences, Rider University, New Jersey, USA ahsan@rider.edu) Ayman

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

A measure of radial asymmetry for bivariate copulas based on Sobolev norm

A measure of radial asymmetry for bivariate copulas based on Sobolev norm A measure of radial asymmetry for bivariate copulas based on Sobolev norm Ahmad Alikhani-Vafa Ali Dolati Abstract The modified Sobolev norm is used to construct an index for measuring the degree of radial

More information

MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES

MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES J. Korean Math. Soc. 47 1, No., pp. 63 75 DOI 1.4134/JKMS.1.47..63 MOMENT CONVERGENCE RATES OF LIL FOR NEGATIVELY ASSOCIATED SEQUENCES Ke-Ang Fu Li-Hua Hu Abstract. Let X n ; n 1 be a strictly stationary

More information

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence

More information

arxiv: v1 [stat.me] 26 May 2014

arxiv: v1 [stat.me] 26 May 2014 arxiv:1405.6751v1 [stat.me] 26 May 2014 THE WEAK LIMITING BEHAVIOR OF THE DE HAAN RESNICK ESTIMATOR OF THE EXPONENT OF A STABLE DISTRIBUTION GANE SAMB LO Abstract. The problem of estimating the exponent

More information

VARIANCE REDUCTION BY SMOOTHING REVISITED BRIAN J E R S K Y (POMONA)

VARIANCE REDUCTION BY SMOOTHING REVISITED BRIAN J E R S K Y (POMONA) PROBABILITY AND MATHEMATICAL STATISTICS Vol. 33, Fasc. 1 2013), pp. 79 92 VARIANCE REDUCTION BY SMOOTHING REVISITED BY ANDRZEJ S. KO Z E K SYDNEY) AND BRIAN J E R S K Y POMONA) Abstract. Smoothing is a

More information

The Convergence Rate for the Normal Approximation of Extreme Sums

The Convergence Rate for the Normal Approximation of Extreme Sums The Convergence Rate for the Normal Approximation of Extreme Sums Yongcheng Qi University of Minnesota Duluth WCNA 2008, Orlando, July 2-9, 2008 This talk is based on a joint work with Professor Shihong

More information

Some New Information Inequalities Involving f-divergences

Some New Information Inequalities Involving f-divergences BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 12, No 2 Sofia 2012 Some New Information Inequalities Involving f-divergences Amit Srivastava Department of Mathematics, Jaypee

More information

Bickel Rosenblatt test

Bickel Rosenblatt test University of Latvia 28.05.2011. A classical Let X 1,..., X n be i.i.d. random variables with a continuous probability density function f. Consider a simple hypothesis H 0 : f = f 0 with a significance

More information

Bahadur representations for bootstrap quantiles 1

Bahadur representations for bootstrap quantiles 1 Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011 LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic

More information

Hyperspacings and the Estimation of Information Theoretic Quantities

Hyperspacings and the Estimation of Information Theoretic Quantities Hyperspacings and the Estimation of Information Theoretic Quantities Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 0003 elm@cs.umass.edu Abstract

More information

Estimation of the Bivariate and Marginal Distributions with Censored Data

Estimation of the Bivariate and Marginal Distributions with Censored Data Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS

A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Statistica Sinica 20 2010, 365-378 A PRACTICAL WAY FOR ESTIMATING TAIL DEPENDENCE FUNCTIONS Liang Peng Georgia Institute of Technology Abstract: Estimating tail dependence functions is important for applications

More information

A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis

A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis A note on the asymptotic distribution of Berk-Jones type statistics under the null hypothesis Jon A. Wellner and Vladimir Koltchinskii Abstract. Proofs are given of the limiting null distributions of the

More information

Strong approximations for resample quantile processes and application to ROC methodology

Strong approximations for resample quantile processes and application to ROC methodology Strong approximations for resample quantile processes and application to ROC methodology Jiezhun Gu 1, Subhashis Ghosal 2 3 Abstract The receiver operating characteristic (ROC) curve is defined as true

More information

Minimum distance tests and estimates based on ranks

Minimum distance tests and estimates based on ranks Minimum distance tests and estimates based on ranks Authors: Radim Navrátil Department of Mathematics and Statistics, Masaryk University Brno, Czech Republic (navratil@math.muni.cz) Abstract: It is well

More information

ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS

ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS AIDA TOMA The nonrobustness of classical tests for parametric models is a well known problem and various

More information

The Laplacian PDF Distance: A Cost Function for Clustering in a Kernel Feature Space

The Laplacian PDF Distance: A Cost Function for Clustering in a Kernel Feature Space The Laplacian PDF Distance: A Cost Function for Clustering in a Kernel Feature Space Robert Jenssen, Deniz Erdogmus 2, Jose Principe 2, Torbjørn Eltoft Department of Physics, University of Tromsø, Norway

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments

On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments UDC 519.2 On probabilities of large and moderate deviations for L-statistics: a survey of some recent developments N. V. Gribkova Department of Probability Theory and Mathematical Statistics, St.-Petersburg

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Trimmed sums of long-range dependent moving averages

Trimmed sums of long-range dependent moving averages Trimmed sums of long-range dependent moving averages Rafal l Kulik Mohamedou Ould Haye August 16, 2006 Abstract In this paper we establish asymptotic normality of trimmed sums for long range dependent

More information

On variable bandwidth kernel density estimation

On variable bandwidth kernel density estimation JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced

More information

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky

A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky A COMPARISON OF POISSON AND BINOMIAL EMPIRICAL LIKELIHOOD Mai Zhou and Hui Fang University of Kentucky Empirical likelihood with right censored data were studied by Thomas and Grunkmier (1975), Li (1995),

More information

Smooth simultaneous confidence bands for cumulative distribution functions

Smooth simultaneous confidence bands for cumulative distribution functions Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang

More information

Lecture Quantitative Finance Spring Term 2015

Lecture Quantitative Finance Spring Term 2015 on bivariate Lecture Quantitative Finance Spring Term 2015 Prof. Dr. Erich Walter Farkas Lecture 07: April 2, 2015 1 / 54 Outline on bivariate 1 2 bivariate 3 Distribution 4 5 6 7 8 Comments and conclusions

More information

Probability Distributions and Estimation of Ali-Mikhail-Haq Copula

Probability Distributions and Estimation of Ali-Mikhail-Haq Copula Applied Mathematical Sciences, Vol. 4, 2010, no. 14, 657-666 Probability Distributions and Estimation of Ali-Mikhail-Haq Copula Pranesh Kumar Mathematics Department University of Northern British Columbia

More information

Does k-th Moment Exist?

Does k-th Moment Exist? Does k-th Moment Exist? Hitomi, K. 1 and Y. Nishiyama 2 1 Kyoto Institute of Technology, Japan 2 Institute of Economic Research, Kyoto University, Japan Email: hitomi@kit.ac.jp Keywords: Existence of moments,

More information

A Goodness-of-fit Test for Copulas

A Goodness-of-fit Test for Copulas A Goodness-of-fit Test for Copulas Artem Prokhorov August 2008 Abstract A new goodness-of-fit test for copulas is proposed. It is based on restrictions on certain elements of the information matrix and

More information

Estimation of Rényi Information Divergence via Pruned Minimal Spanning Trees 1

Estimation of Rényi Information Divergence via Pruned Minimal Spanning Trees 1 Estimation of Rényi Information Divergence via Pruned Minimal Spanning Trees Alfred Hero Dept. of EECS, The university of Michigan, Ann Arbor, MI 489-, USA Email: hero@eecs.umich.edu Olivier J.J. Michel

More information

On a simple construction of bivariate probability functions with fixed marginals 1

On a simple construction of bivariate probability functions with fixed marginals 1 On a simple construction of bivariate probability functions with fixed marginals 1 Djilali AIT AOUDIA a, Éric MARCHANDb,2 a Université du Québec à Montréal, Département de mathématiques, 201, Ave Président-Kennedy

More information

Non Parametric Estimation of Mutual Information through the Entropy of the Linkage

Non Parametric Estimation of Mutual Information through the Entropy of the Linkage Entropy 203, 5, 554-577; doi:0.3390/e52554 Article OPEN ACCESS entropy ISSN 099-4300 www.mdpi.com/ournal/entropy Non Parametric Estimation of Mutual Information through the Entropy of the Linkage Maria

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

The Central Limit Theorem Under Random Truncation

The Central Limit Theorem Under Random Truncation The Central Limit Theorem Under Random Truncation WINFRIED STUTE and JANE-LING WANG Mathematical Institute, University of Giessen, Arndtstr., D-3539 Giessen, Germany. winfried.stute@math.uni-giessen.de

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

Exact Linear Likelihood Inference for Laplace

Exact Linear Likelihood Inference for Laplace Exact Linear Likelihood Inference for Laplace Prof. N. Balakrishnan McMaster University, Hamilton, Canada bala@mcmaster.ca p. 1/52 Pierre-Simon Laplace 1749 1827 p. 2/52 Laplace s Biography Born: On March

More information

Estimation of the quantile function using Bernstein-Durrmeyer polynomials

Estimation of the quantile function using Bernstein-Durrmeyer polynomials Estimation of the quantile function using Bernstein-Durrmeyer polynomials Andrey Pepelyshev Ansgar Steland Ewaryst Rafaj lowic The work is supported by the German Federal Ministry of the Environment, Nature

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

High Breakdown Analogs of the Trimmed Mean

High Breakdown Analogs of the Trimmed Mean High Breakdown Analogs of the Trimmed Mean David J. Olive Southern Illinois University April 11, 2004 Abstract Two high breakdown estimators that are asymptotically equivalent to a sequence of trimmed

More information

Some New Results on Information Properties of Mixture Distributions

Some New Results on Information Properties of Mixture Distributions Filomat 31:13 (2017), 4225 4230 https://doi.org/10.2298/fil1713225t Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Some New Results

More information

Nonparametric regression with martingale increment errors

Nonparametric regression with martingale increment errors S. Gaïffas (LSTA - Paris 6) joint work with S. Delattre (LPMA - Paris 7) work in progress Motivations Some facts: Theoretical study of statistical algorithms requires stationary and ergodicity. Concentration

More information

On a weighted total variation minimization problem

On a weighted total variation minimization problem On a weighted total variation minimization problem Guillaume Carlier CEREMADE Université Paris Dauphine carlier@ceremade.dauphine.fr Myriam Comte Laboratoire Jacques-Louis Lions, Université Pierre et Marie

More information

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and

More information

Optimal global rates of convergence for interpolation problems with random design

Optimal global rates of convergence for interpolation problems with random design Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289

More information

HANDBOOK OF APPLICABLE MATHEMATICS

HANDBOOK OF APPLICABLE MATHEMATICS HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester

More information

EFFICIENT MULTIVARIATE ENTROPY ESTIMATION WITH

EFFICIENT MULTIVARIATE ENTROPY ESTIMATION WITH EFFICIENT MULTIVARIATE ENTROPY ESTIMATION WITH HINTS OF APPLICATIONS TO TESTING SHAPE CONSTRAINTS Richard Samworth, University of Cambridge Joint work with Thomas B. Berrett and Ming Yuan Collaborators

More information

ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY

ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY G.L. Shevlyakov, P.O. Smirnov St. Petersburg State Polytechnic University St.Petersburg, RUSSIA E-mail: Georgy.Shevlyakov@gmail.com

More information

A MODIFICATION OF HILL S TAIL INDEX ESTIMATOR

A MODIFICATION OF HILL S TAIL INDEX ESTIMATOR L. GLAVAŠ 1 J. JOCKOVIĆ 2 A MODIFICATION OF HILL S TAIL INDEX ESTIMATOR P. MLADENOVIĆ 3 1, 2, 3 University of Belgrade, Faculty of Mathematics, Belgrade, Serbia Abstract: In this paper, we study a class

More information

Refining the Central Limit Theorem Approximation via Extreme Value Theory

Refining the Central Limit Theorem Approximation via Extreme Value Theory Refining the Central Limit Theorem Approximation via Extreme Value Theory Ulrich K. Müller Economics Department Princeton University February 2018 Abstract We suggest approximating the distribution of

More information

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin The Annals of Statistics 1996, Vol. 4, No. 6, 399 430 ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE By Michael Nussbaum Weierstrass Institute, Berlin Signal recovery in Gaussian

More information

Statistical inference on Lévy processes

Statistical inference on Lévy processes Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline

More information

Accounting for extreme-value dependence in multivariate data

Accounting for extreme-value dependence in multivariate data Accounting for extreme-value dependence in multivariate data 38th ASTIN Colloquium Manchester, July 15, 2008 Outline 1. Dependence modeling through copulas 2. Rank-based inference 3. Extreme-value dependence

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Highly Robust Variogram Estimation 1. Marc G. Genton 2

Highly Robust Variogram Estimation 1. Marc G. Genton 2 Mathematical Geology, Vol. 30, No. 2, 1998 Highly Robust Variogram Estimation 1 Marc G. Genton 2 The classical variogram estimator proposed by Matheron is not robust against outliers in the data, nor is

More information

Random Bernstein-Markov factors

Random Bernstein-Markov factors Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Asymptotic statistics using the Functional Delta Method

Asymptotic statistics using the Functional Delta Method Quantiles, Order Statistics and L-Statsitics TU Kaiserslautern 15. Februar 2015 Motivation Functional The delta method introduced in chapter 3 is an useful technique to turn the weak convergence of random

More information

Statistical Properties of Numerical Derivatives

Statistical Properties of Numerical Derivatives Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective

More information

Estimation of a quadratic regression functional using the sinc kernel

Estimation of a quadratic regression functional using the sinc kernel Estimation of a quadratic regression functional using the sinc kernel Nicolai Bissantz Hajo Holzmann Institute for Mathematical Stochastics, Georg-August-University Göttingen, Maschmühlenweg 8 10, D-37073

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

On the asymptotic normality of frequency polygons for strongly mixing spatial processes. Mohamed EL MACHKOURI

On the asymptotic normality of frequency polygons for strongly mixing spatial processes. Mohamed EL MACHKOURI On the asymptotic normality of frequency polygons for strongly mixing spatial processes Mohamed EL MACHKOURI Laboratoire de Mathématiques Raphaël Salem UMR CNRS 6085, Université de Rouen France mohamed.elmachkouri@univ-rouen.fr

More information

Exact goodness-of-fit tests for censored data

Exact goodness-of-fit tests for censored data Exact goodness-of-fit tests for censored data Aurea Grané Statistics Department. Universidad Carlos III de Madrid. Abstract The statistic introduced in Fortiana and Grané (23, Journal of the Royal Statistical

More information

A New Estimator for a Tail Index

A New Estimator for a Tail Index Acta Applicandae Mathematicae 00: 3, 2003. 2003 Kluwer Academic Publishers. Printed in the Netherlands. A New Estimator for a Tail Index V. PAULAUSKAS Department of Mathematics and Informatics, Vilnius

More information

Asymptotic distribution of the sample average value-at-risk

Asymptotic distribution of the sample average value-at-risk Asymptotic distribution of the sample average value-at-risk Stoyan V. Stoyanov Svetlozar T. Rachev September 3, 7 Abstract In this paper, we prove a result for the asymptotic distribution of the sample

More information

Estimation of density level sets with a given probability content

Estimation of density level sets with a given probability content Estimation of density level sets with a given probability content Benoît CADRE a, Bruno PELLETIER b, and Pierre PUDLO c a IRMAR, ENS Cachan Bretagne, CNRS, UEB Campus de Ker Lann Avenue Robert Schuman,

More information

SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1

SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 B Technical details B.1 Variance of ˆf dec in the ersmooth case We derive the variance in the case where ϕ K (t) = ( 1 t 2) κ I( 1 t 1), i.e. we take K as

More information

Quantile Approximation of the Chi square Distribution using the Quantile Mechanics

Quantile Approximation of the Chi square Distribution using the Quantile Mechanics Proceedings of the World Congress on Engineering and Computer Science 017 Vol I WCECS 017, October 57, 017, San Francisco, USA Quantile Approximation of the Chi square Distribution using the Quantile Mechanics

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use: This article was downloaded by: [Ferdowsi University of Mashhad] On: 7 June 2010 Access details: Access Details: [subscription number 912974449] Publisher Taylor & Francis Informa Ltd Registered in England

More information

Inference on distributions and quantiles using a finite-sample Dirichlet process

Inference on distributions and quantiles using a finite-sample Dirichlet process Dirichlet IDEAL Theory/methods Simulations Inference on distributions and quantiles using a finite-sample Dirichlet process David M. Kaplan University of Missouri Matt Goldman UC San Diego Midwest Econometrics

More information

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Exact Asymptotics in Complete Moment Convergence for Record Times and the Associated Counting Process

Exact Asymptotics in Complete Moment Convergence for Record Times and the Associated Counting Process A^VÇÚO 33 ò 3 Ï 207 c 6 Chinese Journal of Applied Probability Statistics Jun., 207, Vol. 33, No. 3, pp. 257-266 doi: 0.3969/j.issn.00-4268.207.03.004 Exact Asymptotics in Complete Moment Convergence for

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Adaptive Nonparametric Density Estimators

Adaptive Nonparametric Density Estimators Adaptive Nonparametric Density Estimators by Alan J. Izenman Introduction Theoretical results and practical application of histograms as density estimators usually assume a fixed-partition approach, where

More information

Efficient Estimation for the Partially Linear Models with Random Effects

Efficient Estimation for the Partially Linear Models with Random Effects A^VÇÚO 1 33 ò 1 5 Ï 2017 c 10 Chinese Journal of Applied Probability and Statistics Oct., 2017, Vol. 33, No. 5, pp. 529-537 doi: 10.3969/j.issn.1001-4268.2017.05.009 Efficient Estimation for the Partially

More information

UNIVERSALITY OF THE RIEMANN ZETA FUNCTION: TWO REMARKS

UNIVERSALITY OF THE RIEMANN ZETA FUNCTION: TWO REMARKS Annales Univ. Sci. Budapest., Sect. Comp. 39 (203) 3 39 UNIVERSALITY OF THE RIEMANN ZETA FUNCTION: TWO REMARKS Jean-Loup Mauclaire (Paris, France) Dedicated to Professor Karl-Heinz Indlekofer on his seventieth

More information

RENORMALIZATION OF DYSON S VECTOR-VALUED HIERARCHICAL MODEL AT LOW TEMPERATURES

RENORMALIZATION OF DYSON S VECTOR-VALUED HIERARCHICAL MODEL AT LOW TEMPERATURES RENORMALIZATION OF DYSON S VECTOR-VALUED HIERARCHICAL MODEL AT LOW TEMPERATURES P. M. Bleher (1) and P. Major (2) (1) Keldysh Institute of Applied Mathematics of the Soviet Academy of Sciences Moscow (2)

More information

Nonlinear Analysis 72 (2010) Contents lists available at ScienceDirect. Nonlinear Analysis. journal homepage:

Nonlinear Analysis 72 (2010) Contents lists available at ScienceDirect. Nonlinear Analysis. journal homepage: Nonlinear Analysis 72 (2010) 4298 4303 Contents lists available at ScienceDirect Nonlinear Analysis journal homepage: www.elsevier.com/locate/na Local C 1 ()-minimizers versus local W 1,p ()-minimizers

More information

Oberwolfach Preprints

Oberwolfach Preprints Oberwolfach Preprints OWP 202-07 LÁSZLÓ GYÖFI, HAO WALK Strongly Consistent Density Estimation of egression esidual Mathematisches Forschungsinstitut Oberwolfach ggmbh Oberwolfach Preprints (OWP) ISSN

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

A NOTE ON TESTING THE SHOULDER CONDITION IN LINE TRANSECT SAMPLING

A NOTE ON TESTING THE SHOULDER CONDITION IN LINE TRANSECT SAMPLING A NOTE ON TESTING THE SHOULDER CONDITION IN LINE TRANSECT SAMPLING Shunpu Zhang Department of Mathematical Sciences University of Alaska, Fairbanks, Alaska, U.S.A. 99775 ABSTRACT We propose a new method

More information

NORM DEPENDENCE OF THE COEFFICIENT MAP ON THE WINDOW SIZE 1

NORM DEPENDENCE OF THE COEFFICIENT MAP ON THE WINDOW SIZE 1 NORM DEPENDENCE OF THE COEFFICIENT MAP ON THE WINDOW SIZE Thomas I. Seidman and M. Seetharama Gowda Department of Mathematics and Statistics University of Maryland Baltimore County Baltimore, MD 8, U.S.A

More information