Estimation of the Error Density in a Semiparametric Transformation Model
|
|
- Augustus Morgan
- 5 years ago
- Views:
Transcription
1 Estimation of the Error Density in a Semiparametric Transformation Model Benjamin Colling Université catholique de Louvain Cédric Heuchenne University of Liège and Université catholique de Louvain Rawane Sam Université catholique de Louvain Ingrid Van Keilegom Université catholique de Louvain Novemer 14, 201 Astract Consider the semiparametric transformation model Λ θo (Y) = m(x) + ε, where θ o is an unknown finite dimensional parameter, the functions Λ θo and m are smooth, ε is independent of X, and E(ε) = 0. We propose a kernel-type estimator of the density of the error ε, and prove its asymptotic normality. The estimated errors, which lie at the asis of this estimator, are otained from a profile likelihood estimator of θ o and a nonparametric kernel estimator of m. The practical performance of the proposed density estimator is evaluated in a simulation study. Key Words: Density estimation; Kernel smoothing; Nonparametric regression; Profile likelihood; Transformation model. Running Head: Error density estimation in transformation models Research supported y IAP research network P7/06 of the Belgian Government (Belgian Science Policy), and y the contract Projet d Actions de Recherche Concertées (ARC) 11/16-09 of the Communauté française de Belgique, granted y the Académie universitaire Louvain. Research supported y IAP research network P7/06 of the Belgian Government (Belgian Science Policy). Research supported y the European Research Council under the European Community s Seventh Framework Programme (FP7/ ) / ERC Grant agreement No , y IAP research network P7/06 of the Belgian Government (Belgian Science Policy), and y the contract Projet d Actions de Recherche Concertées (ARC) 11/16-09 of the Communauté française de Belgique, granted y the Académie universitaire Louvain. 1
2 1 Introduction Let (X 1,Y 1 ),...,(X n,y n ) e independent replicates of the random vector (X,Y), where Y is a univariate dependent variale and X is a one-dimensional covariate. We assume that Y and X are related via the semiparametric transformation model Λ θo (Y) = m(x)+ε, (1) where ε is independent of X and has mean zero. We assume that {Λ θ : θ Θ} (with Θ R p compact) is a parametric family of strictly increasing functions defined on an unounded suset D in R, while m is the unknown regression function, elonging to an infinite dimensional parameter set M. We assume that M is a space of functions endowed with the norm M =. We denote θ o Θ and m M for the true unknown finite and infinite dimensional parameters. Define the regression function m θ (x) = E[Λ θ (Y) X = x], for each θ Θ, and let ε θ = ε(θ) = Λ θ (Y) m θ (X). In this paper, we are interested in the estimation of the proaility density function (p.d.f.) f ε of the residual term ε = Λ θo (Y) m(x). To this end, we first otain the estimators θ and m θ of the parameter θ o and the function m θ, and second, form the semiparametric regression residuals ε i ( θ) = Λ θ(y i ) m θ(x i ). To estimate θ o we use a profile likelihood (PL) approach, developed in Linton et al. (2008), whereas m θ is estimated y means of a Nadaraya-Watson-type estimator (Nadaraya (1964), Watson (1964)). To our knowledge, the estimation of the density of ε in model (??) has not yet een investigated in the statistical literature. Estimatingtheerrordensityinthesemiparametrictransformationmodel(SPT)Λ θo (Y) = m(x)+ ε may e very useful in various regression prolems. First, taking transformations of the data may induce normality and error variance homogeneity in the transformed model. So the estimation of the error density in the transformed model may e used for testing these hypotheses; it may also e used for goodness-of-fit tests of a specified error distriution in a parametric or nonparametric regression setting. Some examples can e found in Akritas and Van Keilegom (2001), Cheng and Sun (2008), ut with Λ θo id, i.e. the response is not transformed. Next, the estimation of the error density in the aove model can e useful for testing the symmetry of the residual distriution. See Ahmad and Li (1997), Dette et al. (2002), Neumeyer and Dette (2007) and references therein, in the case Λ θo id. Under this model, Escanciano and Jacho-Chavez (2012) considered the estimation of the (marginal) density of the response Y via the estimation of the error density. Another application of the estimation of the error density in the SPT model is the forecasting of Λ θo (Y) y 2
3 means of the mode approach, since the mode of the p.d.f. of Λ θo (Y) given X = x is m(x)+argmax e R f ε (e), where f ε is the p.d.f. of the error term ε. Taking transformations of the data has een an important part of statistical practice for many years. A major contriution to this methodology was made y Box and Cox (1964), who proposed a parametric power family of transformations that includes the logarithm and the identity. They suggested that the power transformation, when applied to the dependent variale in a linear regression model, might induce normality and homoscedasticity. Lots of effort has een devoted to the investigation of the Box-Cox transformation since its introduction. See, for example, Amemiya (1985), Horowitz (1998), Chen et al. (2002), Shin (2008), and Fitzenerger et al. (2010). Other dependent variale transformations have een suggested, for example, the Zellner and Revankar(1969) transform and the Bickel and Doksum(1981) transform. The transformation methodology has een quite successful and a large literature exists on this topic for parametric models. See Carroll and Ruppert (1988) and Sakia (1992) and references therein. The estimation of (functionals of) the error distriution and density under simplified versions of model (??) has received considerale attention in the statistical literature in recent years. Akritas and Van Keilegom (2001) estimated the cumulative distriution function of the regression error in a heteroscedastic model with univariate covariates. The estimator they proposed is ased on nonparametrically estimated regression residuals. The weak convergence of their estimator was proved. The results otained y Akritas and Van Keilegom (2001) have een generalized y Neumeyer and Van Keilegom (2010) to the case of multivariate covariates. Müller et al. (2004) investigated linear functionals of the error distriution in nonparametric regression. Cheng (2005) estalished the asymptotic normality of an estimator of the error density ased on estimated residuals. The estimator he proposed is constructed y splitting the sample into two parts: the first part is used for the estimation of the residuals, while the second part of the sample is used for the construction of the error density estimator. Efromovich (2005) proposed an adaptive estimator of the error density, ased on a density estimator proposed y Pinsker (1980). Finally, Sam (2011) also considered the estimation of the error density, ut his approach is more closely related to the one in Akritas and Van Keilegom (2001). In order to achieve the ojective of this paper, namely the estimation of the error density under model (??), we first need to estimate the transformation parameter θ o. To this end, we make use of the results in Linton et al. (2008). In the latter paper, the authors first discuss the nonparametric identification of model (??), and second, estimate the transformation parameter θ o under the considered model. For the estimation of this parameter, they propose two approaches. The first approach uses a semiparametric profile likelihood
4 (PL) estimator, while the second is ased on a mean squared distance from independence-estimator (MD) using the estimated distriutions of X, ε and (X, ε). Linton et al. (2008) derived the asymptotic distriutions of their estimators under certain regularity conditions, and proved that oth estimators of θ o are root-n consistent. The authors also showed that, in practice, the performance of the PL method is etter than that of the MD approach. For this reason, the PL method will e considered in this paper for the estimation of θ o. The rest of the paper is organized as follows. Section 2 presents our estimator of the error density and groups some notations and technical assumptions. Section descries the asymptotic results of the paper. A simulation study is given in Section 4, while Section 5 is devoted to some general conclusions. Finally, the proofs of the asymptotic results are collected in Section 6. 2 Definitions and assumptions 2.1 Construction of the estimators The approach proposed here for the estimation of f ε is ased on a two-steps procedure. In a first step, we estimate the finite dimensional parameter θ o. This parameter is estimated y the profile likelihood (PL) method, developed in Linton et al.(2008). The asic idea of this method is to replace all unknown expressions in the likelihood function y their nonparametric kernel estimates. Under model (??), we have P(Y y X) = P(Λ θo (Y) Λ θo (y) X) = P(ε θo Λ θo (y) m θo (X) X) = F ε (Λ θo (y) m θo (X)). Here, F ε (t) = P(ε t), and so f Y X (y x) = f ε (Λ θo (y) m θo (x))λ θ o (y), where f ε and f Y X are the densities of ε, and of Y given X, respectively. Then, the log likelihood function is {logf εθ (Λ θ (Y i ) m θ (X i ))+logλ θ (Y i)}, θ Θ, where f εθ is the density function of ε θ. Now, let ( ) n j=1 Λ Xj x θ(y j )K 1 m θ (x) = n j=1 K 1( Xj x h h ) (2) e the Nadaraya-Watson estimator of m θ (x), and let f εθ (t) = 1 ) ( εi (θ) t K 2. () ng g 4
5 where ε i (θ) = Λ θ (Y i ) m θ (X i ). Here, K 1 and K 2 are kernel functions and h and g are andwidth sequences. Then, the PL estimator of θ o is defined y [ θ = argmax log f ] εθ (Λ θ (Y i ) m θ (X i ))+logλ θ (Y i). (4) θ Θ Recall that m θ (X i ) converges to m θ (X i ) at a slower rate for those X i which are close to the oundary of the support X of the covariate X. That is why we assume implicitly that the proposed estimator (??) of θ o trims the oservations X i outside a suset X 0 of X. Note that we keep the root-n consistency of θ proved in Linton et al. (2008) y trimming the covariates outside X 0. But in this case, the resulting asymptotic variance is different to the one otained in the latter paper. In a second step, we use the aove estimator θ to uild the estimated residuals ε i ( θ) = Λ θ(y i ) m θ(x i ). Then, our proposed estimator f ε (t) of f ε (t) is defined y ( ) f ε (t) = 1 ε i ( θ) t K, (5) n where K is a kernel function and is a andwidth sequence, not necessarily the same as the kernel K 2 and the andwidth g used in (??). Oserve that this estimator is a feasile estimator in the sense that it does not depend on any unknown quantity, as is desirale in practice. This contrasts with the unfeasile ideal kernel estimator f ε (t) = 1 n ( ) εi t K, (6) which depends in particular on the unknown regression errors ε i = ε i (θ o ) = Λ θo (Y i ) m(x i ). It is however intuitively clear that f ε (t) and f ε (t) will e very close for n large enough, as will e illustrated y the results given in Section. 2.2 Notations When there is no amiguity, we use ε and m to indicate ε θo and m θo. Moreover, N(θ o ) represents a neighorhood of θ o. For the kernel K j (j = 1,2,), let µ(k j ) = v 2 K j (v)dv and let K (p) j e the pth derivative of K j. For any function ϕ θ (y), denote ϕ θ (y) = ϕ θ (y)/ θ = ( ϕ θ (y)/ θ 1,..., ϕ θ (y)/ θ p ) t and ϕ θ (y) = ϕ θ(y)/ y. Also, let A = (A t A) 1/2 e the Euclidean norm of any vector A. For any functions m, r, f, ϕ and q, and any θ Θ, let s = ( m,r,f,ϕ,q), s θ = (m θ,ṁ θ,f εθ,f ε θ, f εθ ), ε i (θ, m) = Λ θ (Y i ) m(x i ), and define { n G n (θ,s) = n 1 1 f{ε i (θ, m)} [ ] ϕ{ε i (θ, m)}{ Λ θ (Y i ) r(x i )}+q{ε i (θ, m)} + Λ } θ (Y i) Λ θ (Y, i) 5
6 G(θ,s) = E[G n (θ,s)] and G(θ o,s θo ) = θ G(θ,s θ) θ=θo. 2. Technical assumptions The assumptions we need for the asymptotic results are listed elow for convenient reference. (A1) The function K j (j = 1,2,) is symmetric, has compact support, v k K j (v)dv = 0 for k = 1,...,q j 1 and v qj K j (v)dv 0 for some q j 4, K j is twice continuously differentiale, and K (1) (v)dv = 0. (A2) The andwidth sequences h, g and satisfy nh 2q1 = o(1), ng 2q2 = o(1) (where q 1 and q 2 are defined in (A1)), (n 5 ) 1 = O(1), n h 2 (logh 1 ) 2 and ng 6 (logg 1 ) 2. (A) (i) The support X of the covariate X is a compact suset of R, and X 0 is a suset with non empty interior, whose closure is in the interior of X. (ii) The density f X is ounded away from zero and infinity on X, and has continuous second order partial derivatives on X. (A4) The function m θ (x) is twice continuously differentiale with respect to θ on X N(θ 0 ), and the functions m θ (x) and ṁ θ (x) are q 1 times continuously differentiale with respect to x on X N(θ 0 ). All these derivatives are ounded, uniformly in (x,θ) X N(θ o ). (A5) The error ε = Λ θo (Y) m(x) has finite fourth moment and is independent of X. (A6) The distriution F εθ (t) is q +1 (respectively three) times continuously differentiale with respect to t (respectively θ), and sup θ,t k+l t k θ l θlp p F εθ (t) < for all k and l such that 0 k +l 2, where l = l l p and θ = (θ 1,...,θ p ) t. (A7) The transformation Λ θ (y) is three times continuously differentiale with respect to oth θ and y, and there exists a α > 0 such that E [ sup θ : θ θ α ] k+l Λ θ (Y) < y k θ l θlp p for all θ Θ, and for all k and l such that 0 k + l, where l = l l p and θ = (θ 1,...,θ p ) t. Moreover, sup x X E[ Λ 4 θ o (Y) X = x] <. (A8) For all η > 0, there exists ǫ(η) > 0 such that inf G(θ,s θ) ǫ(η) > 0. θ θ o >η 6
7 Moreover, the matrix G(θ o,s θo ) is non-singular. (A9) (i) E(Λ θo (Y)) = 1, Λ θo (0) = 0 and the set {x X 0 : m (x) 0} has nonempty interior. (ii) Assume that φ(x,t) = Λ θo (Λ 1 θ o (m(x) +t))f ε (t) is continuously differentiale with respect to t for all x and that for all t R and for some δ > 0. sup s: t s δ E φ s (X,s) <. (7) Assumptions (A1), part of (A2), (A)(ii), (A4) and (A6), (A7) and (A8) are used y Linton et al. (2008) to show that the PL estimator θ of θ o is root n-consistent. The differentiaility of K j up to second order imposed in assumption (A1) is used to expand the two-steps kernel estimator f ε (t) in (??) around the unfeasile one f ε (t). Assumptions (A)(ii) and (A4) impose that all the functions to e estimated have ounded derivatives. The last assumption in (A2) is useful for otaining the uniform convergence of the Nadaraya-Watson estimator of m θo in (??) (see for instance Einmahl and Mason (2005)). This assumption is also needed in the study of the difference etween the feasile estimator f ε (t) and the unfeasile estimator f ε (t). Finally, (A9)(i) is needed for identifying the model (see Vanhems and Van Keilegom (2011)). Asymptotic results In this section we are interested in the asymptotic ehavior of the estimator f ε (t). To this end, we first investigate its asymptotic representation, which will e needed to show its asymptotic normality. Theorem 1. Assume (A1)-(A9). Then, f ε (t) f ε (t) = 1 n where R n (t) = o P ( (n) 1/2 ) for all t R. ( ) εi t K f ε (t)+r n (t), This result is important, since it shows that, provided the ias term is negligile, the estimation of θ o and m( ) has asymptotically no effect on the ehavior of the estimator f ε (t). Therefore, this estimator is asymptotically equivalent to the unfeasile estimator f ε (t), ased on the unknown true errors ε 1,...,ε n. Our next result gives the asymptotic normality of the estimator f ε (t). Theorem 2. Assume (A1)-(A9). In addition, assume that n 2q+1 = O(1). Then, ) ( ) d n ( f ε (t) f ε (t) N 0,f ε (t) K(v)dv 2, 7
8 where f ε (t) = f ε (t)+ q q! f(q) ε (t) v q K (v)dv. The proofs of Theorems?? and?? are given in Section??. 4 Simulations In this section, we investigate the performance of our method for different models and different sample sizes. Consider where Λ θ is the Box-Cox (1964) transformation y θ 1 θ, θ 0, Λ θ (y) = log(y), θ = 0, Λ θo (Y) = X sin(πx)+σ e ε, (8) X is uniformly distriuted on the interval [ 1,1], and ε is independent of X. We carry out simulations for two cases : in the first case, ε has a standard normal distriution and, in the second case, the distriution of ε is the mixture of the normal distriutions N( 1.5,0.25) and N(1.5,0.25) with equal weights. To make computations easier, error distriutions are truncated at and (i.e., put to 0 outside the interval [, ]). We study three different model settings. For each of them, 2 = 0 σ e. The other parameters are chosen as follows: Model 1: 0 = 6.5, 1 = 5, σ e = 1.5; Model 2: 0 = 4.5, 1 =.5, σ e = 1; Model : 0 = 2.5, 1 = 2.5, σ e = 0.5. Oursimulationsareperformedforθ 0 = 0,0.5and1. WeusetheEpanechnikovkernelK(x) = 4 ( 1 x 2 ) 1( x 1) for oth the estimator of the regression function and the density function. The results are ased on 100 random samples of size n = 100 and n = 200. For the estimation of θ 0 and f ε (t), we proceed as follows. Let L θ (h,g) = [ log f ] εθ ( ε i (θ,h))+logλ θ(y i ), where ε i (θ,h) = Λ θ (Y i ) m θ (X i,h)and m θ (x,h) denotes m θ (x) constructedwith andwidthh.this function will e maximized with respect to θ for given (optimal) values of (h,g). For each value of θ, h (θ) is otained 8
9 y least squares cross-validation, where h (θ) = argmax h m i,θ (X i ) = (Λ θ (Y i ) m i,θ (X i )) 2, ) n j=1,j i Λ Xj X θ(y j )K( i n j=1,j i K ( Xj X i h and g can e chosen with a classical andwidth selection rule for kernel density estimation. Here, for simplicity, the normal rule is used (ĝ(θ) = (40 π) 1/5 n 1/5 σ ε(θ,h (θ)), where σ ε(θ,h (θ)) is the classical empirical estimator of the standard deviation ased on ε i (θ,h (θ)), i = 1,...,n). The solution h ) θ = argmaxl θ (h (θ),ĝ(θ)) θ is therefore otained iteratively (maximization prolems are solved with the function optimize in R with h [0,2] and θ [ 20,20]) and the estimator of f ε (t) is finally given y ( ) 1 ε i ( θ,h ( θ)) t f ε (t) = K. nĝ( θ) ĝ( θ) Tales??,?? and?? show the mean squared error (MSE) of the estimator f ε (t) of the standardized (pseudo-estimated) error ε = ( Λ θ(y) m θ(x) ) /σ e (with known σ e ), for t = 1, 0 and 1 (respectively t = 1.5, 1, 0, 1 and 1.5) and for the unimodal (respectively imodal) normal error distriution. Tales?? and?? show the integrated mean squared error (IMSE) of the estimator f ε ( ) for oth error distriutions, where the integration is done over the interval [,]. As expected, in oth cases, estimation is etter for the normal density than for the mixture of two normals, and estimation improves when n increases, and in most cases, when σ e decreases. In particular, this can e oservedfrom Tales?? and??. The limiting case θ 0 = 0 (the logarithmic transformation) seems to e more easily captured, especially when the error is normally distriuted. In general, we oserve from Tales??,??,?? that estimation is poorer near local maxima and minima of the density, which is not uncommon for kernel smoothing methods. This also suggests that the choice of the smoothing parameters is important and should e the oject of further investigation. 9
10 Model θ 0 n = 100 n = 200 f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ ( 1) f ǫ (0) f ǫ (1) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale 1: MSE( f ε (t)) for different models, values of t and sample sizes, when f ε ( ) is a standard normal density. 10
11 Model θ 0 n = 100 n = = 6.5 θ 0 = = 5 θ 0 = σ 0 = 1.5 θ 0 = = 4.5 θ 0 = =.5 θ 0 = σ 0 = 1 θ 0 = = 2.5 θ 0 = = 2.5 θ 0 = σ 0 = 0.5 θ 0 = Tale 2: IMSE( f ε ) for different models and sample sizes, when f ε ( ) is a standard normal density. 5 Conclusions In this paper we have studied the estimation of the density of the error in a semiparametric transformation model. The regression function in this model is unspecified (except for some smoothness assumptions), whereas the transformation (of the dependent variale in the model) is supposed to elong to a parametric family of monotone transformations. The proposed estimator is a kernel-type estimator, and we have shown its asymptotic normality. The finite sample performance of the estimator is illustrated y means of a simulation study. It would e interesting to explore various possile applications of the results in this paper. For example, one could use the results on the estimation of the error density to test hypotheses concerning e.g. the normality of the errors, the homoscedasticity of the error variance, or the linearity of the regression function, all of which are important features in the context of transformation models. 6 Proofs Proof of Theorem??. Write f ε (t) f ε (t) = [ f ε (t) f ε (t)]+[ f ε (t) f ε (t)], where f ε (t) = 1 n ) ( εi t K 11
12 Model θ 0 n = 100 f ǫ ( 1.5) f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ (1.5) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale : MSE( f ε (t)) for different models, values of t and n = 100, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. 12
13 Model θ 0 n = 200 f ǫ ( 1.5) f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ (1.5) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale 4: MSE( f ε (t)) for different models, values of t and n = 200, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. 1
14 Model θ 0 n = 100 n = = 6.5 θ 0 = = 5 θ 0 = σ 0 = 1.5 θ 0 = = 4.5 θ 0 = =.5 θ 0 = σ 0 = 1 θ 0 = = 2.5 θ 0 = = 2.5 θ 0 = σ 0 = 0.5 θ 0 = Tale 5: IMSE( f ε ) for different models and sample sizes, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. and ε i = Λ θo (Y i ) m θo (X i ), i = 1,...,n. In a completely similar way as was done for Lemma A.1 in Linton et al. (2008), it can e shown that f ε (t) f ε (t) = 1 n ( ) εi t K f ε (t)+o P ((n) 1/2 ) (9) for all t R. Note that the remainder term in Lemma A.1 in the aove paper equals a sum of i.i.d. terms of mean zero, plus a o P (n 1/2 ) term. Hence, the remainder term in that paper is O P (n 1/2 ), whereas we write o P ((n) 1/2 ) in (??). Therefore,the resultofthe theoremfollowsifweprovethat f ε (t) f ε (t) = o P ((n) 1/2 ). To this end, write f ε (t) f ε (t) 1 = n 2 ( ε i ( θ) ε i (θ o ))K (1) + 1 2n ( ε i ( θ) ε i (θ o )) 2 K (2) ) ( εi (θ o ) t ( ε i (θ o )+β( ε i ( θ) ε i (θ o )) t for some β (0,1). In what follows, we will show that each of the terms aove is o P ((n) 1/2 ). First consider the last term of (??). Since Λ θ (y) and m θ (x) are oth twice continuously differentiale with respect to θ, the second order Taylor expansion gives, for some θ 1 etween θ o and θ (to simplify the notations, we assume ), 14
15 here that p = dim(θ) = 1), ε i ( θ) ε i (θ o ) = Λ θ(y i ) Λ θo (Y i ) ( m θ(x i ) m θo (X i ) ) = ( θ θ o )( Λ θo (Y i ) m θo (X i ))+ 1 2 ( θ θ o ) 2 ( Λ θ1 (Y i ) m θ1 (X i )). Therefore, since θ θ o = o P ((n) 1/2 ) y Theorem 4.1 in Linton et al. (2008) (as efore, we work with a slower rate than what is shown in the latter paper, since this leads to weaker conditions on the andwidths), assumptions (A2) and (A7) imply that 1 n ( ε i ( θ) ε i (θ o )) 2 K (2) ( ) ε i (θ o )+β( ε i ( θ) ε i (θ o )) t ( = O P (n ) 1), which is o P ((n) 1/2 ), since (n 5 ) 1 = O(1) under (A2). For the first term of (??), the decomposition of ε i ( θ) ε i (θ o ) given aove yields 1 n 2 ) ( εi ( ε i ( θ) ε i (θ o ))K (1) (θ o ) t ( Λ θo (Y i ) m θo (X i ))K (1) = ( θ θ o ) n 2 = ( θ θ o ) n 2 ( Λ θo (Y i ) ṁ θo (X i ))K (1) ( εi (θ o ) t ( εi t ) +o P ((n) 1/2 ) where the last equality follows from a Taylor expansion applied to K (1), the fact that m θo (x) ṁ θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ), ) +o P ((n) 1/2 ), (10) uniformly in x X 0 y Lemma??, and the fact that nh (logh 1 ) 1 under (A2). Further, write [ n ( ) ] E ( Λ θo (Y i ) ṁ θo (X i ))K (1) εi t [ ( )] = E Λ θo (Y i )K (1) εi t [ ( )] E[ṁ θo (X i )]E K (1) εi t = A n B n. We will only show that the first term aove is O(n 2 ) for any t R. The proof for the other term is similar. Let ϕ(x,t) = Λ θo (Λ 1 θ o (m(x)+t)) and set φ(x,t) = ϕ(x,t)f ε (t). Then, applying a Taylor expansion to φ(x, ), 15
16 it follows that (for some β (0,1)) A n = [ ( E Λ θo Λ 1 θ o (m(x i )+ε i ) ) K (1) ( e t = n φ(x,e)k (1) = n φ(x,t+v)k (1) = n = n 2 ) f X (x)dxde (v)f X(x)dxdv [ φ(x,t)+v φ ] t (x,t+βv) ( )] εi t v φ t (x,t+βv)k(1) (v)f X(x)dxdv, K (1) (v)f X(x)dxdv since K (1) (v)dv = 0, and this is ounded y Kn2 sup s: t s δ E φ s (X,s) = O(n2 ) y assumption (A9)(ii). Hence, Tcheychev s inequality ensures that ( θ θ o ) 2 ( Λ θo (Y i ) ṁ θo (X i ))K (1) ( ) εi t = ( θ θ o ) n 2 O P (n 2 +(n) 1/2 ) = o P ((n) 1/2 ), since n /2 y (A2). Sustituting this in (??), yields 1 n 2 ) ( εi ( ε i ( θ) ε i (θ o ))K (1) (θ o ) t = o P ((n) 1/2 ), for any t R. This completes the proof. Proof of Theorem??. It follows from Theorem?? that f ε (t) f ε (t) = [ f ε (t) E f ε (t)]+[e f ε (t) f ε (t)]+o P ((n) 1/2 ). (11) The first term on the right hand side of (??) is treated y Lyapounov s Central Limit Theorem (LCT) for triangular arrays (see e.g. Billingsley (1968), Theorem 7.). To this end, let f in (t) = 1 K ( εi t Then, under (A1), (A2) and (A5) it can e easily shown that n E fin (t) E f in (t) Cn 2 f ε (t) K (v) dv +o ( n 2) ( n ) /2 ( = O((n) 1/2 ) = o(1), Var f in (t) n 1 f ε (t) K 2 (v)dv +o( n 1)) /2 ). 16
17 for some C > 0. Hence, the LCT ensures that This gives f ε (t) E f ε (t) Var f ε (t) = f ε (t) E f ε (t) Var f1n(t) ) ( d n ( fε (t) E f ε (t) N 0,f ε (t) n d N (0,1). K 2 (v)dv ). (12) For the second term of (??), straightforward calculations show that E f ε (t) f ε (t) = q q! f(q) ε (t) v q K (v)dv +o( q ). Comining this with (??) and (??), we otain the desired result. Lemma 1. Assume (A1)-(A5) and (A7). Then, sup m θo (x) m θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ), x X 0 sup m θo (x) ṁ θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ). x X 0 Proof. We will only show the proof for m θo (x) ṁ θo (x), the proof for m θo (x) m θo (x) eing very similar. Let c n = (nh) 1/2 (logh 1 ) 1/2, and define r θo (x) = 1 nh j=1 Λ θo (Y j )K 1 ( Xj x h x X 0 ), ṙ θo (x) = E[ r θo (x)], f X (x) = E[ f X (x)], where f X (x) = (nh) 1 n j=1 K 1( Xj x h ). Then, sup m θo (x) ṁ θo (x) sup m θo (x) ṙθ o (x) x X 0 f X (x) + sup ṙθo (x) f X (x)ṁ θo (x). (1) f X (x) x X 0 1 Since E[ Λ 4 θ o (Y) X = x] < uniformly in x X y assumption (A7), a similar proof as was given for Theorem 2 in Einmahl and Mason (2005) ensures that sup m θo (x) ṙθ o (x) f X (x) = O P(c n ). x X 0 Consider now the second term of (??). Since E[ ε(θ o ) X] = 0, where ε(θ o ) = d dθ (Λ θ(y) m θ (X)) θ=θo, we have [ ( )] X x ṙ θo (x) = h 1 E {ṁ θo (X)+ ε(θ o )}K 1 h [ ( )] X x = h 1 E ṁ θo (X)K 1 h = ṁ θo (x+hv)k 1 (v)f X (x+hv)dv, 17
18 from which it follows that ṙ θo (x) f X (x)ṁ θo (x) = [ṁ θo (x+hv) ṁ θo (x)]k 1 (v)f X (x+hv)dv. Hence, a Taylor expansion applied to ṁ θo ( ) yields sup ṙθo (x) f X (x)ṁ θo (x) = O(h q 1 ) = O(c n ), x X 0 since nh 2q1+1 (logh 1 ) 1 = O(1) y (A2). This proves that the second term of (??) is O(c n ), since it can e easily shown that f X (x) is ounded away from 0 and infinity, uniformly in x X 0, using (A)(ii). References [1] Ahmad, I. and Li, Q. (1997). Testing symmetry of an unknown density function y kernel method. Journal of Nonparametric Statistics, 7, [2] Akritas, M.G. and Van Keilegom, I. (2001). Non-parametric estimation of the residual distriution. Scandinavian Journal of Statistics, 28, [] Amemiya, T. (1985). Advanced Econometrics. Harvard University Press, Camridge. [4] Bickel, P.J. and Doksum, K. (1981). An analysis of transformations revisited. Journal of the American Statistical Association, 76, [5] Billingsley, P. (1968). Convergence of Proaility Measures. Wiley, New York. [6] Box, G.E.P. and Cox, D.R. (1964). An analysis of transformations. Journal of the Royal Statistical Society - Series B, 26, [7] Carroll, R.J. and Ruppert, D. (1988). Transformation and Weighting in Regression. Chapman and Hall, New York. [8] Chen, G., Lockhart, R.A. and Stephens, A. (2002). Box-Cox transformations in linear models: Large sample theory and tests of normality (with discussion). Canadian Journal of Statistics, 0, [9] Cheng, F. (2005). Asymptotic distriutions of error density and distriution function estimators in nonparametric regression. Journal of Statistical Planning and Inference, 128,
19 [10] Cheng, F. and Sun, S. (2008). A goodness-of-fit test of the errors in nonlinear autoregressive time series models. Statistics and Proaility Letters, 78, [11] Dette, H., Kusi-Appiah, S. and Neumeyer, N. (2002). Testing symmetry in nonparametric regression models. Journal of Nonparametric Statistics, 14, [12] Efromovich, S. (2005). Estimation of the density of the regression errors. Annals of Statistics,, [1] Einmahl, U. and Mason, D.M. (2005). Uniform in andwidth consistency of kernel-type function estimators. Annals of Statistics,, [14] Escanciano, J.C. and Jacho-Chavez, D. (2012). n-uniformly consistent density estimation in nonparametric regression. Journal of Econometrics, 167, [15] Fitzenerger, B., Wilke, R.A. and Zhang, X. (2010). Implementing Box-Cox quantile regression. Econometric Reviews, 29, [16] Horowitz, J.L. (1998). Semiparametric Methods in Economics. Springer-Verlag, New York. [17] Linton, O., Sperlich, S. and Van Keilegom, I. (2008). Estimation of a semiparametric transformation model. Annals of Statistics, 6, [18] Müller, U.U., Schick, A. and Wefelmeyer, W. (2004). Estimating linear functionals of the error distriution in nonparametric regression. Journal of Statistical Planning and Inference, 119, [19] Nadaraya, E. A. (1964). On estimating regression. Theory of Proaility and its Applications, 9, [20] Neumeyer, N. and Dette, H. (2007). Testing for symmetric error distriution in nonparametric regression models. Statistica Sinica, 17, [21] Neumeyer, N. and Van Keilegom, I. (2010). Estimating the error distriution in nonparametric multiple regression with applications to model testing. Journal of Multivariate Analysis, 101, [22] Pinsker, M.S. (1980). Optimal filtering of a square integrale signal in Gaussian white noise. Prolems of Information Transmission, 16, [2] Sakia, R.M. (1992). The Box-Cox transformation technique: a review. The Statistician, 41,
20 [24] Sam, R. (2011). Nonparametric estimation of the density of regression errors. Comptes Rendus de l Académie des Sciences-Paris, Série I 49, [25] Shin, Y. (2008). Semiparametric estimation of the Box-Cox transformation model. Econometrics Journal, 11, [26] Vanhems, A. and Van Keilegom, I. (2011). Semiparametric transformation model with endogeneity: a control function approach. Journal of Econometrics (under revision). [27] Watson, G.S. (1964). Smooth regression analysis. Sankhya - Series A, 26, [28] Zellner, A. and Revankar, N.S. (1969). Generalized production functions. Reviews of Economic Studies, 6, Postal addresses : Benjamin Colling Cédric Heuchenne Université catholique de Louvain HEC-Management School of the University of Liège, Institute of Statistics Statistique appliquée à la gestion et à l économie Voie du Roman Pays 20 Rue Louvrex 14, Bâtiment N1 148 Louvain-la-Neuve 4000 Liège Belgium Belgium Rawane Sam Ingrid Van Keilegom Centre de recherche du CHUQ/CHUL Université catholique de Louvain 2705, Boulevard Laurier Institute of Statistics G1V 4G2, QC Voie du Roman Pays 20 Quéec 148 Louvain-la-Neuve Canada Belgium 20
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2011/2 ESTIMATION
More informationDEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE
Estimating the error distribution in nonparametric multiple regression with applications to model testing Natalie Neumeyer & Ingrid Van Keilegom Preprint No. 2008-01 July 2008 DEPARTMENT MATHEMATIK ARBEITSBEREICH
More informationNonparametric Estimation of Smooth Conditional Distributions
Nonparametric Estimation of Smooth Conditional Distriutions Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~hansen May 24 Astract This paper considers nonparametric estimation of smooth conditional
More informationEstimation of a semiparametric transformation model in the presence of endogeneity
TSE 654 May 2016 Estimation of a semiparametric transformation model in the presence of endogeneity Anne Vanhems and Ingrid Van Keilegom Estimation of a semiparametric transformation model in the presence
More informationD I S C U S S I O N P A P E R
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S ( I S B A ) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2014/06 Adaptive
More informationEstimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement
Open Journal of Statistics, 07, 7, 834-848 http://www.scirp.org/journal/ojs ISS Online: 6-798 ISS Print: 6-78X Estimating a Finite Population ean under Random on-response in Two Stage Cluster Sampling
More informationEstimation in semiparametric models with missing data
Ann Inst Stat Math (2013) 65:785 805 DOI 10.1007/s10463-012-0393-6 Estimation in semiparametric models with missing data Song Xi Chen Ingrid Van Keilegom Received: 7 May 2012 / Revised: 22 October 2012
More informationNonparametric Econometrics
Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-
More informationSingle Index Quantile Regression for Heteroscedastic Data
Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015
More informationDensity estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas
0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity
More informationGoodness-of-fit tests for the cure rate in a mixture cure model
Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,
More informationI N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S I S B A UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 20/4 SEMIPARAMETRIC
More informationSingle Index Quantile Regression for Heteroscedastic Data
Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR
More informationDISCUSSION PAPER 2016/43
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S ( I S B A ) DISCUSSION PAPER 2016/43 Bounds on Kendall s Tau for Zero-Inflated Continuous
More informationGenerated Covariates in Nonparametric Estimation: A Short Review.
Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated
More informationEVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS
APPLICATIONES MATHEMATICAE 9,3 (), pp. 85 95 Erhard Cramer (Oldenurg) Udo Kamps (Oldenurg) Tomasz Rychlik (Toruń) EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS Astract. We
More informationBootstrap of residual processes in regression: to smooth or not to smooth?
Bootstrap of residual processes in regression: to smooth or not to smooth? arxiv:1712.02685v1 [math.st] 7 Dec 2017 Natalie Neumeyer Ingrid Van Keilegom December 8, 2017 Abstract In this paper we consider
More informationGeneralized Seasonal Tapered Block Bootstrap
Generalized Seasonal Tapered Block Bootstrap Anna E. Dudek 1, Efstathios Paparoditis 2, Dimitris N. Politis 3 Astract In this paper a new lock ootstrap method for periodic times series called Generalized
More informationI N S T I T U T D E S T A T I S T I Q U E
I N S T I T U T D E S T A T I S T I Q U E UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 0610 ESTIMATION OF A SEMIPARAMETRIC TRANSFORMATION MODEL O. LINTON, S. SPERLICH and I. VAN KEILEGOM
More informationHIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO
The Annals of Statistics 2006, Vol. 34, No. 3, 1436 1462 DOI: 10.1214/009053606000000281 Institute of Mathematical Statistics, 2006 HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO BY NICOLAI
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationEcon 582 Nonparametric Regression
Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume
More informationAdditive Isotonic Regression
Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive
More informationComputational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data
Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data Géraldine Laurent 1 and Cédric Heuchenne 2 1 QuantOM, HEC-Management School of
More informationarxiv: v1 [math.st] 4 Apr 2008
The Annals of Statistics 2008, Vol. 36, No. 2, 686 718 DOI: 10.1214/009053607000000848 c Institute of Mathematical Statistics, 2008 arxiv:0804.0719v1 [math.st] 4 Apr 2008 ESTIMATION OF A SEMIPARAMETRIC
More informationHigh Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data
High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department
More informationOberwolfach Preprints
Oberwolfach Preprints OWP 202-07 LÁSZLÓ GYÖFI, HAO WALK Strongly Consistent Density Estimation of egression esidual Mathematisches Forschungsinstitut Oberwolfach ggmbh Oberwolfach Preprints (OWP) ISSN
More informationA Bootstrap Test for Conditional Symmetry
ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School
More informationFinQuiz Notes
Reading 9 A time series is any series of data that varies over time e.g. the quarterly sales for a company during the past five years or daily returns of a security. When assumptions of the regression
More informationTransformation and Smoothing in Sample Survey Data
Scandinavian Journal of Statistics, Vol. 37: 496 513, 2010 doi: 10.1111/j.1467-9469.2010.00691.x Published by Blackwell Publishing Ltd. Transformation and Smoothing in Sample Survey Data YANYUAN MA Department
More informationOn a Nonparametric Notion of Residual and its Applications
On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014
More information41903: Introduction to Nonparametrics
41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific
More informationSupplement to Quantile-Based Nonparametric Inference for First-Price Auctions
Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract
More informationNonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix
Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract
More informationM- and Z- theorems; GMM and Empirical Likelihood Wellner; 5/13/98, 1/26/07, 5/08/09, 6/14/2010
M- and Z- theorems; GMM and Empirical Likelihood Wellner; 5/13/98, 1/26/07, 5/08/09, 6/14/2010 Z-theorems: Notation and Context Suppose that Θ R k, and that Ψ n : Θ R k, random maps Ψ : Θ R k, deterministic
More information13 Endogeneity and Nonparametric IV
13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,
More informationError distribution function for parametrically truncated and censored data
Error distribution function for parametrically truncated and censored data Géraldine LAURENT Jointly with Cédric HEUCHENNE QuantOM, HEC-ULg Management School - University of Liège Friday, 14 September
More informationIntroduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β
Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured
More informationNonparametric Modal Regression
Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric
More informationThe properties of L p -GMM estimators
The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion
More informationMinimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.
Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of
More informationDiscussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon
Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics
More informationNonparametric Regression. Changliang Zou
Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,
More informationA Note on the Scale Efficiency Test of Simar and Wilson
International Journal of Business Social Science Vol. No. 4 [Special Issue December 0] Abstract A Note on the Scale Efficiency Test of Simar Wilson Hédi Essid Institut Supérieur de Gestion Université de
More informationTESTING FOR THE EQUALITY OF k REGRESSION CURVES
Statistica Sinica 17(2007, 1115-1137 TESTNG FOR THE EQUALTY OF k REGRESSON CURVES Juan Carlos Pardo-Fernández, ngrid Van Keilegom and Wenceslao González-Manteiga Universidade de Vigo, Université catholique
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationChapter 2 Inference on Mean Residual Life-Overview
Chapter 2 Inference on Mean Residual Life-Overview Statistical inference based on the remaining lifetimes would be intuitively more appealing than the popular hazard function defined as the risk of immediate
More informationStatistical Properties of Numerical Derivatives
Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective
More informationTesting near or at the Boundary of the Parameter Space (Job Market Paper)
Testing near or at the Boundary of the Parameter Space (Jo Market Paper) Philipp Ketz Brown University Novemer 7, 24 Statistical inference aout a scalar parameter is often performed using the two-sided
More informationOn the Robust Modal Local Polynomial Regression
International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationLocal linear multiple regression with variable. bandwidth in the presence of heteroscedasticity
Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable
More informationNONPARAMETRIC ENDOGENOUS POST-STRATIFICATION ESTIMATION
Statistica Sinica 2011): Preprint 1 NONPARAMETRIC ENDOGENOUS POST-STRATIFICATION ESTIMATION Mark Dahlke 1, F. Jay Breidt 1, Jean D. Opsomer 1 and Ingrid Van Keilegom 2 1 Colorado State University and 2
More informationA New Test in Parametric Linear Models with Nonparametric Autoregressive Errors
A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors By Jiti Gao 1 and Maxwell King The University of Western Australia and Monash University Abstract: This paper considers a
More informationEstimation of the Bivariate and Marginal Distributions with Censored Data
Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationA converse Gaussian Poincare-type inequality for convex functions
Statistics & Proaility Letters 44 999 28 290 www.elsevier.nl/locate/stapro A converse Gaussian Poincare-type inequality for convex functions S.G. Bokov a;, C. Houdre ; ;2 a Department of Mathematics, Syktyvkar
More informationGraduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More informationFIDUCIAL INFERENCE: AN APPROACH BASED ON BOOTSTRAP TECHNIQUES
U.P.B. Sci. Bull., Series A, Vol. 69, No. 1, 2007 ISSN 1223-7027 FIDUCIAL INFERENCE: AN APPROACH BASED ON BOOTSTRAP TECHNIQUES H.-D. HEIE 1, C-tin TÂRCOLEA 2, Adina I. TARCOLEA 3, M. DEMETRESCU 4 În prima
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationMotivational Example
Motivational Example Data: Observational longitudinal study of obesity from birth to adulthood. Overall Goal: Build age-, gender-, height-specific growth charts (under 3 year) to diagnose growth abnomalities.
More informationSmoothing the Nelson-Aalen Estimtor Biostat 277 presentation Chi-hong Tseng
Smoothing the Nelson-Aalen Estimtor Biostat 277 presentation Chi-hong seng Reference: 1. Andersen, Borgan, Gill, and Keiding (1993). Statistical Model Based on Counting Processes, Springer-Verlag, p.229-255
More informationSpiking problem in monotone regression : penalized residual sum of squares
Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where
More informationSubmitted to the Brazilian Journal of Probability and Statistics
Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a
More informationMinimizing a convex separable exponential function subject to linear equality constraint and bounded variables
Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad
More informationA note on L convergence of Neumann series approximation in missing data problems
A note on L convergence of Neumann series approximation in missing data problems Hua Yun Chen Division of Epidemiology & Biostatistics School of Public Health University of Illinois at Chicago 1603 West
More informationThe Bayesian Approach to Multi-equation Econometric Model Estimation
Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation
More informationStatistica Sinica Preprint No: SS
Statistica Sinica Preprint No: SS-017-0013 Title A Bootstrap Method for Constructing Pointwise and Uniform Confidence Bands for Conditional Quantile Functions Manuscript ID SS-017-0013 URL http://wwwstatsinicaedutw/statistica/
More informationSummary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff
Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)
More informationLECTURE 18: NONLINEAR MODELS
LECTURE 18: NONLINEAR MODELS The basic point is that smooth nonlinear models look like linear models locally. Models linear in parameters are no problem even if they are nonlinear in variables. For example:
More informationIdentification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case
Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Arthur Lewbel Boston College Original December 2016, revised July 2017 Abstract Lewbel (2012)
More informationDepth versus Breadth in Convolutional Polar Codes
Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK
More informationNonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity
Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University
More informationClosest Moment Estimation under General Conditions
Closest Moment Estimation under General Conditions Chirok Han and Robert de Jong January 28, 2002 Abstract This paper considers Closest Moment (CM) estimation with a general distance function, and avoids
More informationA732: Exercise #7 Maximum Likelihood
A732: Exercise #7 Maximum Likelihood Due: 29 Novemer 2007 Analytic computation of some one-dimensional maximum likelihood estimators (a) Including the normalization, the exponential distriution function
More informationSemiparametric modeling and estimation of the dispersion function in regression
Semiparametric modeling and estimation of the dispersion function in regression Ingrid Van Keilegom Lan Wang September 4, 2008 Abstract Modeling heteroscedasticity in semiparametric regression can improve
More informationTime Series and Forecasting Lecture 4 NonLinear Time Series
Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations
More informationRobust Backtesting Tests for Value-at-Risk Models
Robust Backtesting Tests for Value-at-Risk Models Jose Olmo City University London (joint work with Juan Carlos Escanciano, Indiana University) Far East and South Asia Meeting of the Econometric Society
More informationSimple Examples. Let s look at a few simple examples of OI analysis.
Simple Examples Let s look at a few simple examples of OI analysis. Example 1: Consider a scalar prolem. We have one oservation y which is located at the analysis point. We also have a ackground estimate
More informationGaussian kernel GARCH models
Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often
More informationUNIVERSITY OF CALIFORNIA Spring Economics 241A Econometrics
DEPARTMENT OF ECONOMICS R. Smith, J. Powell UNIVERSITY OF CALIFORNIA Spring 2006 Economics 241A Econometrics This course will cover nonlinear statistical models for the analysis of cross-sectional and
More informationChapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets
Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence
More informationVariance Function Estimation in Multivariate Nonparametric Regression
Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered
More informationBayesian estimation of bandwidths for a nonparametric regression model with a flexible error density
ISSN 1440-771X Australia Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Bayesian estimation of bandwidths for a nonparametric regression model
More informationTIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION
TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION Jiangping Wang and Dapeng Wu Department of Electrical and Computer Engineering University of Florida, Gainesville, FL 3611 Correspondence author: Prof.
More informationBayesian inference with reliability methods without knowing the maximum of the likelihood function
Bayesian inference with reliaility methods without knowing the maximum of the likelihood function Wolfgang Betz a,, James L. Beck, Iason Papaioannou a, Daniel Strau a a Engineering Risk Analysis Group,
More informationLocal Polynomial Regression
VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based
More informationPENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA
PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA Kasun Rathnayake ; A/Prof Jun Ma Department of Statistics Faculty of Science and Engineering Macquarie University
More informationNonparametric Identification and Estimation of a Transformation Model
Nonparametric and of a Transformation Model Hidehiko Ichimura and Sokbae Lee University of Tokyo and Seoul National University 15 February, 2012 Outline 1. The Model and Motivation 2. 3. Consistency 4.
More informationarxiv: v1 [cs.gt] 4 May 2015
Econometrics for Learning Agents DENIS NEKIPELOV, University of Virginia, denis@virginia.edu VASILIS SYRGKANIS, Microsoft Research, vasy@microsoft.com EVA TARDOS, Cornell University, eva.tardos@cornell.edu
More informationSupplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC)
Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Authors: Alexei C. Ionan, Mei-Yin C. Polley, Lisa M. McShane, Kevin K. Doin Section Page
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationOn Universality of Blow-up Profile for L 2 critical nonlinear Schrödinger Equation
On Universality of Blow-up Profile for L critical nonlinear Schrödinger Equation Frank Merle,, Pierre Raphael Université de Cergy Pontoise Institut Universitaire de France Astract We consider finite time
More informationMerging and splitting endowments in object assignment problems
Merging and splitting endowments in oject assignment prolems Nanyang Bu, Siwei Chen, and William Thomson April 26, 2012 1 Introduction We consider a group of agents, each endowed with a set of indivisile
More informationContributions to extreme-value analysis
Contributions to extreme-value analysis Stéphane Girard INRIA Rhône-Alpes & LJK (team MISTIS). 655, avenue de l Europe, Montbonnot. 38334 Saint-Ismier Cedex, France Stephane.Girard@inria.fr February 6,
More informationIssues on quantile autoregression
Issues on quantile autoregression Jianqing Fan and Yingying Fan We congratulate Koenker and Xiao on their interesting and important contribution to the quantile autoregression (QAR). The paper provides
More informationCritical value of the total debt in view of the debts. durations
Critical value of the total det in view of the dets durations I.A. Molotov, N.A. Ryaova N.V.Pushov Institute of Terrestrial Magnetism, the Ionosphere and Radio Wave Propagation, Russian Academy of Sciences,
More informationSpectrum Opportunity Detection with Weak and Correlated Signals
Specum Opportunity Detection with Weak and Correlated Signals Yao Xie Department of Elecical and Computer Engineering Duke University orth Carolina 775 Email: yaoxie@dukeedu David Siegmund Department of
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationA Local Generalized Method of Moments Estimator
A Local Generalized Method of Moments Estimator Arthur Lewbel Boston College June 2006 Abstract A local Generalized Method of Moments Estimator is proposed for nonparametrically estimating unknown functions
More information