Estimation of the Error Density in a Semiparametric Transformation Model

Size: px
Start display at page:

Download "Estimation of the Error Density in a Semiparametric Transformation Model"

Transcription

1 Estimation of the Error Density in a Semiparametric Transformation Model Benjamin Colling Université catholique de Louvain Cédric Heuchenne University of Liège and Université catholique de Louvain Rawane Sam Université catholique de Louvain Ingrid Van Keilegom Université catholique de Louvain Novemer 14, 201 Astract Consider the semiparametric transformation model Λ θo (Y) = m(x) + ε, where θ o is an unknown finite dimensional parameter, the functions Λ θo and m are smooth, ε is independent of X, and E(ε) = 0. We propose a kernel-type estimator of the density of the error ε, and prove its asymptotic normality. The estimated errors, which lie at the asis of this estimator, are otained from a profile likelihood estimator of θ o and a nonparametric kernel estimator of m. The practical performance of the proposed density estimator is evaluated in a simulation study. Key Words: Density estimation; Kernel smoothing; Nonparametric regression; Profile likelihood; Transformation model. Running Head: Error density estimation in transformation models Research supported y IAP research network P7/06 of the Belgian Government (Belgian Science Policy), and y the contract Projet d Actions de Recherche Concertées (ARC) 11/16-09 of the Communauté française de Belgique, granted y the Académie universitaire Louvain. Research supported y IAP research network P7/06 of the Belgian Government (Belgian Science Policy). Research supported y the European Research Council under the European Community s Seventh Framework Programme (FP7/ ) / ERC Grant agreement No , y IAP research network P7/06 of the Belgian Government (Belgian Science Policy), and y the contract Projet d Actions de Recherche Concertées (ARC) 11/16-09 of the Communauté française de Belgique, granted y the Académie universitaire Louvain. 1

2 1 Introduction Let (X 1,Y 1 ),...,(X n,y n ) e independent replicates of the random vector (X,Y), where Y is a univariate dependent variale and X is a one-dimensional covariate. We assume that Y and X are related via the semiparametric transformation model Λ θo (Y) = m(x)+ε, (1) where ε is independent of X and has mean zero. We assume that {Λ θ : θ Θ} (with Θ R p compact) is a parametric family of strictly increasing functions defined on an unounded suset D in R, while m is the unknown regression function, elonging to an infinite dimensional parameter set M. We assume that M is a space of functions endowed with the norm M =. We denote θ o Θ and m M for the true unknown finite and infinite dimensional parameters. Define the regression function m θ (x) = E[Λ θ (Y) X = x], for each θ Θ, and let ε θ = ε(θ) = Λ θ (Y) m θ (X). In this paper, we are interested in the estimation of the proaility density function (p.d.f.) f ε of the residual term ε = Λ θo (Y) m(x). To this end, we first otain the estimators θ and m θ of the parameter θ o and the function m θ, and second, form the semiparametric regression residuals ε i ( θ) = Λ θ(y i ) m θ(x i ). To estimate θ o we use a profile likelihood (PL) approach, developed in Linton et al. (2008), whereas m θ is estimated y means of a Nadaraya-Watson-type estimator (Nadaraya (1964), Watson (1964)). To our knowledge, the estimation of the density of ε in model (??) has not yet een investigated in the statistical literature. Estimatingtheerrordensityinthesemiparametrictransformationmodel(SPT)Λ θo (Y) = m(x)+ ε may e very useful in various regression prolems. First, taking transformations of the data may induce normality and error variance homogeneity in the transformed model. So the estimation of the error density in the transformed model may e used for testing these hypotheses; it may also e used for goodness-of-fit tests of a specified error distriution in a parametric or nonparametric regression setting. Some examples can e found in Akritas and Van Keilegom (2001), Cheng and Sun (2008), ut with Λ θo id, i.e. the response is not transformed. Next, the estimation of the error density in the aove model can e useful for testing the symmetry of the residual distriution. See Ahmad and Li (1997), Dette et al. (2002), Neumeyer and Dette (2007) and references therein, in the case Λ θo id. Under this model, Escanciano and Jacho-Chavez (2012) considered the estimation of the (marginal) density of the response Y via the estimation of the error density. Another application of the estimation of the error density in the SPT model is the forecasting of Λ θo (Y) y 2

3 means of the mode approach, since the mode of the p.d.f. of Λ θo (Y) given X = x is m(x)+argmax e R f ε (e), where f ε is the p.d.f. of the error term ε. Taking transformations of the data has een an important part of statistical practice for many years. A major contriution to this methodology was made y Box and Cox (1964), who proposed a parametric power family of transformations that includes the logarithm and the identity. They suggested that the power transformation, when applied to the dependent variale in a linear regression model, might induce normality and homoscedasticity. Lots of effort has een devoted to the investigation of the Box-Cox transformation since its introduction. See, for example, Amemiya (1985), Horowitz (1998), Chen et al. (2002), Shin (2008), and Fitzenerger et al. (2010). Other dependent variale transformations have een suggested, for example, the Zellner and Revankar(1969) transform and the Bickel and Doksum(1981) transform. The transformation methodology has een quite successful and a large literature exists on this topic for parametric models. See Carroll and Ruppert (1988) and Sakia (1992) and references therein. The estimation of (functionals of) the error distriution and density under simplified versions of model (??) has received considerale attention in the statistical literature in recent years. Akritas and Van Keilegom (2001) estimated the cumulative distriution function of the regression error in a heteroscedastic model with univariate covariates. The estimator they proposed is ased on nonparametrically estimated regression residuals. The weak convergence of their estimator was proved. The results otained y Akritas and Van Keilegom (2001) have een generalized y Neumeyer and Van Keilegom (2010) to the case of multivariate covariates. Müller et al. (2004) investigated linear functionals of the error distriution in nonparametric regression. Cheng (2005) estalished the asymptotic normality of an estimator of the error density ased on estimated residuals. The estimator he proposed is constructed y splitting the sample into two parts: the first part is used for the estimation of the residuals, while the second part of the sample is used for the construction of the error density estimator. Efromovich (2005) proposed an adaptive estimator of the error density, ased on a density estimator proposed y Pinsker (1980). Finally, Sam (2011) also considered the estimation of the error density, ut his approach is more closely related to the one in Akritas and Van Keilegom (2001). In order to achieve the ojective of this paper, namely the estimation of the error density under model (??), we first need to estimate the transformation parameter θ o. To this end, we make use of the results in Linton et al. (2008). In the latter paper, the authors first discuss the nonparametric identification of model (??), and second, estimate the transformation parameter θ o under the considered model. For the estimation of this parameter, they propose two approaches. The first approach uses a semiparametric profile likelihood

4 (PL) estimator, while the second is ased on a mean squared distance from independence-estimator (MD) using the estimated distriutions of X, ε and (X, ε). Linton et al. (2008) derived the asymptotic distriutions of their estimators under certain regularity conditions, and proved that oth estimators of θ o are root-n consistent. The authors also showed that, in practice, the performance of the PL method is etter than that of the MD approach. For this reason, the PL method will e considered in this paper for the estimation of θ o. The rest of the paper is organized as follows. Section 2 presents our estimator of the error density and groups some notations and technical assumptions. Section descries the asymptotic results of the paper. A simulation study is given in Section 4, while Section 5 is devoted to some general conclusions. Finally, the proofs of the asymptotic results are collected in Section 6. 2 Definitions and assumptions 2.1 Construction of the estimators The approach proposed here for the estimation of f ε is ased on a two-steps procedure. In a first step, we estimate the finite dimensional parameter θ o. This parameter is estimated y the profile likelihood (PL) method, developed in Linton et al.(2008). The asic idea of this method is to replace all unknown expressions in the likelihood function y their nonparametric kernel estimates. Under model (??), we have P(Y y X) = P(Λ θo (Y) Λ θo (y) X) = P(ε θo Λ θo (y) m θo (X) X) = F ε (Λ θo (y) m θo (X)). Here, F ε (t) = P(ε t), and so f Y X (y x) = f ε (Λ θo (y) m θo (x))λ θ o (y), where f ε and f Y X are the densities of ε, and of Y given X, respectively. Then, the log likelihood function is {logf εθ (Λ θ (Y i ) m θ (X i ))+logλ θ (Y i)}, θ Θ, where f εθ is the density function of ε θ. Now, let ( ) n j=1 Λ Xj x θ(y j )K 1 m θ (x) = n j=1 K 1( Xj x h h ) (2) e the Nadaraya-Watson estimator of m θ (x), and let f εθ (t) = 1 ) ( εi (θ) t K 2. () ng g 4

5 where ε i (θ) = Λ θ (Y i ) m θ (X i ). Here, K 1 and K 2 are kernel functions and h and g are andwidth sequences. Then, the PL estimator of θ o is defined y [ θ = argmax log f ] εθ (Λ θ (Y i ) m θ (X i ))+logλ θ (Y i). (4) θ Θ Recall that m θ (X i ) converges to m θ (X i ) at a slower rate for those X i which are close to the oundary of the support X of the covariate X. That is why we assume implicitly that the proposed estimator (??) of θ o trims the oservations X i outside a suset X 0 of X. Note that we keep the root-n consistency of θ proved in Linton et al. (2008) y trimming the covariates outside X 0. But in this case, the resulting asymptotic variance is different to the one otained in the latter paper. In a second step, we use the aove estimator θ to uild the estimated residuals ε i ( θ) = Λ θ(y i ) m θ(x i ). Then, our proposed estimator f ε (t) of f ε (t) is defined y ( ) f ε (t) = 1 ε i ( θ) t K, (5) n where K is a kernel function and is a andwidth sequence, not necessarily the same as the kernel K 2 and the andwidth g used in (??). Oserve that this estimator is a feasile estimator in the sense that it does not depend on any unknown quantity, as is desirale in practice. This contrasts with the unfeasile ideal kernel estimator f ε (t) = 1 n ( ) εi t K, (6) which depends in particular on the unknown regression errors ε i = ε i (θ o ) = Λ θo (Y i ) m(x i ). It is however intuitively clear that f ε (t) and f ε (t) will e very close for n large enough, as will e illustrated y the results given in Section. 2.2 Notations When there is no amiguity, we use ε and m to indicate ε θo and m θo. Moreover, N(θ o ) represents a neighorhood of θ o. For the kernel K j (j = 1,2,), let µ(k j ) = v 2 K j (v)dv and let K (p) j e the pth derivative of K j. For any function ϕ θ (y), denote ϕ θ (y) = ϕ θ (y)/ θ = ( ϕ θ (y)/ θ 1,..., ϕ θ (y)/ θ p ) t and ϕ θ (y) = ϕ θ(y)/ y. Also, let A = (A t A) 1/2 e the Euclidean norm of any vector A. For any functions m, r, f, ϕ and q, and any θ Θ, let s = ( m,r,f,ϕ,q), s θ = (m θ,ṁ θ,f εθ,f ε θ, f εθ ), ε i (θ, m) = Λ θ (Y i ) m(x i ), and define { n G n (θ,s) = n 1 1 f{ε i (θ, m)} [ ] ϕ{ε i (θ, m)}{ Λ θ (Y i ) r(x i )}+q{ε i (θ, m)} + Λ } θ (Y i) Λ θ (Y, i) 5

6 G(θ,s) = E[G n (θ,s)] and G(θ o,s θo ) = θ G(θ,s θ) θ=θo. 2. Technical assumptions The assumptions we need for the asymptotic results are listed elow for convenient reference. (A1) The function K j (j = 1,2,) is symmetric, has compact support, v k K j (v)dv = 0 for k = 1,...,q j 1 and v qj K j (v)dv 0 for some q j 4, K j is twice continuously differentiale, and K (1) (v)dv = 0. (A2) The andwidth sequences h, g and satisfy nh 2q1 = o(1), ng 2q2 = o(1) (where q 1 and q 2 are defined in (A1)), (n 5 ) 1 = O(1), n h 2 (logh 1 ) 2 and ng 6 (logg 1 ) 2. (A) (i) The support X of the covariate X is a compact suset of R, and X 0 is a suset with non empty interior, whose closure is in the interior of X. (ii) The density f X is ounded away from zero and infinity on X, and has continuous second order partial derivatives on X. (A4) The function m θ (x) is twice continuously differentiale with respect to θ on X N(θ 0 ), and the functions m θ (x) and ṁ θ (x) are q 1 times continuously differentiale with respect to x on X N(θ 0 ). All these derivatives are ounded, uniformly in (x,θ) X N(θ o ). (A5) The error ε = Λ θo (Y) m(x) has finite fourth moment and is independent of X. (A6) The distriution F εθ (t) is q +1 (respectively three) times continuously differentiale with respect to t (respectively θ), and sup θ,t k+l t k θ l θlp p F εθ (t) < for all k and l such that 0 k +l 2, where l = l l p and θ = (θ 1,...,θ p ) t. (A7) The transformation Λ θ (y) is three times continuously differentiale with respect to oth θ and y, and there exists a α > 0 such that E [ sup θ : θ θ α ] k+l Λ θ (Y) < y k θ l θlp p for all θ Θ, and for all k and l such that 0 k + l, where l = l l p and θ = (θ 1,...,θ p ) t. Moreover, sup x X E[ Λ 4 θ o (Y) X = x] <. (A8) For all η > 0, there exists ǫ(η) > 0 such that inf G(θ,s θ) ǫ(η) > 0. θ θ o >η 6

7 Moreover, the matrix G(θ o,s θo ) is non-singular. (A9) (i) E(Λ θo (Y)) = 1, Λ θo (0) = 0 and the set {x X 0 : m (x) 0} has nonempty interior. (ii) Assume that φ(x,t) = Λ θo (Λ 1 θ o (m(x) +t))f ε (t) is continuously differentiale with respect to t for all x and that for all t R and for some δ > 0. sup s: t s δ E φ s (X,s) <. (7) Assumptions (A1), part of (A2), (A)(ii), (A4) and (A6), (A7) and (A8) are used y Linton et al. (2008) to show that the PL estimator θ of θ o is root n-consistent. The differentiaility of K j up to second order imposed in assumption (A1) is used to expand the two-steps kernel estimator f ε (t) in (??) around the unfeasile one f ε (t). Assumptions (A)(ii) and (A4) impose that all the functions to e estimated have ounded derivatives. The last assumption in (A2) is useful for otaining the uniform convergence of the Nadaraya-Watson estimator of m θo in (??) (see for instance Einmahl and Mason (2005)). This assumption is also needed in the study of the difference etween the feasile estimator f ε (t) and the unfeasile estimator f ε (t). Finally, (A9)(i) is needed for identifying the model (see Vanhems and Van Keilegom (2011)). Asymptotic results In this section we are interested in the asymptotic ehavior of the estimator f ε (t). To this end, we first investigate its asymptotic representation, which will e needed to show its asymptotic normality. Theorem 1. Assume (A1)-(A9). Then, f ε (t) f ε (t) = 1 n where R n (t) = o P ( (n) 1/2 ) for all t R. ( ) εi t K f ε (t)+r n (t), This result is important, since it shows that, provided the ias term is negligile, the estimation of θ o and m( ) has asymptotically no effect on the ehavior of the estimator f ε (t). Therefore, this estimator is asymptotically equivalent to the unfeasile estimator f ε (t), ased on the unknown true errors ε 1,...,ε n. Our next result gives the asymptotic normality of the estimator f ε (t). Theorem 2. Assume (A1)-(A9). In addition, assume that n 2q+1 = O(1). Then, ) ( ) d n ( f ε (t) f ε (t) N 0,f ε (t) K(v)dv 2, 7

8 where f ε (t) = f ε (t)+ q q! f(q) ε (t) v q K (v)dv. The proofs of Theorems?? and?? are given in Section??. 4 Simulations In this section, we investigate the performance of our method for different models and different sample sizes. Consider where Λ θ is the Box-Cox (1964) transformation y θ 1 θ, θ 0, Λ θ (y) = log(y), θ = 0, Λ θo (Y) = X sin(πx)+σ e ε, (8) X is uniformly distriuted on the interval [ 1,1], and ε is independent of X. We carry out simulations for two cases : in the first case, ε has a standard normal distriution and, in the second case, the distriution of ε is the mixture of the normal distriutions N( 1.5,0.25) and N(1.5,0.25) with equal weights. To make computations easier, error distriutions are truncated at and (i.e., put to 0 outside the interval [, ]). We study three different model settings. For each of them, 2 = 0 σ e. The other parameters are chosen as follows: Model 1: 0 = 6.5, 1 = 5, σ e = 1.5; Model 2: 0 = 4.5, 1 =.5, σ e = 1; Model : 0 = 2.5, 1 = 2.5, σ e = 0.5. Oursimulationsareperformedforθ 0 = 0,0.5and1. WeusetheEpanechnikovkernelK(x) = 4 ( 1 x 2 ) 1( x 1) for oth the estimator of the regression function and the density function. The results are ased on 100 random samples of size n = 100 and n = 200. For the estimation of θ 0 and f ε (t), we proceed as follows. Let L θ (h,g) = [ log f ] εθ ( ε i (θ,h))+logλ θ(y i ), where ε i (θ,h) = Λ θ (Y i ) m θ (X i,h)and m θ (x,h) denotes m θ (x) constructedwith andwidthh.this function will e maximized with respect to θ for given (optimal) values of (h,g). For each value of θ, h (θ) is otained 8

9 y least squares cross-validation, where h (θ) = argmax h m i,θ (X i ) = (Λ θ (Y i ) m i,θ (X i )) 2, ) n j=1,j i Λ Xj X θ(y j )K( i n j=1,j i K ( Xj X i h and g can e chosen with a classical andwidth selection rule for kernel density estimation. Here, for simplicity, the normal rule is used (ĝ(θ) = (40 π) 1/5 n 1/5 σ ε(θ,h (θ)), where σ ε(θ,h (θ)) is the classical empirical estimator of the standard deviation ased on ε i (θ,h (θ)), i = 1,...,n). The solution h ) θ = argmaxl θ (h (θ),ĝ(θ)) θ is therefore otained iteratively (maximization prolems are solved with the function optimize in R with h [0,2] and θ [ 20,20]) and the estimator of f ε (t) is finally given y ( ) 1 ε i ( θ,h ( θ)) t f ε (t) = K. nĝ( θ) ĝ( θ) Tales??,?? and?? show the mean squared error (MSE) of the estimator f ε (t) of the standardized (pseudo-estimated) error ε = ( Λ θ(y) m θ(x) ) /σ e (with known σ e ), for t = 1, 0 and 1 (respectively t = 1.5, 1, 0, 1 and 1.5) and for the unimodal (respectively imodal) normal error distriution. Tales?? and?? show the integrated mean squared error (IMSE) of the estimator f ε ( ) for oth error distriutions, where the integration is done over the interval [,]. As expected, in oth cases, estimation is etter for the normal density than for the mixture of two normals, and estimation improves when n increases, and in most cases, when σ e decreases. In particular, this can e oservedfrom Tales?? and??. The limiting case θ 0 = 0 (the logarithmic transformation) seems to e more easily captured, especially when the error is normally distriuted. In general, we oserve from Tales??,??,?? that estimation is poorer near local maxima and minima of the density, which is not uncommon for kernel smoothing methods. This also suggests that the choice of the smoothing parameters is important and should e the oject of further investigation. 9

10 Model θ 0 n = 100 n = 200 f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ ( 1) f ǫ (0) f ǫ (1) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale 1: MSE( f ε (t)) for different models, values of t and sample sizes, when f ε ( ) is a standard normal density. 10

11 Model θ 0 n = 100 n = = 6.5 θ 0 = = 5 θ 0 = σ 0 = 1.5 θ 0 = = 4.5 θ 0 = =.5 θ 0 = σ 0 = 1 θ 0 = = 2.5 θ 0 = = 2.5 θ 0 = σ 0 = 0.5 θ 0 = Tale 2: IMSE( f ε ) for different models and sample sizes, when f ε ( ) is a standard normal density. 5 Conclusions In this paper we have studied the estimation of the density of the error in a semiparametric transformation model. The regression function in this model is unspecified (except for some smoothness assumptions), whereas the transformation (of the dependent variale in the model) is supposed to elong to a parametric family of monotone transformations. The proposed estimator is a kernel-type estimator, and we have shown its asymptotic normality. The finite sample performance of the estimator is illustrated y means of a simulation study. It would e interesting to explore various possile applications of the results in this paper. For example, one could use the results on the estimation of the error density to test hypotheses concerning e.g. the normality of the errors, the homoscedasticity of the error variance, or the linearity of the regression function, all of which are important features in the context of transformation models. 6 Proofs Proof of Theorem??. Write f ε (t) f ε (t) = [ f ε (t) f ε (t)]+[ f ε (t) f ε (t)], where f ε (t) = 1 n ) ( εi t K 11

12 Model θ 0 n = 100 f ǫ ( 1.5) f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ (1.5) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale : MSE( f ε (t)) for different models, values of t and n = 100, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. 12

13 Model θ 0 n = 200 f ǫ ( 1.5) f ǫ ( 1) f ǫ (0) f ǫ (1) f ǫ (1.5) Bias θ 0 = 0 Var MSE = 6.5 Bias = 5 θ 0 = 0.5 Var σ 0 = 1.5 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 4.5 Bias =.5 θ 0 = 0.5 Var σ 0 = 1 MSE Bias θ 0 = 1 Var MSE Bias θ 0 = 0 Var MSE = 2.5 Bias = 2.5 θ 0 = 0.5 Var σ 0 = 0.5 MSE Bias θ 0 = 1 Var MSE Tale 4: MSE( f ε (t)) for different models, values of t and n = 200, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. 1

14 Model θ 0 n = 100 n = = 6.5 θ 0 = = 5 θ 0 = σ 0 = 1.5 θ 0 = = 4.5 θ 0 = =.5 θ 0 = σ 0 = 1 θ 0 = = 2.5 θ 0 = = 2.5 θ 0 = σ 0 = 0.5 θ 0 = Tale 5: IMSE( f ε ) for different models and sample sizes, when f ε ( ) is a mixture of two normal densities (N( 1.5, 0.25), N(1.5, 0.25)) with equal weights. and ε i = Λ θo (Y i ) m θo (X i ), i = 1,...,n. In a completely similar way as was done for Lemma A.1 in Linton et al. (2008), it can e shown that f ε (t) f ε (t) = 1 n ( ) εi t K f ε (t)+o P ((n) 1/2 ) (9) for all t R. Note that the remainder term in Lemma A.1 in the aove paper equals a sum of i.i.d. terms of mean zero, plus a o P (n 1/2 ) term. Hence, the remainder term in that paper is O P (n 1/2 ), whereas we write o P ((n) 1/2 ) in (??). Therefore,the resultofthe theoremfollowsifweprovethat f ε (t) f ε (t) = o P ((n) 1/2 ). To this end, write f ε (t) f ε (t) 1 = n 2 ( ε i ( θ) ε i (θ o ))K (1) + 1 2n ( ε i ( θ) ε i (θ o )) 2 K (2) ) ( εi (θ o ) t ( ε i (θ o )+β( ε i ( θ) ε i (θ o )) t for some β (0,1). In what follows, we will show that each of the terms aove is o P ((n) 1/2 ). First consider the last term of (??). Since Λ θ (y) and m θ (x) are oth twice continuously differentiale with respect to θ, the second order Taylor expansion gives, for some θ 1 etween θ o and θ (to simplify the notations, we assume ), 14

15 here that p = dim(θ) = 1), ε i ( θ) ε i (θ o ) = Λ θ(y i ) Λ θo (Y i ) ( m θ(x i ) m θo (X i ) ) = ( θ θ o )( Λ θo (Y i ) m θo (X i ))+ 1 2 ( θ θ o ) 2 ( Λ θ1 (Y i ) m θ1 (X i )). Therefore, since θ θ o = o P ((n) 1/2 ) y Theorem 4.1 in Linton et al. (2008) (as efore, we work with a slower rate than what is shown in the latter paper, since this leads to weaker conditions on the andwidths), assumptions (A2) and (A7) imply that 1 n ( ε i ( θ) ε i (θ o )) 2 K (2) ( ) ε i (θ o )+β( ε i ( θ) ε i (θ o )) t ( = O P (n ) 1), which is o P ((n) 1/2 ), since (n 5 ) 1 = O(1) under (A2). For the first term of (??), the decomposition of ε i ( θ) ε i (θ o ) given aove yields 1 n 2 ) ( εi ( ε i ( θ) ε i (θ o ))K (1) (θ o ) t ( Λ θo (Y i ) m θo (X i ))K (1) = ( θ θ o ) n 2 = ( θ θ o ) n 2 ( Λ θo (Y i ) ṁ θo (X i ))K (1) ( εi (θ o ) t ( εi t ) +o P ((n) 1/2 ) where the last equality follows from a Taylor expansion applied to K (1), the fact that m θo (x) ṁ θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ), ) +o P ((n) 1/2 ), (10) uniformly in x X 0 y Lemma??, and the fact that nh (logh 1 ) 1 under (A2). Further, write [ n ( ) ] E ( Λ θo (Y i ) ṁ θo (X i ))K (1) εi t [ ( )] = E Λ θo (Y i )K (1) εi t [ ( )] E[ṁ θo (X i )]E K (1) εi t = A n B n. We will only show that the first term aove is O(n 2 ) for any t R. The proof for the other term is similar. Let ϕ(x,t) = Λ θo (Λ 1 θ o (m(x)+t)) and set φ(x,t) = ϕ(x,t)f ε (t). Then, applying a Taylor expansion to φ(x, ), 15

16 it follows that (for some β (0,1)) A n = [ ( E Λ θo Λ 1 θ o (m(x i )+ε i ) ) K (1) ( e t = n φ(x,e)k (1) = n φ(x,t+v)k (1) = n = n 2 ) f X (x)dxde (v)f X(x)dxdv [ φ(x,t)+v φ ] t (x,t+βv) ( )] εi t v φ t (x,t+βv)k(1) (v)f X(x)dxdv, K (1) (v)f X(x)dxdv since K (1) (v)dv = 0, and this is ounded y Kn2 sup s: t s δ E φ s (X,s) = O(n2 ) y assumption (A9)(ii). Hence, Tcheychev s inequality ensures that ( θ θ o ) 2 ( Λ θo (Y i ) ṁ θo (X i ))K (1) ( ) εi t = ( θ θ o ) n 2 O P (n 2 +(n) 1/2 ) = o P ((n) 1/2 ), since n /2 y (A2). Sustituting this in (??), yields 1 n 2 ) ( εi ( ε i ( θ) ε i (θ o ))K (1) (θ o ) t = o P ((n) 1/2 ), for any t R. This completes the proof. Proof of Theorem??. It follows from Theorem?? that f ε (t) f ε (t) = [ f ε (t) E f ε (t)]+[e f ε (t) f ε (t)]+o P ((n) 1/2 ). (11) The first term on the right hand side of (??) is treated y Lyapounov s Central Limit Theorem (LCT) for triangular arrays (see e.g. Billingsley (1968), Theorem 7.). To this end, let f in (t) = 1 K ( εi t Then, under (A1), (A2) and (A5) it can e easily shown that n E fin (t) E f in (t) Cn 2 f ε (t) K (v) dv +o ( n 2) ( n ) /2 ( = O((n) 1/2 ) = o(1), Var f in (t) n 1 f ε (t) K 2 (v)dv +o( n 1)) /2 ). 16

17 for some C > 0. Hence, the LCT ensures that This gives f ε (t) E f ε (t) Var f ε (t) = f ε (t) E f ε (t) Var f1n(t) ) ( d n ( fε (t) E f ε (t) N 0,f ε (t) n d N (0,1). K 2 (v)dv ). (12) For the second term of (??), straightforward calculations show that E f ε (t) f ε (t) = q q! f(q) ε (t) v q K (v)dv +o( q ). Comining this with (??) and (??), we otain the desired result. Lemma 1. Assume (A1)-(A5) and (A7). Then, sup m θo (x) m θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ), x X 0 sup m θo (x) ṁ θo (x) = O P ((nh) 1/2 (logh 1 ) 1/2 ). x X 0 Proof. We will only show the proof for m θo (x) ṁ θo (x), the proof for m θo (x) m θo (x) eing very similar. Let c n = (nh) 1/2 (logh 1 ) 1/2, and define r θo (x) = 1 nh j=1 Λ θo (Y j )K 1 ( Xj x h x X 0 ), ṙ θo (x) = E[ r θo (x)], f X (x) = E[ f X (x)], where f X (x) = (nh) 1 n j=1 K 1( Xj x h ). Then, sup m θo (x) ṁ θo (x) sup m θo (x) ṙθ o (x) x X 0 f X (x) + sup ṙθo (x) f X (x)ṁ θo (x). (1) f X (x) x X 0 1 Since E[ Λ 4 θ o (Y) X = x] < uniformly in x X y assumption (A7), a similar proof as was given for Theorem 2 in Einmahl and Mason (2005) ensures that sup m θo (x) ṙθ o (x) f X (x) = O P(c n ). x X 0 Consider now the second term of (??). Since E[ ε(θ o ) X] = 0, where ε(θ o ) = d dθ (Λ θ(y) m θ (X)) θ=θo, we have [ ( )] X x ṙ θo (x) = h 1 E {ṁ θo (X)+ ε(θ o )}K 1 h [ ( )] X x = h 1 E ṁ θo (X)K 1 h = ṁ θo (x+hv)k 1 (v)f X (x+hv)dv, 17

18 from which it follows that ṙ θo (x) f X (x)ṁ θo (x) = [ṁ θo (x+hv) ṁ θo (x)]k 1 (v)f X (x+hv)dv. Hence, a Taylor expansion applied to ṁ θo ( ) yields sup ṙθo (x) f X (x)ṁ θo (x) = O(h q 1 ) = O(c n ), x X 0 since nh 2q1+1 (logh 1 ) 1 = O(1) y (A2). This proves that the second term of (??) is O(c n ), since it can e easily shown that f X (x) is ounded away from 0 and infinity, uniformly in x X 0, using (A)(ii). References [1] Ahmad, I. and Li, Q. (1997). Testing symmetry of an unknown density function y kernel method. Journal of Nonparametric Statistics, 7, [2] Akritas, M.G. and Van Keilegom, I. (2001). Non-parametric estimation of the residual distriution. Scandinavian Journal of Statistics, 28, [] Amemiya, T. (1985). Advanced Econometrics. Harvard University Press, Camridge. [4] Bickel, P.J. and Doksum, K. (1981). An analysis of transformations revisited. Journal of the American Statistical Association, 76, [5] Billingsley, P. (1968). Convergence of Proaility Measures. Wiley, New York. [6] Box, G.E.P. and Cox, D.R. (1964). An analysis of transformations. Journal of the Royal Statistical Society - Series B, 26, [7] Carroll, R.J. and Ruppert, D. (1988). Transformation and Weighting in Regression. Chapman and Hall, New York. [8] Chen, G., Lockhart, R.A. and Stephens, A. (2002). Box-Cox transformations in linear models: Large sample theory and tests of normality (with discussion). Canadian Journal of Statistics, 0, [9] Cheng, F. (2005). Asymptotic distriutions of error density and distriution function estimators in nonparametric regression. Journal of Statistical Planning and Inference, 128,

19 [10] Cheng, F. and Sun, S. (2008). A goodness-of-fit test of the errors in nonlinear autoregressive time series models. Statistics and Proaility Letters, 78, [11] Dette, H., Kusi-Appiah, S. and Neumeyer, N. (2002). Testing symmetry in nonparametric regression models. Journal of Nonparametric Statistics, 14, [12] Efromovich, S. (2005). Estimation of the density of the regression errors. Annals of Statistics,, [1] Einmahl, U. and Mason, D.M. (2005). Uniform in andwidth consistency of kernel-type function estimators. Annals of Statistics,, [14] Escanciano, J.C. and Jacho-Chavez, D. (2012). n-uniformly consistent density estimation in nonparametric regression. Journal of Econometrics, 167, [15] Fitzenerger, B., Wilke, R.A. and Zhang, X. (2010). Implementing Box-Cox quantile regression. Econometric Reviews, 29, [16] Horowitz, J.L. (1998). Semiparametric Methods in Economics. Springer-Verlag, New York. [17] Linton, O., Sperlich, S. and Van Keilegom, I. (2008). Estimation of a semiparametric transformation model. Annals of Statistics, 6, [18] Müller, U.U., Schick, A. and Wefelmeyer, W. (2004). Estimating linear functionals of the error distriution in nonparametric regression. Journal of Statistical Planning and Inference, 119, [19] Nadaraya, E. A. (1964). On estimating regression. Theory of Proaility and its Applications, 9, [20] Neumeyer, N. and Dette, H. (2007). Testing for symmetric error distriution in nonparametric regression models. Statistica Sinica, 17, [21] Neumeyer, N. and Van Keilegom, I. (2010). Estimating the error distriution in nonparametric multiple regression with applications to model testing. Journal of Multivariate Analysis, 101, [22] Pinsker, M.S. (1980). Optimal filtering of a square integrale signal in Gaussian white noise. Prolems of Information Transmission, 16, [2] Sakia, R.M. (1992). The Box-Cox transformation technique: a review. The Statistician, 41,

20 [24] Sam, R. (2011). Nonparametric estimation of the density of regression errors. Comptes Rendus de l Académie des Sciences-Paris, Série I 49, [25] Shin, Y. (2008). Semiparametric estimation of the Box-Cox transformation model. Econometrics Journal, 11, [26] Vanhems, A. and Van Keilegom, I. (2011). Semiparametric transformation model with endogeneity: a control function approach. Journal of Econometrics (under revision). [27] Watson, G.S. (1964). Smooth regression analysis. Sankhya - Series A, 26, [28] Zellner, A. and Revankar, N.S. (1969). Generalized production functions. Reviews of Economic Studies, 6, Postal addresses : Benjamin Colling Cédric Heuchenne Université catholique de Louvain HEC-Management School of the University of Liège, Institute of Statistics Statistique appliquée à la gestion et à l économie Voie du Roman Pays 20 Rue Louvrex 14, Bâtiment N1 148 Louvain-la-Neuve 4000 Liège Belgium Belgium Rawane Sam Ingrid Van Keilegom Centre de recherche du CHUQ/CHUL Université catholique de Louvain 2705, Boulevard Laurier Institute of Statistics G1V 4G2, QC Voie du Roman Pays 20 Quéec 148 Louvain-la-Neuve Canada Belgium 20

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2011/2 ESTIMATION

More information

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE Estimating the error distribution in nonparametric multiple regression with applications to model testing Natalie Neumeyer & Ingrid Van Keilegom Preprint No. 2008-01 July 2008 DEPARTMENT MATHEMATIK ARBEITSBEREICH

More information

Nonparametric Estimation of Smooth Conditional Distributions

Nonparametric Estimation of Smooth Conditional Distributions Nonparametric Estimation of Smooth Conditional Distriutions Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~hansen May 24 Astract This paper considers nonparametric estimation of smooth conditional

More information

Estimation of a semiparametric transformation model in the presence of endogeneity

Estimation of a semiparametric transformation model in the presence of endogeneity TSE 654 May 2016 Estimation of a semiparametric transformation model in the presence of endogeneity Anne Vanhems and Ingrid Van Keilegom Estimation of a semiparametric transformation model in the presence

More information

D I S C U S S I O N P A P E R

D I S C U S S I O N P A P E R I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S ( I S B A ) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2014/06 Adaptive

More information

Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement

Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement Open Journal of Statistics, 07, 7, 834-848 http://www.scirp.org/journal/ojs ISS Online: 6-798 ISS Print: 6-78X Estimating a Finite Population ean under Random on-response in Two Stage Cluster Sampling

More information

Estimation in semiparametric models with missing data

Estimation in semiparametric models with missing data Ann Inst Stat Math (2013) 65:785 805 DOI 10.1007/s10463-012-0393-6 Estimation in semiparametric models with missing data Song Xi Chen Ingrid Van Keilegom Received: 7 May 2012 / Revised: 22 October 2012

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S I S B A UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 20/4 SEMIPARAMETRIC

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

DISCUSSION PAPER 2016/43

DISCUSSION PAPER 2016/43 I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S ( I S B A ) DISCUSSION PAPER 2016/43 Bounds on Kendall s Tau for Zero-Inflated Continuous

More information

Generated Covariates in Nonparametric Estimation: A Short Review.

Generated Covariates in Nonparametric Estimation: A Short Review. Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated

More information

EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS

EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS APPLICATIONES MATHEMATICAE 9,3 (), pp. 85 95 Erhard Cramer (Oldenurg) Udo Kamps (Oldenurg) Tomasz Rychlik (Toruń) EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS Astract. We

More information

Bootstrap of residual processes in regression: to smooth or not to smooth?

Bootstrap of residual processes in regression: to smooth or not to smooth? Bootstrap of residual processes in regression: to smooth or not to smooth? arxiv:1712.02685v1 [math.st] 7 Dec 2017 Natalie Neumeyer Ingrid Van Keilegom December 8, 2017 Abstract In this paper we consider

More information

Generalized Seasonal Tapered Block Bootstrap

Generalized Seasonal Tapered Block Bootstrap Generalized Seasonal Tapered Block Bootstrap Anna E. Dudek 1, Efstathios Paparoditis 2, Dimitris N. Politis 3 Astract In this paper a new lock ootstrap method for periodic times series called Generalized

More information

I N S T I T U T D E S T A T I S T I Q U E

I N S T I T U T D E S T A T I S T I Q U E I N S T I T U T D E S T A T I S T I Q U E UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 0610 ESTIMATION OF A SEMIPARAMETRIC TRANSFORMATION MODEL O. LINTON, S. SPERLICH and I. VAN KEILEGOM

More information

HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO

HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO The Annals of Statistics 2006, Vol. 34, No. 3, 1436 1462 DOI: 10.1214/009053606000000281 Institute of Mathematical Statistics, 2006 HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO BY NICOLAI

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Additive Isotonic Regression

Additive Isotonic Regression Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive

More information

Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data

Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data Géraldine Laurent 1 and Cédric Heuchenne 2 1 QuantOM, HEC-Management School of

More information

arxiv: v1 [math.st] 4 Apr 2008

arxiv: v1 [math.st] 4 Apr 2008 The Annals of Statistics 2008, Vol. 36, No. 2, 686 718 DOI: 10.1214/009053607000000848 c Institute of Mathematical Statistics, 2008 arxiv:0804.0719v1 [math.st] 4 Apr 2008 ESTIMATION OF A SEMIPARAMETRIC

More information

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department

More information

Oberwolfach Preprints

Oberwolfach Preprints Oberwolfach Preprints OWP 202-07 LÁSZLÓ GYÖFI, HAO WALK Strongly Consistent Density Estimation of egression esidual Mathematisches Forschungsinstitut Oberwolfach ggmbh Oberwolfach Preprints (OWP) ISSN

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

FinQuiz Notes

FinQuiz Notes Reading 9 A time series is any series of data that varies over time e.g. the quarterly sales for a company during the past five years or daily returns of a security. When assumptions of the regression

More information

Transformation and Smoothing in Sample Survey Data

Transformation and Smoothing in Sample Survey Data Scandinavian Journal of Statistics, Vol. 37: 496 513, 2010 doi: 10.1111/j.1467-9469.2010.00691.x Published by Blackwell Publishing Ltd. Transformation and Smoothing in Sample Survey Data YANYUAN MA Department

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

M- and Z- theorems; GMM and Empirical Likelihood Wellner; 5/13/98, 1/26/07, 5/08/09, 6/14/2010

M- and Z- theorems; GMM and Empirical Likelihood Wellner; 5/13/98, 1/26/07, 5/08/09, 6/14/2010 M- and Z- theorems; GMM and Empirical Likelihood Wellner; 5/13/98, 1/26/07, 5/08/09, 6/14/2010 Z-theorems: Notation and Context Suppose that Θ R k, and that Ψ n : Θ R k, random maps Ψ : Θ R k, deterministic

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

Error distribution function for parametrically truncated and censored data

Error distribution function for parametrically truncated and censored data Error distribution function for parametrically truncated and censored data Géraldine LAURENT Jointly with Cédric HEUCHENNE QuantOM, HEC-ULg Management School - University of Liège Friday, 14 September

More information

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured

More information

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Nonparametric Regression. Changliang Zou

Nonparametric Regression. Changliang Zou Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,

More information

A Note on the Scale Efficiency Test of Simar and Wilson

A Note on the Scale Efficiency Test of Simar and Wilson International Journal of Business Social Science Vol. No. 4 [Special Issue December 0] Abstract A Note on the Scale Efficiency Test of Simar Wilson Hédi Essid Institut Supérieur de Gestion Université de

More information

TESTING FOR THE EQUALITY OF k REGRESSION CURVES

TESTING FOR THE EQUALITY OF k REGRESSION CURVES Statistica Sinica 17(2007, 1115-1137 TESTNG FOR THE EQUALTY OF k REGRESSON CURVES Juan Carlos Pardo-Fernández, ngrid Van Keilegom and Wenceslao González-Manteiga Universidade de Vigo, Université catholique

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Chapter 2 Inference on Mean Residual Life-Overview

Chapter 2 Inference on Mean Residual Life-Overview Chapter 2 Inference on Mean Residual Life-Overview Statistical inference based on the remaining lifetimes would be intuitively more appealing than the popular hazard function defined as the risk of immediate

More information

Statistical Properties of Numerical Derivatives

Statistical Properties of Numerical Derivatives Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective

More information

Testing near or at the Boundary of the Parameter Space (Job Market Paper)

Testing near or at the Boundary of the Parameter Space (Job Market Paper) Testing near or at the Boundary of the Parameter Space (Jo Market Paper) Philipp Ketz Brown University Novemer 7, 24 Statistical inference aout a scalar parameter is often performed using the two-sided

More information

On the Robust Modal Local Polynomial Regression

On the Robust Modal Local Polynomial Regression International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

NONPARAMETRIC ENDOGENOUS POST-STRATIFICATION ESTIMATION

NONPARAMETRIC ENDOGENOUS POST-STRATIFICATION ESTIMATION Statistica Sinica 2011): Preprint 1 NONPARAMETRIC ENDOGENOUS POST-STRATIFICATION ESTIMATION Mark Dahlke 1, F. Jay Breidt 1, Jean D. Opsomer 1 and Ingrid Van Keilegom 2 1 Colorado State University and 2

More information

A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors

A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors By Jiti Gao 1 and Maxwell King The University of Western Australia and Monash University Abstract: This paper considers a

More information

Estimation of the Bivariate and Marginal Distributions with Censored Data

Estimation of the Bivariate and Marginal Distributions with Censored Data Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

A converse Gaussian Poincare-type inequality for convex functions

A converse Gaussian Poincare-type inequality for convex functions Statistics & Proaility Letters 44 999 28 290 www.elsevier.nl/locate/stapro A converse Gaussian Poincare-type inequality for convex functions S.G. Bokov a;, C. Houdre ; ;2 a Department of Mathematics, Syktyvkar

More information

Graduate Econometrics I: Maximum Likelihood I

Graduate Econometrics I: Maximum Likelihood I Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

FIDUCIAL INFERENCE: AN APPROACH BASED ON BOOTSTRAP TECHNIQUES

FIDUCIAL INFERENCE: AN APPROACH BASED ON BOOTSTRAP TECHNIQUES U.P.B. Sci. Bull., Series A, Vol. 69, No. 1, 2007 ISSN 1223-7027 FIDUCIAL INFERENCE: AN APPROACH BASED ON BOOTSTRAP TECHNIQUES H.-D. HEIE 1, C-tin TÂRCOLEA 2, Adina I. TARCOLEA 3, M. DEMETRESCU 4 În prima

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Motivational Example

Motivational Example Motivational Example Data: Observational longitudinal study of obesity from birth to adulthood. Overall Goal: Build age-, gender-, height-specific growth charts (under 3 year) to diagnose growth abnomalities.

More information

Smoothing the Nelson-Aalen Estimtor Biostat 277 presentation Chi-hong Tseng

Smoothing the Nelson-Aalen Estimtor Biostat 277 presentation Chi-hong Tseng Smoothing the Nelson-Aalen Estimtor Biostat 277 presentation Chi-hong seng Reference: 1. Andersen, Borgan, Gill, and Keiding (1993). Statistical Model Based on Counting Processes, Springer-Verlag, p.229-255

More information

Spiking problem in monotone regression : penalized residual sum of squares

Spiking problem in monotone regression : penalized residual sum of squares Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad

More information

A note on L convergence of Neumann series approximation in missing data problems

A note on L convergence of Neumann series approximation in missing data problems A note on L convergence of Neumann series approximation in missing data problems Hua Yun Chen Division of Epidemiology & Biostatistics School of Public Health University of Illinois at Chicago 1603 West

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

Statistica Sinica Preprint No: SS

Statistica Sinica Preprint No: SS Statistica Sinica Preprint No: SS-017-0013 Title A Bootstrap Method for Constructing Pointwise and Uniform Confidence Bands for Conditional Quantile Functions Manuscript ID SS-017-0013 URL http://wwwstatsinicaedutw/statistica/

More information

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)

More information

LECTURE 18: NONLINEAR MODELS

LECTURE 18: NONLINEAR MODELS LECTURE 18: NONLINEAR MODELS The basic point is that smooth nonlinear models look like linear models locally. Models linear in parameters are no problem even if they are nonlinear in variables. For example:

More information

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case

Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Identification and Estimation Using Heteroscedasticity Without Instruments: The Binary Endogenous Regressor Case Arthur Lewbel Boston College Original December 2016, revised July 2017 Abstract Lewbel (2012)

More information

Depth versus Breadth in Convolutional Polar Codes

Depth versus Breadth in Convolutional Polar Codes Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK

More information

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University

More information

Closest Moment Estimation under General Conditions

Closest Moment Estimation under General Conditions Closest Moment Estimation under General Conditions Chirok Han and Robert de Jong January 28, 2002 Abstract This paper considers Closest Moment (CM) estimation with a general distance function, and avoids

More information

A732: Exercise #7 Maximum Likelihood

A732: Exercise #7 Maximum Likelihood A732: Exercise #7 Maximum Likelihood Due: 29 Novemer 2007 Analytic computation of some one-dimensional maximum likelihood estimators (a) Including the normalization, the exponential distriution function

More information

Semiparametric modeling and estimation of the dispersion function in regression

Semiparametric modeling and estimation of the dispersion function in regression Semiparametric modeling and estimation of the dispersion function in regression Ingrid Van Keilegom Lan Wang September 4, 2008 Abstract Modeling heteroscedasticity in semiparametric regression can improve

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Robust Backtesting Tests for Value-at-Risk Models

Robust Backtesting Tests for Value-at-Risk Models Robust Backtesting Tests for Value-at-Risk Models Jose Olmo City University London (joint work with Juan Carlos Escanciano, Indiana University) Far East and South Asia Meeting of the Econometric Society

More information

Simple Examples. Let s look at a few simple examples of OI analysis.

Simple Examples. Let s look at a few simple examples of OI analysis. Simple Examples Let s look at a few simple examples of OI analysis. Example 1: Consider a scalar prolem. We have one oservation y which is located at the analysis point. We also have a ackground estimate

More information

Gaussian kernel GARCH models

Gaussian kernel GARCH models Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often

More information

UNIVERSITY OF CALIFORNIA Spring Economics 241A Econometrics

UNIVERSITY OF CALIFORNIA Spring Economics 241A Econometrics DEPARTMENT OF ECONOMICS R. Smith, J. Powell UNIVERSITY OF CALIFORNIA Spring 2006 Economics 241A Econometrics This course will cover nonlinear statistical models for the analysis of cross-sectional and

More information

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Bayesian estimation of bandwidths for a nonparametric regression model with a flexible error density

Bayesian estimation of bandwidths for a nonparametric regression model with a flexible error density ISSN 1440-771X Australia Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Bayesian estimation of bandwidths for a nonparametric regression model

More information

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION Jiangping Wang and Dapeng Wu Department of Electrical and Computer Engineering University of Florida, Gainesville, FL 3611 Correspondence author: Prof.

More information

Bayesian inference with reliability methods without knowing the maximum of the likelihood function

Bayesian inference with reliability methods without knowing the maximum of the likelihood function Bayesian inference with reliaility methods without knowing the maximum of the likelihood function Wolfgang Betz a,, James L. Beck, Iason Papaioannou a, Daniel Strau a a Engineering Risk Analysis Group,

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA Kasun Rathnayake ; A/Prof Jun Ma Department of Statistics Faculty of Science and Engineering Macquarie University

More information

Nonparametric Identification and Estimation of a Transformation Model

Nonparametric Identification and Estimation of a Transformation Model Nonparametric and of a Transformation Model Hidehiko Ichimura and Sokbae Lee University of Tokyo and Seoul National University 15 February, 2012 Outline 1. The Model and Motivation 2. 3. Consistency 4.

More information

arxiv: v1 [cs.gt] 4 May 2015

arxiv: v1 [cs.gt] 4 May 2015 Econometrics for Learning Agents DENIS NEKIPELOV, University of Virginia, denis@virginia.edu VASILIS SYRGKANIS, Microsoft Research, vasy@microsoft.com EVA TARDOS, Cornell University, eva.tardos@cornell.edu

More information

Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC)

Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Authors: Alexei C. Ionan, Mei-Yin C. Polley, Lisa M. McShane, Kevin K. Doin Section Page

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

On Universality of Blow-up Profile for L 2 critical nonlinear Schrödinger Equation

On Universality of Blow-up Profile for L 2 critical nonlinear Schrödinger Equation On Universality of Blow-up Profile for L critical nonlinear Schrödinger Equation Frank Merle,, Pierre Raphael Université de Cergy Pontoise Institut Universitaire de France Astract We consider finite time

More information

Merging and splitting endowments in object assignment problems

Merging and splitting endowments in object assignment problems Merging and splitting endowments in oject assignment prolems Nanyang Bu, Siwei Chen, and William Thomson April 26, 2012 1 Introduction We consider a group of agents, each endowed with a set of indivisile

More information

Contributions to extreme-value analysis

Contributions to extreme-value analysis Contributions to extreme-value analysis Stéphane Girard INRIA Rhône-Alpes & LJK (team MISTIS). 655, avenue de l Europe, Montbonnot. 38334 Saint-Ismier Cedex, France Stephane.Girard@inria.fr February 6,

More information

Issues on quantile autoregression

Issues on quantile autoregression Issues on quantile autoregression Jianqing Fan and Yingying Fan We congratulate Koenker and Xiao on their interesting and important contribution to the quantile autoregression (QAR). The paper provides

More information

Critical value of the total debt in view of the debts. durations

Critical value of the total debt in view of the debts. durations Critical value of the total det in view of the dets durations I.A. Molotov, N.A. Ryaova N.V.Pushov Institute of Terrestrial Magnetism, the Ionosphere and Radio Wave Propagation, Russian Academy of Sciences,

More information

Spectrum Opportunity Detection with Weak and Correlated Signals

Spectrum Opportunity Detection with Weak and Correlated Signals Specum Opportunity Detection with Weak and Correlated Signals Yao Xie Department of Elecical and Computer Engineering Duke University orth Carolina 775 Email: yaoxie@dukeedu David Siegmund Department of

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

A Local Generalized Method of Moments Estimator

A Local Generalized Method of Moments Estimator A Local Generalized Method of Moments Estimator Arthur Lewbel Boston College June 2006 Abstract A local Generalized Method of Moments Estimator is proposed for nonparametrically estimating unknown functions

More information