Nonparametric Modal Regression

Size: px
Start display at page:

Download "Nonparametric Modal Regression"

Transcription

1 Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric modal regression is distinguished from the conventional nonparametric regression in that it measures the center using the most likely conditional values rather than the conditional average. We propose to use local polynomial regression to estimate the nonparametric modal regression. The asymptotic properties of the resulting estimator are investigated. We further develop a modal expectationmaximization (EM) algorithm for the proposed model. The simulation result confirms the theoretic findings. The proposed methodology is further illustrated via the analysis of a real data example. Some key words:: Linear regression; M-estimator; Modal regression; Mode; Robust regression;. Introduction Given a random sample (x i, y i ), i =,..., n}, where x i is a p-dimension column vector, suppose f(y x) is the conditional density function of Y given x i. For the conventional nonparametric regression models, the mean of f(y x) is usually used to investigate the relationship between Y and X. When the distribution is highly skewed, it is well known that the mode provides a more meaningful location estimator than the mean. Several authors have made efforts to identify the modes of population distributions. See, for example, Silverman (98); Muller and Sawitzki (99); Scott (992); Friedman and Fisher (999); Chaudhuri and Marron (999); Fisher and

2 Marron (200); Davies and Kovac (2004); Hall, Minnotte, and Zhang (2004); Ray and Lindsay (2005); Yao and Lindsay (2009). In this article, we propose a nonparametric modal regression model that aims to estimate the mode of f(y x) for any given x. Modal regression measures the center using the most likely conditional values rather than the conditional average used by the traditional regression methods. We propose to use local polynomial regression to estimate the nonparametric modal regression. Sampling properties of the proposed estimation procedure are systematically studied. We further develop a modal expectation-maximization (EM) algorithm for the proposed model. The simulation result confirms the theoretic findings. The proposed methodology is further illustrated via the analysis of a real data example. We further develop a modal EM algorithm for the modal linear regression. The extensive Monte Carlo simulation study is conducted to assess the finite sample performance of the proposed procedure and to compare our MLR estimator with the LSE for the symmetric error density. The rest of this article is organized as follows. In Section 2, we introduce our new nonparametric modal regression model and the estimation procedure based on local polynomial regression. The asymptotic properties of the resulting estimator are also provided. In Section 3, A Monte Carlo simulation study is conducted, and a real data application is used to illustrate the proposed methodology. Some discussions are given in Section 4. Technical conditions and proofs are given in the Appendix. 2. Nonparametric Modal regression In this section, we will introduce the nonparametric modal regression model and propose to use local polynomial regression to estimate the nonparametric modal regression. An EM type algorithm is also proposed to estimate the unknown modal parameters. In addition, we will also study the asymptotic properties of the proposed 2

3 estimator. 2.. Introduction to the models and estimation procedure Suppose that (x, y ),..., (x n, y n ) are an independent and identically distributed random sample from f(x, y). The modal regression is defined as m(x) = arg max f(y x), (2.) y where m( ) is an unknown nonparametric smoothing function to be estimated. Let ϵ = y m(x). Denote by g(t x) the conditional density of ϵ given X = x. Based on the model assumption (2.), one can know that g(t x) is maximized at 0 for any x. If g(t x) is symmetric about 0, the m(x) will be the same as the conventional regression function E(Y X = x). In this article, we propose an estimation procedure for the nonparametric modal regression m(x). Since f(y x) = f(x, y)/f(x), finding the mode of f(y x) is equivalent to finding the mode of f(x, y) with x fixed. Suppose f(x, y) is estimated by the kernel density estimator, i.e. ˆf(x, y) = n K h (x i x)ϕ h2 (y i y), where K h (x) = h K(x/h) and ϕ h (t) = h ϕ(t/h) are the symmetric kernel functions and (h, h 2 ) are the bandwidths. A natural estimation procedure is to estimate m(x 0 ) by ˆm(x 0 ) = arg max y n K h (x i x 0 )ϕ h2 (y i y). (2.2) Note that the modal regression (2.2) only uses one term (the intercept) for the conditional mode, like the local constant estimator (Nadaraya, 964; Watson, 964). It is known that the local linear estimator is superior to the local constant one (Fan 3

4 and Gijbels, 996). So we may want to extend the idea of local constant modal regression (2.2) to the local linear case, or more generally to the local polynomial case. For x in a neighborhood of x 0, we approximate m(x) p j=0 m (j) (x 0 ) (x x 0 ) j j! p β j (x x 0 ) j, j=0 where β j = m (j) (x 0 )/j!. Our local polynomial modal regression (LPMR) estimation procedure is to maximize over θ = (β 0,..., β p ) l(θ) n ) p K h (x i x 0 )ϕ h2 (y i β j (x i x 0 ) j. (2.3) j=0 For ease of computation, we use the standard normal density for ϕ(t) throughout this proposal. See (2.4) below. Denote the maximizer of l(θ) to be ˆθ = ( ˆβ 0,, ˆβ p ). Then the estimator of the v-th derivative of m(x), m (v) (x), will be ˆm (v) (x 0 ) = v! ˆβ v, for v = 0,, p. Specifically, when p = and v = 0, we refer to this method as local linear modal regression (LLMR) Modal expectation-maximization algorithm Note that (2.3) does not have an explicit solution. In this section, we propose to extend the modal expectation-maximization (MEM) algorithm, proposed by Li et al. (2007), to maximize (2.3). Similar to an EM algorithm, the MEM algorithm consists of two steps: E-step and M-step. Let θ (0) be the initial value. Starting with k = 0: E-Step: Update π(j θ (k) ) π(j θ (k) ) = K h (x j x 0 )ϕ h2 y j } p l=0 β(k) l (x j x 0 ) l [K h (x i x 0 )ϕ h2 y i }], j =,..., n. p l=0 β(k) l (x i x 0 ) l 4

5 M-Step: Update θ (k+) θ (k+) = arg max θ [π(j θ (k) ) log ϕ h2 y j j= }] p β l (x j x 0 ) l = (X T W k X) X T W k Y (2.4) l=0 since ϕ( ) is the density function of a standard normal distribution. Here X = (x,..., x n) T with x i =, x i x 0,, (x i x 0 ) p } T, W k is an n n diagonal matrix with diagonal elements π(j θ (k) )s, and Y = (y,..., y n ) T. The MEM algorithm is to iterate the E-step and the M-step until the algorithm converges. The ascending property of the proposed MEM algorithm can be established along the lines of Li et al. (2007). Note that the converged value may depend on the starting-point as the usual EM algorithm, and there is no guarantee that the algorithm will converge to the global optimal solution. Therefore, it is prudent to run the algorithm from several starting-points and choose the best local optima found. The closed form solution for θ (k+) is one of the benefits of using normal density function ϕ h ( ) in (2.3). From the MEM algorithm, it can be seen that the major difference between the conventional local polynomial regression (LPR) and LPMR lies in the E-step. The contribution of observation (x i, y i ) to the LPR depends on the weight K h (x i x 0 ), which in turn depends on how close x i is to x 0 only. On the other hand, the weight in the LPMR depends on both how close x i is to x 0 and how close y i is to the regression curve. This weight scheme allows the LPMR to downweight the observations further away from the regression curve to achieve adaptive robustness Theoretical properties We first establish the convergent rate of the LPMR estimator in the following theorem, whose proof can be found in the Appendix. 5

6 Theorem 2.. Under the regularity conditions (A) (A5) in the Appendix, there exists a consistent local maximizer ˆθ of (2.3) such that h v ˆmv (x 0 ) m (v) (x 0 ) } = Op (nh h 3 2) /2 + h p+ + h 2 2}, v = 0,,..., p, where ˆm v (x 0 ) = v! ˆβ v is the estimate of m (v) (x 0 ) and m (v) (x 0 ) is the v th derivative of m(x) at x 0. The proof of Theorem 2. is given in the Appendix. To derive the asymptotic bias and variance of the LPMR estimator, we need the following notation. The moments of K and K 2 are denoted respectively by µ j = t j K(t)dt and ν j = t j K 2 (t)dt. Let S, S, and S be (p + ) (p + ) matrix with (j, l)-element µ j+l 2, µ j+l, and ν j+l 2, respectively, and c p, c p, and c p be p vector with j-th element µ p+j, µ p+j+, and µ j, respectively. Furthermore, let e v+ = (0,..., 0,, 0,..., 0) T, a p vector with on the (v + ) th position. Theorem 2.2. Under the regularity conditions (A) (A5) in the Appendix, the asymptotic variance of ˆm v (x 0 ), given in Theorem 2., is given by Var ˆm v (x 0 )} = e T v+s S S v! 2 g(0 x 0 ) ν e v+ f(x 0 )g (0 x 0 ) 2 nh +2v + o h 3 p ()}. (2.5) 2 where ν = t 2 ϕ 2 (t)dt. The asymptotic bias of ˆm v (x 0 ), denoted by b v (x 0 ), for p v odd is given by b v (x 0 ) = e T v+s h p+ v v! (p + )! m(p+) (x 0 )c p g (0 x 0 )v!h 2 2 2g (0 x 0 )h v 6 c p } + o()} (2.6)

7 Furthermore, the asymptotic bias for p v even is [ b v (x 0 ) =e T v+s h p+2 v } v! c p m (p+2) (x 0 ) + (p + 2)m (p+) (x 0 ) Γ (x 0 ) (p + 2)! Γ(x 0 ) ] g (0 x 0 )v!h 2 2 c 2g (0 x 0 )h v p + o()}, (2.7) provided that m (p+2) ( ) are continuous in a neighborhood of x 0 and nh 3 h 5 2. The proof of Theorem 2.2 is given in the Appendix. Similar to the LPR, the second term in (2.7) often creates extra bias. Thus, it is preferable to use odd values of p v in practice. Therefore, it is consistent with the selection order of p for the LPR (?). Theorem 2.3. Under the regularity conditions (A) (A5) in the Appendix, the estimate ˆm v (x 0 ), given in Theorem 2., has the following asymptotic distribution ˆm v (x 0 ) m (v) (x 0 ) b v (x 0 ) Var ˆmv (x 0 )} D N(0, ). The proof of Theorem 2.3 is given in the Appendix. The optimal bandwidth h = O p ( 2 7p + 9 ), h 2 = O p ( p + 7p + 9 ). Specially, when p = and v = 0, the asymptotic variance of ˆm(x 0 ) is Var ˆm(x 0 )} g(0 x 0 ) νν 0 nh h 3 2g (0 x 0 ) 2 f(x 0 ) + o p()} (2.8) and the asymptotic bias is b(x 0 ) 2 m (x 0 )µ 2 h 2 g (0 x 0 )h 2 2 2g (0 x 0 ) (2.9) 7

8 The optimal bandwidth h = O p ( /8), h 2 = O p ( /8) Bandwidth selection 3. Simulation Study and Application In this section, we will conduct Monte Carlo simulation to assess the performance of our proposed LPMR. We also demonstrate the LPMR using one real data application. 3.. Simulation study 3.2. An application Appendix APPENDIX: PROOFS The following technical conditions are imposed in this section. They are not the weakest possible conditions, but they are imposed to facilitate the proofs. Technical Conditions: (A) The m(x) has continuous (p + ) th derivative at the point x 0. (A2) The f(x) is bounded and has continuous first derivative at the point x 0 and f(x 0 ) > 0. (A3) g (0 x) = 0, g (0 x) < 0, g (v) (t x) is bounded in a neighbor of x 0 and has continuous first derivative at the point x 0 as a function of x, for v = 0,..., 4. 8

9 (A4) K( ) is a symmetric (about 0) probability density with compact support [, ]. (A5) The bandwidths h and h 2 tend to 0 such that nh h 5 2 and h p+ /h 2 0. Denote X i =, (X i x 0 )/h,..., (X i x 0 ) p /h p } T, H = diag, h,..., h p }, θ = (β 0, β,..., β p ) T, θ = Hθ, R(X i ) = m(x i ) p j=0 β j(x i x 0 ) j, and K i = K h (x i x 0 ), where β j = m (j) (x 0 )/j!, j = 0,,..., p. Proof of Theorem 2.. Denote α n = (nh h 3 2) /2 + h p+ + h 2 2. It is sufficient to show that for any given η > 0, there exists a large constant c such that P sup l(θ + α n µ) < l(θ )} η, µ =c (A.) where l(θ) is defined in (2.3). By using Taylor expansion, it follows that l(θ + α n µ) l(θ ) = n = n K i ϕh2 (ϵ i + R(X i ) α n µ T Xi ) ϕ h2 (ϵ i + R(X i )) } K i ϕ h 2 (ϵ i + R(X i ))α n µ T X i + 2 ϕ h 2 (ϵ i + R(X i ))α 2 n(µ T X i ) 2 } 6 ϕ h 2 (z i )αn(µ 3 T Xi ) 3 ) I + I 2 + I 3, (A.2) where z i is between ϵ i + R(X i ) and ϵ i + R(X i ) + α n µ T X i. Note that ϕ h 2 (t) = t ( ) t ϕ, ϕ h 3 h 2 h 2 (t) = ( ) ( ) t 2 t ϕ, and ϕ 2 h 3 2 h 2 h 2 h 2 (t) = ( ) } 3 ( ) 3t t t ϕ. 2 h 4 2 h 2 h 2 h 2 9

10 If a n (x) = o p (h 2 ), and g (v) (t x) is bounded in a neighbor of x 0, we have Eϕ h 2 (ϵ + a n (x)) X = x} = h 2 tϕ(t)g(th + a n (x) x)dt = h 2 tϕ(t)g(th 2 x) + g (th 2 x)a n (x) + g (z x)a 2 n(x)}dt } g (0 x) = h g (0 x)a n (x) + o p ()}. 2 If h p+ /h 2 0, by directly calculating the mean and variance, we obtain E(I ) = α n E [ ] K i Eϕ h 2 (ϵ i + R(X i )) X i }µ T Xi ( + o()) g = α n µ T (0 x 0 ) 2 = O α n c ( )} h h p+ ; f(x 0 ) c p h 2 2 g (0 x 0 )c p f(x 0 ) m(p+) (x 0 ) (p + )! Var(I ) = n αnvar[k 2 i ϕ h 2 ϵ i + R(X i )}(µ T Xi )] = n αnµ } 2 T g(0 x 0 )f(x 0 )ν 0 S h 3 2 h µ } h p+ = O(α 2 n(nh h 3 2) c 2 ), (A.3) where c p = (µ 0, µ,..., µ p ) T. Hence I = O α n c ( )} h h p+ + αn co p ((nh h 3 2 ) /2 ) = O p (cαn). 2 If nh 5 2h, similar to (A.3), we can prove I 2 = n I 3 = n } 2 K iϕ h 2 (ϵ i + R(X i ))αnµ 2 T Xi Xi T µ = αng 2 (0 x 0 )f(x 0 )µ T Sµ( + o p ()), } 6 K iϕ h 2 (z i )αn(µ 3 T Xi ) 3 = o p (αn). 2 (A.4) 0

11 Noticing that S is a positive matrix, µ = c, and g (0 x 0 ) < 0, we can choose c large enough such that I 2 dominates both I and I 3 with probability at least η. Thus (A.) holds. Therefore, with probability approaching (wpa), there exists a local maximizer ˆθ such that ˆθ θ α n c. Based on the definition of θ, we can get, wpa, h v ˆmv (x 0 ) m (v) (x 0 ) } = Op (nh h 3 2) /2 + h p+ + h 2 2}. Define W n = Xi K i ϕ h 2 (ϵ i ). We have the following asymptotic representation. (A.5) Lemma A.. Under conditions (A) (A6), it follows that ˆθ θ = h p+ m (p+) (x 0 ) S c p ( + o p ()) + (p + )! S W n nf (x 0, h 2 )f(x 0 ) ( + o p()). (A.6) Proof. Let p ˆγ i = R(X i ) ( ˆβ j β j )(X i x 0 ) j = R(X i ) (ˆθ θ ) T X. j=0 Then p Y i ˆβ j (X i x 0 ) j = ϵ i + ˆγ i. j=0 The solution ˆθ satisfies the equation Xi K i ϕ h 2 (ϵ i + ˆγ i ) = X i K i ϕ h 2 (ϵ i ) + ϕ h 2 (ϵ i )ˆγ i + 2 ϕ h 2 (ϵ )ˆγ 2 } = 0, (A.7) where ϵ is between ϵ i and ϵ i + ˆγ i. Note that the second term on the left hand side of

12 (A.7) is K i ϕ h 2 (ϵ i )R(X i )Xi K i ϕ h 2 (ϵ i )Xi Xi (ˆθ θ ) J + J 2. (A.8) From the proof of (A.3), we have J = nh p+ g m (p+) (x 0 ) (0 x 0 )f(x 0 )c p + o p (nh p+ ), (p + )! and J 2 = ng (0 x 0 )f(x 0 )S( + o p ())(ˆθ θ ). From the Theorem 2., we know ˆθ θ = O p h p+ + h (nh h 3 2) /2 }, hence sup ˆγ i sup R(X i ) + (ˆθ θ ) T X i: X i x 0 /h i: X i x 0 /h = O p (h p+ + ˆθ θ ) = O p ( ˆθ θ ). (A.9) (m (p+) (x) is bounded) Also, we have E ( ) ( ) j x K i (X i x 0 ) j /h} j x0 x x0 = K f(x)dx h h h = µ j f(x 0 ) + o(). (A.0) Based on (A.4), (A.9), and ˆθ θ = O p (α n ) } } K iˆγ i 2 ϕ h n 2 (ϵ )(X i x 0 ) j /h j = K i α n nϕ 2 h 2 (ϵ )(X i x 0 ) j /h j = o p ( α n ) 2

13 Hence for the third term on the left-hand side of (A.7), K iˆγ 2 i X i ϕ h 2 (ϵ ) = o p (nα n ) = o p (J 2 ). Then, it follows from (A.5) and (A.7) that ˆθ θ = ng (0 x 0 )f(x 0 )S} nh p+ g m (p+) (x 0 ) (0 x 0 )f(x 0 )c p ( + o p ()) (p + )! + ng (0 x 0 )f(x 0 )S} W n ( + o p ()), =h p+ m (p+) (x 0 ) S c p ( + o p ()) + (p + )! S W n ng (0 x 0 )f(x 0 ) ( + o p()). (A.) Proof of Theorem 2.2. Based on (A.5) and the condition (A6), we can easily get ( ) E n W n = E Xi K i ϕ h 2 (ϵ i ) } = g (0 x 0 ) f(x 0 )c 2 ph o p (h 2 2). Similar to the proof of A.3, we have E K 2 i ϕ h 2 (ϵ i ) 2 (X i x 0 ) j /h j } = (h h 3 2) ν j νg(0 x 0 )f(x 0 ) + o()}. where ν = ϕ 2 (t)t 2 dt. So Cov(W n /n) = (nh h 3 2) νg(0 x 0 )f(x 0 )S ( + o()). (A.2) Based on the result (A.6), the asymptotic bias b v (x 0 ) and variance of ˆm v (x 0 ) are naturally given by b v (x 0 ) = e T v+s h p+ v v! (p + )! m(p+) (x 0 )c p g (0 x 0 )v!h 2 2 2g (0 x 0 )h v c p } ( + o()) 3

14 and Var ˆm v (x 0 )} = v! 2 g(0 x 0 ) ν nh 3 2h +2v f(x 0 )g (0 x 0 ) 2 et v+s S S e v+ ( + o()). Noting that µ j = 0 for odd j, by some simple calculation, we can know the (v + ) th element of S c p is zero for p v even. So we need higher order expansion of asymptotic bias for p v even. Following the similar arguments of Theorem 2., if nh 3 h 5 2 (make the root of variance order less than bias order), we can prove [ J = nh p+ m (p+) (x 0 ) Γ(x 0 )c p (p + )! +h c p Γ (x 0 ) m(p+) (x 0 ) (p + )! J 2 = n Γ(x 0 )S + h SΓ (x 0 ) + Γ(x 0 ) m(p+2) (x 0 ) (p + 2)! } + o p ()}(ˆθ θ ), }] + o p ()}, where J and J 2 is defined in (A.8) and Γ(x) = g (0 x)f(x). Then, it follows from (A.7) that } [ ˆθ θ = h p+ Γ(x 0 )S + h SΓ m (p+) (x 0 ) (x 0 ) Γ(x 0 )c p (p + )! +h c p Γ (x 0 ) m(p+) (x 0 ) (p + )! + Γ(x 0 ) m(p+2) (x 0 ) (p + 2)! S W n + ng (0 x 0 )f(x 0 ) ( + o p()) m = h p+ (p+) (x 0 ) S c p + h b (x 0 ) (p + )! } }] + o p ()} ( + o p ()) + S W n nγ(x 0 ) ( + o p()), 4

15 where b (x 0 ) =Γ (x 0 )S c } p Γ (x 0 ) m(p+) (x 0 ) + Γ(x 0 ) m(p+2) (x 0 ) (p + )! (p + 2)! Γ (x 0 )Γ (x 0 ) m(p+) (x 0 ) S SS c p. (p + )! For p v even, since the (v + ) th element of S c p and S SS c p is zero, the asymptotic bias b v (x 0 ) of ˆm v (x 0 ) are naturally given by [ b v (x 0 ) =e T v+s h p+2 v v! c p (p + 2)! ] g (0 x 0 )v!h 2 2 c 2g (0 x 0 )h v p m (p+2) (x 0 ) + (p + 2)m (p+) (x 0 ) Γ (x 0 ) Γ(x 0 ) + o()}. } Proof of Theorem 2.3. It is sufficient to show that W n h h 3 2/nW n D N(0, D) (A.3) where D = νg(0 x 0 )f(x 0 )S, because using Slutsky s theorem, it follows from (A.6), (A.3), and Theorem 2.2 that ˆm v (x 0 ) m (v) (x 0 ) b v (x 0 ) Var ˆmv (x 0 )} D N(0, ). Next we show (A.3). For any unit vector d R p+, we prove d T Cov(W n)d} 2 d T W n d T E(W n)} D N(0, ). 5

16 Let ξ i = h h 3 2/nK i ϕ h 2 (ϵ i )d T X i. Then d T W n = n ξ i. We check the Lyapunov s condition. Based on (A.2), we can get Cov(W n) = νg(0 x 0 )f(x 0 )S ( + o()) and Var(d T W nd) = d T Cov(W n)d = νg(0 x 0 )f(x 0 )d T S d( + o()). So we only need to prove ne ξ 3 0. Noticing that (d X i ) 2 d 2 X i 2, ϕ ( ) is bounded, and K( ) has compact support, ne ξ 3 O(nn 3/2 h 3/2 h 9/2 2 ) = O(n /2 h 3/2 h 9/2 2 ) p ( ) 3j E K3 ϕ h 2 (ϵ ) 3 X x 0 j=0 p j=0 h O(h 2 h 5 2 ) = O(nh h 2 ) /2 } 0 So the asymptotic normality for W n holds with covariance matrix νg(0 x 0 )f(x 0 )S. 6

17 References Chatterjee, S. and Price, B. (977). Regression Analysis by Example. John Wiley and Sons, New York. Chaudhuri, P. and Marron, J. S. (999). Sizer for Exploration of Structures in Curves. Journal of the American Statistical Association, 94, Davies, P. L. and Kovac, A. (2004). Densities, Spectral Densities and Modality. Annals of Statistics, 32, Donoho, D.L. (982). Breakdown Properties of Multivariate Location Estimators. Ph.D. Qualifying paper, Department of Statistics, Harvard University. Donoho, D.L. and Huber, P.J. (983). The Notion of Breakdown Point. In: Doksum, K., Hodges, J.L. (Editors), Festchrift in Honor of Erich Lehman. Wadsworth, Belmont, CA. Fan, J. and Gijbels, I. (996). Local Polynomial Modelling and Its Applications. Chapman and Hall, London. Fan, J. and Jiang, J. (2000). Variable Bandwidth and One-Step Local M-Estimator. Science in China, 43, Fisher, N. I. and Marron, J. S. (200). Mode Testing via The Excess Mass Estimate. Biometrika, 88, Friedman J. H. and Fisher, N. I. (999). Bump Hunting in High-Dimensional Data. Statistics and Computing, 9, Hall, P., Minnotte, M. C., and Zhang, C. (2004). Bump Hunting with Non-Gaussian Kernels. Annals of Statistics, 32,

18 Hample, F. R. (97). A General Quantitative Definition of Robustness. Annals of Mathematiccal Statistics, 42, Hample, F. R. (974). The Influence Curve and Its Role in Robust Estimation. Journal of the American Statistical Association, 69, Hample, F. R. (986). Robust Statistics. Wiley, New York. Huber, P. J. (98). Robust Statistics. Wiley, New York. Li, J., Ray, S., and Lindsay, B. G. (2007). A Nonparametric Statistical Approach to Clustering via Mode Identification. Journal of Machine Learning Research, 8(8), Muller, D. W. and Sawitzki, G. (99). Excess Mass Estimates and Tests for Multimodality. Journal of the American Statistical Association, 86, Nadaraya, E. A. (964). On estimating regression. Theory of Probability Applied, 0, Parzen, E. (962). On estimation of a probability density function and mode. Annals of Mathematical Statistics, 33, Ray, S. and Lindsay, B. G. (2005). The Topography of Multivariate Normal Mixtures. Annals of Statistics, 33, Silverman, B. W. (98). Using Kernel Density Estimates to Investigate Multimodality. Journal of Royal Statistical Society, B43, Silverman, B. W. (986). Density Estimation for Statistics and Data Analysis. Chapman and Hall: London. 8

19 Scott, D. W. (992). Multivariate Density Estimation: Theory, Practice and Visualization. New York: Wiley. Watson, G. S. (964). Smooth regression analysis. Sankhya, Ser. A, 26, Yao, W. and Lindsay, B. G. (2009). Bayesian Mixture Labelling by Highest Posterior Density. Journal of American Statistical Association, 04,

Local Modal Regression

Local Modal Regression Local Modal Regression Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Bruce G. Lindsay and Runze Li Department of Statistics, The Pennsylvania

More information

Nonparametric and Varying Coefficient Modal. Regression

Nonparametric and Varying Coefficient Modal. Regression Nonparametric and Varying Coefficient Modal arxiv:1602.06609v1 [stat.me] 22 Feb 2016 Regression Weixin Yao and Sijia Xiang Summary In this article, we propose a new nonparametric data analysis tool, which

More information

A New Regression Model: Modal Linear Regression

A New Regression Model: Modal Linear Regression A New Regression Model: Modal Linear Regression Weixin Yao and Longhai Li Abstract The population mode has been considered as an important feature of data and it is traditionally estimated based on some

More information

On the Robust Modal Local Polynomial Regression

On the Robust Modal Local Polynomial Regression International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

Function of Longitudinal Data

Function of Longitudinal Data New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li Abstract This paper develops a new estimation of nonparametric regression functions for

More information

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data ew Local Estimation Procedure for onparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li The Pennsylvania State University Technical Report Series #0-03 College of Health and Human

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

The EM Algorithm for the Finite Mixture of Exponential Distribution Models

The EM Algorithm for the Finite Mixture of Exponential Distribution Models Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 2, 57-64 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.312133 The EM Algorithm for the Finite Mixture of Exponential Distribution

More information

ON SOME TWO-STEP DENSITY ESTIMATION METHOD

ON SOME TWO-STEP DENSITY ESTIMATION METHOD UNIVESITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XLIII 2005 ON SOME TWO-STEP DENSITY ESTIMATION METHOD by Jolanta Jarnicka Abstract. We introduce a new two-step kernel density estimation method,

More information

A nonparametric method of multi-step ahead forecasting in diffusion processes

A nonparametric method of multi-step ahead forecasting in diffusion processes A nonparametric method of multi-step ahead forecasting in diffusion processes Mariko Yamamura a, Isao Shoji b a School of Pharmacy, Kitasato University, Minato-ku, Tokyo, 108-8641, Japan. b Graduate School

More information

Nonparametric Regression. Changliang Zou

Nonparametric Regression. Changliang Zou Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,

More information

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Statistica Sinica 18(2008), 515-534 DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Kani Chen 1, Jianqing Fan 2 and Zhezhen Jin 3 1 Hong Kong University of Science and Technology,

More information

NONPARAMETRIC MIXTURE OF REGRESSION MODELS

NONPARAMETRIC MIXTURE OF REGRESSION MODELS NONPARAMETRIC MIXTURE OF REGRESSION MODELS Huang, M., & Li, R. The Pennsylvania State University Technical Report Series #09-93 College of Health and Human Development The Pennsylvania State University

More information

An Empirical Characteristic Function Approach to Selecting a Transformation to Normality

An Empirical Characteristic Function Approach to Selecting a Transformation to Normality Communications for Statistical Applications and Methods 014, Vol. 1, No. 3, 13 4 DOI: http://dx.doi.org/10.5351/csam.014.1.3.13 ISSN 87-7843 An Empirical Characteristic Function Approach to Selecting a

More information

Test for Discontinuities in Nonparametric Regression

Test for Discontinuities in Nonparametric Regression Communications of the Korean Statistical Society Vol. 15, No. 5, 2008, pp. 709 717 Test for Discontinuities in Nonparametric Regression Dongryeon Park 1) Abstract The difference of two one-sided kernel

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

Transformation and Smoothing in Sample Survey Data

Transformation and Smoothing in Sample Survey Data Scandinavian Journal of Statistics, Vol. 37: 496 513, 2010 doi: 10.1111/j.1467-9469.2010.00691.x Published by Blackwell Publishing Ltd. Transformation and Smoothing in Sample Survey Data YANYUAN MA Department

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

On variable bandwidth kernel density estimation

On variable bandwidth kernel density estimation JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced

More information

Local Polynomial Modelling and Its Applications

Local Polynomial Modelling and Its Applications Local Polynomial Modelling and Its Applications J. Fan Department of Statistics University of North Carolina Chapel Hill, USA and I. Gijbels Institute of Statistics Catholic University oflouvain Louvain-la-Neuve,

More information

Discussion Paper No. 28

Discussion Paper No. 28 Discussion Paper No. 28 Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on the Circle Yasuhito Tsuruta Masahiko Sagae Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

A Comparison of Robust Estimators Based on Two Types of Trimming

A Comparison of Robust Estimators Based on Two Types of Trimming Submitted to the Bernoulli A Comparison of Robust Estimators Based on Two Types of Trimming SUBHRA SANKAR DHAR 1, and PROBAL CHAUDHURI 1, 1 Theoretical Statistics and Mathematics Unit, Indian Statistical

More information

Local Polynomial Estimation for Sensitivity Analysis on Models With Correlated Inputs

Local Polynomial Estimation for Sensitivity Analysis on Models With Correlated Inputs Local Polynomial Estimation for Sensitivity Analysis on Models With Correlated Inputs Sébastien Da Veiga, François Wahl, Fabrice Gamboa To cite this version: Sébastien Da Veiga, François Wahl, Fabrice

More information

Label Switching and Its Simple Solutions for Frequentist Mixture Models

Label Switching and Its Simple Solutions for Frequentist Mixture Models Label Switching and Its Simple Solutions for Frequentist Mixture Models Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Abstract The label switching

More information

Local Polynomial Wavelet Regression with Missing at Random

Local Polynomial Wavelet Regression with Missing at Random Applied Mathematical Sciences, Vol. 6, 2012, no. 57, 2805-2819 Local Polynomial Wavelet Regression with Missing at Random Alsaidi M. Altaher School of Mathematical Sciences Universiti Sains Malaysia 11800

More information

Adaptive Kernel Estimation of The Hazard Rate Function

Adaptive Kernel Estimation of The Hazard Rate Function Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

Choosing the Summary Statistics and the Acceptance Rate in Approximate Bayesian Computation

Choosing the Summary Statistics and the Acceptance Rate in Approximate Bayesian Computation Choosing the Summary Statistics and the Acceptance Rate in Approximate Bayesian Computation COMPSTAT 2010 Revised version; August 13, 2010 Michael G.B. Blum 1 Laboratoire TIMC-IMAG, CNRS, UJF Grenoble

More information

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

Nonparametric Inference via Bootstrapping the Debiased Estimator

Nonparametric Inference via Bootstrapping the Debiased Estimator Nonparametric Inference via Bootstrapping the Debiased Estimator Yen-Chi Chen Department of Statistics, University of Washington ICSA-Canada Chapter Symposium 2017 1 / 21 Problem Setup Let X 1,, X n be

More information

1 Empirical Likelihood

1 Empirical Likelihood Empirical Likelihood February 3, 2016 Debdeep Pati 1 Empirical Likelihood Empirical likelihood a nonparametric method without having to assume the form of the underlying distribution. It retains some of

More information

Automatic Local Smoothing for Spectral Density. Abstract. This article uses local polynomial techniques to t Whittle's likelihood for spectral density

Automatic Local Smoothing for Spectral Density. Abstract. This article uses local polynomial techniques to t Whittle's likelihood for spectral density Automatic Local Smoothing for Spectral Density Estimation Jianqing Fan Department of Statistics University of North Carolina Chapel Hill, N.C. 27599-3260 Eva Kreutzberger Department of Mathematics University

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Optimal Estimation of a Nonsmooth Functional

Optimal Estimation of a Nonsmooth Functional Optimal Estimation of a Nonsmooth Functional T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania http://stat.wharton.upenn.edu/ tcai Joint work with Mark Low 1 Question Suppose

More information

Empirical likelihood based modal regression

Empirical likelihood based modal regression Stat Papers (2015) 56:411 430 DOI 10.1007/s00362-014-0588-4 REGULAR ARTICLE Empirical likelihood based modal regression Weihua Zhao Riquan Zhang Yukun Liu Jicai Liu Received: 24 December 2012 / Revised:

More information

"ZERO-POINT" IN THE EVALUATION OF MARTENS HARDNESS UNCERTAINTY

ZERO-POINT IN THE EVALUATION OF MARTENS HARDNESS UNCERTAINTY "ZERO-POINT" IN THE EVALUATION OF MARTENS HARDNESS UNCERTAINTY Professor Giulio Barbato, PhD Student Gabriele Brondino, Researcher Maurizio Galetto, Professor Grazia Vicario Politecnico di Torino Abstract

More information

Convergence rates for uniform confidence intervals based on local polynomial regression estimators

Convergence rates for uniform confidence intervals based on local polynomial regression estimators Journal of Nonparametric Statistics ISSN: 1048-5252 Print) 1029-0311 Online) Journal homepage: http://www.tandfonline.com/loi/gnst20 Convergence rates for uniform confidence intervals based on local polynomial

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

KANSAS STATE UNIVERSITY

KANSAS STATE UNIVERSITY ROBUST MIXTURES OF REGRESSION MODELS by XIUQIN BAI M.S., Kansas State University, USA, 2010 AN ABSTRACT OF A DISSERTATION submitted in partial fulfillment of the requirements for the degree DOCTOR OF PHILOSOPHY

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 Last week... supervised and unsupervised methods need adaptive

More information

Finite Population Sampling and Inference

Finite Population Sampling and Inference Finite Population Sampling and Inference A Prediction Approach RICHARD VALLIANT ALAN H. DORFMAN RICHARD M. ROYALL A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane

More information

SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT

SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT Statistica Sinica 22 (2012), 27-46 doi:http://dx.doi.org/10.5705/ss.2010.062 SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT J. Cao and W. Yao Simon Fraser University and Kansas

More information

Multiscale Exploratory Analysis of Regression Quantiles using Quantile SiZer

Multiscale Exploratory Analysis of Regression Quantiles using Quantile SiZer Multiscale Exploratory Analysis of Regression Quantiles using Quantile SiZer Cheolwoo Park Thomas C. M. Lee Jan Hannig April 6, 2 Abstract The SiZer methodology proposed by Chaudhuri & Marron (999) is

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Han Liu John Lafferty Larry Wasserman Statistics Department Computer Science Department Machine Learning Department Carnegie Mellon

More information

Smooth functions and local extreme values

Smooth functions and local extreme values Smooth functions and local extreme values A. Kovac 1 Department of Mathematics University of Bristol Abstract Given a sample of n observations y 1,..., y n at time points t 1,..., t n we consider the problem

More information

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Statistica Sinica 19 (2009), 71-81 SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Song Xi Chen 1,2 and Chiu Min Wong 3 1 Iowa State University, 2 Peking University and

More information

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE

DEPARTMENT MATHEMATIK ARBEITSBEREICH MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE Estimating the error distribution in nonparametric multiple regression with applications to model testing Natalie Neumeyer & Ingrid Van Keilegom Preprint No. 2008-01 July 2008 DEPARTMENT MATHEMATIK ARBEITSBEREICH

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES

THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES REVSTAT Statistical Journal Volume 5, Number 1, March 2007, 1 17 THE BREAKDOWN POINT EXAMPLES AND COUNTEREXAMPLES Authors: P.L. Davies University of Duisburg-Essen, Germany, and Technical University Eindhoven,

More information

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES Liping Zhu, Mian Huang, & Runze Li The Pennsylvania State University Technical Report Series #10-104 College of Health and Human Development

More information

Computation of an efficient and robust estimator in a semiparametric mixture model

Computation of an efficient and robust estimator in a semiparametric mixture model Journal of Statistical Computation and Simulation ISSN: 0094-9655 (Print) 1563-5163 (Online) Journal homepage: http://www.tandfonline.com/loi/gscs20 Computation of an efficient and robust estimator in

More information

Pattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods

Pattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods Pattern Recognition and Machine Learning Chapter 6: Kernel Methods Vasil Khalidov Alex Kläser December 13, 2007 Training Data: Keep or Discard? Parametric methods (linear/nonlinear) so far: learn parameter

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

LOCAL FITTING WITH A POWER BASIS

LOCAL FITTING WITH A POWER BASIS REVSTAT Statistical Journal Volume 2, Number 2, November 2004 LOCAL FITTING WITH A POWER BASIS Authors: Jochen Einbeck Department of Statistics, Ludwig Maximilians University Munich, Germany einbeck@stat.uni-muenchen.de)

More information

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University Integrated Likelihood Estimation in Semiparametric Regression Models Thomas A. Severini Department of Statistics Northwestern University Joint work with Heping He, University of York Introduction Let Y

More information

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Robert V. Breunig Centre for Economic Policy Research, Research School of Social Sciences and School of

More information

Local Whittle Likelihood Estimators and Tests for non-gaussian Linear Processes

Local Whittle Likelihood Estimators and Tests for non-gaussian Linear Processes Local Whittle Likelihood Estimators and Tests for non-gaussian Linear Processes By Tomohito NAITO, Kohei ASAI and Masanobu TANIGUCHI Department of Mathematical Sciences, School of Science and Engineering,

More information

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression Local regression I Patrick Breheny November 1 Patrick Breheny STA 621: Nonparametric Statistics 1/27 Simple local models Kernel weighted averages The Nadaraya-Watson estimator Expected loss and prediction

More information

Kernel Density Based Linear Regression Estimate

Kernel Density Based Linear Regression Estimate Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency

More information

A Perturbation Technique for Sample Moment Matching in Kernel Density Estimation

A Perturbation Technique for Sample Moment Matching in Kernel Density Estimation A Perturbation Technique for Sample Moment Matching in Kernel Density Estimation Arnab Maity 1 and Debapriya Sengupta 2 Abstract The fundamental idea of kernel smoothing technique can be recognized as

More information

Statistica Sinica Preprint No: SS

Statistica Sinica Preprint No: SS Statistica Sinica Preprint No: SS-017-0013 Title A Bootstrap Method for Constructing Pointwise and Uniform Confidence Bands for Conditional Quantile Functions Manuscript ID SS-017-0013 URL http://wwwstatsinicaedutw/statistica/

More information

Nonparametric Regression. Badr Missaoui

Nonparametric Regression. Badr Missaoui Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

MULTIVARIATE TECHNIQUES, ROBUSTNESS

MULTIVARIATE TECHNIQUES, ROBUSTNESS MULTIVARIATE TECHNIQUES, ROBUSTNESS Mia Hubert Associate Professor, Department of Mathematics and L-STAT Katholieke Universiteit Leuven, Belgium mia.hubert@wis.kuleuven.be Peter J. Rousseeuw 1 Senior Researcher,

More information

Modification and Improvement of Empirical Likelihood for Missing Response Problem

Modification and Improvement of Empirical Likelihood for Missing Response Problem UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

A Derivation of the EM Updates for Finding the Maximum Likelihood Parameter Estimates of the Student s t Distribution

A Derivation of the EM Updates for Finding the Maximum Likelihood Parameter Estimates of the Student s t Distribution A Derivation of the EM Updates for Finding the Maximum Likelihood Parameter Estimates of the Student s t Distribution Carl Scheffler First draft: September 008 Contents The Student s t Distribution The

More information

Expectation Maximization and Mixtures of Gaussians

Expectation Maximization and Mixtures of Gaussians Statistical Machine Learning Notes 10 Expectation Maximiation and Mixtures of Gaussians Instructor: Justin Domke Contents 1 Introduction 1 2 Preliminary: Jensen s Inequality 2 3 Expectation Maximiation

More information

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i NADARAYA WATSON ESTIMATE JAN 0, 2006: version 2 DATA: (x i, Y i, i =,..., n. ESTIMATE E(Y x = m(x by n i= ˆm (x = Y ik ( x i x n i= K ( x i x EXAMPLES OF K: K(u = I{ u c} (uniform or box kernel K(u = u

More information

Single Equation Linear GMM with Serially Correlated Moment Conditions

Single Equation Linear GMM with Serially Correlated Moment Conditions Single Equation Linear GMM with Serially Correlated Moment Conditions Eric Zivot November 2, 2011 Univariate Time Series Let {y t } be an ergodic-stationary time series with E[y t ]=μ and var(y t )

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units Sahar Z Zangeneh Robert W. Keener Roderick J.A. Little Abstract In Probability proportional

More information

Statistical Properties of Numerical Derivatives

Statistical Properties of Numerical Derivatives Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective

More information

A Novel Nonparametric Density Estimator

A Novel Nonparametric Density Estimator A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with

More information

Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures

Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures David Hunter Pennsylvania State University, USA Joint work with: Tom Hettmansperger, Hoben Thomas, Didier Chauveau, Pierre Vandekerkhove,

More information

Local Polynomial Regression with Application to Sea Surface Temperatures

Local Polynomial Regression with Application to Sea Surface Temperatures Clemson University TigerPrints All Theses Theses 8-2012 Local Polynomial Regression with Application to Sea Surface Temperatures Michael Finney Clemson University, finney2@clemson.edu Follow this and additional

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling

More information

Tobit and Interval Censored Regression Model

Tobit and Interval Censored Regression Model Global Journal of Pure and Applied Mathematics. ISSN 0973-768 Volume 2, Number (206), pp. 98-994 Research India Publications http://www.ripublication.com Tobit and Interval Censored Regression Model Raidani

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

3 Nonparametric Density Estimation

3 Nonparametric Density Estimation 3 Nonparametric Density Estimation Example: Income distribution Source: U.K. Family Expenditure Survey (FES) 1968-1995 Approximately 7000 British Households per year For each household many different variables

More information

SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT. J. Cao and W. Yao. Simon Fraser University and Kansas State University

SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT. J. Cao and W. Yao. Simon Fraser University and Kansas State University SEMIPARAMETRIC MIXTURE OF BINOMIAL REGRESSION WITH A DEGENERATE COMPONENT J. Cao and W. Yao Simon Fraser University and Kansas State University Abstract: Many datasets contain a large number of zeros,

More information

Local Modelling with A Priori Known Bounds Using Direct Weight Optimization

Local Modelling with A Priori Known Bounds Using Direct Weight Optimization Local Modelling with A Priori Known Bounds Using Direct Weight Optimization Jacob Roll, Alexander azin, Lennart Ljung Division of Automatic Control Department of Electrical Engineering Linköpings universitet,

More information

Nonparametric Density Estimation

Nonparametric Density Estimation Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted

More information

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram

More information

JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODEL

JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODEL 1 / 25 JEREMY TAYLOR S CONTRIBUTIONS TO TRANSFORMATION MODELS DEPT. OF STATISTICS, UNIV. WISCONSIN, MADISON BIOMEDICAL STATISTICAL MODELING. CELEBRATION OF JEREMY TAYLOR S OF 60TH BIRTHDAY. UNIVERSITY

More information

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University

More information