Nonparametric Estimation of Smooth Conditional Distributions

Size: px
Start display at page:

Download "Nonparametric Estimation of Smooth Conditional Distributions"

Transcription

1 Nonparametric Estimation of Smooth Conditional Distriutions Bruce E. Hansen University of Wisconsin May 24 Astract This paper considers nonparametric estimation of smooth conditional distriution functions (CDFs) using kernel smoothing methods. We propose estimation using a new smoothed local linear (SLL) estimator. Estimation ias is reduced through the use of a local linear estimator rather than local averaging. Estimation variance is reduced through the use of smoothing. Asymptotic analysis of mean integrated squared error (MISE) reveals the form of these efficiency gains, and their magnitudes are demonstrated in numerical simulations. Considerale attention is devoted to the development of a plugin rule for andwidths which minimize estimates of the asymptotic MISE. We illustrate the estimation method with an application to the U.S. quarterly GDP growth rate. Research supported y the National Science Foundation. Department of Economics, 118 Oservatory Drive, University of Wisconsin, Madison, WI 5376

2 1 Introduction This paper considers nonparametric estimation of smooth conditional distriution functions (CDFs). As the CDF is the conditional expectation of an indicator function, it may e estimated y nonparametric regression estimation methods. A recent paper which exploits advances in nonparametric regression for estimation of the CDF is Hall, Wolff and Yao (1999). See also Fan and Yao (23). Standard nonparametric regression estimates of the CDF, however, are not smooth and therefore are inefficient. If the dependent variale in these regressions the indicator function is replaced y a smooth function, estimation variance can e reduced at the cost of an increase in ias. If the degree of smoothing is carefully chosen, it can result in a lowered finite sample mean integrated squared error (MISE). The advantages of smoothed estimation of distriution functions has een discussed in the statistics literature. For references see Hansen (24). The extension to CDF estimation considered here is new. Our analysis concerns a pair of random variales (Y,) R R that have a smooth joint distriution. Let f(y) and g(x) denote the marginal densities of Y and, and let f (y x) denote the conditional density of Y given x. The conditional distriution function is Define as well the derivatives F (y x) Z y f (u x) du. F (s) (y x) s F (y x) xs f (s) (y x) s f (y x) xs f (y x) f (y x). y Given a random sample {Y 1, 1,..., Y n, n } from this distriution, the goal is to estimate F (y x) at a fixed value x. To simplify the notation we will sometimes suppress dependence on x. Let Φ(s) and φ(s) denote the standard normal distriution and density functions and let Φ σ (s) Φ(s/σ) and φ σ (s) σ 1 φ(s/σ). Let φ (m) (s) (d m /ds m ) φ(s) and φ σ (m) (s) (d m /ds m ) φ σ (s) σ (m+1) φ (m) (s/σ) denote the derivatives of φ(s) and φ σ (s). In particular, φ (1) (s) sφ(s) and φ (2) (s) s 2 1 φ(s). Our estimators will make use of two univariate kernel functions w(s) and k(s). We assume that oth are symmetric and second-order, are normalized to have a unit variance (e.g. R s2 w(s)ds 1), and have a finite sixth moment (e.g. R s6 w(s)ds < ). Let R R w(s)2 ds denote the roughness of w. Let Z s K(s) k(u)du 1

3 denote the integrated kernel of k and define the constant ψ 2 sk(s)k(s)ds >. (1) For our applications, we set oth k(s) and w(s) to equal the normal kernel φ(s), in which case R 1/2 π, ψ 1/ π and K(s) Φ(s). Section 2 presents unsmoothed estimators. Section 3 introduces our smoothed estimators. Section 4 develops plug-in andwidth rules for the local linear and smoothed local linear estimator. Section 5 presents a numerical simulation of performance. Section 6 is a simple application to the U.S. quarterly real GDP growth rate. Section 7 concludes. The proof of Theorem 1 is presented in the Appendix. Gauss code which implements the estimators and andwidth methods is availale on the author s wepage 2 Unsmoothed Estimators Note that F (y x) E (1 (Y i y) i x), andthusitisaregressionfunction. ThustheCDFmaye estimated y regression of 1 (Y i y) on i. A simple nonparametric estimator is the Nadaraya-Watson (NW) estimator, and takes the form of a local average ˆF (y x) P n i1 w i1 (Y i y) P n i1 w i (2) where w i 1 µ x w i (3) are kernel weights and is a andwidth. Reduced ias may e achieved using a local linear (LL) estimator. Local linear estimation is a special case of local polynomial regression, and is typically recommended for estimation mean. For a general treatment see Fan and Gijels (1996). The estimator is the intercept from a weighted least-squares regression of 1 (Y i y) on (x i ) using the weights w i, and can e written as P n i1 F (y x) w i 1 (Y i y) P n i1 w i wi w i ³1 ˆβ (x i ) Ã n ˆβ i1 w i (x i ) 2! 1 Ã n i1! w i (x i ). (4) (5) AsshowninTheorem1() of Hall, Wolff and Yao (1999), if cn 1/5 then ³ 2 R (F (y x)(1 F (y x))) E F (y x) F (y x) ' + 4 g(x)n 4 F (2) (y x) 2 + O ³n 6/5. 2

4 Integrating over y we otain the MISE ³ 2 RV E F (y x) F (y x) dy gn + 4 V 1 + O ³n 6/5 4 (6) where g g(x) V V 1 F (y x)(1 F (y x)) dy F (2) (y x) 2 dy. The first term on the right side of (6) is the integrated asymptotic variance, and the second is the integrated squared asymptotic ias. This ias term is simplified relative to the NW estimator, and this is the source of the improved performance of F over ˆF. Hall, Wolff and Yao (1999) point out that the local linear estimator F has two undesirale properties. It may e nonmonotonic in y, and is not constrained to lie in [, 1]. Thus the estimator F is not a CDF. The prolem occurs when the weights (5) are negative, which occurs when ˆβ (x i ) > 1. A simple solution 1 is to modify wi to exclude negative values. Thus we redefine the weights as follows ( wi ˆβ (x i ) > 1 w i ³1 ˆβ (7) (x i ) ˆβ (x i ) 1 This modification is asymptotically negligile, ut constrains F to e a valid CDF. Using the modified weights (7) our estimator is a slight modification of the local linear estimator, however we will continue to refer to it as the local linear (LL) estimator for simplicity. 3 Smoothed Estimators For some andwidth h> let K h (s) K(s/h) and consider the integrated kernel smooth K h (y Y i ). For small h it can e viewed as a smooth approximation to the indicator function 1 (Y i y). Smoothed CDF estimators replace the indicator function 1 (Y i y) in (2) and (4) with K h (y Y i ), yielding the smoothed Nadaraya-Watson (SNW) estimator ˆF h, (y x) P n i1 w ik h (y Y i ) P n i1 w i 1 Hall, Wolff and Yao (1999) proposed an alternative solution, introducing weights to the Nadaraya-Watson estimator which are selected y an empirical likelihood criterion. Their estimator is asymptotically equivalent to ours, ut computationally more complicated. 3

5 and the smoothed local linear (SLL) estimator F h, (y x) P n i1 w i K h (y Y i ) P n. i1 w i The weights w i and w i are definedin(3)and(7),respectively. Theorem 1 Assume that F (y x) and g(x) are continuously differentiale up to the fourth order in oth y and x. If cn 1/5 and h O() as n, then ³ 2 R E F h, (y x) F (y x) dy gn (V V 1 hψ)+4 h2 2 V 2 + h4 V 3 + O ³n 6/ (8) where g g(x) V V 1 V 2 V 3 F (y x)(1 F (y x)) dy F (2) (y x) 2 dy f (y x) f (2) (y x) dy f (y x) 2 dy. The MISE in (8) is a generalization of (6) to allow for h. The MISE is strictly decreasing at h (since ψ > ) demonstrating that the asymptotic MISE is minimized for strictly positive h. Thus there are (at least asymptotic) efficiency gains from smoothing. 4 Plug-In Bandwidth Selection We suggest selecting the andwidths (h, ) y the plug-in method. Plug-in andwidths are the values which minimize an estimate of the asymptotic MISE (8). Estimation of the latter requires estimation of the parameters g, V, V 1,V 2, and V 2. In this section we discuss estimation of these components, and implementation of the plug-in andwidth rule. Readers who are interested in the plug-in algorithm ut not the motivational details can jump directly to Section 4.5, which summarizes the implementation details. 4.1 Kernel Estimation of g The normal kernel density estimator is ĝ 1 n n φ r (x i ) (9) j1 4

6 where r is a andwidth. An estimate of the asymptotic MSE-minimizing choice for r is ˆr 2 π ĝ s (x) ³ ĝ (2) t (x) 2 n 1/5 (1) where ĝ s (x) 1 n ĝ (2) t (x) 1 n n φ s (x i ) j1 n j1 φ (2) t (x i ) and s and t are preliminary andwidths. Letting ˆσ x denote the sample standard deviation for i, we suggest the Gaussian reference rules s 1.6ˆσ x n 1/5 and t.94ˆσ x n 1/9, which are MISE-optimal for estimation of the density and its second derivative when g is Gaussian. 4.2 Estimation of V The parameter V is a function of the conditional distriution F (y x). A parametric estimator of the latter is the intercept from a p th order polynomial regression of 1 (Y i y) on ( i x). For p let 1 1 x ( 1 x) p p..., (11) 1 n x ( n x) p and let 1 (Y y) denote the stacked 1 (Y i y). Then the estimator can e written as ˆF (y x) δ 1 p p p 1 (Y y) (12) where δ 1 (1 ) is the first unit vector. Given (12), an estimate of V is ³ ˆV ˆF (y x) 1 ˆF (y x) dy Z δ 1 1 p p p 1 (Y >y) 1 (Y y) dy p 1 p p δ1 δ 1 p p p Y Y + p p p 1 δ1 (13) where (a) + a1 (a ) for scalar a, and is an element-y-element operation for matrices. A nonparametric estimator of V can e ased on a local linear regression. Given a andwidth q, let w i φ q (x i ) denote Gaussian kernel weights and set W diag{w i }. Then the estimator is F (y x) δ 1 1 W 1 1 W 1 (Y y) 5

7 and the estimate of V is Ṽ δ 1 1 W 1 1 W Y Y + W 1 1 W 1 1 δ1. (14) The andwidth q which minimizes the asymptotic mean integrated squared error of F (y x) is µ V q 2 πgv 1 n which depends on the unknowns g, V and V 1. (See equation (8) with h.) Using the estimates ĝ from (9), ˆV from (13), and ˆV 1 from (17) elow, we otain a plug-in andwidth ˆq.78 Ã ˆV ĝ ˆV 1 n 1/5! 1/5. (15) 4.3 Estimation of V 1 The parameter V 1 is a function of the second derivative F (2) (y x) whichcaneestimatedypolynomial regression. For some p 2 let p e defined as in (11). The estimator is ˆF (2) (y x) δ 3 1 p p p 1 (Y y) (16) where δ 3 is the third unit vector (a one in the third place, and zeros elsewhere). To calculate V 2 from (16), oserve that for any c>max Y i Z c 1 (Y y) 1 (Y y) dy Dc max Y,Y where D is an n n matrix of ones, and oserve that δ 3 p p p D. Thus Z c Z ˆF (2) (y x) 2 dy δ c 3 1 p p p 1 (Y y) 1 (Y y) dy p 1 p p δ3 δ 3 p p p Dc max Y,Y p p p 1 δ3 δ 3 1 p p p max Y,Y p 1 p p δ3 where the max operator is element-y-element. Note that this expression is invariant to c>max Y i, thus we define ˆV 1 δ 3 1 p p p max Y,Y p 1 p p δ3. (17) This estimator was also used in (15) to construct a plug-in rule for the nonparametric estimate Ṽ. A nonparametric estimator of V 1 can e ased on a local cuic regression. (While any local polynomial regression of order two or aove is consistent, local cuic regression is recommend for estimation of a regression second derivative.) Given a andwidth q 1, let w 1i φ q1 (x i ) denote Gaussian kernel 6

8 weights and set W 1 diag{w 1i }. The estimator of F (2) (y x) is F (2) (y x) δ 3 3 W W 1 1 (Y y) (18) and the estimate of V 1 is Ṽ 1 δ 3 3 W W 1 max Y,Y W W δ3. (19) Using Fan and Gijels (1996, Theorem 3.1 and equation (3.2)) the andwidth which minimizes the asymptotic mean integrated squared error of ˆF (2) (y x) is where The constant V 4 may e estimated y µ V 1/9 q gv 4 n V 4 F (4) (y x) 2 dy. ˆV 4 δ 5 p p p max Y,Y p p p 1 δ5 (2) where p 4 and δ 5 is the fifth unit vector, suggesting the plug-in andwidth ˆq Ã! 1/9 ˆV. (21) ĝ ˆV 4 n 4.4 Estimation of V 2 and V 3 If f (y x) does not depend on x then we have the simplification V 3 R f (y) 2 dy which is a wellstudied estimation prolem. For a andwidth a let e the n 1 vector of ones and ˆf a (y) 1 n n j1 φ (1) a (y Y i)φ (1) a y Y 1 e a Gaussian kernel estimate of f (y). An estimate of V 3 in this case is ˆf a (y)2 dy φ (2) a φ (1) a (y Y ) φ(1) y Y dy 1 a Y Y 1 (22) where a 2a. The second equality is equation (7.1) ofmarronandwand(1992): φ (r) a (y Y i ) φ (s) a (y Y j ) dy ( 1) r φ (r+s) 2a (Y i Y j ). (23) 7

9 The estimator (22) is the Jones and Sheather (1991) estimator. Hansen (24) descries a multiple plug-in method for selection of a. We recommend this choice and hold it fixed for the remainder of this su-section and the next. Now consider estimation of V 3 allowing for dependence on x. Note that the estimate of f (y) in (24) is a regression of φ (1) a (y Y i) on a constant. Similarly, we can estimate f (y x) as the intercept in a polynomial regression on x i x. For p definedin(11) let The associated estimate of V 3 is ˆV 3 ˆf a (y x)2 dy Z δ 1 1 p p p δ 1 ˆf a (y x) δ 1 1 p p p φ (1) a (y Y ). (24) p p p φ (2) a φ (1) a (y Y ) φ(1) y Y dy p 1 p p δ1 a Y Y p p p 1 δ1. (25) Alternatively, we could use a fully nonparametric estimator such as a local linear estimator. However, the optimal andwidths depend on derivatives of the conditional density that are difficult to estimate. We thus do not consider further complications. In addition, using an analog of (16) and (24), a simple estimator of V 2 is ˆV 2 ˆf (2) a (y x) ˆf a (y) dy Z δ 3 1 p p p φ a (y Y ) φ a y Y dy p 1 p p δ1 where the third equality uses (23). δ 3 p p p φ a Y Y p p p 1 δ1. (26) 4.5 Plug-In Bandwidth Our plug-in estimate of (8) is M (h, ) R ³Ṽ hψ + 4 Ṽ 1 2 h2 2 ˆV 2 + h4 ˆV 3 ĝn (27) The plug-in andwidth pair ( h, ) jointly minimize M (h, ). A closed-form solution is not availale, so the function needs to e numerically minimized. One caveat is that we must constrain h<ψ/ṽ to ensure a sensile solution. (When h>ψ/ṽ then M is strictly increasing in so unconstrained minimization sets which is not sensile.) As a practical matter it makes sense to ound h more sustantially, we suggest the constraint h ψ/2ṽ,andthisisdoneinournumericalapplications. The plug-in andwidth for the unsmoothed estimator F is found y setting h and minimizing 8

10 to find ˆ ³ RṼ/ĝṼ 1 n 1/6. To summarize, the computation algorithm is as follows. First, ĝ is calculated from (9) using (1). Second, for some p 4, ˆV, ˆV 1 and ˆV 4 are calculated using (13), (17) and (2). Third, using the andwidth (15), Ṽ is calculated using (14). Fourth, using the andwidth (21), Ṽ 1 is calculated using (19). Fifth, the andwidth a is calculated using the method of Hansen (24), allowing ˆV 2 and ˆV 3 to e calculated using (26) and (25). Finally, ĝ, Ṽ,Ṽ 1, ˆV 2, ˆV 3 are set into (27) which is numerically minimized over (h, ) to otain ( h, ). 5 Simulation The performance of the CDF estimators is investigated in a simple simulation experiment using ivariate data. The variales i are generated as iid N(, 1). The variale Y i is generated as Y i i + e i where e i is iid, independent of i. The error e i is distriuted as G, one of the first nine mixture-normal distriutions from Marron and Wand (1992). Consequently, the true CDF for Y i is G(y x). Foreach choice of G we generated 1, independent samples of size n 5, 1 and 3. ForeachsampleweestimatedtheCDFatx 1 and y { 2., 1.5, 1.,.5,.,.5, 1., 1.5, 2.}, using the NW, SNW, LL and SLL estimators. Our measure of precision is the exact mean-squared error, averaged over these values of y (as an approximation to the MISE). The standard normal kernel was used for all estimates. We first astracted from the issue of andwidth selection and evaluated the precision of the four estimates using the infeasile finite-sample-optimal andwidth, calculated y quasi-newton minimization. The results are presented in the first three columns of Tale 1. Wehavetakentheunsmoothed NW estimator as a aseline, and have divided the MISE of the other estimators y that of the NW estimator. Thus the numers in Tale 1 represent the relative improvement attained relative to the NW estimator. The results show that across models and sample sizes, the SLL estimator achieves the lowest MISE. In each case, using local linear estimation rather than local averaging reduces the MISE, and using smoothing lowers the MISE. We next investigated the performance of our plug-in andwidths ˆ for the LL estimator and ( h, ) for the SLL estimator. On each sample descried aove we calculated these plug-in andwidths and CDF estimators, and evaluated the MISE across the 1, samples. The results are presented in the final two columns of Tale 1. As efore, the MISE is normalized y the MISE of the infeasile NW estimator. Thus the numers represented the relative improvement attained y the feasile plug-in estimators relative to this infeasile estimator. In each case, the plug-in SLL estimator has lower MISE than the plug-in LL estimator. In some cases the difference is minor, in other case it is quite sustantial (the largest gains are for Model #5, where the MISE is reduced y aout 25%). In each case the MISE is higher than if the infeasile optimal andwidth had een used, in some cases sustantially higher, suggesting that further improvements in andwidth choice may e possile. 9

11 Overall, the simulation evidence confirms the implications of the asymptotic analysis. We can recommend use of the SLL estimator (with our proposed plug-in andwidth) for empirical practice. 6 GDP Growth Forecasting We illustrate the method with a simple application to the U.S. real GDP growth rate. Let GDP i denote the level of quarterly real GDP and let y i 4(ln(GDP i ) ln (GDP i 1 )) denote the annualized quarterly growth rate. Set x i y i 1. We restrict the sample to run from first third quarter of 1984 to the first quarter of 24 (79 oservations). This choice is consistent with the decline in output volatility documented y McConnell and Perez-Quiros (2). We estimated the CDF using the local linear (LL) and smoothed local linear (SLL) estimators for y i given x i x, fixing x to equal 2% and 4.5%, which are approximately the 25 th and 75 th quantiles of the unconditional distriution. The CDF was evaluated at 51 evenly spaced increments etween -2 and 7, and the andwidths were calculated y the plug-in method descried in Section 4.5. To provide a contrast, we also estimated the CDF using a homoskedastic Gaussian AR(1). Figure 1 displays the three CDFs for the case of x 2. and Figure 2 for the case x 4.5. In oth cases, the three estimates are meaningfully different from one another. The parametric estimate has a fairly different location and variance. The two nonparametric estimates have the same asic shape, ut the unsmoothed local linear estimate is visually quite erratic, and the smoothed local linear estimator appears to apply an intuitively correct degree of smoothing. It is also interesting to note that the LL estimator uses a plug-in andwidth of ˆ 1.59 in Figure 1 and ˆ 2.13 in Figure 2, while the SLL estimator uses 1.44, h.53 and 2. and h.23 respectively. In oth cases, the plug-in rule selects similar values for for the two estimators (with that for SLL slightly smaller), and selects a consideraly smaller value for h than. 7 Conclusion We have shown how to comine local linear methods and smoothing to produce good nonparametric estimates of the conditional distriution function for ivariate data. We have shown that meaningful improvements in estimation efficiency can e achieved y using these techniques. Furthermore, we have derived data-ased plug-in rules for andwidth selection which appear to work in practice. These tools should prove useful for empirical application. This analysis has een confined to ivariate data. An important extension would to the case of multivariate data. In particular, this extension would require the study and development of feasile andwidth selection rules. 1

12 8 Appendix: Proof of Theorem 1 The proof of Theorem 1 is ased on the following set of expansions. Lemma 1 Let g (s) ds dx s g(x) and K h (Y i)k h (y Y i ) F (y x). Ew i g(x)+ 2 2 g(2) (x)+o( 4 ) (28) E (w i (x i )) 2 g (1) (x)+o( 4 ) (29) E ³w 2 i (x i ) 2 g(x)+o( 4 ) (3) E (w i Kh(Y i )) 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x) (31) + h2 2 f (y x) g(x)+o 4 E (w i (x i ) Kh(Y i )) 2 F (1) (y x) g(x)+o 4 (32) E w 2 i K h(y i ) 2 Rg(x)(F (y x)(1 F (y x)) hψf (y x)) +O () (33) E ³w 2 i 2 (x i ) 2 Kh(Y i ) O() (34) Proof. Throughout, we make use of the assumption that h O() to simplify the ounds. The derivations make use of the following Taylor series expansions which are valid under the assumption that the functions are fourth order differentiale: g(x v) g(x) vg (1) (x)+ v2 2 2 g(2) (x) v3 3 6 g(3) (x)+o( 4 ) (35) F (y hu v) F (y v) huf (y v)+ h2 u 2 2 f (y v) h3 u 3 6 f (y v)+o(h 4 ) (36) F (y x v) F (y x) vf (1) (y x)+ 2 v 2 2 F (2) (y x) 3 v 3 6 F (3) (y x)+o( 4 ) (37) f (y x v) f (y x) v y f (1) (y x)+o( 2 ). (38) To show (28), y a change of variales and (35) Ew i 1 w µ x v g(v)dv w (v) g(x v)dv w (v) µg(x) vg (1) (x)+ v2 2 2 g(2) (x) v3 3 6 g(3) (x) dv + O( 4 ) g(x)+ 2 2 g(2) (x)+o( 4 ). 11

13 Note that this makes use of the facts that R w (v) dy 1, R w (v) v, R w (v) v2 1 and R w (v) v3. Similarly E (w i (x i )) (x v) 1 µ x v w vw (v) g(x v)dv g(v)dv vw (v) µg(x) vg (1) (x)+ v2 2 2 g(2) (x) dv + O( 4 ) 2 g (1) (x)+o( 4 ) yielding (29) and E ³w 2 i (x i ) (x v) 2 1 w 2 2 g(x)+o( 4 ) µ x v v 2 w (v) g (x v) dv g(v)dv yielding (3). To show (31) and(32),first oserve that y integration y parts, a change of variales and (36) K h (y u) f (u v) du +O(h 4 ) k h (y u) F (u v) du k (u) F (y hu v) du k (u) µf (y v) huf (y v)+ h2 u 2 2 f (y v) h3 u 3 6 f (y v) du F (y v)+ h2 2 f (y v)+o(h 4 ). (39) 12

14 By a change of variales, and using (35) and (37), F (y v) 1 µ x v w g(v)dv F (y x v) g(x v)w (v) dv µf (y x) vf (1) (y x)+ 2 v 2 2 F (2) (y x) 3 v 3. µg(x) vg (1) (x)+ 2 v 2 2 g(2) (x) v3 3 6 g(3) (x) w(v)+o( 4 ) 6 F (3) (y x) F (y x) g(x)+ 2 2 F (y x) g(2) (x)+ 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x)+o( 4 ). (4) Similarly, y a change of variales and (35) and (38), f (y v) 1 µ Z x v w g(v)dv f (y x v) g(x v)w(v)dv f (y x) g(x)+o( 2 ). (41) Together, (28), (39), (4) and (41) show that which is (31). E (w i Kh(Y i )) E (w i K h (Y i )) E (w i ) F (y x) µ 1 x v w K h (y u) f (u v) g(v)dudv µ g(x)+ 2 2 g(2) (x) F (y x)+o( 4 ) F (y v) 1 µ µ x v w g(v)dv g(x)+ 2 2 g(2) (x) F (y x) g(v)dv + O( 4 ) + h2 2 f (y v) 1 w µ x v h2 2 f (y x) g(x)+ 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x) +O 4 13

15 Similarly, using (29), (39) and similar derivations, E (w i (x i ) Kh(Y i )) E (w i (x i ) K h (Y i )) E (w i (x i )) F (y x) (x v) 1 µ x v w K h (y u) f (u v) g(v)dvdu + 2 g (1) (x)f (y x)+o( 4 ) vf (y x v) g(x v)w (v) dv + h2 2 vf (y x v) g(x v)w (v) dv + 2 g (1) (x)f (y x)+o( 4 ) 2 F (1) (y x) g(x)+o 4 which is (32). To show (33) first oserve that y a change of variales and (35), E Z wi 2 µ 1 x v 2 2 w g (v) dv 1 w (v) 2 g (x v) dv 1 ³ w (v) 2 g(x) vg (1) (x) dv + O() Rg(x) + O(). (42) Second, using (39), a change of variales, (35) and (37), E wi 2 K h (Y i ) µ 1 x v 2 w K h (y u) f (u v) g(v)dvdu µ 1 x v 2 µ h 2 2 w g(v)f (y v) dv + O 1 1 w (v) 2 g(x v)f (y x v) dv + O () ³ ³ w (v) 2 g(x) vg (1) (x) F (y x) vf (1) (y x) dv + O( 2 ) Rg(x)F (y x) + O(). (43) 14

16 Third, y integration y parts, a change of variales, and (36) K h (y u) 2 f (u v) du 2 K h (y u) k h (y u) F (u v) du 2 2 K (u) k (u) F (y hu v) du K (u) k (u)(f (y v) huf (y v)) du + O(h 2 ) F (y v) hψf (y v)+o(h 2 ). (44) The final equality uses the definition (1) and the fact that the symmetry of k(u) aout zero implies 2K (u) k (u) du. Fourth, using (44), (35), (37) and (38), E wi 2 K h (Y i ) 2 µ 1 x v 2 2 w K h (y u) 2 f (u v) g(v)dvdu µ 1 x v 2 µ h 2 2 w (F (y v) hψf (y v)) g(v)dv + O 1 w (v) 2 (F (y x v) hψf (y x v)) g(x v)dv + O () Rg(x)(F (y x) hψf (y x)) + O() (45) Together, (42), (43) and (45) show E wi 2 Kh(Y i ) 2 E wi 2 K h (Y i ) 2 2E wi 2 K h (Y i ) F (y x)+e wi 2 F (y x) 2 Rg(x)(F (y x) hψf (y x)) 2Rg(x)F (y x)2 + Rg(x)F (y x) 2 + O () Rg(x)(F (y x)(1 F (y x)) hψf (y x)) + O () which is (33). (34) is shown similarly. ProofofTheorem1:From equations (29)-(3), Eˆβ x ³ E ³w i (x i ) 2 1 (E (wi (x i ))) + O 2 g(x) 1 g (1) (x)+o 2. 15

17 Comined with (28), this yields Next, using equations (31) and(32), Together with (46) we find Ewi Ew i EˆβE (w i (x i )) + O( 2 ) (46) f(x)+o( 2 ). E (w i K h(y i )) E (w i K h(y i )) EˆβE (w i (x i ) K h(y i )) + O F (2) (y x) g(x)+ h2 2 f (y x) g(x)+o 4. ³ E F h, (y x) F (y x) Ew i K h (y Y i) + O Ew i ³ 2 F (2) (y x)+h 2 f (y x) + O 4. Similarly, using (33) and (34) Var³ F h, (y x) 1 Var(wi K h (y Y i)) n (Ewi )2 µ 1 E (w i Kh (y Y i)) 2 n g(x) 2 + O n R (F (y x)(1 F (y x)) hψf (y x)) g(x)n + O µ O ³n 6/5. n Together, these expressions and the assumption that cn 1/5 estalish that ³ 2 E F h, (y x) F (y x) R (F (y x)(1 F (y x)) ψhf (y x) f(x)) g(x)n + h2 2 f (y x) F (2) (y x) f (y x) F (2) (y x) h4 f (y x) O ³n 6/5. Integrating over y we otain ³ Z R ³ 2 R F (y x)(1 F (y x)) dy ψh R f (y x) dyf(x) E F h, (y x) F (y x) dy g(x)n + h4 f (y x) 2 dy + h2 2 f (y x) F (2) (y x) dy F (2) (y x) 2 dy + O ³n 6/5 4 16

18 which equals the expression in (8) using the equivalence f (y x) F (2) (y x) dy f (y x) f (2) (y x) dy which holds y integration y parts. This completes the proof. 17

19 Tale 1: Normalized MISE Conditional Distriution Function Estimators n 5 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal n 1 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal n 3 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal

20 References [1] Fan, Jianqing and Irene Gijels (1996): Local Polynomial Modelling and Its Applications: NewYork: Chapman and Hall. [2] Fan, Jianqing and Qiwei Yao (23): Nonlinear Time Series: Nonparametric and Parametric Methods. NewYork:Springer. [3] Hall, Peter, Rodney C. L. Wolff, andqiweiyao(1999): Methods for Estimating a Conditional Distriution Function, Journal of the American Statistical Association, 94, [4] Hansen, Bruce E. (24): Bandwidth Selection for Nonparametric Distriution Estimation, manuscript, University of Wisconsin. [5] Jones, M.C. and S. J. Sheather (1991): Using non-stochastic terms to advantage in kernel-ased estimation of integrated squared density derivatives, Statistics and Proaility Letters, 11, [6] Marron, J.S. and M.P. Wand (1992): Exact mean integrated squared error, Annals of Statistics, 2, [7] McConnell, M.M. and G. Perez-Quiros (2): Output fluctuations in the United States: What has changed since the early 198s? American Economic Review, 9,

21

Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement

Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement Open Journal of Statistics, 07, 7, 834-848 http://www.scirp.org/journal/ojs ISS Online: 6-798 ISS Print: 6-78X Estimating a Finite Population ean under Random on-response in Two Stage Cluster Sampling

More information

Estimation of cumulative distribution function with spline functions

Estimation of cumulative distribution function with spline functions INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

A Novel Nonparametric Density Estimator

A Novel Nonparametric Density Estimator A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)

I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2011/2 ESTIMATION

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

Adaptive Kernel Estimation of The Hazard Rate Function

Adaptive Kernel Estimation of The Hazard Rate Function Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

Nonparametric Regression. Changliang Zou

Nonparametric Regression. Changliang Zou Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

A nonparametric method of multi-step ahead forecasting in diffusion processes

A nonparametric method of multi-step ahead forecasting in diffusion processes A nonparametric method of multi-step ahead forecasting in diffusion processes Mariko Yamamura a, Isao Shoji b a School of Pharmacy, Kitasato University, Minato-ku, Tokyo, 108-8641, Japan. b Graduate School

More information

FinQuiz Notes

FinQuiz Notes Reading 9 A time series is any series of data that varies over time e.g. the quarterly sales for a company during the past five years or daily returns of a security. When assumptions of the regression

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

Asymptotic distributions of nonparametric regression estimators for longitudinal or functional data

Asymptotic distributions of nonparametric regression estimators for longitudinal or functional data Journal of Multivariate Analysis 98 (2007) 40 56 www.elsevier.com/locate/jmva Asymptotic distriutions of nonparametric regression estimators for longitudinal or functional data Fang Yao Department of Statistics,

More information

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Robert V. Breunig Centre for Economic Policy Research, Research School of Social Sciences and School of

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation Preface Nonparametric econometrics has become one of the most important sub-fields in modern econometrics. The primary goal of this lecture note is to introduce various nonparametric and semiparametric

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

Each element of this set is assigned a probability. There are three basic rules for probabilities:

Each element of this set is assigned a probability. There are three basic rules for probabilities: XIV. BASICS OF ROBABILITY Somewhere out there is a set of all possile event (or all possile sequences of events which I call Ω. This is called a sample space. Out of this we consider susets of events which

More information

EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS

EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS APPLICATIONES MATHEMATICAE 9,3 (), pp. 85 95 Erhard Cramer (Oldenurg) Udo Kamps (Oldenurg) Tomasz Rychlik (Toruń) EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS Astract. We

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

Simple Examples. Let s look at a few simple examples of OI analysis.

Simple Examples. Let s look at a few simple examples of OI analysis. Simple Examples Let s look at a few simple examples of OI analysis. Example 1: Consider a scalar prolem. We have one oservation y which is located at the analysis point. We also have a ackground estimate

More information

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Kuangyu Wen & Ximing Wu Texas A&M University Info-Metrics Institute Conference: Recent Innovations in Info-Metrics October

More information

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships

More information

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

Real option valuation for reserve capacity

Real option valuation for reserve capacity Real option valuation for reserve capacity MORIARTY, JM; Palczewski, J doi:10.1016/j.ejor.2016.07.003 For additional information aout this pulication click this link. http://qmro.qmul.ac.uk/xmlui/handle/123456789/13838

More information

Estimation of the Error Density in a Semiparametric Transformation Model

Estimation of the Error Density in a Semiparametric Transformation Model Estimation of the Error Density in a Semiparametric Transformation Model Benjamin Colling Université catholique de Louvain Cédric Heuchenne University of Liège and Université catholique de Louvain Rawane

More information

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

Chaos and Dynamical Systems

Chaos and Dynamical Systems Chaos and Dynamical Systems y Megan Richards Astract: In this paper, we will discuss the notion of chaos. We will start y introducing certain mathematical concepts needed in the understanding of chaos,

More information

Local Polynomial Modelling and Its Applications

Local Polynomial Modelling and Its Applications Local Polynomial Modelling and Its Applications J. Fan Department of Statistics University of North Carolina Chapel Hill, USA and I. Gijbels Institute of Statistics Catholic University oflouvain Louvain-la-Neuve,

More information

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

Teruko Takada Department of Economics, University of Illinois. Abstract

Teruko Takada Department of Economics, University of Illinois. Abstract Nonparametric density estimation: A comparative study Teruko Takada Department of Economics, University of Illinois Abstract Motivated by finance applications, the objective of this paper is to assess

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Uniform Convergence Rates for Nonparametric Estimation

Uniform Convergence Rates for Nonparametric Estimation Uniform Convergence Rates for Nonparametric Estimation Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~bansen October 2004 Preliminary and Incomplete Abstract Tis paper presents a set of rate

More information

Luis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model

Luis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model Luis Manuel Santana Gallego 100 Appendix 3 Clock Skew Model Xiaohong Jiang and Susumu Horiguchi [JIA-01] 1. Introduction The evolution of VLSI chips toward larger die sizes and faster clock speeds makes

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad

More information

Transformation-based Nonparametric Estimation of Multivariate Densities

Transformation-based Nonparametric Estimation of Multivariate Densities Transformation-based Nonparametric Estimation of Multivariate Densities Meng-Shiuh Chang Ximing Wu March 9, 2013 Abstract We propose a probability-integral-transformation-based estimator of multivariate

More information

Nonparametric Density Estimation

Nonparametric Density Estimation Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Weak bidders prefer first-price (sealed-bid) auctions. (This holds both ex-ante, and once the bidders have learned their types)

Weak bidders prefer first-price (sealed-bid) auctions. (This holds both ex-ante, and once the bidders have learned their types) Econ 805 Advanced Micro Theory I Dan Quint Fall 2007 Lecture 9 Oct 4 2007 Last week, we egan relaxing the assumptions of the symmetric independent private values model. We examined private-value auctions

More information

Web-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach

Web-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach 1 Web-based Supplementary Material for Dependence Calibration in Conditional Copulas: A Nonparametric Approach Elif F. Acar, Radu V. Craiu, and Fang Yao Web Appendix A: Technical Details The score and

More information

Nonparametric confidence intervals. for receiver operating characteristic curves

Nonparametric confidence intervals. for receiver operating characteristic curves Nonparametric confidence intervals for receiver operating characteristic curves Peter G. Hall 1, Rob J. Hyndman 2, and Yanan Fan 3 5 December 2003 Abstract: We study methods for constructing confidence

More information

DEPARTMENT OF ECONOMICS

DEPARTMENT OF ECONOMICS ISSN 089-64 ISBN 978 0 7340 405 THE UNIVERSITY OF MELBOURNE DEPARTMENT OF ECONOMICS RESEARCH PAPER NUMBER 06 January 009 Notes on the Construction of Geometric Representations of Confidence Intervals of

More information

HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO

HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO The Annals of Statistics 2006, Vol. 34, No. 3, 1436 1462 DOI: 10.1214/009053606000000281 Institute of Mathematical Statistics, 2006 HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO BY NICOLAI

More information

Gaussian kernel GARCH models

Gaussian kernel GARCH models Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often

More information

Nonparametric Function Estimation with Infinite-Order Kernels

Nonparametric Function Estimation with Infinite-Order Kernels Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density

More information

Local Polynomial Wavelet Regression with Missing at Random

Local Polynomial Wavelet Regression with Missing at Random Applied Mathematical Sciences, Vol. 6, 2012, no. 57, 2805-2819 Local Polynomial Wavelet Regression with Missing at Random Alsaidi M. Altaher School of Mathematical Sciences Universiti Sains Malaysia 11800

More information

Conditional density estimation: an application to the Ecuadorian manufacturing sector. Abstract

Conditional density estimation: an application to the Ecuadorian manufacturing sector. Abstract Conditional density estimation: an application to the Ecuadorian manufacturing sector Kim Huynh Indiana University David Jacho-Chavez Indiana University Abstract This note applies conditional density estimation

More information

A note on inconsistent families of discrete multivariate distributions

A note on inconsistent families of discrete multivariate distributions Ghosh et al. Journal of Statistical Distriutions and Applications 2017 4:7 DOI 10.1186/s40488-017-0061-8 RESEARCH Open Access A note on inconsistent families of discrete multivariate distriutions Sugata

More information

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram

More information

1. Define the following terms (1 point each): alternative hypothesis

1. Define the following terms (1 point each): alternative hypothesis 1 1. Define the following terms (1 point each): alternative hypothesis One of three hypotheses indicating that the parameter is not zero; one states the parameter is not equal to zero, one states the parameter

More information

Smooth simultaneous confidence bands for cumulative distribution functions

Smooth simultaneous confidence bands for cumulative distribution functions Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

Polynomial Degree and Finite Differences

Polynomial Degree and Finite Differences CONDENSED LESSON 7.1 Polynomial Degree and Finite Differences In this lesson, you Learn the terminology associated with polynomials Use the finite differences method to determine the degree of a polynomial

More information

Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC)

Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Authors: Alexei C. Ionan, Mei-Yin C. Polley, Lisa M. McShane, Kevin K. Doin Section Page

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

Spiking problem in monotone regression : penalized residual sum of squares

Spiking problem in monotone regression : penalized residual sum of squares Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where

More information

Random fraction of a biased sample

Random fraction of a biased sample Random fraction of a biased sample old models and a new one Statistics Seminar Salatiga Geurt Jongbloed TU Delft & EURANDOM work in progress with Kimberly McGarrity and Jilt Sietsma September 2, 212 1

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Does k-th Moment Exist?

Does k-th Moment Exist? Does k-th Moment Exist? Hitomi, K. 1 and Y. Nishiyama 2 1 Kyoto Institute of Technology, Japan 2 Institute of Economic Research, Kyoto University, Japan Email: hitomi@kit.ac.jp Keywords: Existence of moments,

More information

The variance for partial match retrievals in k-dimensional bucket digital trees

The variance for partial match retrievals in k-dimensional bucket digital trees The variance for partial match retrievals in k-dimensional ucket digital trees Michael FUCHS Department of Applied Mathematics National Chiao Tung University January 12, 21 Astract The variance of partial

More information

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis A Combined Adaptive-Mixtures/Plug-In Estimator of Multivariate Probability Densities 1 J. Cwik and J. Koronacki Institute of Computer Science, Polish Academy of Sciences Ordona 21, 01-237 Warsaw, Poland

More information

Bandwidth selection for kernel conditional density

Bandwidth selection for kernel conditional density Bandwidth selection for kernel conditional density estimation David M Bashtannyk and Rob J Hyndman 1 10 August 1999 Abstract: We consider bandwidth selection for the kernel estimator of conditional density

More information

Expansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2

Expansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2 Least squares: Mathematical theory Below we provide the "vector space" formulation, and solution, of the least squares prolem. While not strictly necessary until we ring in the machinery of matrix algera,

More information

Lecture 6 January 15, 2014

Lecture 6 January 15, 2014 Advanced Graph Algorithms Jan-Apr 2014 Lecture 6 January 15, 2014 Lecturer: Saket Sourah Scrie: Prafullkumar P Tale 1 Overview In the last lecture we defined simple tree decomposition and stated that for

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

Smooth functions and local extreme values

Smooth functions and local extreme values Smooth functions and local extreme values A. Kovac 1 Department of Mathematics University of Bristol Abstract Given a sample of n observations y 1,..., y n at time points t 1,..., t n we consider the problem

More information

SPECIFICATION TESTING FOR ERRORS-IN-VARIABLES MODELS

SPECIFICATION TESTING FOR ERRORS-IN-VARIABLES MODELS SPECIFICATION TESTING FOR ERRORS-IN-VARIABLES MODELS TAISUKE OTSU AND LUKE TAYLOR Astract. This paper considers specification testing for regression models with errors-invariales and proposes a test statistic

More information

Essential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March.

Essential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March. Essential Maths 1 The information in this document is the minimum assumed knowledge for students undertaking the Macquarie University Masters of Applied Finance, Graduate Diploma of Applied Finance, and

More information

INTRODUCTION. 2. Characteristic curve of rain intensities. 1. Material and methods

INTRODUCTION. 2. Characteristic curve of rain intensities. 1. Material and methods Determination of dates of eginning and end of the rainy season in the northern part of Madagascar from 1979 to 1989. IZANDJI OWOWA Landry Régis Martial*ˡ, RABEHARISOA Jean Marc*, RAKOTOVAO Niry Arinavalona*,

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA)

ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) General information Availale time: 2.5 hours (150 minutes).

More information

Asymmetric least squares estimation and testing

Asymmetric least squares estimation and testing Asymmetric least squares estimation and testing Whitney Newey and James Powell Princeton University and University of Wisconsin-Madison January 27, 2012 Outline ALS estimators Large sample properties Asymptotic

More information

A converse Gaussian Poincare-type inequality for convex functions

A converse Gaussian Poincare-type inequality for convex functions Statistics & Proaility Letters 44 999 28 290 www.elsevier.nl/locate/stapro A converse Gaussian Poincare-type inequality for convex functions S.G. Bokov a;, C. Houdre ; ;2 a Department of Mathematics, Syktyvkar

More information

Fixed-b Inference for Testing Structural Change in a Time Series Regression

Fixed-b Inference for Testing Structural Change in a Time Series Regression Fixed- Inference for esting Structural Change in a ime Series Regression Cheol-Keun Cho Michigan State University imothy J. Vogelsang Michigan State University August 29, 24 Astract his paper addresses

More information

Sulphur dioxide Number of factories

Sulphur dioxide Number of factories 18 Nonparametric smoothing Nonparametric smoothing or nonparametric function estimation grew enormously in the 1980s. The word `nonparametric' has nothing to do with the classical rank-ased methods, such

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE

A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE BRAC University Journal, vol. V1, no. 1, 2009, pp. 59-68 A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE Daniel F. Froelich Minnesota State University, Mankato, USA and Mezbahur

More information

Nonparametric Identication of a Binary Random Factor in Cross Section Data and

Nonparametric Identication of a Binary Random Factor in Cross Section Data and . Nonparametric Identication of a Binary Random Factor in Cross Section Data and Returns to Lying? Identifying the Effects of Misreporting When the Truth is Unobserved Arthur Lewbel Boston College This

More information

Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E.

Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E. Forecasting Lecture 3 Structural Breaks Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, 2013 1 / 91 Bruce E. Hansen Organization Detection

More information

" $ CALCULUS 2 WORKSHEET #21. t, y = t + 1. are A) x = 0, y = 0 B) x = 0 only C) x = 1, y = 0 D) x = 1 only E) x= 0, y = 1

 $ CALCULUS 2 WORKSHEET #21. t, y = t + 1. are A) x = 0, y = 0 B) x = 0 only C) x = 1, y = 0 D) x = 1 only E) x= 0, y = 1 CALCULUS 2 WORKSHEET #2. The asymptotes of the graph of the parametric equations x = t t, y = t + are A) x = 0, y = 0 B) x = 0 only C) x =, y = 0 D) x = only E) x= 0, y = 2. What are the coordinates of

More information

Local linear multivariate. regression with variable. bandwidth in the presence of. heteroscedasticity

Local linear multivariate. regression with variable. bandwidth in the presence of. heteroscedasticity Model ISSN 1440-771X Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Local linear multivariate regression with variable bandwidth in the presence

More information

NCoVaR Granger Causality

NCoVaR Granger Causality NCoVaR Granger Causality Cees Diks 1 Marcin Wolski 2 1 Universiteit van Amsterdam 2 European Investment Bank Bank of Italy Rome, 26 January 2018 The opinions expressed herein are those of the authors and

More information

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Efficient estimation of a semiparametric dynamic copula model

Efficient estimation of a semiparametric dynamic copula model Efficient estimation of a semiparametric dynamic copula model Christian Hafner Olga Reznikova Institute of Statistics Université catholique de Louvain Louvain-la-Neuve, Blgium 30 January 2009 Young Researchers

More information

Depth versus Breadth in Convolutional Polar Codes

Depth versus Breadth in Convolutional Polar Codes Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK

More information

Nonparametric estimation and symmetry tests for

Nonparametric estimation and symmetry tests for Nonparametric estimation and symmetry tests for conditional density functions Rob J Hyndman 1 and Qiwei Yao 2 6 July 2001 Abstract: We suggest two improved methods for conditional density estimation. The

More information

Bayesian inference with reliability methods without knowing the maximum of the likelihood function

Bayesian inference with reliability methods without knowing the maximum of the likelihood function Bayesian inference with reliaility methods without knowing the maximum of the likelihood function Wolfgang Betz a,, James L. Beck, Iason Papaioannou a, Daniel Strau a a Engineering Risk Analysis Group,

More information

Spectrum Opportunity Detection with Weak and Correlated Signals

Spectrum Opportunity Detection with Weak and Correlated Signals Specum Opportunity Detection with Weak and Correlated Signals Yao Xie Department of Elecical and Computer Engineering Duke University orth Carolina 775 Email: yaoxie@dukeedu David Siegmund Department of

More information

Data-Based Choice of Histogram Bin Width. M. P. Wand. Australian Graduate School of Management. University of New South Wales.

Data-Based Choice of Histogram Bin Width. M. P. Wand. Australian Graduate School of Management. University of New South Wales. Data-Based Choice of Histogram Bin Width M. P. Wand Australian Graduate School of Management University of New South Wales 13th May, 199 Abstract The most important parameter of a histogram is the bin

More information