Nonparametric Estimation of Smooth Conditional Distributions
|
|
- Shona O’Neal’
- 6 years ago
- Views:
Transcription
1 Nonparametric Estimation of Smooth Conditional Distriutions Bruce E. Hansen University of Wisconsin May 24 Astract This paper considers nonparametric estimation of smooth conditional distriution functions (CDFs) using kernel smoothing methods. We propose estimation using a new smoothed local linear (SLL) estimator. Estimation ias is reduced through the use of a local linear estimator rather than local averaging. Estimation variance is reduced through the use of smoothing. Asymptotic analysis of mean integrated squared error (MISE) reveals the form of these efficiency gains, and their magnitudes are demonstrated in numerical simulations. Considerale attention is devoted to the development of a plugin rule for andwidths which minimize estimates of the asymptotic MISE. We illustrate the estimation method with an application to the U.S. quarterly GDP growth rate. Research supported y the National Science Foundation. Department of Economics, 118 Oservatory Drive, University of Wisconsin, Madison, WI 5376
2 1 Introduction This paper considers nonparametric estimation of smooth conditional distriution functions (CDFs). As the CDF is the conditional expectation of an indicator function, it may e estimated y nonparametric regression estimation methods. A recent paper which exploits advances in nonparametric regression for estimation of the CDF is Hall, Wolff and Yao (1999). See also Fan and Yao (23). Standard nonparametric regression estimates of the CDF, however, are not smooth and therefore are inefficient. If the dependent variale in these regressions the indicator function is replaced y a smooth function, estimation variance can e reduced at the cost of an increase in ias. If the degree of smoothing is carefully chosen, it can result in a lowered finite sample mean integrated squared error (MISE). The advantages of smoothed estimation of distriution functions has een discussed in the statistics literature. For references see Hansen (24). The extension to CDF estimation considered here is new. Our analysis concerns a pair of random variales (Y,) R R that have a smooth joint distriution. Let f(y) and g(x) denote the marginal densities of Y and, and let f (y x) denote the conditional density of Y given x. The conditional distriution function is Define as well the derivatives F (y x) Z y f (u x) du. F (s) (y x) s F (y x) xs f (s) (y x) s f (y x) xs f (y x) f (y x). y Given a random sample {Y 1, 1,..., Y n, n } from this distriution, the goal is to estimate F (y x) at a fixed value x. To simplify the notation we will sometimes suppress dependence on x. Let Φ(s) and φ(s) denote the standard normal distriution and density functions and let Φ σ (s) Φ(s/σ) and φ σ (s) σ 1 φ(s/σ). Let φ (m) (s) (d m /ds m ) φ(s) and φ σ (m) (s) (d m /ds m ) φ σ (s) σ (m+1) φ (m) (s/σ) denote the derivatives of φ(s) and φ σ (s). In particular, φ (1) (s) sφ(s) and φ (2) (s) s 2 1 φ(s). Our estimators will make use of two univariate kernel functions w(s) and k(s). We assume that oth are symmetric and second-order, are normalized to have a unit variance (e.g. R s2 w(s)ds 1), and have a finite sixth moment (e.g. R s6 w(s)ds < ). Let R R w(s)2 ds denote the roughness of w. Let Z s K(s) k(u)du 1
3 denote the integrated kernel of k and define the constant ψ 2 sk(s)k(s)ds >. (1) For our applications, we set oth k(s) and w(s) to equal the normal kernel φ(s), in which case R 1/2 π, ψ 1/ π and K(s) Φ(s). Section 2 presents unsmoothed estimators. Section 3 introduces our smoothed estimators. Section 4 develops plug-in andwidth rules for the local linear and smoothed local linear estimator. Section 5 presents a numerical simulation of performance. Section 6 is a simple application to the U.S. quarterly real GDP growth rate. Section 7 concludes. The proof of Theorem 1 is presented in the Appendix. Gauss code which implements the estimators and andwidth methods is availale on the author s wepage 2 Unsmoothed Estimators Note that F (y x) E (1 (Y i y) i x), andthusitisaregressionfunction. ThustheCDFmaye estimated y regression of 1 (Y i y) on i. A simple nonparametric estimator is the Nadaraya-Watson (NW) estimator, and takes the form of a local average ˆF (y x) P n i1 w i1 (Y i y) P n i1 w i (2) where w i 1 µ x w i (3) are kernel weights and is a andwidth. Reduced ias may e achieved using a local linear (LL) estimator. Local linear estimation is a special case of local polynomial regression, and is typically recommended for estimation mean. For a general treatment see Fan and Gijels (1996). The estimator is the intercept from a weighted least-squares regression of 1 (Y i y) on (x i ) using the weights w i, and can e written as P n i1 F (y x) w i 1 (Y i y) P n i1 w i wi w i ³1 ˆβ (x i ) Ã n ˆβ i1 w i (x i ) 2! 1 Ã n i1! w i (x i ). (4) (5) AsshowninTheorem1() of Hall, Wolff and Yao (1999), if cn 1/5 then ³ 2 R (F (y x)(1 F (y x))) E F (y x) F (y x) ' + 4 g(x)n 4 F (2) (y x) 2 + O ³n 6/5. 2
4 Integrating over y we otain the MISE ³ 2 RV E F (y x) F (y x) dy gn + 4 V 1 + O ³n 6/5 4 (6) where g g(x) V V 1 F (y x)(1 F (y x)) dy F (2) (y x) 2 dy. The first term on the right side of (6) is the integrated asymptotic variance, and the second is the integrated squared asymptotic ias. This ias term is simplified relative to the NW estimator, and this is the source of the improved performance of F over ˆF. Hall, Wolff and Yao (1999) point out that the local linear estimator F has two undesirale properties. It may e nonmonotonic in y, and is not constrained to lie in [, 1]. Thus the estimator F is not a CDF. The prolem occurs when the weights (5) are negative, which occurs when ˆβ (x i ) > 1. A simple solution 1 is to modify wi to exclude negative values. Thus we redefine the weights as follows ( wi ˆβ (x i ) > 1 w i ³1 ˆβ (7) (x i ) ˆβ (x i ) 1 This modification is asymptotically negligile, ut constrains F to e a valid CDF. Using the modified weights (7) our estimator is a slight modification of the local linear estimator, however we will continue to refer to it as the local linear (LL) estimator for simplicity. 3 Smoothed Estimators For some andwidth h> let K h (s) K(s/h) and consider the integrated kernel smooth K h (y Y i ). For small h it can e viewed as a smooth approximation to the indicator function 1 (Y i y). Smoothed CDF estimators replace the indicator function 1 (Y i y) in (2) and (4) with K h (y Y i ), yielding the smoothed Nadaraya-Watson (SNW) estimator ˆF h, (y x) P n i1 w ik h (y Y i ) P n i1 w i 1 Hall, Wolff and Yao (1999) proposed an alternative solution, introducing weights to the Nadaraya-Watson estimator which are selected y an empirical likelihood criterion. Their estimator is asymptotically equivalent to ours, ut computationally more complicated. 3
5 and the smoothed local linear (SLL) estimator F h, (y x) P n i1 w i K h (y Y i ) P n. i1 w i The weights w i and w i are definedin(3)and(7),respectively. Theorem 1 Assume that F (y x) and g(x) are continuously differentiale up to the fourth order in oth y and x. If cn 1/5 and h O() as n, then ³ 2 R E F h, (y x) F (y x) dy gn (V V 1 hψ)+4 h2 2 V 2 + h4 V 3 + O ³n 6/ (8) where g g(x) V V 1 V 2 V 3 F (y x)(1 F (y x)) dy F (2) (y x) 2 dy f (y x) f (2) (y x) dy f (y x) 2 dy. The MISE in (8) is a generalization of (6) to allow for h. The MISE is strictly decreasing at h (since ψ > ) demonstrating that the asymptotic MISE is minimized for strictly positive h. Thus there are (at least asymptotic) efficiency gains from smoothing. 4 Plug-In Bandwidth Selection We suggest selecting the andwidths (h, ) y the plug-in method. Plug-in andwidths are the values which minimize an estimate of the asymptotic MISE (8). Estimation of the latter requires estimation of the parameters g, V, V 1,V 2, and V 2. In this section we discuss estimation of these components, and implementation of the plug-in andwidth rule. Readers who are interested in the plug-in algorithm ut not the motivational details can jump directly to Section 4.5, which summarizes the implementation details. 4.1 Kernel Estimation of g The normal kernel density estimator is ĝ 1 n n φ r (x i ) (9) j1 4
6 where r is a andwidth. An estimate of the asymptotic MSE-minimizing choice for r is ˆr 2 π ĝ s (x) ³ ĝ (2) t (x) 2 n 1/5 (1) where ĝ s (x) 1 n ĝ (2) t (x) 1 n n φ s (x i ) j1 n j1 φ (2) t (x i ) and s and t are preliminary andwidths. Letting ˆσ x denote the sample standard deviation for i, we suggest the Gaussian reference rules s 1.6ˆσ x n 1/5 and t.94ˆσ x n 1/9, which are MISE-optimal for estimation of the density and its second derivative when g is Gaussian. 4.2 Estimation of V The parameter V is a function of the conditional distriution F (y x). A parametric estimator of the latter is the intercept from a p th order polynomial regression of 1 (Y i y) on ( i x). For p let 1 1 x ( 1 x) p p..., (11) 1 n x ( n x) p and let 1 (Y y) denote the stacked 1 (Y i y). Then the estimator can e written as ˆF (y x) δ 1 p p p 1 (Y y) (12) where δ 1 (1 ) is the first unit vector. Given (12), an estimate of V is ³ ˆV ˆF (y x) 1 ˆF (y x) dy Z δ 1 1 p p p 1 (Y >y) 1 (Y y) dy p 1 p p δ1 δ 1 p p p Y Y + p p p 1 δ1 (13) where (a) + a1 (a ) for scalar a, and is an element-y-element operation for matrices. A nonparametric estimator of V can e ased on a local linear regression. Given a andwidth q, let w i φ q (x i ) denote Gaussian kernel weights and set W diag{w i }. Then the estimator is F (y x) δ 1 1 W 1 1 W 1 (Y y) 5
7 and the estimate of V is Ṽ δ 1 1 W 1 1 W Y Y + W 1 1 W 1 1 δ1. (14) The andwidth q which minimizes the asymptotic mean integrated squared error of F (y x) is µ V q 2 πgv 1 n which depends on the unknowns g, V and V 1. (See equation (8) with h.) Using the estimates ĝ from (9), ˆV from (13), and ˆV 1 from (17) elow, we otain a plug-in andwidth ˆq.78 Ã ˆV ĝ ˆV 1 n 1/5! 1/5. (15) 4.3 Estimation of V 1 The parameter V 1 is a function of the second derivative F (2) (y x) whichcaneestimatedypolynomial regression. For some p 2 let p e defined as in (11). The estimator is ˆF (2) (y x) δ 3 1 p p p 1 (Y y) (16) where δ 3 is the third unit vector (a one in the third place, and zeros elsewhere). To calculate V 2 from (16), oserve that for any c>max Y i Z c 1 (Y y) 1 (Y y) dy Dc max Y,Y where D is an n n matrix of ones, and oserve that δ 3 p p p D. Thus Z c Z ˆF (2) (y x) 2 dy δ c 3 1 p p p 1 (Y y) 1 (Y y) dy p 1 p p δ3 δ 3 p p p Dc max Y,Y p p p 1 δ3 δ 3 1 p p p max Y,Y p 1 p p δ3 where the max operator is element-y-element. Note that this expression is invariant to c>max Y i, thus we define ˆV 1 δ 3 1 p p p max Y,Y p 1 p p δ3. (17) This estimator was also used in (15) to construct a plug-in rule for the nonparametric estimate Ṽ. A nonparametric estimator of V 1 can e ased on a local cuic regression. (While any local polynomial regression of order two or aove is consistent, local cuic regression is recommend for estimation of a regression second derivative.) Given a andwidth q 1, let w 1i φ q1 (x i ) denote Gaussian kernel 6
8 weights and set W 1 diag{w 1i }. The estimator of F (2) (y x) is F (2) (y x) δ 3 3 W W 1 1 (Y y) (18) and the estimate of V 1 is Ṽ 1 δ 3 3 W W 1 max Y,Y W W δ3. (19) Using Fan and Gijels (1996, Theorem 3.1 and equation (3.2)) the andwidth which minimizes the asymptotic mean integrated squared error of ˆF (2) (y x) is where The constant V 4 may e estimated y µ V 1/9 q gv 4 n V 4 F (4) (y x) 2 dy. ˆV 4 δ 5 p p p max Y,Y p p p 1 δ5 (2) where p 4 and δ 5 is the fifth unit vector, suggesting the plug-in andwidth ˆq Ã! 1/9 ˆV. (21) ĝ ˆV 4 n 4.4 Estimation of V 2 and V 3 If f (y x) does not depend on x then we have the simplification V 3 R f (y) 2 dy which is a wellstudied estimation prolem. For a andwidth a let e the n 1 vector of ones and ˆf a (y) 1 n n j1 φ (1) a (y Y i)φ (1) a y Y 1 e a Gaussian kernel estimate of f (y). An estimate of V 3 in this case is ˆf a (y)2 dy φ (2) a φ (1) a (y Y ) φ(1) y Y dy 1 a Y Y 1 (22) where a 2a. The second equality is equation (7.1) ofmarronandwand(1992): φ (r) a (y Y i ) φ (s) a (y Y j ) dy ( 1) r φ (r+s) 2a (Y i Y j ). (23) 7
9 The estimator (22) is the Jones and Sheather (1991) estimator. Hansen (24) descries a multiple plug-in method for selection of a. We recommend this choice and hold it fixed for the remainder of this su-section and the next. Now consider estimation of V 3 allowing for dependence on x. Note that the estimate of f (y) in (24) is a regression of φ (1) a (y Y i) on a constant. Similarly, we can estimate f (y x) as the intercept in a polynomial regression on x i x. For p definedin(11) let The associated estimate of V 3 is ˆV 3 ˆf a (y x)2 dy Z δ 1 1 p p p δ 1 ˆf a (y x) δ 1 1 p p p φ (1) a (y Y ). (24) p p p φ (2) a φ (1) a (y Y ) φ(1) y Y dy p 1 p p δ1 a Y Y p p p 1 δ1. (25) Alternatively, we could use a fully nonparametric estimator such as a local linear estimator. However, the optimal andwidths depend on derivatives of the conditional density that are difficult to estimate. We thus do not consider further complications. In addition, using an analog of (16) and (24), a simple estimator of V 2 is ˆV 2 ˆf (2) a (y x) ˆf a (y) dy Z δ 3 1 p p p φ a (y Y ) φ a y Y dy p 1 p p δ1 where the third equality uses (23). δ 3 p p p φ a Y Y p p p 1 δ1. (26) 4.5 Plug-In Bandwidth Our plug-in estimate of (8) is M (h, ) R ³Ṽ hψ + 4 Ṽ 1 2 h2 2 ˆV 2 + h4 ˆV 3 ĝn (27) The plug-in andwidth pair ( h, ) jointly minimize M (h, ). A closed-form solution is not availale, so the function needs to e numerically minimized. One caveat is that we must constrain h<ψ/ṽ to ensure a sensile solution. (When h>ψ/ṽ then M is strictly increasing in so unconstrained minimization sets which is not sensile.) As a practical matter it makes sense to ound h more sustantially, we suggest the constraint h ψ/2ṽ,andthisisdoneinournumericalapplications. The plug-in andwidth for the unsmoothed estimator F is found y setting h and minimizing 8
10 to find ˆ ³ RṼ/ĝṼ 1 n 1/6. To summarize, the computation algorithm is as follows. First, ĝ is calculated from (9) using (1). Second, for some p 4, ˆV, ˆV 1 and ˆV 4 are calculated using (13), (17) and (2). Third, using the andwidth (15), Ṽ is calculated using (14). Fourth, using the andwidth (21), Ṽ 1 is calculated using (19). Fifth, the andwidth a is calculated using the method of Hansen (24), allowing ˆV 2 and ˆV 3 to e calculated using (26) and (25). Finally, ĝ, Ṽ,Ṽ 1, ˆV 2, ˆV 3 are set into (27) which is numerically minimized over (h, ) to otain ( h, ). 5 Simulation The performance of the CDF estimators is investigated in a simple simulation experiment using ivariate data. The variales i are generated as iid N(, 1). The variale Y i is generated as Y i i + e i where e i is iid, independent of i. The error e i is distriuted as G, one of the first nine mixture-normal distriutions from Marron and Wand (1992). Consequently, the true CDF for Y i is G(y x). Foreach choice of G we generated 1, independent samples of size n 5, 1 and 3. ForeachsampleweestimatedtheCDFatx 1 and y { 2., 1.5, 1.,.5,.,.5, 1., 1.5, 2.}, using the NW, SNW, LL and SLL estimators. Our measure of precision is the exact mean-squared error, averaged over these values of y (as an approximation to the MISE). The standard normal kernel was used for all estimates. We first astracted from the issue of andwidth selection and evaluated the precision of the four estimates using the infeasile finite-sample-optimal andwidth, calculated y quasi-newton minimization. The results are presented in the first three columns of Tale 1. Wehavetakentheunsmoothed NW estimator as a aseline, and have divided the MISE of the other estimators y that of the NW estimator. Thus the numers in Tale 1 represent the relative improvement attained relative to the NW estimator. The results show that across models and sample sizes, the SLL estimator achieves the lowest MISE. In each case, using local linear estimation rather than local averaging reduces the MISE, and using smoothing lowers the MISE. We next investigated the performance of our plug-in andwidths ˆ for the LL estimator and ( h, ) for the SLL estimator. On each sample descried aove we calculated these plug-in andwidths and CDF estimators, and evaluated the MISE across the 1, samples. The results are presented in the final two columns of Tale 1. As efore, the MISE is normalized y the MISE of the infeasile NW estimator. Thus the numers represented the relative improvement attained y the feasile plug-in estimators relative to this infeasile estimator. In each case, the plug-in SLL estimator has lower MISE than the plug-in LL estimator. In some cases the difference is minor, in other case it is quite sustantial (the largest gains are for Model #5, where the MISE is reduced y aout 25%). In each case the MISE is higher than if the infeasile optimal andwidth had een used, in some cases sustantially higher, suggesting that further improvements in andwidth choice may e possile. 9
11 Overall, the simulation evidence confirms the implications of the asymptotic analysis. We can recommend use of the SLL estimator (with our proposed plug-in andwidth) for empirical practice. 6 GDP Growth Forecasting We illustrate the method with a simple application to the U.S. real GDP growth rate. Let GDP i denote the level of quarterly real GDP and let y i 4(ln(GDP i ) ln (GDP i 1 )) denote the annualized quarterly growth rate. Set x i y i 1. We restrict the sample to run from first third quarter of 1984 to the first quarter of 24 (79 oservations). This choice is consistent with the decline in output volatility documented y McConnell and Perez-Quiros (2). We estimated the CDF using the local linear (LL) and smoothed local linear (SLL) estimators for y i given x i x, fixing x to equal 2% and 4.5%, which are approximately the 25 th and 75 th quantiles of the unconditional distriution. The CDF was evaluated at 51 evenly spaced increments etween -2 and 7, and the andwidths were calculated y the plug-in method descried in Section 4.5. To provide a contrast, we also estimated the CDF using a homoskedastic Gaussian AR(1). Figure 1 displays the three CDFs for the case of x 2. and Figure 2 for the case x 4.5. In oth cases, the three estimates are meaningfully different from one another. The parametric estimate has a fairly different location and variance. The two nonparametric estimates have the same asic shape, ut the unsmoothed local linear estimate is visually quite erratic, and the smoothed local linear estimator appears to apply an intuitively correct degree of smoothing. It is also interesting to note that the LL estimator uses a plug-in andwidth of ˆ 1.59 in Figure 1 and ˆ 2.13 in Figure 2, while the SLL estimator uses 1.44, h.53 and 2. and h.23 respectively. In oth cases, the plug-in rule selects similar values for for the two estimators (with that for SLL slightly smaller), and selects a consideraly smaller value for h than. 7 Conclusion We have shown how to comine local linear methods and smoothing to produce good nonparametric estimates of the conditional distriution function for ivariate data. We have shown that meaningful improvements in estimation efficiency can e achieved y using these techniques. Furthermore, we have derived data-ased plug-in rules for andwidth selection which appear to work in practice. These tools should prove useful for empirical application. This analysis has een confined to ivariate data. An important extension would to the case of multivariate data. In particular, this extension would require the study and development of feasile andwidth selection rules. 1
12 8 Appendix: Proof of Theorem 1 The proof of Theorem 1 is ased on the following set of expansions. Lemma 1 Let g (s) ds dx s g(x) and K h (Y i)k h (y Y i ) F (y x). Ew i g(x)+ 2 2 g(2) (x)+o( 4 ) (28) E (w i (x i )) 2 g (1) (x)+o( 4 ) (29) E ³w 2 i (x i ) 2 g(x)+o( 4 ) (3) E (w i Kh(Y i )) 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x) (31) + h2 2 f (y x) g(x)+o 4 E (w i (x i ) Kh(Y i )) 2 F (1) (y x) g(x)+o 4 (32) E w 2 i K h(y i ) 2 Rg(x)(F (y x)(1 F (y x)) hψf (y x)) +O () (33) E ³w 2 i 2 (x i ) 2 Kh(Y i ) O() (34) Proof. Throughout, we make use of the assumption that h O() to simplify the ounds. The derivations make use of the following Taylor series expansions which are valid under the assumption that the functions are fourth order differentiale: g(x v) g(x) vg (1) (x)+ v2 2 2 g(2) (x) v3 3 6 g(3) (x)+o( 4 ) (35) F (y hu v) F (y v) huf (y v)+ h2 u 2 2 f (y v) h3 u 3 6 f (y v)+o(h 4 ) (36) F (y x v) F (y x) vf (1) (y x)+ 2 v 2 2 F (2) (y x) 3 v 3 6 F (3) (y x)+o( 4 ) (37) f (y x v) f (y x) v y f (1) (y x)+o( 2 ). (38) To show (28), y a change of variales and (35) Ew i 1 w µ x v g(v)dv w (v) g(x v)dv w (v) µg(x) vg (1) (x)+ v2 2 2 g(2) (x) v3 3 6 g(3) (x) dv + O( 4 ) g(x)+ 2 2 g(2) (x)+o( 4 ). 11
13 Note that this makes use of the facts that R w (v) dy 1, R w (v) v, R w (v) v2 1 and R w (v) v3. Similarly E (w i (x i )) (x v) 1 µ x v w vw (v) g(x v)dv g(v)dv vw (v) µg(x) vg (1) (x)+ v2 2 2 g(2) (x) dv + O( 4 ) 2 g (1) (x)+o( 4 ) yielding (29) and E ³w 2 i (x i ) (x v) 2 1 w 2 2 g(x)+o( 4 ) µ x v v 2 w (v) g (x v) dv g(v)dv yielding (3). To show (31) and(32),first oserve that y integration y parts, a change of variales and (36) K h (y u) f (u v) du +O(h 4 ) k h (y u) F (u v) du k (u) F (y hu v) du k (u) µf (y v) huf (y v)+ h2 u 2 2 f (y v) h3 u 3 6 f (y v) du F (y v)+ h2 2 f (y v)+o(h 4 ). (39) 12
14 By a change of variales, and using (35) and (37), F (y v) 1 µ x v w g(v)dv F (y x v) g(x v)w (v) dv µf (y x) vf (1) (y x)+ 2 v 2 2 F (2) (y x) 3 v 3. µg(x) vg (1) (x)+ 2 v 2 2 g(2) (x) v3 3 6 g(3) (x) w(v)+o( 4 ) 6 F (3) (y x) F (y x) g(x)+ 2 2 F (y x) g(2) (x)+ 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x)+o( 4 ). (4) Similarly, y a change of variales and (35) and (38), f (y v) 1 µ Z x v w g(v)dv f (y x v) g(x v)w(v)dv f (y x) g(x)+o( 2 ). (41) Together, (28), (39), (4) and (41) show that which is (31). E (w i Kh(Y i )) E (w i K h (Y i )) E (w i ) F (y x) µ 1 x v w K h (y u) f (u v) g(v)dudv µ g(x)+ 2 2 g(2) (x) F (y x)+o( 4 ) F (y v) 1 µ µ x v w g(v)dv g(x)+ 2 2 g(2) (x) F (y x) g(v)dv + O( 4 ) + h2 2 f (y v) 1 w µ x v h2 2 f (y x) g(x)+ 2 2 F (2) (y x) g(x)+ 2 F (1) (y x) g (1) (x) +O 4 13
15 Similarly, using (29), (39) and similar derivations, E (w i (x i ) Kh(Y i )) E (w i (x i ) K h (Y i )) E (w i (x i )) F (y x) (x v) 1 µ x v w K h (y u) f (u v) g(v)dvdu + 2 g (1) (x)f (y x)+o( 4 ) vf (y x v) g(x v)w (v) dv + h2 2 vf (y x v) g(x v)w (v) dv + 2 g (1) (x)f (y x)+o( 4 ) 2 F (1) (y x) g(x)+o 4 which is (32). To show (33) first oserve that y a change of variales and (35), E Z wi 2 µ 1 x v 2 2 w g (v) dv 1 w (v) 2 g (x v) dv 1 ³ w (v) 2 g(x) vg (1) (x) dv + O() Rg(x) + O(). (42) Second, using (39), a change of variales, (35) and (37), E wi 2 K h (Y i ) µ 1 x v 2 w K h (y u) f (u v) g(v)dvdu µ 1 x v 2 µ h 2 2 w g(v)f (y v) dv + O 1 1 w (v) 2 g(x v)f (y x v) dv + O () ³ ³ w (v) 2 g(x) vg (1) (x) F (y x) vf (1) (y x) dv + O( 2 ) Rg(x)F (y x) + O(). (43) 14
16 Third, y integration y parts, a change of variales, and (36) K h (y u) 2 f (u v) du 2 K h (y u) k h (y u) F (u v) du 2 2 K (u) k (u) F (y hu v) du K (u) k (u)(f (y v) huf (y v)) du + O(h 2 ) F (y v) hψf (y v)+o(h 2 ). (44) The final equality uses the definition (1) and the fact that the symmetry of k(u) aout zero implies 2K (u) k (u) du. Fourth, using (44), (35), (37) and (38), E wi 2 K h (Y i ) 2 µ 1 x v 2 2 w K h (y u) 2 f (u v) g(v)dvdu µ 1 x v 2 µ h 2 2 w (F (y v) hψf (y v)) g(v)dv + O 1 w (v) 2 (F (y x v) hψf (y x v)) g(x v)dv + O () Rg(x)(F (y x) hψf (y x)) + O() (45) Together, (42), (43) and (45) show E wi 2 Kh(Y i ) 2 E wi 2 K h (Y i ) 2 2E wi 2 K h (Y i ) F (y x)+e wi 2 F (y x) 2 Rg(x)(F (y x) hψf (y x)) 2Rg(x)F (y x)2 + Rg(x)F (y x) 2 + O () Rg(x)(F (y x)(1 F (y x)) hψf (y x)) + O () which is (33). (34) is shown similarly. ProofofTheorem1:From equations (29)-(3), Eˆβ x ³ E ³w i (x i ) 2 1 (E (wi (x i ))) + O 2 g(x) 1 g (1) (x)+o 2. 15
17 Comined with (28), this yields Next, using equations (31) and(32), Together with (46) we find Ewi Ew i EˆβE (w i (x i )) + O( 2 ) (46) f(x)+o( 2 ). E (w i K h(y i )) E (w i K h(y i )) EˆβE (w i (x i ) K h(y i )) + O F (2) (y x) g(x)+ h2 2 f (y x) g(x)+o 4. ³ E F h, (y x) F (y x) Ew i K h (y Y i) + O Ew i ³ 2 F (2) (y x)+h 2 f (y x) + O 4. Similarly, using (33) and (34) Var³ F h, (y x) 1 Var(wi K h (y Y i)) n (Ewi )2 µ 1 E (w i Kh (y Y i)) 2 n g(x) 2 + O n R (F (y x)(1 F (y x)) hψf (y x)) g(x)n + O µ O ³n 6/5. n Together, these expressions and the assumption that cn 1/5 estalish that ³ 2 E F h, (y x) F (y x) R (F (y x)(1 F (y x)) ψhf (y x) f(x)) g(x)n + h2 2 f (y x) F (2) (y x) f (y x) F (2) (y x) h4 f (y x) O ³n 6/5. Integrating over y we otain ³ Z R ³ 2 R F (y x)(1 F (y x)) dy ψh R f (y x) dyf(x) E F h, (y x) F (y x) dy g(x)n + h4 f (y x) 2 dy + h2 2 f (y x) F (2) (y x) dy F (2) (y x) 2 dy + O ³n 6/5 4 16
18 which equals the expression in (8) using the equivalence f (y x) F (2) (y x) dy f (y x) f (2) (y x) dy which holds y integration y parts. This completes the proof. 17
19 Tale 1: Normalized MISE Conditional Distriution Function Estimators n 5 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal n 1 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal n 3 Optimal Bandwidth Plug-In Bandwidths Error Density SNW LL SLL LL SLL Gaussian Skewed Unimodal Strongly Skewed Kurtotic Unimodal Outlier Bimodal Separated Bimodal Asymmetric Bimodal Trimodal
20 References [1] Fan, Jianqing and Irene Gijels (1996): Local Polynomial Modelling and Its Applications: NewYork: Chapman and Hall. [2] Fan, Jianqing and Qiwei Yao (23): Nonlinear Time Series: Nonparametric and Parametric Methods. NewYork:Springer. [3] Hall, Peter, Rodney C. L. Wolff, andqiweiyao(1999): Methods for Estimating a Conditional Distriution Function, Journal of the American Statistical Association, 94, [4] Hansen, Bruce E. (24): Bandwidth Selection for Nonparametric Distriution Estimation, manuscript, University of Wisconsin. [5] Jones, M.C. and S. J. Sheather (1991): Using non-stochastic terms to advantage in kernel-ased estimation of integrated squared density derivatives, Statistics and Proaility Letters, 11, [6] Marron, J.S. and M.P. Wand (1992): Exact mean integrated squared error, Annals of Statistics, 2, [7] McConnell, M.M. and G. Perez-Quiros (2): Output fluctuations in the United States: What has changed since the early 198s? American Economic Review, 9,
21
Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement
Open Journal of Statistics, 07, 7, 834-848 http://www.scirp.org/journal/ojs ISS Online: 6-798 ISS Print: 6-78X Estimating a Finite Population ean under Random on-response in Two Stage Cluster Sampling
More informationEstimation of cumulative distribution function with spline functions
INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative
More informationDensity estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas
0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity
More informationTime Series and Forecasting Lecture 4 NonLinear Time Series
Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations
More informationEcon 582 Nonparametric Regression
Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume
More informationA Novel Nonparametric Density Estimator
A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with
More informationLocal Polynomial Regression
VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based
More informationI N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A)
I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S (I S B A) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2011/2 ESTIMATION
More informationO Combining cross-validation and plug-in methods - for kernel density bandwidth selection O
O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem
More informationAdaptive Kernel Estimation of The Hazard Rate Function
Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationNonparametric Regression. Changliang Zou
Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,
More informationNonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix
Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract
More informationNonparametric Econometrics
Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-
More informationA nonparametric method of multi-step ahead forecasting in diffusion processes
A nonparametric method of multi-step ahead forecasting in diffusion processes Mariko Yamamura a, Isao Shoji b a School of Pharmacy, Kitasato University, Minato-ku, Tokyo, 108-8641, Japan. b Graduate School
More informationFinQuiz Notes
Reading 9 A time series is any series of data that varies over time e.g. the quarterly sales for a company during the past five years or daily returns of a security. When assumptions of the regression
More informationA Bootstrap Test for Conditional Symmetry
ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School
More informationAsymptotic distributions of nonparametric regression estimators for longitudinal or functional data
Journal of Multivariate Analysis 98 (2007) 40 56 www.elsevier.com/locate/jmva Asymptotic distriutions of nonparametric regression estimators for longitudinal or functional data Fang Yao Department of Statistics,
More informationDo Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods
Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Robert V. Breunig Centre for Economic Policy Research, Research School of Social Sciences and School of
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationMinimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.
Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of
More informationPreface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation
Preface Nonparametric econometrics has become one of the most important sub-fields in modern econometrics. The primary goal of this lecture note is to introduce various nonparametric and semiparametric
More informationLocal linear multiple regression with variable. bandwidth in the presence of heteroscedasticity
Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable
More informationEach element of this set is assigned a probability. There are three basic rules for probabilities:
XIV. BASICS OF ROBABILITY Somewhere out there is a set of all possile event (or all possile sequences of events which I call Ω. This is called a sample space. Out of this we consider susets of events which
More informationEVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS
APPLICATIONES MATHEMATICAE 9,3 (), pp. 85 95 Erhard Cramer (Oldenurg) Udo Kamps (Oldenurg) Tomasz Rychlik (Toruń) EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS Astract. We
More informationOn a Nonparametric Notion of Residual and its Applications
On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014
More informationSimple Examples. Let s look at a few simple examples of OI analysis.
Simple Examples Let s look at a few simple examples of OI analysis. Example 1: Consider a scalar prolem. We have one oservation y which is located at the analysis point. We also have a ackground estimate
More informationSpatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood
Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Kuangyu Wen & Ximing Wu Texas A&M University Info-Metrics Institute Conference: Recent Innovations in Info-Metrics October
More informationModel-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego
Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships
More informationNonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction
Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(
More informationNonparametric Regression
Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y
More informationReal option valuation for reserve capacity
Real option valuation for reserve capacity MORIARTY, JM; Palczewski, J doi:10.1016/j.ejor.2016.07.003 For additional information aout this pulication click this link. http://qmro.qmul.ac.uk/xmlui/handle/123456789/13838
More informationEstimation of the Error Density in a Semiparametric Transformation Model
Estimation of the Error Density in a Semiparametric Transformation Model Benjamin Colling Université catholique de Louvain Cédric Heuchenne University of Liège and Université catholique de Louvain Rawane
More informationIntroduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β
Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured
More informationNonparametric Methods
Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis
More informationChaos and Dynamical Systems
Chaos and Dynamical Systems y Megan Richards Astract: In this paper, we will discuss the notion of chaos. We will start y introducing certain mathematical concepts needed in the understanding of chaos,
More informationLocal Polynomial Modelling and Its Applications
Local Polynomial Modelling and Its Applications J. Fan Department of Statistics University of North Carolina Chapel Hill, USA and I. Gijbels Institute of Statistics Catholic University oflouvain Louvain-la-Neuve,
More informationNonparametric Modal Regression
Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric
More information13 Endogeneity and Nonparametric IV
13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,
More informationTeruko Takada Department of Economics, University of Illinois. Abstract
Nonparametric density estimation: A comparative study Teruko Takada Department of Economics, University of Illinois Abstract Motivated by finance applications, the objective of this paper is to assess
More informationAsymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands
Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department
More informationUniform Convergence Rates for Nonparametric Estimation
Uniform Convergence Rates for Nonparametric Estimation Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~bansen October 2004 Preliminary and Incomplete Abstract Tis paper presents a set of rate
More informationLuis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model
Luis Manuel Santana Gallego 100 Appendix 3 Clock Skew Model Xiaohong Jiang and Susumu Horiguchi [JIA-01] 1. Introduction The evolution of VLSI chips toward larger die sizes and faster clock speeds makes
More informationDiscussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon
Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics
More informationMinimizing a convex separable exponential function subject to linear equality constraint and bounded variables
Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad
More informationTransformation-based Nonparametric Estimation of Multivariate Densities
Transformation-based Nonparametric Estimation of Multivariate Densities Meng-Shiuh Chang Ximing Wu March 9, 2013 Abstract We propose a probability-integral-transformation-based estimator of multivariate
More informationNonparametric Density Estimation
Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted
More informationWEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract
Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of
More informationWeak bidders prefer first-price (sealed-bid) auctions. (This holds both ex-ante, and once the bidders have learned their types)
Econ 805 Advanced Micro Theory I Dan Quint Fall 2007 Lecture 9 Oct 4 2007 Last week, we egan relaxing the assumptions of the symmetric independent private values model. We examined private-value auctions
More informationWeb-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach
1 Web-based Supplementary Material for Dependence Calibration in Conditional Copulas: A Nonparametric Approach Elif F. Acar, Radu V. Craiu, and Fang Yao Web Appendix A: Technical Details The score and
More informationNonparametric confidence intervals. for receiver operating characteristic curves
Nonparametric confidence intervals for receiver operating characteristic curves Peter G. Hall 1, Rob J. Hyndman 2, and Yanan Fan 3 5 December 2003 Abstract: We study methods for constructing confidence
More informationDEPARTMENT OF ECONOMICS
ISSN 089-64 ISBN 978 0 7340 405 THE UNIVERSITY OF MELBOURNE DEPARTMENT OF ECONOMICS RESEARCH PAPER NUMBER 06 January 009 Notes on the Construction of Geometric Representations of Confidence Intervals of
More informationHIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO
The Annals of Statistics 2006, Vol. 34, No. 3, 1436 1462 DOI: 10.1214/009053606000000281 Institute of Mathematical Statistics, 2006 HIGH-DIMENSIONAL GRAPHS AND VARIABLE SELECTION WITH THE LASSO BY NICOLAI
More informationGaussian kernel GARCH models
Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often
More informationNonparametric Function Estimation with Infinite-Order Kernels
Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density
More informationLocal Polynomial Wavelet Regression with Missing at Random
Applied Mathematical Sciences, Vol. 6, 2012, no. 57, 2805-2819 Local Polynomial Wavelet Regression with Missing at Random Alsaidi M. Altaher School of Mathematical Sciences Universiti Sains Malaysia 11800
More informationConditional density estimation: an application to the Ecuadorian manufacturing sector. Abstract
Conditional density estimation: an application to the Ecuadorian manufacturing sector Kim Huynh Indiana University David Jacho-Chavez Indiana University Abstract This note applies conditional density estimation
More informationA note on inconsistent families of discrete multivariate distributions
Ghosh et al. Journal of Statistical Distriutions and Applications 2017 4:7 DOI 10.1186/s40488-017-0061-8 RESEARCH Open Access A note on inconsistent families of discrete multivariate distriutions Sugata
More informationECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd
ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram
More information1. Define the following terms (1 point each): alternative hypothesis
1 1. Define the following terms (1 point each): alternative hypothesis One of three hypotheses indicating that the parameter is not zero; one states the parameter is not equal to zero, one states the parameter
More informationSmooth simultaneous confidence bands for cumulative distribution functions
Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang
More informationVariance Function Estimation in Multivariate Nonparametric Regression
Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered
More informationPolynomial Degree and Finite Differences
CONDENSED LESSON 7.1 Polynomial Degree and Finite Differences In this lesson, you Learn the terminology associated with polynomials Use the finite differences method to determine the degree of a polynomial
More informationSupplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC)
Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC) Authors: Alexei C. Ionan, Mei-Yin C. Polley, Lisa M. McShane, Kevin K. Doin Section Page
More information41903: Introduction to Nonparametrics
41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific
More informationSpiking problem in monotone regression : penalized residual sum of squares
Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where
More informationRandom fraction of a biased sample
Random fraction of a biased sample old models and a new one Statistics Seminar Salatiga Geurt Jongbloed TU Delft & EURANDOM work in progress with Kimberly McGarrity and Jilt Sietsma September 2, 212 1
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationDoes k-th Moment Exist?
Does k-th Moment Exist? Hitomi, K. 1 and Y. Nishiyama 2 1 Kyoto Institute of Technology, Japan 2 Institute of Economic Research, Kyoto University, Japan Email: hitomi@kit.ac.jp Keywords: Existence of moments,
More informationThe variance for partial match retrievals in k-dimensional bucket digital trees
The variance for partial match retrievals in k-dimensional ucket digital trees Michael FUCHS Department of Applied Mathematics National Chiao Tung University January 12, 21 Astract The variance of partial
More informationJ. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis
A Combined Adaptive-Mixtures/Plug-In Estimator of Multivariate Probability Densities 1 J. Cwik and J. Koronacki Institute of Computer Science, Polish Academy of Sciences Ordona 21, 01-237 Warsaw, Poland
More informationBandwidth selection for kernel conditional density
Bandwidth selection for kernel conditional density estimation David M Bashtannyk and Rob J Hyndman 1 10 August 1999 Abstract: We consider bandwidth selection for the kernel estimator of conditional density
More informationExpansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2
Least squares: Mathematical theory Below we provide the "vector space" formulation, and solution, of the least squares prolem. While not strictly necessary until we ring in the machinery of matrix algera,
More informationLecture 6 January 15, 2014
Advanced Graph Algorithms Jan-Apr 2014 Lecture 6 January 15, 2014 Lecturer: Saket Sourah Scrie: Prafullkumar P Tale 1 Overview In the last lecture we defined simple tree decomposition and stated that for
More informationSome Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model
Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;
More informationSmooth functions and local extreme values
Smooth functions and local extreme values A. Kovac 1 Department of Mathematics University of Bristol Abstract Given a sample of n observations y 1,..., y n at time points t 1,..., t n we consider the problem
More informationSPECIFICATION TESTING FOR ERRORS-IN-VARIABLES MODELS
SPECIFICATION TESTING FOR ERRORS-IN-VARIABLES MODELS TAISUKE OTSU AND LUKE TAYLOR Astract. This paper considers specification testing for regression models with errors-invariales and proposes a test statistic
More informationEssential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March.
Essential Maths 1 The information in this document is the minimum assumed knowledge for students undertaking the Macquarie University Masters of Applied Finance, Graduate Diploma of Applied Finance, and
More informationINTRODUCTION. 2. Characteristic curve of rain intensities. 1. Material and methods
Determination of dates of eginning and end of the rainy season in the northern part of Madagascar from 1979 to 1989. IZANDJI OWOWA Landry Régis Martial*ˡ, RABEHARISOA Jean Marc*, RAKOTOVAO Niry Arinavalona*,
More informationModelling Non-linear and Non-stationary Time Series
Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September
More informationERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA)
ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) General information Availale time: 2.5 hours (150 minutes).
More informationAsymmetric least squares estimation and testing
Asymmetric least squares estimation and testing Whitney Newey and James Powell Princeton University and University of Wisconsin-Madison January 27, 2012 Outline ALS estimators Large sample properties Asymptotic
More informationA converse Gaussian Poincare-type inequality for convex functions
Statistics & Proaility Letters 44 999 28 290 www.elsevier.nl/locate/stapro A converse Gaussian Poincare-type inequality for convex functions S.G. Bokov a;, C. Houdre ; ;2 a Department of Mathematics, Syktyvkar
More informationFixed-b Inference for Testing Structural Change in a Time Series Regression
Fixed- Inference for esting Structural Change in a ime Series Regression Cheol-Keun Cho Michigan State University imothy J. Vogelsang Michigan State University August 29, 24 Astract his paper addresses
More informationSulphur dioxide Number of factories
18 Nonparametric smoothing Nonparametric smoothing or nonparametric function estimation grew enormously in the 1980s. The word `nonparametric' has nothing to do with the classical rank-ased methods, such
More informationECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications
ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty
More informationA NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE
BRAC University Journal, vol. V1, no. 1, 2009, pp. 59-68 A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE Daniel F. Froelich Minnesota State University, Mankato, USA and Mezbahur
More informationNonparametric Identication of a Binary Random Factor in Cross Section Data and
. Nonparametric Identication of a Binary Random Factor in Cross Section Data and Returns to Lying? Identifying the Effects of Misreporting When the Truth is Unobserved Arthur Lewbel Boston College This
More informationCentral Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E.
Forecasting Lecture 3 Structural Breaks Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, 2013 1 / 91 Bruce E. Hansen Organization Detection
More information" $ CALCULUS 2 WORKSHEET #21. t, y = t + 1. are A) x = 0, y = 0 B) x = 0 only C) x = 1, y = 0 D) x = 1 only E) x= 0, y = 1
CALCULUS 2 WORKSHEET #2. The asymptotes of the graph of the parametric equations x = t t, y = t + are A) x = 0, y = 0 B) x = 0 only C) x =, y = 0 D) x = only E) x= 0, y = 2. What are the coordinates of
More informationLocal linear multivariate. regression with variable. bandwidth in the presence of. heteroscedasticity
Model ISSN 1440-771X Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Local linear multivariate regression with variable bandwidth in the presence
More informationNCoVaR Granger Causality
NCoVaR Granger Causality Cees Diks 1 Marcin Wolski 2 1 Universiteit van Amsterdam 2 European Investment Bank Bank of Italy Rome, 26 January 2018 The opinions expressed herein are those of the authors and
More informationBoundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta
Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationEfficient estimation of a semiparametric dynamic copula model
Efficient estimation of a semiparametric dynamic copula model Christian Hafner Olga Reznikova Institute of Statistics Université catholique de Louvain Louvain-la-Neuve, Blgium 30 January 2009 Young Researchers
More informationDepth versus Breadth in Convolutional Polar Codes
Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK
More informationNonparametric estimation and symmetry tests for
Nonparametric estimation and symmetry tests for conditional density functions Rob J Hyndman 1 and Qiwei Yao 2 6 July 2001 Abstract: We suggest two improved methods for conditional density estimation. The
More informationBayesian inference with reliability methods without knowing the maximum of the likelihood function
Bayesian inference with reliaility methods without knowing the maximum of the likelihood function Wolfgang Betz a,, James L. Beck, Iason Papaioannou a, Daniel Strau a a Engineering Risk Analysis Group,
More informationSpectrum Opportunity Detection with Weak and Correlated Signals
Specum Opportunity Detection with Weak and Correlated Signals Yao Xie Department of Elecical and Computer Engineering Duke University orth Carolina 775 Email: yaoxie@dukeedu David Siegmund Department of
More informationData-Based Choice of Histogram Bin Width. M. P. Wand. Australian Graduate School of Management. University of New South Wales.
Data-Based Choice of Histogram Bin Width M. P. Wand Australian Graduate School of Management University of New South Wales 13th May, 199 Abstract The most important parameter of a histogram is the bin
More information