EXACT MEAN INTEGRATED SQUARED ERROR OF HIGHER ORDER KERNEL ESTIMATORS

Size: px
Start display at page:

Download "EXACT MEAN INTEGRATED SQUARED ERROR OF HIGHER ORDER KERNEL ESTIMATORS"

Transcription

1 Econometric Theory, 2, 2005, Printed in the United States of America+ DOI: 0+070S EXACT MEAN INTEGRATED SQUARED ERROR OF HIGHER ORDER KERNEL ESTIMATORS BRUCE E. HANSEN University of Wisconsin The exact mean integrated suared error ~MISE! of the nonparametric kernel density estimator is derived for the asymptotically optimal smooth polynomial kernels of Müller ~984, Annals of Statistics 2, ! and the trapezoid kernel of Politis and Romano ~999, Journal of Multivariate Analysis 68, 25! and is used to contrast their finite-sample efficiency with the higher order Gaussian kernels of Wand and Schucany ~990 Canadian Journal of Statistics 8, !+ We find that these three kernels have similar finite-sample efficiency+ Of greater importance is the choice of kernel order, as we find that kernel order can have a major impact on finite-sample MISE, even in small samples, but the optimal kernel order depends on the unknown density function+ We propose selecting the kernel order by the criterion of minimax regret, where the regret ~the loss relative to the infeasible optimum! is maximized over the class of two-component mixturenormal density functions+ This minimax regret rule produces a kernel that is a function of sample size only and uniformly bounds the regret below 2% over this density class+ The paper also provides new analytic results for the smooth polynomial kernels, including their characteristic function+. INTRODUCTION This paper is concerned with the choice of kernel for univariate nonparametric density estimation+ A variety of kernel functions have been proposed+ Some references include Parzen ~962!, Epanechnikov ~969!, Müller ~984!, Silverman ~986!, Wand and Schucany ~990!, Scott ~992!, Politis and Romano ~999!, and Abadir and Lawford ~2004!+ A common measure to evaluate the efficiency of a density estimator is mean integrated suared error ~MISE!+ As asymptotic MISE ~AMISE! is easier to calculate than exact MISE, it has been common to analyze AMISE and adopt kernel functions that minimize AMISE+ Marron and Wand ~992! refocused attention on the exact finite-sample MISE in their study of the higher order This research was supported in part by the National Science Foundation+ I thank Oliver Linton and a referee for helpful comments and suggestions that improved the paper+ Address correspondence to Bruce E+ Hansen, Department of Economics, 80 Observatory Drive, University of Wisconsin, Madison, WI 53706, USA; bhansen@ssc+wisc+edu Cambridge University Press $

2 032 BRUCE E. HANSEN Gaussian kernels of Wand and Schucany ~990!, showing that AMISE can be uite misleading in many cases of interest+ We extend this literature by developing new expressions for the exact MISE of several important kernel density estimators, including the asymptotically optimal smooth symmetric polynomial kernels of Müller ~984!, the infinite-order Dirichlet kernel, and the trapezoid kernel of Politis and Romano ~999!+ We compute the MISE when the sampling densities are normal mixtures as in Marron and Wand ~992!+ As an important by-product, we obtain convenient expressions for the characteristic functions of the smooth polynomial kernels, a class that includes the Epanechnikov, the biweight, and most higher order polynomial kernels used in empirical practice+ Kernel characteristic functions are useful for many purposes, including computationally efficient density estimation as discussed in Silverman ~986!+ Our calculations show that the choice between the higher order Gaussian kernels and the smooth polynomial kernels has only a small impact on MISE, whereas the choice of kernel order has a large impact on MISE+ Efficiency gains ~and losse from higher order kernels can be uite substantial even in small samples+ The practical problem is that the ideal kernel order depends on the unknown density+ To provide a practical solution to the problem of kernel order selection, we suggest the decision theoretic criterion of minimax regret+ This criterion picks the kernel order to minimize the maximal percentage deviation between the actual MISE and the infeasible optimal MISE over a set of candidate density functions+ The minimax regret kernel order is a function only of sample size and thus can be implemented in practice+ We calculate the minimax regret rule by maximizing the regret over all two-component mixture-normal densities+ We find that the minimax regret kernel order is increasing in sample size, and this rule uniformly bounds the regret below 2%+ We should mention two caveats at the outset+ First, higher order kernel density estimators are not constrained to be nonnegative and thus may violate the properties of a density function+ A practical solution to this problem ~when it arise is to remove the negative part of the estimator and rescale the positive part to integrate to one+ As shown by Hall and Murison ~993! this does not affect the AMISE+ However, the impact of this procedure on the finite-sample MISE is unclear+ Second, our MISE calculations implicitly assume that the density has support on the entire real line ~our analysis is for mixture-normal densitie+ The existence of a boundary can have substantial effects on the estimation problem and hence on the MISE+ The remainder of the paper is organized as follows+ Section 2 introduces notation+ Section 3 presents new results concerning smooth polynomial kernels+ Section 4 computes the characteristic functions and roughness of these kernels+ Section 5 presents the exact MISE of the density estimates, when the underlying density is a mixture of normals as in Marron and Wand ~992!+ Section 6

3 MISE OF HIGHER ORDER KERNEL ESTIMATORS 033 presents a numerical analysis of the MISE of the various kernels+ We find that the MISE can be greatly influenced by the order of the kernel, even in small samples, and that the selection of kernel order is essential to accurate estimation+ We find that the higher order Gaussian kernels of Wand and Schucany ~990! are the most reasonable candidates for finite-sample efficient estimation+ Section 7 proposes the selection of kernel order using minimax regret+ Section 8 is a brief conclusion+ Proofs are presented in the Appendix+ Further numerical results and details, in addition to the compute code for the calculations, are available at 2. NOTATION For any function g, let g ~m! ~x! ~d m 0dx m!g~x! denote its mth derivative+ Let f~x!, f s ~x! s f~x0, and f ~m! s ~x! s m f ~m! ~x0 denote the standard normal density, the scaled normal density, and its mth derivative, respectively+ We call K~x! a symmetric kernel function if K~x! K~x! and * R K~x! dx + Following Müller ~984!, we say that K is s-smooth if K ~ is absolutely continuous on R+ If the kernel has bounded support, this reuires that the ~s!th derivative euals zero at the boundaries+ For integer r, a kernel function K is of order 2r if * x k K~x! dx 0 for all k 2r yet * x 2r K~x! dx 0+ In particular, a kernel is second-order if * x 2 K~x! dx 0+ Pochhammer s symbol is denoted by n ~a! n ) ~a j! j0 G~a n! + G~a! A useful result is Legendre s duplication formula: ~2m!! 2m m!2 + 2m ) We will be making extensive use of special functions, for which a useful reference is Magnus, Oberhettinger, and Soni ~966!+ Many can be represented as generalized hypergeometric functions ~a! j +++~a p! j x p F ~a,+++,a p ; c,+++,c ; x! j j0 ~c! j +++~c! j j! + 2) For a review of the latter see Abadir ~999!+ In particular, for integer m 0, our results will make use of the Gegenbauer polynomials C l m ~! k G~l m k!~2x! m2k G~l! k0 k!~m 2k!! 3)

4 034 BRUCE E. HANSEN and the spherical Bessel function j m ~x! k0 2 x m 3 2m ~! k 3 k! 2mk 0 F x m2k m; x ) The Gegenbauer polynomials have several hypergeometric representations, including 2~l! x l C 2m ~x! ~! m m ~m!! 2 F m, m l; 3 2 ; x 2+ 5) See Magnus et al+ ~966, p+ 220!+ 3. SMOOTH POLYNOMIAL KERNELS Using a random sample with n real observations X,+++,X n drawn independently from a distribution with density f ~x!, the nonparametric Rosenblatt kernel estimator of f ~x! is f n ~x! n K x X i, nh i h where K is a kernel function and h is a bandwidth+ The MISE of f n is MISE n ~h, K, f! E~ f n ~x! f ~x!! 2 dx+ For the moment, restrict attention to s-smooth, second-order kernels with bounded support+ Müller ~984! and Granovsky and Müller ~99! found that the MISE of f ~ n is minimized by setting K to eual 2s M s ~x! ~ x 2! s + 6) This class includes the uniform ~s 0!, Epanechnikov ~s!, biweight ~s 2!, and triweight ~s 3!+ Furthermore, as s diverges, the optimal smooth kernel approaches the Gaussian kernel, as

5 MISE OF HIGHER ORDER KERNEL ESTIMATORS 035 lim sr s M2s M2s M x f~x!+ Now consider the more general case of s-smooth bounded-support kernels of order 2r+ Müller ~984! and Granovsky and Müller ~99! found that the MISE of f n ~ is minimized by using a polynomial kernel of order s r inx 2 + We now present several alternative representations for this kernel+ THEOREM + Müller s s-smooth, 2rth-order kernel is M 2r, s ~x! B r, s ~x!m s ~x!, 7) where B r, s ~x! 2r ~!k r sr 2 s r k ~s! r k0 k!~r k!! 3 2k x 2k 8) ~! r Furthermore s 2 m0 2r ~s! rs 2 x C s02 2r ~x! 9) r s 2m2m ~! m ~s m!! 2 C s02 2m ~x!+ 0) r M 2r, s ~x! ~! m m0 2ms m!2 2m 22ms M ~2m! 2ms ~x!+ ) Euation ~8! gives an explicit formula for the kernel, whereas ~9! ~! give alternatives+ For numerical computation, the formula ~7! ~8! is numerically reliable+

6 G E I E 036 BRUCE E. HANSEN Granovsky and Müller ~99! examined the limit as s r + They showed that lim sr M2s M 2r, s x M2s G 2r ~x! ~!r f~2r! ~x! 2 r ~r!!x + 2) Here G 2r are the higher order Gaussian kernels of Wand and Schucany ~990! and Marron and Wand ~992!+ We now show that as the order 2r increases, both classes of kernels approach the infinite-order Dirichlet kernel D~x! sin~x!0~px!+ lim rr lim rr THEOREM 2+ 2r M 2r, s THEOREM 3+ x 2r D~x!+ 3) 2r M2r 2 G x D~x!+ 4) M2r 2 4. CHARACTERISTIC FUNCTION AND ROUGHNESS For any function g, let g~t! * exp~itx!g~x! dx denote the characteristic function of g~x!+ For example, the characteristic function of the normal density is f~t! exp~t 2 02!+ For the higher order Gaussian kernels, the characteristic function G 2r ~t! exp~t 2 02! r k0 t 2k 02 k k! is given by Wand and Schucany ~990, p+ 200!+ However, the characteristic function has not been calculated previously for the smooth polynomial kernels M 2r, s + We now show that they consist of linear combinations of spherical Bessel functions as defined in ~4!+ THEOREM 4+ M s ~t! 2s 3 2 s t j s ~t!+ THEOREM 5+ MG 2r, s ~t! 2 Mp 2 r ts m0 a s ~m!j s2m ~t!, 5)

7 where a s ~m! G 2 m s 2 2m s m! + 6) For computational purposes, when t is large, the series ~4! can be numerically unstable, in which case it is preferable to compute j m ~t! using the recurrence 2m j m ~t! j m ~t! j m ~t! 7) t with the initial conditions MISE OF HIGHER ORDER KERNEL ESTIMATORS 037 j 0 ~t! sin~t! t 8) and j ~t! sin~t! t 2 cos~t! + t Combining this with Parseval s euality, we can use ~5! to calculate the roughness of M 2r, s + THEOREM 6+ R~M 2r, s! M 2r, s ~x! 2 dx ~2! p r r l0 m0 ~6l m6 s!a s ~l!a s ~m! ~l m!~m l! 2 l m 2s + Politis and Romano ~999! introduced a class of infinite-order general flattop kernels that have flat characteristic functions at the origin+ Their preferred kernel is l 2 ~x! 2~sin2 ~x! sin 2 ~x02!! px 2, which has the trapezoidal characteristic function ld 2 ~t! ~2 6t 6! ~ 6t 6!,

8 038 BRUCE E. HANSEN where ~t! max~t,0!+ We thus call l 2 ~t! the trapezoid kernel+ It is easy to calculate that R~l 2! 40~3p!+ Figure a displays the four kernels M 4,8, M 0,8, M 20,8, and l 2 + The kernels have been rescaled and normalized so that all have eual roughness+ The basic shapes of the kernels are similar, with the waviness of the kernels increasing in kernel order+ The second panel displays the characteristic functions of the same kernels ~again normalized to have eual roughnes+ Here it is easier to see the Figure. ~a! Polynomial and Trapezoid Kernel Functions+ ~b! Polynomial and Trapezoid Characteristic Functions+

9 D D E E E MISE OF HIGHER ORDER KERNEL ESTIMATORS 039 contrasts+ As the kernel order increases, the characteristic functions are increasingly flat at the origin, with the high-order polynomial characteristic function developing a significant negative lobe+ 5. EXACT MISE The exact MISE of the kernel density estimator f n can be written as MISE n ~h, K, f! R~K! nh ni 2 ~h, K, f! 2I ~h, K, f! I 0 ~ f!, 9) where I j ~h, K, f! p 0 K~ht! j f D~t! 2 dt ph 0 K~t! j D f h t 2 dt 20) with K and fd denoting the characteristic functions of K and f+ Euation ~9! is shown, for example, in euation ~3+! of Chiu ~99!+ We write I 0 ~h, K, f! I 0 ~ f! because it is independent of K and h+ Marron and Wand ~992! proposed a test set of mixture-normal density functions for which ~9! is feasible to calculate+ The Marron Wand densities take the form f ~x! w k f sk ~x m k!+ 2) k The characteristic function of ~2! is f ~t! w k e s k 2 t 2 02 exp~itm k! k with suare f ~t! 2 i k w i w k e s ik 2 t 2 02 cos~tm ik!, 22) where s 2 ik s 2 i s 2 k and m ik m i m k + We follow Marron and Wand and restrict attention to this class of density functions for our calculation of exact MISE+ From ~22! and euation ~A+7! in the Appendix, or Theorem 2+ of Marron and Wand ~992!, we find that I 0 ~ f! i k w i w k f sik ~m ik!+ To calculate MISE n ~h, K, f! we also need the integrals I ~h, K, f! and I 2 ~h, K, f!, which may be calculated by numerical or analytic integration+ Although we use numerical integration to calculate MISE in the next section we use analytic integration to verify their accuracy+ Our first result is the integrals for the smooth polynomial kernels+

10 040 BRUCE E. HANSEN THEOREM 7+ For the mixture-normal densities 2), r I ~h, M 2r, s, f! ~! m a s ~m!u s ~h, m!, m0 r r I 2 ~h, M 2r, s, f! l0 m0 ~! lm a s ~l!a s ~m!v s ~h, l, m!, where a s ~m! are defined in 6), and U s ~h, m! i k w i w k a0 2 h 2m2a f ~2m2a! sik ~m ik! a!gs 2m a 3 2, 23) V s ~h, l, m! i k a0 w i w k ~2a 2s 2m 2l!! 2 h 2l2m2a f ~2l2m2a! sik ~m ik! ~a 2s 2m 2l!!Ga s 2l 3 2Ga s 2m 3 2a! + 24) Our second result is the integrals for the trapezoid and Dirichlet kernels+ This extends the work of Davis ~98!, who made similar calculations for D when f f+ THEOREM 8+ For the mixture-normal densities 2), I ~h, l 2, f! 2C 02 2 h hc 2 h C 0 h hc h, I 2 ~h, l 2, f! 4C 02 2 h 4hC 2 h h 2 C h 3C 02 h 4hC h h 2 C 302 h, I ~h, D, f! I 2 ~h, D, f! C 02 h,

11 D MISE OF HIGHER ORDER KERNEL ESTIMATORS 04 where C h ~u! 2p i k w i w k 2 h s ik 2h a0 ga h, u 2 2 s ik a! 2a 2 m 2 a ik 2s ik 2 25) and x g~a, x! t 0 a e t dt is the incomplete gamma function. Theorems 7 and 8 express the integrals I ~h, K, f! and I 2 ~h, K, f! as convergent infinite series+ Methods to numerically evaluate infinite series and other special functions are discussed in detail in Zhang and Jin ~996!+ As they discuss, numerical evaluation of infinite series can fail as a result of accumulated round-off error when there are alternating signs, which occurs in our expressions+ To investigate their accuracy, we numerically compared the expressions in Theorems 7 and 8 with numerical integration for the Marron Wand densities+ We used Gauss Legendre uadrature with 8,000 grid points T # where T was selected so that the numerical integral of p * T 0 f ~t! 2 dt was within 0+0% of the exact value * f ~x! 2 dx+ The results were plotted against h+ ~See Hansen, 2003+! We found that in some cases the terms in the series expansions grow sufficiently large such that the numerical results are unreliable+ This occurred for the polynomial kernels for large h, r, and0or s and also for the trapezoid kernel for very small h in a few cases+ We concluded that numerical integration was more reliable for evaluation of I ~h, K, f! and I 2 ~h, K, f! for these kernels+ 6. ANALYSIS OF MISE For each kernel we computed MISE n ~h, K, f! onagridof3,000 values of h over the * #, where h * is the Silverman rule-of-thumb bandwidth for a sample size of 50+ Plots of MISE n ~h, K, f! against h for each model and kernel indicated that this region appeared to contain the global minimum for all cases+ ~See Hansen, 2003+! For given K and f the optimal finite-sample MISE is MISE n ~K, f! min MISE n ~h, K, f!+ h For the kernel classes G 2r and M 2r, s, MISE n ~K, f! was plotted as a function of r+ Figures 2 and 3 display these plots for the Marron Wand densities ~the Gaussian density! and 3 ~strongly skewed density!+ These densities are shown

12 042 BRUCE E. HANSEN Figure 2. MISE, Density + here because are they are extreme cases regarding the gain or loss from the use of higher order kernels+ Plots for other densities can be seen in Hansen ~2003!+ Plotted are the MISE for the Gaussian kernels G 2r and the smooth polynomial kernels M 2r,, M 2r,5, and M 2r,8, for 2r 2,+++,20, and for the sample sizes of n 50, 200, 500, and,000+ The dotted flat lines are the MISE of D and l 2,

13 MISE OF HIGHER ORDER KERNEL ESTIMATORS 043 Figure 3. MISE, Density 3+ which are independent of r+ For each plot, the MISE n ~K, f! is normalized relative to min K MISE n ~K, f!+ First examine the flat-top kernels D and l 2 + In nearly all cases, the trapezoid kernel l 2 does much better than the Dirichlet kernel D, and in most cases l 2 performs uite well relative to other kernel choices+ However, it never has as low MISE as the best Gaussian or polynomial kernel+

14 044 BRUCE E. HANSEN Second, compare the smooth polynomial kernels M 2r, s across the smoothness index s+ We observe that increasing the smoothness order typically improves the MISE, but the difference diminishes in larger samples+ Indeed, for very large samples ~not shown!, the best MISE is for small s, but the gain is minimal+ There also appears to be little gain achieved by increasing s above s 8 ~again not shown!, so setting s 8 emerges as a practical recommendation+ Third, compare the MISE of the smooth polynomial kernel M 2r,8 and the Gaussian kernel G 2r + In most cases the difference in MISE is uite small, with the Gaussian kernels achieving lower MISE in small samples and the reverse in large samples+ Examining the plots of MISE n ~K, f!, it is clear that there is no uniformly optimal choice of kernel K ~including kernel order! across sample sizes n and densities f+ For any given f there may be a best choice of kernel among those we consider, but this choice will vary across f+ A data-dependent method such as cross-validation can possibly be used to select kernel order, which might enable estimation to adapt to the unknown density f+ This is a difficult problem, however, and is not pursued further in this paper+ 7. MINIMAX REGRET KERNELS Let K denote a class of kernels, such as the Gaussian class $G 2r ; r % or the polynomial class $M 2r, s ; r ; s %+ The problem raised at the end of the previous section is how to select a kernel K K when it is known that the finite-sample optimal choice varies with the true underlying density f+ This section proposes a pragmatic solution based on minimax regret, a concept in statistical decision theory dating back to Savage ~95!+ For any density f and sample size n, there is a best possible kernel choice that can be defined as K n ~ f! argmin ln MISE * n ~K, f!+ KK Note that we implicitly define our loss function as ln MISE n * ~K, f! to eliminate scale effects+ The regret associated with the use of an arbitrary kernel K is Regret n ~ f, K! ln MISE n ~K, f! ln MISE n ~K n ~ f!, f! ln MISE n~k, f! MISE n ~K n ~ f!, f!, the percentage deviation of the realized MISE from the optimal value+ For a compact class of density functions F the maximal regret associated with K is the maximum over f F Regret n ~K! max Regret n ~ f, K!+ ff

15 The minimax regret rule is to pick K to minimize the maximum regret K * n argmin Regret n ~K! KK with the associated minimized regret Regret n * Regret n ~K n *! min KK MISE OF HIGHER ORDER KERNEL ESTIMATORS 045 max Regret n ~ f, K!+ ff Note that whereas the small-sample optimal kernel K n ~ f! depends on the unknown density f in addition to n, the minimax regret kernel K * n only depends on the sample size n+ The kernel K * n guarantees that if the true density f belongs to F, the percentage deviation of the MISE from the optimum is bounded by Regret * n + For the density class F we use the Marron Wand mixture-normal class ~2! with 2: f ~x! w s f x m ~ w! s s 2 f x m 2+ s 2 Many of the Marron Wand test densities are in this class, and two normal components are sufficient to generate most density shapes of interest+ Because the regret criterion is invariant to location and scale, we select m 2 and s 2 so that EX 0 and Var~X!, so the density class is fully described by the remaining parameters ~w, m, s!+ We use a grid over these parameters, varying each in increments of 0+ over all feasible values+ This results in a density class with,887 elements+ Setting K to eual the Gaussian kernel class G 2r, Figure 4 plots Regret n ~G 2r! as a function of 2r for n 25, 00, 400, and 800+ The minimax regret solution r * n locates its lowest value for any curve+ We can see that this solution r * n varies from 2 to 8 for the four sample sizes+ We calculated the minimax regret r * n for the Gaussian kernels G 2r as a function of sample size, and we present the findings in Table + The optimal kernel order r * n is strictly increasing with n+ The magnitudes may be surprising to many readers, as the minimax kernel orders are higher than typically used in practice+ Let G * n denote the Gaussian higher order kernel with r r * n, and similarly M * n, s + These are feasible kernel choices as they vary only with sample size+ We now compare the regret of the feasible kernel choices f, G * n, M * n,2, and l 2 + Figure 5 plots Regret * n for these kernels as a function of n where the general kernel class includes G 2r, M 2r,2, and l 2 + We see that the standard normal kernel f has regret that is increasing in n and is clearly dominated by the other choices except when the sample size is very small+ The trapezoid kernel l 2 has high

16 046 BRUCE E. HANSEN Figure 4. Regret of Gaussian Kernels+ regret for small samples but has comparable regret with the optimal Gaussian and polynomial kernels for large samples+ The optimal Gaussian and polynomial kernels have similar regret for most sample sizes, with that for G n * slightly lower+ Because G n * is somewhat easier to manipulate numerically, it is our recommendation for empirical practice+ A sensible alternative choice is the Table. Minimax regret Gaussian kernel order by sample size: Regret calculated over twocomponent normal mixture densities 2r n * n 2 n 40 4 n 40 6 n n n n,500 4 n 2,900 6 n 6,000 8 n 6, n 34,000

17 MISE OF HIGHER ORDER KERNEL ESTIMATORS 047 Figure 5. Regret across Kernel Classes+ trapezoidal kernel l 2, as long as n 500 ~its regret is bounded below 2% for such sample+ Finally, from Figure 5 we find that the regret of G n * is uniformly bounded below 2%+ This means that this kernel choice guarantees that regardless of f ~in the class of two-component mixture-normal distribution the MISE n * ~G n *, f! cannot differ from the optimal MISE n * ~K n ~ f!, f! by more than 2%+ 8. CONCLUSION We have found that finite-sample MISE varies considerably across kernel order 2r and underlying density f+ If information about f is known this can be used to help pick the kernel order+ Otherwise, the minimax regret solution picks the kernel order to minimize the worst-case regret+ We find that by focusing on the class of mixture-normal densities, we are able to deduce a practical rule for kernel order selection that bounds the regret below 2% for any sample size+ Our analysis points to the excellent performance of the higher order Gaussian and smooth polynomial kernels+ An important caveat is that minimax regret is defined relative to a density class F, and we restricted attention to two-component normal mixtures+ If another set F is selected, the minimax regret kernel can change+ Further investigation

18 048 BRUCE E. HANSEN of the sensitivity of these recommendations to alternative sets F would be a useful subject of future research+ REFERENCES Abadir, K+M+ ~999! An introduction to hypergeometric functions for economists+ Econometric Reviews 3, Abadir, K+M+ &S+ Lawford ~2004! Optimal asymmetric kernels+ Economics Letters 83, Chiu, S+-T+ ~99! Bandwidth selection for kernel density estimation+ Annals of Statistics 9, Davis, K+B+ ~98! Mean integrated suare error properties of density estimates+ Annals of Statistics 5, Epanechnikov, V+I+ ~969! Non-parametric estimation of a multivariate probability density+ Theory of Probability and Its Applications 4, Granovsky, B+L+ &H+-G+ Müller ~99! Optimizing kernel methods: A unifying variational principle+ International Statistical Review 59, Hall, P+ &R+D+ Murison ~993! Correcting the negativity of high-order kernel density estimators+ Journal of Multivariate Analysis 47, Hansen, B+E+ ~2003! Appendix: Exact mean integrated suared error of higher-order kernel estimators+ www+ssc+wisc+edu0;bhansen0papers0mise+html+ Magnus, W+, F+ Oberhettinger, &R+P+ Soni ~966! Formulas and Theorems for the Special Functions of Mathematical Physics+ Springer-Verlag+ Marron, J+S+ &M+P+ Wand ~992! Exact mean integrated suared error+ Annals of Statistics 20, Müller, H+-G+ ~984! Smooth optimum kernel estimators of densities, regression curves and modes+ Annals of Statistics 2, Parzen, E+ ~962! On estimation of a probability density and mode+ Annals of Mathematical Statistics 35, Politis, D+N+ &J+P+ Romano ~999! Multivariate density estimation with general flat-top kernels of infinite order+ Journal of Multivariate Analysis 68, 25+ Savage, L+J+ ~95! The theory of statistical decision+ Journal of the American Statistical Association 46, Scott, D+W+ ~992! Multivariate Density Estimation+ Wiley+ Silverman, B+W+ ~986! Density Estimation for Statistics and Data Analysis+ Chapman and Hall+ Tranter, C+J+ ~968! Bessel Functions with some Physical Applications+ Hart+ Wand, M+P+ &W+R+ Schucany ~990! Gaussian-based kernels+ Canadian Journal of Statistics 8, Zhang, S+ &J+ Jin ~996! Computation of Special Functions+ Wiley+ APPENDIX: Proofs Proof of Theorem. Müller ~984, Thm+ 2+4! and Granovsky and Müller ~99, Cor+ 2! show that the optimal kernel is the uniue 2rth-order kernel of the form ~7! with B r, s ~x! a polynomial of degree r inx 2 + Because ~8! takes this form, it is sufficient to show that B r, s ~x!m s ~x! is a valid 2rth-order kernel, that ~8! ~0! are euivalent, and ~!+ We start by showing the euivalence of ~8! ~0!+ First, because for k m, ~m! k ~! k m!0~m k!!, then

19 MISE OF HIGHER ORDER KERNEL ESTIMATORS 049 2r ~!k r sr 2 s r x k 2k ~s! r k0 k!~r k!! 3 2k 2r sr ~s! r ~r!! k0 k ~ r! r 2 s r x k 2k k! 3 2k 2 3 2r 2 s r ~s! r ~r!! 2 F r, 2 s r ; 3 2 ; x 2 r ~! 2r ~s! rs x C s02 2r ~x!, where the final euality is ~5!+ This shows the euality of ~8! and ~9!+ Second, let l s 2 _ + The Gegenbauer polynomials C m l ~x! satisfy the recurrence ~Magnus et al+, 966, p+ 222! ~ j!c l j Thus for any j, 3 2j ~s j!! C l 2j~x! ~x! ~2j 2s!xC l j ~x! ~ j 2C l j ~x!+ 3 2j2j s 2 ~s j!! xc l 2j ~x! By back substitution and the fact C l ~x! ~2s!x, we find 3 r 2r ~s r!! x l C 2r ~x! ~! r ~s m!! m 3 2j ~s j!! C l 2j~x!+ ~! m 3 2m2m s 2 C l 2m ~x! ~! r x C l ~x! m s r 2m2m ~! r 2 ~! ~s m!! m0 2 C l 2m ~x!+

20 050 BRUCE E. HANSEN Hence r ~! 2r ~s! rs x C s02 2r ~x! s 2 m0 C s02 2m ~x!, m s r 2m2m 2 ~! ~s m!! showing the euivalence of ~9! and ~0!+ We now show that B r, s ~x!m s ~x! is a valid 2rth-order kernel+ First, using ~0! and the facts that C l 0 ~x! and * C l m ~x!m s ~x! dx 0 for all m, B r, s ~x!m s ~x! dx s 2 m0 2 m s r 2m2m ~! ~!0 20s 2 s 2 ~s m!! M s ~x! dx + C l 2m ~x!m s ~x! dx Thus B r, s ~x!m s ~x! is a valid kernel+ Second, using ~9! and the orthogonal properties of the Gegenbauer polynomials, for any j 2r x j B r, s ~x!m s ~x! dx ~! r 2r ~s! rs 2 x j C s02 2r ~x!m s ~x! dx 0+ Thus B r, s ~x!m s ~x! is of order 2r as reuired+ Finally, we demonstrate ~!+ Rodrigues formula for the Gegenbauer polynomials ~Magnus et al+, 966, p+ 22! is ~2m 2! C s02 2m ~x! 2 2m ~2m!!~2!~2m! ~ x d 2m 2! s dx ~ x 2! 2ms + 2m Using ~! and ~6!, C s02 2m ~x!m s ~x! 2ms~m!s 2 ~2m!! 22ms M ~2m! 2ms ~x!+

21 Combined with ~0! this yields M r, s ~x! s 2 m0 s 2 m0 MISE OF HIGHER ORDER KERNEL ESTIMATORS 05 r ~! m r ~! m 2m2m s 2 ~s m!! 2m2m s 2 ~s m!! C s02 2m ~x!m s ~x! 2ms~m!s 2 ~2m!! 22ms M ~2m! 2ms ~x! r ~! m m0 2ms m!2 2m 22ms M ~2m! 2ms ~x!, which is ~!+ Proof of Theorem 2. Let J v ~x! denote the Bessel functions and P m ~a, b! ~x! denote the Jacobi polynomials+ ~See, e+g+, Ch+ 3 and Sect+ 5+2 of Magnus et al+, 966+! In particular, J s02 ~x! 2x 02 j p s ~x! where j s ~x! is the spherical Bessel function defined in ~4!+ Using results from pp+ 29 and 20 of Magnus et al+ ~966! and ~!, 2 2r ~r!!gr s ~! r x l C 2r 2 ~x! ~! r Gr s 2 Gs 2r 2 P r ~2r!!Gs 2 P ~s,02! r ~2x 2! A.) ~02, ~ 2x 2!+ A.2) By Magnus et al+ ~966, p+ 26!, as r r Mr P ~02, r x 2 2r 2 r 2 x 02J 02 ~x! 2 p j 0~x! 2 sin~x! 02 p 02 x + A.3) The eualities are ~A+! and ~8!+ Furthermore, using Stirling s formula we can show that

22 052 BRUCE E. HANSEN lim mr G 2 m m 02 G~m! + A.4) Thus combining ~9! with ~A+2!, and using the limits in ~A+3! and ~A+4!, we obtain as r r Gr s 2 2r B r, s x 2r r 2rGs 3 2G~s r! Furthermore, as r r M s 2r x Thus as r r x 2r M 2r, s Gs 3 2 sin~x! p 02 x + 2s x 2 2 s r 4r 2r 2r B r, s r establishing ~3!+ 2rM x s2r x Gs 3 2 sin~x! p 02 x D~x!, ~02, P r x 2 2 2r 2s + 2s Proof of Theorem 3. Let He m ~x! denote the hermite polynomials+ By the Rodrigues formula and a limit property for the Hermite polynomials ~Magnus et al+, 966, pp+ 252, 255!, as r r, M2r 2 G 2r x M2r 2 establishing ~4!+ ~! r f ~2r! x M2r 2 2 r ~r!!x ~! r He 2r x M2r 2f x M2r 2 r sin x px, 2 r ~r!!x

23 H G 0 Proof of Theorem 4. From Magnus et al+ ~966, p+ 40! MISE OF HIGHER ORDER KERNEL ESTIMATORS 053 cos~tx!~ x 2! s dx 2 s t p 2t J s02~t! 2 s j t s ~t!, the second euality using ~A+!+ Then by the definition of the Fourier transform H M s ~x! 2s e itx ~ x 2! s dx 3 2s 0 cos~tx!~ x 2! s dx 3 2s 2 tsj s ~t!+ Proof of Theorem 5. Using ~!, the derivative property of Fourier transformations, and Theorem 4, r M 2r, s ~x! ~! m m0 2ms m!2 2m 22ms M ~2m! H 2ms ~x! r m0 2ms m!2 2m 22ms t 2m M 2ms ~t! as stated+ r m0 2ms m!2 2m 22ms 3 22ms 2m t 2 2ms j t 2ms ~t! Mp 2 2 r G 2 m s 2 2m s ts m0 j 2ms ~t! m!

24 G G 054 BRUCE E. HANSEN 4 Proof of Theorem 6. By ~A+! and Magnus et al+ ~966, p+ 99!, p 0 2 2s t 2 t 0 j s2l ~t!j s2m ~t! dt 2s J s2l02 ~t!j s2m02 ~t! dt G~2s!G 2 l m G 3 2 l m 2s G~ l m G~ l m ~l m!~m l! 2 l m 2s ~2! 6l m6 s + A.5) By Parseval s euality, Theorem 5, and ~A+5!, R~M 2r, s! M 2r, s ~x! 2 dx M 2r, s ~t! 2 dt 2p p l0 r r m0 0 6l m6 s a s ~l!a s ~m! 4 p 0 2 t 2s j s2l ~t!j s2m ~t! dt ~2! p r r l0 m0 ~6l m6 s! a s ~l!a s ~m! ~l m!~m l! 2 l m 2s + Proof of Theorem 7. Using expressions ~5! and ~22! we find I ~h, M 2r, s! p 0 r a s ~m! m0 r M 2r, s ~ht! f D~t! 2 dt 2 a s ~m! m0 p i k ht 2 s j s2m ~ht! f D~t! 2 dt 2 w i w k 2 s j p ht s2m ~ht!e s ik 2 t 2 02 cos~tm ik! dt+ A.6)

25 G MISE OF HIGHER ORDER KERNEL ESTIMATORS 055 Using Magnus et al+ ~966, p+ 402! we find that p 0 t 2m e s 2 t 2 02 cos~tm! dt ~!m M2ps 2m exp m2 2s 2He 2m m s ~! m f s ~2m! ~m!+ A.7) Using ~A+!, ~4!, and ~A+7!, 2 2 s e p ht s 2 t 2 cos~tm!j s2m ~ht! dt 2 s02 e p 0 ht s 2 t 2 cos~tm!j s2m02 ~ht! dt a0 a0 ~! a Gs 2m a 3 2a! 22m2a h e s 2 t 2 cos~tm!t 2m2a dt p 0 ~! m a!gs 2m a 2 h ma 2 3 f ~2m2a! s ~m!+ Combining this with ~A+6! we obtain ~23!+ Next, I 2 ~h, M 2r, s! p 0 M 2r, s ~ht! 2 f D~t! 2 dt r r l0 m0 a s ~l!a s ~m! r r l0 m0 0 4 a s ~l!a s ~m! 2 ht 2s p 2 0 i k 2 ht 2s w i w k 4 p 2 j s2l ~ht!j s2m ~ht! f D~t! 2 dt j s2l ~ht!j s2m ~ht!e s ik 2 t 2 02 cos~tm ik! dt+ A.8) By euation ~2+20! of Tranter ~968! 4 p j s2l~t!j s2m ~t! 2 t J s2l02~t!j s2m02 ~t! a0 ~! a ~2a 2s 2l 2m!!~t02! 2l2m2s2a a!ga s 2l 3 2Ga s 2m 3 2~a 2s 2l 2m!! +

26 056 BRUCE E. HANSEN Combining this with ~A+7! we obtain 4 ht 2 2s j s2l ~ht!j s2m ~ht!e s 2 t 2 02 cos~tm! dt p 2 0 ~! a ~2a 2s 2l 2m!! a0 a!ga s 2l 2Ga 3 s 2m 2~a 3 2s 2l 2m!! 2 h 2l2m2a p 0 t 2l2m2a e s 2 t 2 02 cos~tm! dt ~! a ~2a 2s 2l 2m!! a0 a!ga s 2l 2Ga 3 s 2m 2~a 3 2s 2l 2m!! 2 h 2l2m2a f ~2l2m2a! s ~m!+ Combining this with ~A+8! we obtain ~24! as reuired+ Proof of Theorem 8. By a series expansion and ~! ~x cos~x! 2! a a0 ~2a!! a0 ~x 2! a! a + 2 2a 2a Thus for any h 2 _ 20 u t 2h cos~tm!exp s 2 t 2 dt 2 2 a0 2h ~m 2! a! a u t 0 2h2a exp s 2 t 2 dt 2 2 2a 2a s 2h a0 a! 2a 2 aga m2 h, u 2 s 2 + 2s 2 Summing, we find p 0 u t 2h f D~t! 2 dt 2p i 2p i k k w i w k 20 w i w k 2 h u s ik 2h a0 t 2h cos~tm ik!exp s ik 2 t 2 dt 2 a! 2a m 2 a ik 2s ik 2 ga h, u 2 2 s ik 2 C h ~u!+

27 D D E MISE OF HIGHER ORDER KERNEL ESTIMATORS 057 Hence I ~h, D! I 2 ~h, D! p 0 D~ht! f D~t! 2 dt p 0 0h f D~t! 2 dt C ~0h!, I ~h, l 2! p 0 l 2 ~ht! f D~t! 2 dt p 0 20h ~2 ht! f D~t! 2 dt 0h ~ ht! f D~t! 2 dt p 0 2C 02 ~20h! hc ~20h! C 02 ~0h! hc ~0h!, and I 2 ~h, l 2! p 0 l 2 ~ht! 2 f D~t! 2 dt p 4t t 2!~t 2! ~3 4t t 2!~t!# f D~t! 2 dt 4C 02 ~20h! 4hC ~20h! h 2 C 302 ~20h! 3C 02 ~0h! 4hC ~0h! h 2 C 302 ~0h!+

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Estimation of cumulative distribution function with spline functions

Estimation of cumulative distribution function with spline functions INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative

More information

Averaging Estimators for Regressions with a Possible Structural Break

Averaging Estimators for Regressions with a Possible Structural Break Averaging Estimators for Regressions with a Possible Structural Break Bruce E. Hansen University of Wisconsin y www.ssc.wisc.edu/~bhansen September 2007 Preliminary Abstract This paper investigates selection

More information

A Novel Nonparametric Density Estimator

A Novel Nonparametric Density Estimator A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with

More information

Smooth simultaneous confidence bands for cumulative distribution functions

Smooth simultaneous confidence bands for cumulative distribution functions Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang

More information

Discussion Paper No. 28

Discussion Paper No. 28 Discussion Paper No. 28 Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on the Circle Yasuhito Tsuruta Masahiko Sagae Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on

More information

ON SOME TWO-STEP DENSITY ESTIMATION METHOD

ON SOME TWO-STEP DENSITY ESTIMATION METHOD UNIVESITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XLIII 2005 ON SOME TWO-STEP DENSITY ESTIMATION METHOD by Jolanta Jarnicka Abstract. We introduce a new two-step kernel density estimation method,

More information

A Note on the Differential Equations with Distributional Coefficients

A Note on the Differential Equations with Distributional Coefficients MATEMATIKA, 24, Jilid 2, Bil. 2, hlm. 115 124 c Jabatan Matematik, UTM. A Note on the Differential Equations with Distributional Coefficients Adem Kilicman Department of Mathematics, Institute for Mathematical

More information

A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE

A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE BRAC University Journal, vol. V1, no. 1, 2009, pp. 59-68 A NOTE ON THE CHOICE OF THE SMOOTHING PARAMETER IN THE KERNEL DENSITY ESTIMATE Daniel F. Froelich Minnesota State University, Mankato, USA and Mezbahur

More information

arxiv: v1 [stat.co] 26 May 2009

arxiv: v1 [stat.co] 26 May 2009 MAXIMUM LIKELIHOOD ESTIMATION FOR MARKOV CHAINS arxiv:0905.4131v1 [stat.co] 6 May 009 IULIANA TEODORESCU Abstract. A new approach for optimal estimation of Markov chains with sparse transition matrices

More information

arxiv: v1 [physics.comp-ph] 22 Jul 2010

arxiv: v1 [physics.comp-ph] 22 Jul 2010 Gaussian integration with rescaling of abscissas and weights arxiv:007.38v [physics.comp-ph] 22 Jul 200 A. Odrzywolek M. Smoluchowski Institute of Physics, Jagiellonian University, Cracov, Poland Abstract

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

Nonparametric Density Estimation

Nonparametric Density Estimation Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Bias Correction and Higher Order Kernel Functions

Bias Correction and Higher Order Kernel Functions Bias Correction and Higher Order Kernel Functions Tien-Chung Hu 1 Department of Mathematics National Tsing-Hua University Hsinchu, Taiwan Jianqing Fan Department of Statistics University of North Carolina

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Adaptive Nonparametric Density Estimators

Adaptive Nonparametric Density Estimators Adaptive Nonparametric Density Estimators by Alan J. Izenman Introduction Theoretical results and practical application of histograms as density estimators usually assume a fixed-partition approach, where

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation Preface Nonparametric econometrics has become one of the most important sub-fields in modern econometrics. The primary goal of this lecture note is to introduce various nonparametric and semiparametric

More information

CORRECTION OF DENSITY ESTIMATORS WHICH ARE NOT DENSITIES

CORRECTION OF DENSITY ESTIMATORS WHICH ARE NOT DENSITIES CORRECTION OF DENSITY ESTIMATORS WHICH ARE NOT DENSITIES I.K. Glad\ N.L. Hjort 1 and N.G. Ushakov 2 * November 1999 1 Department of Mathematics, University of Oslo, Norway 2Russian Academy of Sciences,

More information

1 Review of di erential calculus

1 Review of di erential calculus Review of di erential calculus This chapter presents the main elements of di erential calculus needed in probability theory. Often, students taking a course on probability theory have problems with concepts

More information

Nonparametric Function Estimation with Infinite-Order Kernels

Nonparametric Function Estimation with Infinite-Order Kernels Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density

More information

A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers

A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers 6th St.Petersburg Workshop on Simulation (2009) 1-3 A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers Ansgar Steland 1 Abstract Sequential kernel smoothers form a class of procedures

More information

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density

More information

Defect Detection using Nonparametric Regression

Defect Detection using Nonparametric Regression Defect Detection using Nonparametric Regression Siana Halim Industrial Engineering Department-Petra Christian University Siwalankerto 121-131 Surabaya- Indonesia halim@petra.ac.id Abstract: To compare

More information

PROBLEMS AND SOLUTIONS

PROBLEMS AND SOLUTIONS Econometric Theory, 18, 2002, 193 194+ Printed in the United States of America+ PROBLEMS AND SOLUTIONS PROBLEMS 02+1+1+ LS and BLUE Are Algebraically Identical, proposed by Richard W+ Farebrother+ Let

More information

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates. LEGENDRE POLYNOMIALS AND APPLICATIONS We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.. Legendre equation: series solutions The Legendre equation is

More information

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

Bandwith selection based on a special choice of the kernel

Bandwith selection based on a special choice of the kernel Bandwith selection based on a special choice of the kernel Thomas Oksavik Master of Science in Physics and Mathematics Submission date: June 2007 Supervisor: Nikolai Ushakov, MATH Norwegian University

More information

Miscellanea Kernel density estimation and marginalization consistency

Miscellanea Kernel density estimation and marginalization consistency Biometrika (1991), 78, 2, pp. 421-5 Printed in Great Britain Miscellanea Kernel density estimation and marginalization consistency BY MIKE WEST Institute of Statistics and Decision Sciences, Duke University,

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis

J. Cwik and J. Koronacki. Institute of Computer Science, Polish Academy of Sciences. to appear in. Computational Statistics and Data Analysis A Combined Adaptive-Mixtures/Plug-In Estimator of Multivariate Probability Densities 1 J. Cwik and J. Koronacki Institute of Computer Science, Polish Academy of Sciences Ordona 21, 01-237 Warsaw, Poland

More information

A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions

A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions Yu Yvette Zhang Abstract This paper is concerned with economic analysis of first-price sealed-bid auctions with risk

More information

Introduction to Curve Estimation

Introduction to Curve Estimation Introduction to Curve Estimation Density 0.000 0.002 0.004 0.006 700 800 900 1000 1100 1200 1300 Wilcoxon score Michael E. Tarter & Micheal D. Lock Model-Free Curve Estimation Monographs on Statistics

More information

Waves on 2 and 3 dimensional domains

Waves on 2 and 3 dimensional domains Chapter 14 Waves on 2 and 3 dimensional domains We now turn to the studying the initial boundary value problem for the wave equation in two and three dimensions. In this chapter we focus on the situation

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces 9.520: Statistical Learning Theory and Applications February 10th, 2010 Reproducing Kernel Hilbert Spaces Lecturer: Lorenzo Rosasco Scribe: Greg Durrett 1 Introduction In the previous two lectures, we

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Reproducing Kernel Hilbert Spaces Class 03, 15 February 2006 Andrea Caponnetto

Reproducing Kernel Hilbert Spaces Class 03, 15 February 2006 Andrea Caponnetto Reproducing Kernel Hilbert Spaces 9.520 Class 03, 15 February 2006 Andrea Caponnetto About this class Goal To introduce a particularly useful family of hypothesis spaces called Reproducing Kernel Hilbert

More information

On prediction and density estimation Peter McCullagh University of Chicago December 2004

On prediction and density estimation Peter McCullagh University of Chicago December 2004 On prediction and density estimation Peter McCullagh University of Chicago December 2004 Summary Having observed the initial segment of a random sequence, subsequent values may be predicted by calculating

More information

Akaike Information Criterion to Select the Parametric Detection Function for Kernel Estimator Using Line Transect Data

Akaike Information Criterion to Select the Parametric Detection Function for Kernel Estimator Using Line Transect Data Journal of Modern Applied Statistical Methods Volume 12 Issue 2 Article 21 11-1-2013 Akaike Information Criterion to Select the Parametric Detection Function for Kernel Estimator Using Line Transect Data

More information

SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1

SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 B Technical details B.1 Variance of ˆf dec in the ersmooth case We derive the variance in the case where ϕ K (t) = ( 1 t 2) κ I( 1 t 1), i.e. we take K as

More information

Variance Function Estimation in Multivariate Nonparametric Regression

Variance Function Estimation in Multivariate Nonparametric Regression Variance Function Estimation in Multivariate Nonparametric Regression T. Tony Cai 1, Michael Levine Lie Wang 1 Abstract Variance function estimation in multivariate nonparametric regression is considered

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

The choice of weights in kernel regression estimation

The choice of weights in kernel regression estimation Biometrika (1990), 77, 2, pp. 377-81 Printed in Great Britain The choice of weights in kernel regression estimation BY THEO GASSER Zentralinstitut fur Seelische Gesundheit, J5, P.O.B. 5970, 6800 Mannheim

More information

MATH 118, LECTURES 27 & 28: TAYLOR SERIES

MATH 118, LECTURES 27 & 28: TAYLOR SERIES MATH 8, LECTURES 7 & 8: TAYLOR SERIES Taylor Series Suppose we know that the power series a n (x c) n converges on some interval c R < x < c + R to the function f(x). That is to say, we have f(x) = a 0

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces Reproducing Kernel Hilbert Spaces Lorenzo Rosasco 9.520 Class 03 February 12, 2007 About this class Goal To introduce a particularly useful family of hypothesis spaces called Reproducing Kernel Hilbert

More information

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim An idea how to solve some of the problems 5.2-2. (a) Does not converge: By multiplying across we get Hence 2k 2k 2 /2 k 2k2 k 2 /2 k 2 /2 2k 2k 2 /2 k. As the series diverges the same must hold for the

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

A nonparametric method of multi-step ahead forecasting in diffusion processes

A nonparametric method of multi-step ahead forecasting in diffusion processes A nonparametric method of multi-step ahead forecasting in diffusion processes Mariko Yamamura a, Isao Shoji b a School of Pharmacy, Kitasato University, Minato-ku, Tokyo, 108-8641, Japan. b Graduate School

More information

Asymptotics of Integrals of. Hermite Polynomials

Asymptotics of Integrals of. Hermite Polynomials Applied Mathematical Sciences, Vol. 4, 010, no. 61, 04-056 Asymptotics of Integrals of Hermite Polynomials R. B. Paris Division of Complex Systems University of Abertay Dundee Dundee DD1 1HG, UK R.Paris@abertay.ac.uk

More information

5.4 Bessel s Equation. Bessel Functions

5.4 Bessel s Equation. Bessel Functions SEC 54 Bessel s Equation Bessel Functions J (x) 87 # with y dy>dt, etc, constant A, B, C, D, K, and t 5 HYPERGEOMETRIC ODE At B (t t )(t t ), t t, can be reduced to the hypergeometric equation with independent

More information

be the set of complex valued 2π-periodic functions f on R such that

be the set of complex valued 2π-periodic functions f on R such that . Fourier series. Definition.. Given a real number P, we say a complex valued function f on R is P -periodic if f(x + P ) f(x) for all x R. We let be the set of complex valued -periodic functions f on

More information

MATHEMATICAL FORMULAS AND INTEGRALS

MATHEMATICAL FORMULAS AND INTEGRALS MATHEMATICAL FORMULAS AND INTEGRALS ALAN JEFFREY Department of Engineering Mathematics University of Newcastle upon Tyne Newcastle upon Tyne United Kingdom Academic Press San Diego New York Boston London

More information

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Han Liu John Lafferty Larry Wasserman Statistics Department Computer Science Department Machine Learning Department Carnegie Mellon

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

The Laplace driven moving average a non-gaussian stationary process

The Laplace driven moving average a non-gaussian stationary process The Laplace driven moving average a non-gaussian stationary process 1, Krzysztof Podgórski 2, Igor Rychlik 1 1 Mathematical Sciences, Mathematical Statistics, Chalmers 2 Centre for Mathematical Sciences,

More information

6 Lecture 6b: the Euler Maclaurin formula

6 Lecture 6b: the Euler Maclaurin formula Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Fall 217 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 217 March 26, 218 6 Lecture 6b: the Euler Maclaurin formula

More information

Polynomial Solutions of Nth Order Nonhomogeneous Differential Equations

Polynomial Solutions of Nth Order Nonhomogeneous Differential Equations Polynomial Solutions of th Order onhomogeneous Differential Equations Lawrence E. Levine Ray Maleh Department of Mathematical Sciences Stevens Institute of Technology Hoboken,J 07030 llevine@stevens-tech.edu

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

A COMPLETE ASYMPTOTIC SERIES FOR THE AUTOCOVARIANCE FUNCTION OF A LONG MEMORY PROCESS. OFFER LIEBERMAN and PETER C. B. PHILLIPS

A COMPLETE ASYMPTOTIC SERIES FOR THE AUTOCOVARIANCE FUNCTION OF A LONG MEMORY PROCESS. OFFER LIEBERMAN and PETER C. B. PHILLIPS A COMPLETE ASYMPTOTIC SERIES FOR THE AUTOCOVARIANCE FUNCTION OF A LONG MEMORY PROCESS BY OFFER LIEBERMAN and PETER C. B. PHILLIPS COWLES FOUNDATION PAPER NO. 1247 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS

More information

Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain.

Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain. Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain. For example f(x) = 1 1 x = 1 + x + x2 + x 3 + = ln(1 + x) = x x2 2

More information

On integral representations of q-gamma and q beta functions

On integral representations of q-gamma and q beta functions On integral representations of -gamma and beta functions arxiv:math/3232v [math.qa] 4 Feb 23 Alberto De Sole, Victor G. Kac Department of Mathematics, MIT 77 Massachusetts Avenue, Cambridge, MA 239, USA

More information

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces Reproducing Kernel Hilbert Spaces Lorenzo Rosasco 9.520 Class 03 February 11, 2009 About this class Goal To introduce a particularly useful family of hypothesis spaces called Reproducing Kernel Hilbert

More information

Homework and Computer Problems for Math*2130 (W17).

Homework and Computer Problems for Math*2130 (W17). Homework and Computer Problems for Math*2130 (W17). MARCUS R. GARVIE 1 December 21, 2016 1 Department of Mathematics & Statistics, University of Guelph NOTES: These questions are a bare minimum. You should

More information

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Robert V. Breunig Centre for Economic Policy Research, Research School of Social Sciences and School of

More information

Power Series Solutions to the Legendre Equation

Power Series Solutions to the Legendre Equation Department of Mathematics IIT Guwahati The Legendre equation The equation (1 x 2 )y 2xy + α(α + 1)y = 0, (1) where α is any real constant, is called Legendre s equation. When α Z +, the equation has polynomial

More information

Positive data kernel density estimation via the logkde package for R

Positive data kernel density estimation via the logkde package for R Positive data kernel density estimation via the logkde package for R Andrew T. Jones 1, Hien D. Nguyen 2, and Geoffrey J. McLachlan 1 which is constructed from the sample { i } n i=1. Here, K (x) is a

More information

INTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES

INTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES INTRODUCTION TO REAL ANALYSIS II MATH 433 BLECHER NOTES. As in earlier classnotes. As in earlier classnotes (Fourier series) 3. Fourier series (continued) (NOTE: UNDERGRADS IN THE CLASS ARE NOT RESPONSIBLE

More information

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity Songnian Chen a, Xun Lu a, Xianbo Zhou b and Yahong Zhou c a Department of Economics, Hong Kong University

More information

Study # 1 11, 15, 19

Study # 1 11, 15, 19 Goals: 1. Recognize Taylor Series. 2. Recognize the Maclaurin Series. 3. Derive Taylor series and Maclaurin series representations for known functions. Study 11.10 # 1 11, 15, 19 f (n) (c)(x c) n f(c)+

More information

The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1

The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1 Journal of Theoretical Probability. Vol. 10, No. 1, 1997 The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1 Jan Rosinski2 and Tomasz Zak Received June 20, 1995: revised September

More information

Closest Moment Estimation under General Conditions

Closest Moment Estimation under General Conditions Closest Moment Estimation under General Conditions Chirok Han and Robert de Jong January 28, 2002 Abstract This paper considers Closest Moment (CM) estimation with a general distance function, and avoids

More information

Numerical Analysis: Interpolation Part 1

Numerical Analysis: Interpolation Part 1 Numerical Analysis: Interpolation Part 1 Computer Science, Ben-Gurion University (slides based mostly on Prof. Ben-Shahar s notes) 2018/2019, Fall Semester BGU CS Interpolation (ver. 1.00) AY 2018/2019,

More information

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES

SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Statistica Sinica 19 (2009), 71-81 SMOOTHED BLOCK EMPIRICAL LIKELIHOOD FOR QUANTILES OF WEAKLY DEPENDENT PROCESSES Song Xi Chen 1,2 and Chiu Min Wong 3 1 Iowa State University, 2 Peking University and

More information

More on Estimation. Maximum Likelihood Estimation.

More on Estimation. Maximum Likelihood Estimation. More on Estimation. In the previous chapter we looked at the properties of estimators and the criteria we could use to choose between types of estimators. Here we examine more closely some very popular

More information

Section Taylor and Maclaurin Series

Section Taylor and Maclaurin Series Section.0 Taylor and Maclaurin Series Ruipeng Shen Feb 5 Taylor and Maclaurin Series Main Goal: How to find a power series representation for a smooth function us assume that a smooth function has a power

More information

MATH 5640: Fourier Series

MATH 5640: Fourier Series MATH 564: Fourier Series Hung Phan, UMass Lowell September, 8 Power Series A power series in the variable x is a series of the form a + a x + a x + = where the coefficients a, a,... are real or complex

More information

Math 341: Probability Seventeenth Lecture (11/10/09)

Math 341: Probability Seventeenth Lecture (11/10/09) Math 341: Probability Seventeenth Lecture (11/10/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams

More information

Optimal Polynomial Control for Discrete-Time Systems

Optimal Polynomial Control for Discrete-Time Systems 1 Optimal Polynomial Control for Discrete-Time Systems Prof Guy Beale Electrical and Computer Engineering Department George Mason University Fairfax, Virginia Correspondence concerning this paper should

More information

Smooth functions and local extreme values

Smooth functions and local extreme values Smooth functions and local extreme values A. Kovac 1 Department of Mathematics University of Bristol Abstract Given a sample of n observations y 1,..., y n at time points t 1,..., t n we consider the problem

More information

Density Estimation (II)

Density Estimation (II) Density Estimation (II) Yesterday Overview & Issues Histogram Kernel estimators Ideogram Today Further development of optimization Estimating variance and bias Adaptive kernels Multivariate kernel estimation

More information

PARTIAL DIFFERENTIAL EQUATIONS. Lecturer: D.M.A. Stuart MT 2007

PARTIAL DIFFERENTIAL EQUATIONS. Lecturer: D.M.A. Stuart MT 2007 PARTIAL DIFFERENTIAL EQUATIONS Lecturer: D.M.A. Stuart MT 2007 In addition to the sets of lecture notes written by previous lecturers ([1, 2]) the books [4, 7] are very good for the PDE topics in the course.

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

COURSE Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method

COURSE Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method COURSE 7 3. Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method The presence of derivatives in the remainder difficulties in applicability to practical problems

More information

arxiv: v1 [stat.me] 25 Mar 2019

arxiv: v1 [stat.me] 25 Mar 2019 -Divergence loss for the kernel density estimation with bias reduced Hamza Dhaker a,, El Hadji Deme b, and Youssou Ciss b a Département de mathématiques et statistique,université de Moncton, NB, Canada

More information

The Simplex Method: An Example

The Simplex Method: An Example The Simplex Method: An Example Our first step is to introduce one more new variable, which we denote by z. The variable z is define to be equal to 4x 1 +3x 2. Doing this will allow us to have a unified

More information

Computation Of Asymptotic Distribution. For Semiparametric GMM Estimators. Hidehiko Ichimura. Graduate School of Public Policy

Computation Of Asymptotic Distribution. For Semiparametric GMM Estimators. Hidehiko Ichimura. Graduate School of Public Policy Computation Of Asymptotic Distribution For Semiparametric GMM Estimators Hidehiko Ichimura Graduate School of Public Policy and Graduate School of Economics University of Tokyo A Conference in honor of

More information

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b) Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)

More information

Bessel s and legendre s equations

Bessel s and legendre s equations Chapter 12 Bessel s and legendre s equations 12.1 Introduction Many linear differential equations having variable coefficients cannot be solved by usual methods and we need to employ series solution method

More information

EXPANSION OF ANALYTIC FUNCTIONS IN TERMS INVOLVING LUCAS NUMBERS OR SIMILAR NUMBER SEQUENCES

EXPANSION OF ANALYTIC FUNCTIONS IN TERMS INVOLVING LUCAS NUMBERS OR SIMILAR NUMBER SEQUENCES EXPANSION OF ANALYTIC FUNCTIONS IN TERMS INVOLVING LUCAS NUMBERS OR SIMILAR NUMBER SEQUENCES PAUL F. BYRD San Jose State College, San Jose, California 1. INTRODUCTION In a previous article [_ 1J, certain

More information

A note on multivariate Gauss-Hermite quadrature

A note on multivariate Gauss-Hermite quadrature A note on multivariate Gauss-Hermite quadrature Peter Jäckel 1th May 5 1 Introduction Gaussian quadratures are an ingenious way to approximate the integral of an unknown function f(x) over a specified

More information

Math 308 Exam I Practice Problems

Math 308 Exam I Practice Problems Math 308 Exam I Practice Problems This review should not be used as your sole source of preparation for the exam. You should also re-work all examples given in lecture and all suggested homework problems..

More information

MA22S3 Summary Sheet: Ordinary Differential Equations

MA22S3 Summary Sheet: Ordinary Differential Equations MA22S3 Summary Sheet: Ordinary Differential Equations December 14, 2017 Kreyszig s textbook is a suitable guide for this part of the module. Contents 1 Terminology 1 2 First order separable 2 2.1 Separable

More information

THE HYPERBOLIC METRIC OF A RECTANGLE

THE HYPERBOLIC METRIC OF A RECTANGLE Annales Academiæ Scientiarum Fennicæ Mathematica Volumen 26, 2001, 401 407 THE HYPERBOLIC METRIC OF A RECTANGLE A. F. Beardon University of Cambridge, DPMMS, Centre for Mathematical Sciences Wilberforce

More information

MTH310 EXAM 2 REVIEW

MTH310 EXAM 2 REVIEW MTH310 EXAM 2 REVIEW SA LI 4.1 Polynomial Arithmetic and the Division Algorithm A. Polynomial Arithmetic *Polynomial Rings If R is a ring, then there exists a ring T containing an element x that is not

More information

Kernel Density Estimation

Kernel Density Estimation Kernel Density Estimation and Application in Discriminant Analysis Thomas Ledl Universität Wien Contents: Aspects of Application observations: 0 Which distribution? 0?? 0.0 0. 0. 0. 0.0 0. 0. 0 0 0.0

More information

Two special equations: Bessel s and Legendre s equations. p Fourier-Bessel and Fourier-Legendre series. p

Two special equations: Bessel s and Legendre s equations. p Fourier-Bessel and Fourier-Legendre series. p LECTURE 1 Table of Contents Two special equations: Bessel s and Legendre s equations. p. 259-268. Fourier-Bessel and Fourier-Legendre series. p. 453-460. Boundary value problems in other coordinate system.

More information