Kernel Density Based Linear Regression Estimate

Size: px
Start display at page:

Download "Kernel Density Based Linear Regression Estimate"

Transcription

1 Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency compared to te maximum likeliood estimate (MLE. In tis article, we propose a kernel density based regression estimate (KDRE tat is adaptive to te unknown error distribution. Te key idea is to approximate te likeliood function by using a nonparametric kernel density estimate of te error density based on some initial parameter estimate. Te proposed estimate is sown to be asymptotically as efficient as te oracle MLE wic assumes te error density were known. In addition, we propose an EM type algoritm to maximize te estimated likeliood function and sow tat te KDRE can be considered as an iterated weigted least squares estimate, wic provides us some insigts on te adaptiveness of KDRE to te unknown error distribution. Our Monte Carlo simulation studies sow tat, wile comparable to te traditional LSE for normal errors, te proposed estimation procedure can ave substantial efficiency gain for non-normal errors. Moreover, te efficiency gain can be acieved even for a small sample size. Key words: EM algoritm, Kernel density estimate, Least squares estimate, Linear regression, Maximum likeliood estimate. Department of Statistics, Kansas State University, Manattan, KS wxyao@ksu.edu Department of Statistics, Te Pennsylvania State University, University Park, PA zuz13@stat.psu.edu 1

2 1 Introduction Linear regression models are widely used to investigate te relationsip between several variables. Suppose (x 1, y 1,..., (x n, y n are sampled from te regression model y = x T β + ϵ, (1.1 were x is a p-dimensional vector of covariates independent of te error ϵ wit E(ϵ = 0. Te well-known least squares estimate (LSE of β is β = argmin β (y i x T i β 2. (1.2 For normally distributed errors, β is exactly te maximum likeliood estimate (MLE. However, β will lose some efficiency wen te error is not normally distributed. Terefore, it is desirable to ave an estimate tat can be adaptive to te unknown error distribution. Te idea of adaptiveness is not new. Beran (1974 and Stone (1975 considered adaptive estimation for location model. Bickel (1982, Scick (1993, Yuan and De Gooijer (2007, and Yuan (2010 extended te adaptive idea to regression and some oter models. Linton and Xiao (2007 furter applied te adaptive idea to nonparametric regression estimate. Wang and Yao (2012 applied te adaptive idea to dimension reduction. Empirical likeliood tecniques (Owen, 1988, 2001 ave also been used for regression problems to adaptively construct te confidence intervals and testing statistics witout any parametric assumption for te error density. However, empirical likeliood regression can t provide te efficient point regression estimates by adaptively using te unknown error density information. In tis article, we propose an adaptive kernel density based regression estimate (KDRE. Te basic idea is to estimate te error density by kernel density estimate based on some initial parameter estimate and ten estimate te regression parameters by maximizing te estimated likeliood function. Our proposed estimation procedure uses similar kernel error idea of Stone (1975 and Linton and Xiao (2007 to gain te adaptiveness based on some initial consistent estimate. However, Linton and Xiao (2007 mainly deals wit nonparametric regression, te current paper deals wit 2

3 parametric regression. We prove tat te proposed estimate is asymptotically as efficient as te oracle MLE, wic assumes te error density were known. Terefore, our proposed estimate can adapt to different error distributions. In addition, we propose a novel EM algoritm to maximize te estimated likeliood function and sow tat te KDRE can be viewed as an iterated weigted least squares estimate, wic provides us some insigts on wy te KDRE can adapt to te unknown error distribution. To examine te finite sample performance, we conduct a Monte Carlo simulation study based on a wide range of error densities, including eavy-tail error, multiple-modal error, and skewed error density. Our simulation study confirms our teoretical finding. Our main claims are as follows. 1. Te KDRE is comparable to te traditional LSE wen te error is normal. 2. Te KDRE is more efficient tan te LSE wen te error is not normal. Te efficiency gain can be substantial even for a small sample size. Te remainder of tis paper is organized as follows. In section 2, we introduce te new estimation procedure and prove its asymptotic oracle property. In addition, an EM type algoritm is introduced to maximize te estimated likeliood function. Numerical comparisons are conducted in Section 3. Summary and discussion are given in Section 4. Tecnical proofs are gatered in te Appendix. 2 Kernel Density Based Regression Estimate 2.1 Te new estimation metod Let f(t be te marginal density of ϵ in (1.1. If f(t is known, instead of using te LSE, we can better estimate β in (1.1 by maximizing te log-likeliood log f(y i x T i β. (2.1 In practice, owever, f(t is often unknown and tus (2.1 is not directly applicable. To attenuate tis, denote by β an initial estimate of β, suc as te LSE in (1.2. Based on te residuals ϵ i = y i x T β, i we can estimate f(t by te kernel density 3

4 estimate, denoted by f(t, as f(t = 1 n K (t ϵ j, (2.2 j=1 were K (t = 1 K(t/, K( is a kernel density, and is te tuning parameter. In tis article, we use te Gaussian kernel for K(. Replacing f( in (2.1 wit f(, we ten propose te kernel density based regression parameter estimate (KDRE as were Q(β is te estimated likeliood function Q(β = log f(y i x T i β = ˆβ = argmax Q(β, (2.3 β log { 1 n ( } K yi x T i β ϵ j. (2.4 Here we use leave-one-out kernel density estimate for f(ϵ i to remove te estimation bias; see also Yuan and Gooijer (2007 and Linton and Xiao (2007. j i Te above estimation procedure can be easily extended to te nonlinear regression by replacing x T i β in (2.4 wit te assumed nonlinear function. 2.2 Asymptotic result Let β 0 be te true value of β. Ten we ave te following asymptotic oracle results for our proposed estimate ˆβ. Teorem 2.1. Assume tat Assumptions C1 C5 in te Appendix old. As n, n(ˆβ β0 d N ( 0, V 1, (2.5 were ϵ = y x T β 0, V = I β0 M, 1 M = lim n n { } f x i x T i = E(xx T (ϵ 2, I β0 = E. (2.6 f(ϵ 2 Remark 1: By te above teorem, te proposed estimate ˆβ in (2.4 as root n 4

5 convergence rate and its asymptotic distribution does not depend on te kernel K( or te bandwidt, altoug te kernel density estimator wit slower convergence rate is involved. In addition, ˆβ as te same asymptotic variance as tat of te infeasible oracle MLE, wic assumes f( were known. Remark 2: In (2.4, if we replace te objective function log f( by anoter objective function, say ρ( wit E {ρ (ϵ} = 0 (te LSE corresponds to ρ(ϵ = ϵ 2, ten te resulting estimate as limiting variance v ρ = [ ] E{ρ (ϵ 2 1 } E{ρ (ϵ} M. 2 Based on te classical Cramér-Rao inequality tat [ E{ρ (ϵ 2 } E{ρ (ϵ} 2 ] 1 I 1 β 0, we ave v ρ [I β0 M] 1. Terefore, te objective functions we used in (2.4 is optimal in te sense tat te proposed estimate is asymptotically efficient. Remark 3: Our proposed metod can also be applied to nonlinear regression model and similar oracle properties can also be establised as in Teorem 2.1. Remark 4: Yuan and De Gooijer (2007 proposed estimating β by maximizing log [ 1 n { K r(yi x T i β r(y j x T j β }], (2.7 j i were r( is some monotone nonlinear function, suc as r(z = e z /(1 + e z. Here r( is used to avoid te cancelation of te intercept term in β. Note tat te asymptotic variance in (2.5 is te same as tat in Yuan and De Gooijer (2007 wit r(t = t, wic is efficient. One main advantage of teir metod is tat it does not require an initial estimate. However, te asymptotic variance of teir estimator depends on te coice of r( and generally does not reac te Cramér-Rao lower bound [I β0 M] 1 for a nonlinear function of r(. Note tat wen r(t = t in (2.7, altoug te intercept term, denoted by β 0, will be canceled, te slope parameter, denoted by β 1, will remain estimable. Let β 1 be 5

6 its estimate. In (2.5, let V 1 = ( V 11 V 12 V 21 V 22, were V 11 is a scalar. Based on te result of Yuan and De Gooijer (2007, we know tat β 1 is still an efficient estimate and as te asymptotic distribution n( β1 β 1 d N ( 0, V 22. Let x = (1, x T T. Based on te slope estimate β 1, we can simply estimate β 0 by β 0 = Ȳ x i T β1. Note tat β 0 can be considered as an LSE for model y i x i T β1 = β 0 + ϵ i after we fix β 1 at β 1. Denote by KDRE1 te resulting estimate ( β 0, β 1. Based on some standard calculations (te sketcy of te proof is given at te end of Appendix, we can get te asymptotic distribution for β 0 : n( β0 β 0 d N(0, σ 2, were [ σ 2 = var ϵ i f (ϵ i { } ] E(x T V 21 + E(x T V 22 x i. f(ϵ i Note tat generally β 0 does not reac te Cramér-Rao lower bound and te efficiency loss depends on te true error density f(ϵ. However, one nice feature of suc estimate is tat it doesn t require an initial estimate. In addition, it does not require to coose a nonlinear function r(. 2.3 Computations: an EM algoritm Note tat te objective function (2.4 as a mixture form. In tis section, we propose an EM algoritm to maximize it. Te proposed EM algoritm can be similarly used to find β 1 by maximizing (2.7 wen r(t = t. Let β (0 be an initial parameter estimate, suc as te LSE. We ten update te parameter estimate according to te algoritm below. Algoritm 2.1. At (k + 1t step, we calculate te following E and M steps: 6

7 E-Step: Calculate te classification probabilities, p (k+1 ij = K (y i x T i β (k ϵ j l i K (y i x T i β(k ϵ l K (y i x T i β (k ϵ j, j i, (2.8 M-Step: Update β (k+1, β (k+1 = argmax β = argmin β { } p (k+1 ij log K (y i x T i β ϵ j j i { } p (k+1 ij (y i x T i β ϵ j 2, (2.9 j i wic as explicit solutions, since K ( is a Gaussian kernel density. From te M step (2.9, te KDRE can be considered as a weigted least squares estimate, wic minimizes te weigted squared difference between te new residual y i x T i β and te initial residual ϵ j for all 1 i j n. Based on te weigts in (2.8, one knows tat if jt observation is an isolated outlier (i.e., ϵ j is large, ten te weigts p (k+1 ij will be also small. will be small for i j and tus te effect of ϵ j on updating β (k+1 By Teorem 2.2 below, te Algoritm 2.1 is truly an EM algoritm and as te monotone property for te objective function (2.4. Teorem 2.2. Te objective function (2.4 is non-decreasing after eac iteration of Algoritm 2.1, i.e., Q(β (k+1 Q(β (k, until a fixed point is reaced. 3 Simulation Studies In tis section, we use a simulation study to compare te proposed KDRE and KDRE1 wit te traditional LSE for linear regression models wit different types of error densities. For te proposed estimate, we use te rule-of-tumb bandwidt = 1.06n 1/5ˆσ for te kernel density estimate of f(ϵ, were ˆσ is te sample standard deviation of te initial residual ϵ i = y i x T β i and β is te LSE. Better estimates migt be obtained if we use some more sopisticated bandwidt for kernel density estimate. See, for example, Seater and Jones (1991 and Raykar and Duraiswami (2006. In addition, 7

8 we can also use cross validation metod to selection te bandwidt, wic focuses on te performance of regression estimate directly instead of density estimate. We generate independent and identically distributed data {(x i, y i, i = 1,..., n} from te model Y = 1 + 3X + ϵ, were X U(0, 1, te uniform distribution on [0, 1]. For te error density, we consider te following six coices (all ave standard deviation around 1: Case 1: ϵ N(0, 1, normal error. Case 2: ϵ U( 2, 2, te uniform distribution on [ 2, 2], sort-tail error. Case 3: ϵ t 3 / 3, t-distribution wit 3 degrees of freedom, eavy-tail error. Case 4: ϵ 0.95N(0, N(0, 3.5 2, contaminated normal error. Te 5% data from N(0, are most likely to be outliers. Case 5: ϵ 0.5N( 1, N(1, 0.5 2, multi-modal error. Case 6: ϵ 0.3N( 1.4, N(0.6, 0.4 2, skewed error. Here, we also used te Case 6 to ceck ow our metod performed compared wit LSE wen te error is not symmetric. We estimate te regression parameters using KDRE, KDRE1, and te traditional LSE. Based on 1000 replicates, Tables 1 2 report te mean squared errors (MSE of te parameter estimates for intercept and slope, respectively, for sample size n = 30, 100, 300, and 600. Te rigtmost two columns contain te relative efficiency of KDRE and KDRE1 wen compared to te LSE. For example, RE(KDRE=MSE(LSE/MSE(KDRE. From te Case 2 to Case 6 in Tables 1 2, we can see tat KDRE and KDRE1 are muc more efficient tan te LSE wen te error is not normal (for bot symmetric and skewed error densities. Moreover, te efficiency gain can be substantial even for a small sample size. In addition, wen te error is normal, KDRE is comparable to te LSE and works better tan KDRE1 especially for small sample size. However, for Case 6 skewed error densities, KDRE1 works better tan KDRE, altoug bot of tem ave muc better performance tan LSE. In addition, for large sample size, te performances of KDRE and KDRE1 are 8

9 almost te same, even for intercept estimate, altoug KDRE as some teoretical advantage over KDRE1. Note tat KDRE1 is simpler witout first estimating te error data. 4 Summary In tis article, we proposed an adaptive linear regression estimate by maximizing an estimated likeliood function, in wic te error density is estimated by kernel density estimate. Te proposed estimate can adapt to unknown error density and is asymptotically equivalent to te oracle MLE. Using te proposed EM algoritm, te computation is quick and stable. Our extensive simulation studies sow tat te proposed metod outperforms te LSE in te presence of non-normal errors. Altoug developed for linear regression models, te same idea can be easily extended to nonlinear regression cases. Te asymptotic oracle property follows similarly. In addition, our proposed EM algoritm can be also used to estimate te adaptive nonparametric regression of Linton and Xiao (2007 and te semiparametric regression of Yuan and De Gooijer (2007. Future researc directions include extensions to oter regression models suc as varying coefficient partially linear models and nonparametric additive models. 5 Appendix: Proofs Te following conditions are imposed to facilitate te proof. C1. {ϵ i } and {x i } are i.i.d. and mutually independent wit E(ϵ i = 0, E( ϵ i 3 <. Additionally, te predictors x i ave bounded support and. C2. Te density f( of ϵ is symmetric about 0 and as bounded continuous derivatives up to order 4. Let l(ϵ = log f(ϵ. Assume E{l (ϵ 2 + l (ϵ + l (ϵ } <. C3. Te kernel K( is symmetric, as bounded support, and are four times continuously differentiable. C4. As n, n 4 and n

10 C5. For te initial estimate β of β 0, assume β β 0 = O p (n 1/2. Te condition C1 can guarantee tat te least squares estimate is consistent and as root n convergence rate. Te condition C2 is used to guarantee te adaptiveness of our proposed estimate. If lim n n 1 n x i = 0, ten te symmetric condition of f(ϵ can be removed. 5.1 Proof of Teorem 2.1 We follow a similar strategy in Linton and Xiao (2007. Note tat te maximizer ˆβ in (2.3 is te solution of te score function 1 n f (y i x T i β f(y i x T i β x i, (5.1 were f (t is te derivative of f(t in (2.2. For tecnical reason, we will consider anoter trimmed version of ˆβ as te solution of S(β = 0, were S(β = 1 n f (y i x T i β f(y i x T i β x ig b ( f(ϵ i. (5.2 Here 0, x < b; G b (x = x b g b(tdt, b x 2b; 1, x > 2b. were g b (t is any density function wit support on [b, 2b] suc tat G b (t is four times continuously differentiable on [b, 2b]. In te following proof, we assume tat b = r, were 0 < r < 1/2. In practice, wen b is small, te difference between te original estimate and te trimmed one is negligible. By Taylor s expansion, tere exists β suc tat β β 0 ˆβ β 0 and S(ˆβ = S(β 0 + S(β 0 β (ˆβ β (ˆβ β 0 T 2 S(β β β T (ˆβ β 0. Te desired result ten follows from Lemmas below. 10

11 Lemma 5.1. For f in (2.2, we ave te uniform consistency results sup f(t f(t = O p [ 2 + t sup f (t f (t = O p [ 2 + t { } ] 1/2 log(n, (5.3 n { } ] 1/2 log(n. (5.4 n 3 Proof. Denote by f (k te kt derivative of f wit te convention f (0 = f. Let ˇf (k (t = 1 n k+1 j=1 ( t K (k ϵj, k = 0, 1, 2, 3, be te traditional kernel density derivative estimator of f (k (. By Silverman (1978, { } } 1/2 sup ˇf log(n (k (t f (k (t = O p { 2 +. (5.5 t n 2k+1 Since x i as bounded support and β β 0 = O p (n 1/2, ϵ j ϵ j = x T j (β 0 β = O p (n 1/2, uniformly over j. By Taylor s expansion, for some ϵ j between ϵ j and ϵ j, f(t ˇf(t = 1 n 2 j + 1 6n 4 ( t K ϵj (ϵ j ϵ j + 1 j K ( t ϵ j (ϵ j ϵ j 3 2n 3 j ( t K ϵj (ϵ j ϵ j 2 =O p (1/ n + O p (1/nO p {1 + log(n/(n 5 } + O p (1/n 3/2 O p (1/ 4, uniformly, entailing (5.3 via Condition C4 and (5.5. Similarly, (5.4 follows. Lemma 5.2. Let V be defined as in (2.6. Ten S(β 0 / β p V. Proof. For notational convenience we write f i = f(ϵ i, f i = f (ϵ i, f i = f (ϵ i, f i = f(ϵ i, f i = f (ϵ i, = f (ϵ i. Note tat f i S(β 0 β = 1 n f i 2 G f i 2 b ( f i x i x T i + 1 n f i f i G b ( f i x i x T i + 1 n f 2 i f i g b ( f i x i x T i = A + B + C. 11

12 It suffices to prove A p V, B p 0, and C p 0. First, we consider A. Let i = f i f i, i = f i f i, δ n = 2 + log(n/(n, and δ n = 2 + log(n/(n 3. By Lemma 5.1, max i i = O p (δ n and max i i = O p (δ n. By definition, sup x G b (x/x k 1/b k, k 0. So, by te boundedness of f i, f i, { f i 2 G f i 2 b ( f f i 2 i = fi 2 + i( f i + f i f 2 i + if i 2 (f i + f } i fi 2 f G i 2 b ( f i = f i 2 G fi 2 b ( f i + O p(δ n + O p(δ n f 2 i. b 2 b 2 fi 2 By Condition C2, f 2 i /f 2 i is integrable, so we ave A = 1 n f i 2 G fi 2 b ( f i x i x T i + O p ( δ n b 2. (5.6 By Condition C2 and te Dominated Convergence Teorem, as b 0, { f 2 E i f 2 i } { f 2 (1 G b (f i E i f 2 i } I(f(ϵ i < 2b = o(1. Note tat max 1 i n G b ( f i G b (f i = o p (1. Terefore, by decomposing G b ( f i in (5.6 into 1 + {G b (f i 1} + {G b ( f i G b (f i }, it is easily seen tat A p V. Next, we consider B. Tere exists ξ between 0 and ( f f/f suc tat { } f 1 (ϵ = f 1 (ϵ (1 + ξ 2 f 2 (ϵ f(ϵ f(ϵ. (5.7 Using te latter identity, we ave B = 1 f (ϵ i n f(ϵ i G b( f i x i x T i + 1 f (ϵ i f (ϵ i G b ( n f(ϵ i f i x i x T i 1 { f(ϵ i f(ϵ i } f (ϵ i G n f(ϵ i 2 (1 + ξ i 2 b ( f i x i x T i =B 1 + B 2 + B 3. (5.8 12

13 Similar to te proof of A in (5.6, we can get B 1 = o p (1. Note tat B 2 1 n f(ϵ i j=1 ( K ϵi ϵ ( j x T j β x T j β 0 G b (f i x i x T i. Elementary calculations sow tat { ( } t ϵ E K (k = k+1 K(zf (k (t + zdz, k = 1, 2, 3. Let k 1 (ϵ i, ϵ j = f(ϵ i K It can be easily sown tat, for distinct i, j, k, l, ( ϵi ϵ j G b (f i. E {k 1 (ϵ i, ϵ j } = O(b 1, E { k 2 1(ϵ i, ϵ j } = O(b 2 7, E {k 1 (ϵ i, ϵ j k 1 (ϵ i, ϵ l } = O(b 2 Tus, calculating te first two moments based on te result of U-statistics, we ave B 2 = O p (1/ n O p (b 1 O p (1/ n 2 b / nb 2 = o p (1. Tat B 3 = o p (1 follows from { max f(ϵ [ i f(ϵ i } 1 i n f(ϵ i 2 (1 + ξ i G b( f 2 i = O p 2 + Finally, we consider C. Note tat { } ] 1/2 1 n log(1/ b 2 = o p (1. C = 1 f (ϵ i 2 n f(ϵ i g b( f i x i x T i + 1 f (ϵ i 2 f (ϵ i 2 g b ( n f(ϵ i f i x i x T i 1 f(ϵ i f(ϵ i n f(ϵ i 2 (1 + ξ i g b( f 2 i x i x T i =C 1 + C 2 + C 3. Based on te uniform convergency results in Lemma 5.1 and g b ( = O(b 1, we can 13

14 easily get C 2 = o p (1 and C 3 = o p (1. By te Dominated Convergence Teorem, { } f (ϵ i 2 E f(ϵ i g b(f i { } f max {g (ϵ i 2 b(xx}e x f 2 (ϵ i I(b f(ϵ i 2b 0, wic, along wit te argument in te proof of A in (5.6, gives C 1 = o p (1. Lemma 5.3. Let V be defined as in (2.6. Ten n S(β 0 Proof. By (5.7, n S(β0 = 1 n 1 n d N(0, V. f (ϵ i f(ϵ i x ig b ( f(ϵ i + 1 f (ϵ i f (ϵ i x i G b ( n f(ϵ i f(ϵ i f (ϵ i f { } (ϵ i f(ϵ f(ϵ x (1 + ξ 2 f(ϵ i 3 i G b ( f(ϵ i =J 1 + J 2 + J 3. By te tecnique in Lemma 5.2 and Lemma S2 of Linton and Xiao (2007, It remains to prove J 2 J 1 = 1 n f (ϵ i f(ϵ i x d i + o p (1 N(0, V. (5.9 p p 0 and J 3 0. Decompose J 2 as J 2 = 1 n = J 21 + J 22. f (ϵ i ˇf (ϵ i f(ϵ i x i G b ( f(ϵ i + 1 n ˇf (ϵ i f (ϵ i x i G b ( f(ϵ i f(ϵ i Note tat (J 21 a 1 n n 3 j=1 =O p (1/ 1 n n n 3 1 f(ϵ i K j=1 ( ϵi ϵ j 1 f(ϵ i K x T j ( β β 0 X ia G b (f(ϵ i ( ϵi ϵ j x T j X ia G b (f(ϵ i. Similar to te proof of B 2 in (5.8, by calculating te first two moments of (J 21 a 14

15 using te results of U-statistics, we ave E{(J 21 a } = O( 2 b 1 and var{(j 21 a } = O(1/ nb 4. Terefore, (J 21 a = o p (1. Note tat J 22 1 n = 1 n + 1 n ˇf (ϵ i f (ϵ i x i G b (f(ϵ i f(ϵ i { ( K ϵ i ϵ j f(ϵ i (n 2 1 n j=1 =J 22A + J 22B, (n 2 1 n j=1 E ik ( ϵ i ϵ j f(ϵ i Ei K ( ϵ i ϵ j } x i G b (f(ϵ i f (ϵ i x i G b (f(ϵ i were E i is te conditional expectation given ϵ i. Similar to te proof of B 2 in (5.8 and te proof tecniques in te Lemma S2 of Linton and Xiao (2007, we can prove E(J 22A = 0 and var{j 22A } = o(1. Terefore J 22A = o p (1. Similarly, we can prove J 22B = o p (1 and J 3 = o p (1. Lemma S(β / β β T = o p ( n. Proof. It follows from te same argument in Lemmas and we omit te details. 5.2 Proof of Teorem 2.2 Let Z (k+1 i P be a random variable suc tat { Z (k+1 i = K ( y i x T i β (k+1 ϵ j /K ( y i x T i β (k ϵ j } = p (k+1 ij, j i. 15

16 By Jensen s inequality, we ave Q(β (k+1 Q(β (k j i K (y i x Ti β (k+1 ϵ j = log j i K (y i x Ti β(k ϵ j ( = log K y i x T i β (k+1 ϵ j p(k+1 ij ( j i K y i x T i β(k ϵ j { } = log E(Z (k+1 i E log(z (k+1 i. By te M-step of Algoritm 2.1, te desired result follows from { } E log(z (k+1 i = j i p (k+1 ij ( K log y i x T i β (k+1 ϵ j ( K y i x T i β(k ϵ j 0. Sketc of te proof of asymptotic distribution of β 0 : Let x = (1, x T T. Note tat β 0 = ȳ x T β1 = β 0 + x T β 1 + ϵ x T β1 = β 0 + x T (β 1 β 1 + ϵ Terefore, n( β0 β 0 = x T n(β 1 β 1 + n ϵ In addition, we know n(β1 β 1 = 1 n f (ϵ i f(ϵ i (V 21 + V 22 x i + o p (1. 16

17 Terefore, n( β0 β 0 = 1 n d N(0, σ 2, { ϵ i f } (ϵ i f(ϵ i ( x T V 21 + x T V 22 x i were [ σ 2 = var ϵ i f (ϵ i { } ] E(x T V 21 + E(x T V 22 x i. f(ϵ i 6 Acknowledgements Te autors are grateful to te editors and te referee for teir insigtful comments and suggestions, wic greatly improved tis article. In addition, te metod of KDRE1 is based on te referee s suggestion. References Beran, R. (1978. Asymptotically efficient adaptive rank estimates in location models. Annals of Statistics, 2, Bickel, P. J. (1982. On adaptive estimation. Annals of Statistics, 10, Linton, O. and Xiao, Z. (2007. A nonparametric regression estimator tat adapts to error distribution of unknown form. Econometric Teory, 23, Owen, A. B. (1988. Empirical likeliood ratio confidence intervals for a single functional. Biometrika, 75, Owen, A. B. (2001. Empirical Likeliood. New York: Capman & Hall/CRC. Raykar, V. C. and Duraiswami, R. (2006. Fast optimal bandwidt selection for kernel density estimation. In proceedings of te sixt SIAM International Conference on Data Mining, Betesda, April 2006, Scick, A. (1993. On efficient estimation in regression models. Annals of Statistics, 21,

18 Seater, S. J. and Jones, M. C. (1991. A reliable data-based bandwidt selection metod for kernel density estimation. Journal of Royal Statistical Society, B, 53, Silverman, B. W. (1978. Weak and strong uniform consistency of te kernel estimate of density and its derivatives. Annals of Statistics, 6, Stone, C. (1975. Adaptive maximum likeliood estimation of a location parameters. Annals of Statistics, 3, Wang, Q. and Yao, W. (2012. An adaptive estimation of MAVE. Journal of Multivariate Analysis, 104, Yuan, A. and De Gooijer, J. G. (2007. Semiparametric regression wit kernel error model. Scandinavian Journal of Statistics, 34, Yuan, A. (2010. Semiparametric inference wit kernel likeliood. Journal of Nonparametric Statistics, 21,

19 Table 1: Simulation Results for te Intercept Estimates. Mean(MSE Error Distribution n LSE KDRE KDRE1 RE(KDRE RE(KDRE N(0, (Standard normal U( 2, (Sort-tail distribution t 3 / (Heavy-tail distribution N(0, N(0, (Contaminated normal N( 1, N(1, (Multi-modal distribution N( 1.4, N(0.6, (Skewed distribution

20 Table 2: Simulation Results for te Slope Estimates. Mean(MSE Error Distribution n LSE KDRE KDRE1 RE(KDRE RE(KDRE N(0, (Standard normal U( 2, (Sort-tail distribution t 3 / (Heavy-tail distribution N(0, N(0, (Contaminated normal N( 1, N(1, (Multi-modal distribution N( 1.4, N(0.6, (Skewed distribution

7 Semiparametric Methods and Partially Linear Regression

7 Semiparametric Methods and Partially Linear Regression 7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).

More information

EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS

EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Statistica Sinica 24 2014, 395-414 doi:ttp://dx.doi.org/10.5705/ss.2012.064 EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Jun Sao 1,2 and Seng Wang 3 1 East Cina Normal University,

More information

Financial Econometrics Prof. Massimo Guidolin

Financial Econometrics Prof. Massimo Guidolin CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis

More information

The Priestley-Chao Estimator

The Priestley-Chao Estimator Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are

More information

Fast optimal bandwidth selection for kernel density estimation

Fast optimal bandwidth selection for kernel density estimation Fast optimal bandwidt selection for kernel density estimation Vikas Candrakant Raykar and Ramani Duraiswami Dept of computer science and UMIACS, University of Maryland, CollegePark {vikas,ramani}@csumdedu

More information

Fast Exact Univariate Kernel Density Estimation

Fast Exact Univariate Kernel Density Estimation Fast Exact Univariate Kernel Density Estimation David P. Hofmeyr Department of Statistics and Actuarial Science, Stellenbosc University arxiv:1806.00690v2 [stat.co] 12 Jul 2018 July 13, 2018 Abstract Tis

More information

Bandwidth Selection in Nonparametric Kernel Testing

Bandwidth Selection in Nonparametric Kernel Testing Te University of Adelaide Scool of Economics Researc Paper No. 2009-0 January 2009 Bandwidt Selection in Nonparametric ernel Testing Jiti Gao and Irene Gijbels Bandwidt Selection in Nonparametric ernel

More information

Chapter 1. Density Estimation

Chapter 1. Density Estimation Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f

More information

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties

More information

DEPARTMENT MATHEMATIK SCHWERPUNKT MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE

DEPARTMENT MATHEMATIK SCHWERPUNKT MATHEMATISCHE STATISTIK UND STOCHASTISCHE PROZESSE U N I V E R S I T Ä T H A M B U R G A note on residual-based empirical likeliood kernel density estimation Birte Musal and Natalie Neumeyer Preprint No. 2010-05 May 2010 DEPARTMENT MATHEMATIK SCHWERPUNKT

More information

On Local Linear Regression Estimation of Finite Population Totals in Model Based Surveys

On Local Linear Regression Estimation of Finite Population Totals in Model Based Surveys American Journal of Teoretical and Applied Statistics 2018; 7(3): 92-101 ttp://www.sciencepublisinggroup.com/j/ajtas doi: 10.11648/j.ajtas.20180703.11 ISSN: 2326-8999 (Print); ISSN: 2326-9006 (Online)

More information

Nonparametric density estimation for linear processes with infinite variance

Nonparametric density estimation for linear processes with infinite variance Ann Inst Stat Mat 2009) 61:413 439 DOI 10.1007/s10463-007-0149-x Nonparametric density estimation for linear processes wit infinite variance Tosio Honda Received: 1 February 2006 / Revised: 9 February

More information

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i NADARAYA WATSON ESTIMATE JAN 0, 2006: version 2 DATA: (x i, Y i, i =,..., n. ESTIMATE E(Y x = m(x by n i= ˆm (x = Y ik ( x i x n i= K ( x i x EXAMPLES OF K: K(u = I{ u c} (uniform or box kernel K(u = u

More information

232 Calculus and Structures

232 Calculus and Structures 3 Calculus and Structures CHAPTER 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS FOR EVALUATING BEAMS Calculus and Structures 33 Copyrigt Capter 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS 17.1 THE

More information

Applications of the van Trees inequality to non-parametric estimation.

Applications of the van Trees inequality to non-parametric estimation. Brno-06, Lecture 2, 16.05.06 D/Stat/Brno-06/2.tex www.mast.queensu.ca/ blevit/ Applications of te van Trees inequality to non-parametric estimation. Regular non-parametric problems. As an example of suc

More information

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Order of Accuracy. ũ h u Ch p, (1)

Order of Accuracy. ũ h u Ch p, (1) Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical

More information

Boosting Kernel Density Estimates: a Bias Reduction. Technique?

Boosting Kernel Density Estimates: a Bias Reduction. Technique? Boosting Kernel Density Estimates: a Bias Reduction Tecnique? Marco Di Marzio Dipartimento di Metodi Quantitativi e Teoria Economica, Università di Cieti-Pescara, Viale Pindaro 42, 65127 Pescara, Italy

More information

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER*

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER* EO BOUNDS FO THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BADLEY J. LUCIE* Abstract. Te expected error in L ) attimet for Glimm s sceme wen applied to a scalar conservation law is bounded by + 2 ) ) /2 T

More information

Basic Nonparametric Estimation Spring 2002

Basic Nonparametric Estimation Spring 2002 Basic Nonparametric Estimation Spring 2002 Te following topics are covered today: Basic Nonparametric Regression. Tere are four books tat you can find reference: Silverman986, Wand and Jones995, Hardle990,

More information

New Distribution Theory for the Estimation of Structural Break Point in Mean

New Distribution Theory for the Estimation of Structural Break Point in Mean New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University

More information

New families of estimators and test statistics in log-linear models

New families of estimators and test statistics in log-linear models Journal of Multivariate Analysis 99 008 1590 1609 www.elsevier.com/locate/jmva ew families of estimators and test statistics in log-linear models irian Martín a,, Leandro Pardo b a Department of Statistics

More information

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example, NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing

More information

Math 161 (33) - Final exam

Math 161 (33) - Final exam Name: Id #: Mat 161 (33) - Final exam Fall Quarter 2015 Wednesday December 9, 2015-10:30am to 12:30am Instructions: Prob. Points Score possible 1 25 2 25 3 25 4 25 TOTAL 75 (BEST 3) Read eac problem carefully.

More information

Poisson Equation in Sobolev Spaces

Poisson Equation in Sobolev Spaces Poisson Equation in Sobolev Spaces OcMountain Dayligt Time. 6, 011 Today we discuss te Poisson equation in Sobolev spaces. It s existence, uniqueness, and regularity. Weak Solution. u = f in, u = g on

More information

Exam 1 Review Solutions

Exam 1 Review Solutions Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),

More information

1. Questions (a) through (e) refer to the graph of the function f given below. (A) 0 (B) 1 (C) 2 (D) 4 (E) does not exist

1. Questions (a) through (e) refer to the graph of the function f given below. (A) 0 (B) 1 (C) 2 (D) 4 (E) does not exist Mat 1120 Calculus Test 2. October 18, 2001 Your name Te multiple coice problems count 4 points eac. In te multiple coice section, circle te correct coice (or coices). You must sow your work on te oter

More information

Kernel Smoothing and Tolerance Intervals for Hierarchical Data

Kernel Smoothing and Tolerance Intervals for Hierarchical Data Clemson University TigerPrints All Dissertations Dissertations 12-2016 Kernel Smooting and Tolerance Intervals for Hierarcical Data Cristoper Wilson Clemson University, cwilso6@clemson.edu Follow tis and

More information

Artificial Neural Network Model Based Estimation of Finite Population Total

Artificial Neural Network Model Based Estimation of Finite Population Total International Journal of Science and Researc (IJSR), India Online ISSN: 2319-7064 Artificial Neural Network Model Based Estimation of Finite Population Total Robert Kasisi 1, Romanus O. Odiambo 2, Antony

More information

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk

More information

A Simple Matching Method for Estimating Sample Selection Models Using Experimental Data

A Simple Matching Method for Estimating Sample Selection Models Using Experimental Data ANNALS OF ECONOMICS AND FINANCE 6, 155 167 (2005) A Simple Matcing Metod for Estimating Sample Selection Models Using Experimental Data Songnian Cen Te Hong Kong University of Science and Tecnology and

More information

Homework 1 Due: Wednesday, September 28, 2016

Homework 1 Due: Wednesday, September 28, 2016 0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

EFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING

EFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING Statistica Sinica 13(2003), 641-653 EFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING J. K. Kim and R. R. Sitter Hankuk University of Foreign Studies and Simon Fraser University Abstract:

More information

Bootstrap confidence intervals in nonparametric regression without an additive model

Bootstrap confidence intervals in nonparametric regression without an additive model Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap

More information

Exponential Concentration for Mutual Information Estimation with Application to Forests

Exponential Concentration for Mutual Information Estimation with Application to Forests Exponential Concentration for Mutual Information Estimation wit Application to Forests Han Liu Department of Operations Researc and Financial Engineering Princeton University, NJ 8544 anliu@princeton.edu

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics 1

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics 1 Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics 1 By Jiti Gao 2 and Maxwell King 3 Abstract We propose a simultaneous model specification procedure for te conditional

More information

A Jump-Preserving Curve Fitting Procedure Based On Local Piecewise-Linear Kernel Estimation

A Jump-Preserving Curve Fitting Procedure Based On Local Piecewise-Linear Kernel Estimation A Jump-Preserving Curve Fitting Procedure Based On Local Piecewise-Linear Kernel Estimation Peiua Qiu Scool of Statistics University of Minnesota 313 Ford Hall 224 Curc St SE Minneapolis, MN 55455 Abstract

More information

Function Composition and Chain Rules

Function Composition and Chain Rules Function Composition and s James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 8, 2017 Outline 1 Function Composition and Continuity 2 Function

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator. Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative

More information

A New Diagnostic Test for Cross Section Independence in Nonparametric Panel Data Model

A New Diagnostic Test for Cross Section Independence in Nonparametric Panel Data Model e University of Adelaide Scool of Economics Researc Paper No. 2009-6 October 2009 A New Diagnostic est for Cross Section Independence in Nonparametric Panel Data Model Jia Cen, Jiti Gao and Degui Li e

More information

(a) At what number x = a does f have a removable discontinuity? What value f(a) should be assigned to f at x = a in order to make f continuous at a?

(a) At what number x = a does f have a removable discontinuity? What value f(a) should be assigned to f at x = a in order to make f continuous at a? Solutions to Test 1 Fall 016 1pt 1. Te grap of a function f(x) is sown at rigt below. Part I. State te value of eac limit. If a limit is infinite, state weter it is or. If a limit does not exist (but is

More information

Solution. Solution. f (x) = (cos x)2 cos(2x) 2 sin(2x) 2 cos x ( sin x) (cos x) 4. f (π/4) = ( 2/2) ( 2/2) ( 2/2) ( 2/2) 4.

Solution. Solution. f (x) = (cos x)2 cos(2x) 2 sin(2x) 2 cos x ( sin x) (cos x) 4. f (π/4) = ( 2/2) ( 2/2) ( 2/2) ( 2/2) 4. December 09, 20 Calculus PracticeTest s Name: (4 points) Find te absolute extrema of f(x) = x 3 0 on te interval [0, 4] Te derivative of f(x) is f (x) = 3x 2, wic is zero only at x = 0 Tus we only need

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Department of Econometrics and Business Statistics

Department of Econometrics and Business Statistics ISSN 1440-771X Australia Department of Econometrics and Business Statistics ttp://www.buseco.monas.edu.au/depts/ebs/pubs/wpapers/ Bayesian Bandwidt Estimation in Nonparametric Time-Varying Coefficient

More information

Mathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative

Mathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative Matematics 5 Workseet 11 Geometry, Tangency, and te Derivative Problem 1. Find te equation of a line wit slope m tat intersects te point (3, 9). Solution. Te equation for a line passing troug a point (x

More information

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

The Verlet Algorithm for Molecular Dynamics Simulations

The Verlet Algorithm for Molecular Dynamics Simulations Cemistry 380.37 Fall 2015 Dr. Jean M. Standard November 9, 2015 Te Verlet Algoritm for Molecular Dynamics Simulations Equations of motion For a many-body system consisting of N particles, Newton's classical

More information

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to

More information

UNIMODAL KERNEL DENSITY ESTIMATION BY DATA SHARPENING

UNIMODAL KERNEL DENSITY ESTIMATION BY DATA SHARPENING Statistica Sinica 15(2005), 73-98 UNIMODAL KERNEL DENSITY ESTIMATION BY DATA SHARPENING Peter Hall 1 and Kee-Hoon Kang 1,2 1 Australian National University and 2 Hankuk University of Foreign Studies Abstract:

More information

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms More on generalized inverses of partitioned matrices wit anaciewicz-scur forms Yongge Tian a,, Yosio Takane b a Cina Economics and Management cademy, Central University of Finance and Economics, eijing,

More information

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series Lawrence D. Brown, Pilip A. Ernst, Larry Sepp, and Robert Wolpert August 27, 2015 Abstract We consider te class,

More information

f a h f a h h lim lim

f a h f a h h lim lim Te Derivative Te derivative of a function f at a (denoted f a) is f a if tis it exists. An alternative way of defining f a is f a x a fa fa fx fa x a Note tat te tangent line to te grap of f at te point

More information

arxiv: v1 [math.pr] 28 Dec 2018

arxiv: v1 [math.pr] 28 Dec 2018 Approximating Sepp s constants for te Slepian process Jack Noonan a, Anatoly Zigljavsky a, a Scool of Matematics, Cardiff University, Cardiff, CF4 4AG, UK arxiv:8.0v [mat.pr] 8 Dec 08 Abstract Slepian

More information

Uniform Convergence Rates for Nonparametric Estimation

Uniform Convergence Rates for Nonparametric Estimation Uniform Convergence Rates for Nonparametric Estimation Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~bansen October 2004 Preliminary and Incomplete Abstract Tis paper presents a set of rate

More information

Uniform Consistency for Nonparametric Estimators in Null Recurrent Time Series

Uniform Consistency for Nonparametric Estimators in Null Recurrent Time Series Te University of Adelaide Scool of Economics Researc Paper No. 2009-26 Uniform Consistency for Nonparametric Estimators in Null Recurrent Time Series Jiti Gao, Degui Li and Dag Tjøsteim Te University of

More information

ch (for some fixed positive number c) reaching c

ch (for some fixed positive number c) reaching c GSTF Journal of Matematics Statistics and Operations Researc (JMSOR) Vol. No. September 05 DOI 0.60/s4086-05-000-z Nonlinear Piecewise-defined Difference Equations wit Reciprocal and Cubic Terms Ramadan

More information

Differentiation in higher dimensions

Differentiation in higher dimensions Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for Second-Order Elliptic Problems

Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for Second-Order Elliptic Problems Applied Matematics, 06, 7, 74-8 ttp://wwwscirporg/journal/am ISSN Online: 5-7393 ISSN Print: 5-7385 Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for

More information

Bootstrap prediction intervals for Markov processes

Bootstrap prediction intervals for Markov processes arxiv: arxiv:0000.0000 Bootstrap prediction intervals for Markov processes Li Pan and Dimitris N. Politis Li Pan Department of Matematics University of California San Diego La Jolla, CA 92093-0112, USA

More information

MAT 145. Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points

MAT 145. Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points MAT 15 Test #2 Name Solution Guide Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points Use te grap of a function sown ere as you respond to questions 1 to 8. 1. lim f (x) 0 2. lim

More information

158 Calculus and Structures

158 Calculus and Structures 58 Calculus and Structures CHAPTER PROPERTIES OF DERIVATIVES AND DIFFERENTIATION BY THE EASY WAY. Calculus and Structures 59 Copyrigt Capter PROPERTIES OF DERIVATIVES. INTRODUCTION In te last capter you

More information

OPTIMAL PREDICTION UNDER ASYMMETRIC LOSS 1. By Peter F. Christoffersen and Francis X. Diebold 2 1. INTRODUCTION

OPTIMAL PREDICTION UNDER ASYMMETRIC LOSS 1. By Peter F. Christoffersen and Francis X. Diebold 2 1. INTRODUCTION OPTIMAL PREDICTION UNDER ASYMMETRIC LOSS 1 By Peter F. Cristoffersen and Francis X. Diebold Cristoffersen P. and Diebold F.X. (1997) "Optimal Prediction Under Asymmetric Loss" Econometric Teory 13 808-817.

More information

WEIGHTED KERNEL ESTIMATORS IN NONPARAMETRIC BINOMIAL REGRESSION

WEIGHTED KERNEL ESTIMATORS IN NONPARAMETRIC BINOMIAL REGRESSION WEIGHTED KEREL ESTIMATORS I OPARAMETRIC BIOMIAL REGRESSIO Hidenori OKUMURA and Kanta AITO Department of Business management and information science, Cugoku Junior College, Okayama, 70-097 Japan Department

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x

More information

Local Orthogonal Polynomial Expansion (LOrPE) for Density Estimation

Local Orthogonal Polynomial Expansion (LOrPE) for Density Estimation Local Ortogonal Polynomial Expansion (LOrPE) for Density Estimation Alex Trindade Dept. of Matematics & Statistics, Texas Tec University Igor Volobouev, Texas Tec University (Pysics Dept.) D.P. Amali Dassanayake,

More information

Section 15.6 Directional Derivatives and the Gradient Vector

Section 15.6 Directional Derivatives and the Gradient Vector Section 15.6 Directional Derivatives and te Gradient Vector Finding rates of cange in different directions Recall tat wen we first started considering derivatives of functions of more tan one variable,

More information

Taylor Series and the Mean Value Theorem of Derivatives

Taylor Series and the Mean Value Theorem of Derivatives 1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential

More information

Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution

Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution International Journal of Clinical Medicine Researc 2016; 3(5): 76-80 ttp://www.aascit.org/journal/ijcmr ISSN: 2375-3838 Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution

More information

Math 212-Lecture 9. For a single-variable function z = f(x), the derivative is f (x) = lim h 0

Math 212-Lecture 9. For a single-variable function z = f(x), the derivative is f (x) = lim h 0 3.4: Partial Derivatives Definition Mat 22-Lecture 9 For a single-variable function z = f(x), te derivative is f (x) = lim 0 f(x+) f(x). For a function z = f(x, y) of two variables, to define te derivatives,

More information

VARIANCE ESTIMATION FOR COMBINED RATIO ESTIMATOR

VARIANCE ESTIMATION FOR COMBINED RATIO ESTIMATOR Sankyā : Te Indian Journal of Statistics 1995, Volume 57, Series B, Pt. 1, pp. 85-92 VARIANCE ESTIMATION FOR COMBINED RATIO ESTIMATOR By SANJAY KUMAR SAXENA Central Soil and Water Conservation Researc

More information

Preconditioning in H(div) and Applications

Preconditioning in H(div) and Applications 1 Preconditioning in H(div) and Applications Douglas N. Arnold 1, Ricard S. Falk 2 and Ragnar Winter 3 4 Abstract. Summarizing te work of [AFW97], we sow ow to construct preconditioners using domain decomposition

More information

Online Learning: Bandit Setting

Online Learning: Bandit Setting Online Learning: Bandit Setting Daniel asabi Summer 04 Last Update: October 0, 06 Introduction [TODO Bandits. Stocastic setting Suppose tere exists unknown distributions ν,..., ν, suc tat te loss at eac

More information

Chapter 5 FINITE DIFFERENCE METHOD (FDM)

Chapter 5 FINITE DIFFERENCE METHOD (FDM) MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY APPLICATIONES MATHEMATICAE 36, (29), pp. 2 Zbigniew Ciesielski (Sopot) Ryszard Zieliński (Warszawa) POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY Abstract. Dvoretzky

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

Continuity and Differentiability of the Trigonometric Functions

Continuity and Differentiability of the Trigonometric Functions [Te basis for te following work will be te definition of te trigonometric functions as ratios of te sides of a triangle inscribed in a circle; in particular, te sine of an angle will be defined to be te

More information

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these. Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra

More information

The Complexity of Computing the MCD-Estimator

The Complexity of Computing the MCD-Estimator Te Complexity of Computing te MCD-Estimator Torsten Bernolt Lerstul Informatik 2 Universität Dortmund, Germany torstenbernolt@uni-dortmundde Paul Fiscer IMM, Danisc Tecnical University Kongens Lyngby,

More information

2.8 The Derivative as a Function

2.8 The Derivative as a Function .8 Te Derivative as a Function Typically, we can find te derivative of a function f at many points of its domain: Definition. Suppose tat f is a function wic is differentiable at every point of an open

More information

Introduction to Derivatives

Introduction to Derivatives Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))

More information

3.4 Worksheet: Proof of the Chain Rule NAME

3.4 Worksheet: Proof of the Chain Rule NAME Mat 1170 3.4 Workseet: Proof of te Cain Rule NAME Te Cain Rule So far we are able to differentiate all types of functions. For example: polynomials, rational, root, and trigonometric functions. We are

More information

Analysis of Solar Generation and Weather Data in Smart Grid with Simultaneous Inference of Nonlinear Time Series

Analysis of Solar Generation and Weather Data in Smart Grid with Simultaneous Inference of Nonlinear Time Series Te First International Worksop on Smart Cities and Urban Informatics 215 Analysis of Solar Generation and Weater Data in Smart Grid wit Simultaneous Inference of Nonlinear Time Series Yu Wang, Guanqun

More information

1 Introduction to Optimization

1 Introduction to Optimization Unconstrained Convex Optimization 2 1 Introduction to Optimization Given a general optimization problem of te form min x f(x) (1.1) were f : R n R. Sometimes te problem as constraints (we are only interested

More information

Ratio estimation using stratified ranked set sample

Ratio estimation using stratified ranked set sample METRON - International Journal of Statistics 003, vol. LXI, n. 1, pp. 75-90 HANI M. SAMAWI MAHMOUD I. SIAM Ratio estimation using stratified ranked set sample Summary - Ratio estimation metod is used to

More information

Deconvolution problems in density estimation

Deconvolution problems in density estimation Deconvolution problems in density estimation Dissertation zur Erlangung des Doktorgrades Dr. rer. nat. der Fakultät für Matematik und Wirtscaftswissenscaften der Universität Ulm vorgelegt von Cristian

More information

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h Lecture Numerical differentiation Introduction We can analytically calculate te derivative of any elementary function, so tere migt seem to be no motivation for calculating derivatives numerically. However

More information

Pre-Calculus Review Preemptive Strike

Pre-Calculus Review Preemptive Strike Pre-Calculus Review Preemptive Strike Attaced are some notes and one assignment wit tree parts. Tese are due on te day tat we start te pre-calculus review. I strongly suggest reading troug te notes torougly

More information

Complexity of Decoding Positive-Rate Reed-Solomon Codes

Complexity of Decoding Positive-Rate Reed-Solomon Codes Complexity of Decoding Positive-Rate Reed-Solomon Codes Qi Ceng 1 and Daqing Wan 1 Scool of Computer Science Te University of Oklaoma Norman, OK73019 Email: qceng@cs.ou.edu Department of Matematics University

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Kernel Density Estimation

Kernel Density Estimation Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative

More information

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul

More information

Numerical analysis of a free piston problem

Numerical analysis of a free piston problem MATHEMATICAL COMMUNICATIONS 573 Mat. Commun., Vol. 15, No. 2, pp. 573-585 (2010) Numerical analysis of a free piston problem Boris Mua 1 and Zvonimir Tutek 1, 1 Department of Matematics, University of

More information

Estimation of boundary and discontinuity points in deconvolution problems

Estimation of boundary and discontinuity points in deconvolution problems Estimation of boundary and discontinuity points in deconvolution problems A. Delaigle 1, and I. Gijbels 2, 1 Department of Matematics, University of California, San Diego, CA 92122 USA 2 Universitair Centrum

More information