SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1

Size: px
Start display at page:

Download "SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1"

Transcription

1 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 B Technical details B.1 Variance of ˆf dec in the ersmooth case We derive the variance in the case where ϕ K (t) = ( 1 t 2) κ I( 1 t 1), i.e. we take K as in (A3), and assume that ϕ U (t) = a(t) exp ( t α), where r, α > and a denotes a symmetric, real-valued function satisfying a() = 1 and a(t) ξ t α 1 as t, (B.1) with < α 1 < and ξ >. That is, we put γ = 1, α 2 = α 1 and d = d 1 in (3.2). The more general setting at (3.2) obtains similarly. Assume too that the distribution of which f W is the density has finite variance: w 2 f W (w) dw <. (B.2) Recall the definition of K U at (2.2). In this notation, it is well known that the asymptotic variance of ˆf dec is given by Var ˆfdec (x) } = n 1 K 2 U f W (x) n 1 E ˆf dec (x) } 2. Theorem 3. If (B.1) (B.2) hold then, for each x, K 2 U f W (x) as h. 2 κ ξ 1 Γ(κ + 1) h (κ+1)α+α 1 exp ( h α) } 2 2π α κ+1 cos(x y)} 2 f W (y) dy (B.3) Proof. It is notationally convenient to put b = a 1. Note that 2π K U (x) = = h 1 cos(tx) ( 1 t 2) κ b(t/h) exp ( t/h α ) dt 1/h cos(hux) 1 (hu) 2} κ b(u) exp ( u α ) du = h I 1 (x; h), (B.4)

2 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 2 say. Let ϵ = ϵ(h) decrease to zero slowly as h, at a rate that we shall address shortly, and observe that, by (B.1), 1/h I 1(x; h) cos(hux) 1 (hu) 2} κ ( ) b(u) exp u α du (1 ϵ)/h (1 ϵ)/h C } 1 1 (hu) 2 κ (1 + u ) α 1 exp ( u α) du (1 ϵ)/h C 2 (1 + u ) α 1 exp ( u α) du C 3 h α 1 exp [ (1 ϵ)/h} α], (B.5) where, here and below, C 1, C 2,... denote generic positive constants not depending on h or x. Note too that 1/h (1 ϵ)/h cos(hux) 1 (hu) 2} κ b(u) exp ( u α ) du cos(x) 1/h C 3 ϵ x (1 ϵ)/h 1/h (1 ϵ)/h } 1 (hu) 2 κ ( ) b(u) exp u α du 1 (hu) 2 } κ (1 + u ) α 1 exp ( u α) du. (B.6) Since α > then, writing o u (1) for a generic function of u that satisfies u ϵ/h α o u (1) = o(1), and taking ϵ to decrease to zero so slowly that ϵ/h α, we have: 1/h 1 (hu) 2 } κ b(u) exp ( u α ) du (B.7) (1 ϵ)/h = ϵ/h 2 hv (hv) 2 } κ b ( h 1 v ) exp ( h 1 v α) dv ϵ/h = (2h) κ v ( κ 1 1 hv) κ ( 2 b h 1 v ) exp ( h 1 v α) dv ξ 1 (2h) κ h α 1 = ξ 1 (2h) κ h α 1 ϵ/h ϵ/h v κ exp ( h 1 v α) dv v κ exp ( h α 1 hv α) dv

3 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 3 ϵ/h α = 2 κ ξ 1 h (κ+1)α+α 1 1 = 2 κ ξ 1 h (κ+1)α+α 1 1 ϵ/h α u κ exp ( h α 1 h α u α) du u κ exp h α (1 α h α u) + o u (1) } du = 2 κ ξ 1 h (κ+1)α+α1 1 exp ( h α) ϵ/h α u κ exp α u + o u (1) } du 2 κ ξ 1 h (κ+1)α+α1 1 exp ( h α) u κ exp( α u) du (B.8) = 2κ ξ 1 Γ(κ + 1) α κ+1 h (κ+1)α+α 1 1 exp ( h α). (B.9) Combining (B.6) and (B.9) we deduce that 1/h cos(hux) 1 (hu) 2} κ ( ) b(u) exp u α du (1 ϵ)/h cos(x) + o unif (1)} 2κ ξ 1 Γ(κ + 1) α κ+1 C 4 ϵ x h (κ+1)α+α 1 1 exp ( h α), h (κ+1)α+α1 1 exp ( h α) (B.1) where, here and below, o unif (1) and O unif (1) denote quantities that depend on x and equal o(1) and O(1), respectively, uniformly in < x <, as h. Together, (B.5) and (B.1) imply that I 1 (x; h) = cos(x) 2κ ξ 1 Γ(κ + 1) α κ+1 h (κ+1)α+α 1 1 exp ( h α) + o unif (1) + O unif (1) ϵ x } h (κ+1)α+α 1 1 exp ( h α) + O unif (1) h α 1 exp [ (1 ϵ)/h} α]. (B.11) Together, (B.4) and (B.11) imply that, if ϵ decreases to zero slowly as h, KU 2 f W (x) = KU(x 2 y) f W (y) dy = h2 I (2π) 1(x 2 y; h) f 2 W (y) dy 2 κ ξ 1 Γ(κ + 1) = h (κ+1)α+α 1 exp ( h α) } 2 2π α κ+1 cos(x y)} 2 f W (y) dy + o[ h (κ+1)α+α 1 exp ( h α)} ] 2. (B.12) (Here we have used (B.2).) This result is equivalent to (B.3).

4 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 4 B.2 Main steps in derivation of convergence rate of ˆf rat (see (3.13)) in the case n 1/2 η Theorem 4 below demonstrates that, under the assumption that ˆθ = θ 1 + O p (n 1/2 ), the asymptotic bias of ˆfrat (x ˆθ) equals that of ˆfrat (x θ 1 ), and the error about the mean of ˆf rat (x ˆθ) equals that of ˆf dec (x) plus negligible terms. Write E fx for expectation when the distribution of the data W j has density f U f X ; and, for any random variable R with finite mean, let (1 E fx ) R denote R E fx (R). Theorem 4. If (3.1), (3.6) and (B1) (B6) in section 3.5 hold, then, for each ϵ >, ˆf rat (x ˆθ) (1 E fx ) ˆf dec (x) E fx ˆfrat (x θ 1 ) } = O p n ( ϵ nh 2α 1) } 1/2 + n 1/2 + h r s+1. (B.13) Moreover, the remainder term on the right-hand side is of the stated order uniformly in densities f X for which (3.1), (B1) (B6) in section 3.5, and (3.6) hold, in the sense that lim lim C n f X F > C P fx [ ˆf rat (x ˆθ) (1 E fx ) ˆf dec (x) E fx ˆfrat (x θ 1 ) } n ϵ ( nh 2α 1) 1/2 + n 1/2 + h r s+1 } ] =. (B.14) Our next result describes the bias of ˆf rat ( θ 1 ). Let s(ψ) be as in (3.15). Theorem 5. If (3.1), and (B1) (B6) in section 3.5 hold, then θ Θ h β s(ψ) } 1 EfX ˆfrat (x θ 1 )} f X (x) = O(1) as n. (B.15) Let Var fx denote the variance operator when the data W j have density f U f X. If (3.1), (B1), (B2) and (B3) hold, then standard deconvolution results imply that f X F Var fx ˆfdec (x) } = O (nh 2α+1 ) 1}, (B.16)

5 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 5 where the order is exact when f W (x) is nonzero. Hence, the remainder on the righthand side of (B.13) is negligibly small relative to the term (1 E fx ) ˆf dec (x) on the left-hand side, for any x such that f W (x) is nonzero. Combining (B.14), (B.15) and (B.16) we deduce that, under the conditions of Theorem 5, if the term in h r s+1 in (B.14) can be ignored then [ lim lim P fx ˆfrat (x) f X (x) (nh > C ) }] 2α+1 1/2 + h β s(ψ) =. C n θ Θ (B.17) Under the assumptions for Theorem 5, if ψ represents a sequence of functions, and we take h min1, (nη 2 ) 1/(2α+2β+1) }, then the minimiser, with respect to h, of the term within braces (i.e. the coefficient of C) on the left-hand side of (B.17) is of size (η 2α+1 /n β ) 1/(2α+2β+1) in the case n 1/2 η, in which instance h is asymptotic to a constant multiple of (nη 2 ) 1/(2α+2β+1). The term in h r s+1 in (B.14) can be ignored if h r s+1 = O(η 2α+1 /n β ) 1/(2α+2β+1) }, or equivalently, since h is asymptotic to a constant multiple of (nη 2 ) 1/(2α+2β+1), if n 1 = O ( η (2α+2r 2s+3)/(r s+1 β)) (B.18) (provided that r > s 1 + β). For example, if η = O(n t ) where t (, 1 2 ) and 2 t α + β (1 2 t) (r s + 1) t, then (B.18) holds, in which case h r s+1 can be ignored in (B.2) and therefore property (3.13) follows from (B.17). B.3 Proof of Theorem 4 Observe that 2π f(x θ) L x (w θ, h) =f(x θ) =f(x θ) e it(x w) = ϕ U (t) dt e itw ϕ K x( θ,h)(t) dt ϕ U (t) e itw ϕ U (t) dt e it(x hu) K(u) f(x hu θ) du e ihtu f(x θ) K(u) f(x hu θ) du

6 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 6 = r k= h k e it(x w) k! g k(x θ) ϕ U (t) dt + h r+1 e it(x w) ϕ U (t) dt Since γ h,x has s derivatives then s integrations by parts give: e ihtu ω(hu, x θ) u r+1 K(u) du = (iht) s Therefore, e it(x w) ϕ U (t) dt e ihtu ω(hu, x θ) u r+1 K(u) du 1 dt γ h,x (u θ) du + ϕ U (t) 1 e ihtu u k K(u) du e ihtu ω(hu, x θ) u r+1 K(u) du. t >1 e ihtu γ (s) h,x (u θ) du. Combining this bound with (B.19), and noting (B6), we deduce that: r f(x θ) L h k g k (x θ) x(w θ, h) <w< θ Θ 2πk! k= 1 B 4 h r+1 dt ϕ U (t) + 1 (B.19) dt γ (s) ht s h,x ϕ U (t) (u θ) du. e it(x w) ϕ U (t) dt t >1 dt ht s ϕ U (t) e ihtu u k K(u) du }, (B.2) uniformly in < h C 1, where B 4 > is a constant. Property (3.1) implies that the first integral on the right-hand side of (B.2) is finite, and (3.1) and (B6) imply that the second integral is bounded above by t >1 dt ht s ϕ U (t) C (1 + t ) α 1 dt = O ( h s). t >1 ht s These results, (B.2) and the property e ihtu u k K(u) du = ( i) k ϕ (k) K ( ht), entail: f(x θ) L x(w θ, h) <w< θ Θ r ( ih) k g k (x θ) e it(w x) 2πk! ϕ U (t) ϕ K (k) (ht) dt = O( h r s+1). (B.21) k=

7 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 7 Note that (B.21), and (B.22) and (B.24) below, involve only the densities f( θ) and f U, which we take to be fixed; these formulae do not involve f X, which can vary with n. Let s k = ( 1) k/2 if k is even, and s k = ( 1) (k 1)/2 if k is odd. Since the distribution of U is symmetric then ϕ U is symmetric, and so, defining cos(tu) ϕ U (t) ϕ K (k) (ht) dt if k is even ψ k (u, h) = sin(tu) ϕ U (t) ϕ K (k) (ht) dt if k is odd we have: <w< θ Θ Hence, defining f(x θ) L x(w θ, h) 1 2π r k= s k h k k! Ψ k (x) = 1 n g k (x θ) ψ k (w x, h) = O( h r s+1). n ψ k (W j x, h), j=1 (B.22) (B.23) we have: for a constant B 5 >, not depending on h, n or the data W 1,..., W n, ˆf rat (x θ) 1 r s k h k g k (x θ) 2π k! Ψ k (x) B 5 h r s+1, (B.24) θ Θ k= where the right-hand side denotes a deterministic quantity of the stated size. Define and note that, by (B2), L k (u) = L k (u) B 6 h α <u< exp(itu) ϕ U (t/h) ϕ K (k) (t) dt (1 + t ) α ϕk (k) (t) dt B7 h α, (B.25) where B 6, B 7 > are constants not depending on h or n. Using (3.1), (B1), (B2) and (B.25) we deduce that, for each integer m 2, h m E ψ k (W x, h) m} E [ L k (W x)/h} m] L k m 2 E [ L k (W x)/h} 2]

8 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 8 =h L k m 2 L k (u) 2 f W (x + hu) du B 8 h ( B 7 h α) m 2 L k (u) 2 du =B 9 (m) h 1 (m 2)α (2π) 1 ϕ (k) K (t) ϕ U (t/h) B 1 (m) h 1 (m 2)α (1 + t/h ) α ϕ K (k) (t) } 2 dt =O ( h 1 mα), (B.26) 2 dt where the bounds apply uniformly in x I, and B 8 = f W, B 9 (m) = B m 2 7 B 8, and B l (m) or B l, for l 1, denote constants not depending on h or n. (The second identity in the string at (B.26) follows using Parseval s identity. Note that if (B1) holds then f X (x) is bounded uniformly in x and n, from which it follows that the same is true for f W (x).) Therefore, E ψ k (W x, h) m} = O ( h 1 m(α+1)). (B.27) Let 1 j r, and recall the definition of Ψ k at (B.23). Replacing m by either 2 or 2m in the bound at (B.27), we have for each m 1, using Rosenthal s identity (see e.g. Hall and Heyde (198), p. 23): E (1 E) Ψj (x) ( [ 2m B 11 (m) n 1 E ψ k (W x, h) 2}] m + n (2m 1) E ψ k (W x, h) 2m}) (nh 2α+1 B 12 (m) ) m ( + h α nh α+1) } (2m 1) B 13 (m) ( nh 2α+1) m, uniformly in x I, where the last inequality is valid whenever nh 1. Hence, if I(n) is a set of at most n B values x I, then, for each ϵ >, P (n) (1 E) Ψj (x) ( > n ϵ nh 2α+1) } 1/2

9 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 9 n B (1 P E) Ψj (x) ( > n ϵ nh 2α+1) } 1/2 (n) n B n ϵ ( nh 2α+1) 1/2 } 2m E (1 E) Ψj (x) 2m B 14 (m) n B 2mϵ. (B.28) Observe too that if 1 k r and x, x I, then, using (3.1), (B2) and (B3), Ψk (x) Ψ k (x ) 1 n n j=1 C 11 n 1 x x ψk (W j x, h) ψ k (W j x, h) n j=1 t (1 + t ) α ϕk (k) (ht) dt B 15 x x h (α+2) B 16 x x n (α+2)/(2α+1). The last inequality follows from the fact that, by (B3), nh 2α+1 is bounded away from. Therefore, if I(n) represents a grid in I with edge width n 1 (α+2)/(2α+1), and if for each x I we define x to be the point in I(n) nearest to x, then and therefore Ψk (x) Ψ k (x ) B17 n 1 (α+2)/(2α+1) n (α+2)/(2α+1) = B 17 n 1, (1 E) Ψk (x) Ψ k (x ) } 2 B 17 n 1. Hence the following version of (B.28), with I(n) there replaced by I, holds: for each ϵ >, P (1 E) Ψ j (x) > n ( ϵ nh 2α+1) } 1/2. Combining (B.24) and (B.29) we deduce that, for each ϵ >, θ Θ (B.29) ˆf rat (x θ) g (x θ) Ψ (x) 1 r s k h k g k (x θ) E Ψ k (x)} 2π k! k=1 = O p n ( ϵ nh 2α 1) } 1/2 + h r s+1. (B.3) Note too that e it(w x) E ϕ U (t) } e ϕ (k) itx ϕ X (t) ϕ U (t) K (ht) dt = ϕ U (t) ϕ K (k) (ht) dt

10 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 = e itx ϕ X (t) ϕ K (k) (ht) dt = 2π i k f X (x hu) u k K(u) du, which implies that E Ψk (x) } = 2π s k f X (x hu) u k K(u) du. (B.31) Furthermore, g (x θ) 1. Hence, by (B.3), ˆf rat (x θ) Ψ r (x) θ Θ where ϵ > is arbitrary. k=1 h k k! g k(x θ) f X (x hu) u k K(u) du = O p n ϵ ( nh 2α 1) 1/2 + h r s+1 }, (B.32) Assumption (B5) implies that g k (x θ), and its first derivative with respect to θ, are bounded uniformly in x I and θ Θ. Using this result, (B5) and (3.6) we deduce that: max gk (x ˆθ) g k (x θ 1 ) ( ) = Op n 1/2. 1 k r Conditions (B1) and (B2) imply that f X (x hu) u k K(u) du is bounded uniformly in x and h, for k =,..., r. Combining these results with (B.32), and recalling that g (x θ 1 ) 1, we deduce that, for each ϵ >, ˆf rat (x ˆθ) (1 E) Ψ r (x) h k k! g k(x θ 1 ) f X (x hu) u k K(u) du k= = O p n ( ϵ nh 2α 1) } 1/2 + n 1/2 + h r s+1. (B.33) Result (B.22) implies that f(x θ 1) EL x (W θ 1, h)} r k= h k k! g k(x θ 1 ) f X (x hu) u k K(u) du} = O( h r s+1). (B.34) Since Ψ = ˆf dec where ˆf dec is as defined at (2.6), and E ˆfrat (x θ 1 ) } = f(x θ 1 ) EL x (W θ 1, h)},

11 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 11 where ˆf rat (x θ) is as defined at (2.3), then (B.33) and (B.34) together imply (B.13). The results mentioned in the second and third sentences of this paragraph, together with (B.24), (B.29), (B.31) and (B.34), similarly imply (B.14). B.4 Proof of Theorem 5 Noting that f X = f( θ 1 ) + ψ, and that, by construction, EL x (W θ 1, h)} = EK x (W θ 1, h)}, we have: EL x (W θ 1, h)} = K x (u θ 1, h) du 1 e itu ϕ X (t) dt 2π = 1 ( x u ) fx (u) K h h f(u θ 1 ) du = K(u) f(x hu θ 1) + ψ(x hu) du f(x hu θ 1 ) ψ(x hu) =1 + K(u) f(x hu θ 1 ) du. Therefore, in notation introduced in section 3.5, E ˆfrat (x θ 1 )} f X (x) = = K(u) f(x θ 1 ) f(x hu θ 1 ) ψ(x hu) du ψ(x) K(u) ψ(x hu) ψ(x)} du + h r+1 γ h,x (u θ 1 ) ψ(x hu) du + r k=1 h k k! g k(x θ 1 ) u k K(u) ψ(x hu) du (B.35) =O h β s(ψ) }, (B.36) uniformly in θ Θ and x I. This is equivalent to (B.15). The last identity in (B.36) follows on using the moment condition u j K(u) du =, for j = 1,..., β 1.

12 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 12 C Details of implementation C.1 Using a ridge when computing ˆf rat If U has a Laplace distribution with scale parameter λ, so that ϕ U (t) = (1 + λ 2 t 2 ) 1, then if K and f( θ) are twice differentiable, we can write L x (u θ, h) = 1 2π e itu (1 + λ 2 t 2 ) ϕ Kx (t θ, h)dt = K x (u θ, h) λ 2 K x(u θ, h). More generally, if U has the distribution of an N-fold Laplace convolution, then L x is a linear combination of derivatives of K x of the form K x (k) (W j θ, h) = k K(x u)/h} = u k hf(u θ) u=wj k+1 l=1 g l (W j x, h, θ) f l (W j θ), for some functions g l which are sums and products of positive powers of K(x W j )/h}, f(w j θ) and their derivatives. In practice, the denominators f l (W j θ) are often too close to zero for some W j s, which makes the estimator work rather poorly. To avoid this problem, we can use a ridge parameter in the denominators. We tried several approaches to ridging in the particular Laplace error case, and found that the following approach performed well in practice: replace f l ( θ) by maxf l ( θ), δ} where δ > is a ridge parameter. In our numerical work, we used this approach with δ =.4 f( θ). C.2 Details of SIMEX bandwidth for ˆf rat It follows from our theoretical results that, under regularity conditions, the asymptotic mean integrated squared error (AMISE) of ˆf rat is equal to AMISE ˆfrat ( θ) } = h 4 µ 2 2(K)Rr f( θ)}/4 + (2πnh) 1 ϕ K (t) 2 ϕ U (t/h) 2 dt, where µ 2 (K) = x 2 K(x) dx, r = f X /f( θ), and we used the notation R(f) = f 2. We suggest choosing h for ˆf rat by minimising a SIMEX estimator of the AMISE;

13 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 13 see Cook and Stefanski (1994) and Stefanski and Cook (1995) for an introduction to SIMEX. The idea of SIMEX methods is that, in some way, the relation between data from f W and f X can be mimicked by that between data from f W (1) f W f U and f W (2) f W (1) f U. Now, quantities related to f W are easy to estimate from the data, since we have a sample of W i s and can generate data from f W (1) be exploited to estimate unknown quantities related to f X. and f W (2). This can Let h, h 1 and h 2 be the bandwidths that minimise, respectively, AMISE ˆfrat ( θ) }, AMISE ˆfrat,W ( θ) } and AMISE ˆfrat,W (1)( θ) }, where ˆf rat,w and ˆf rat,w (1) denote the ratio estimators of, respectively, f W and f W (1) computed from a sample of size n having a density, respectively f W (1) and f W (2). Then, extending the SIMEX-based bandwidth of Delaigle and Hall (28) to our problem, h /h 1 can be mimicked by h 1 /h 2, and hence h can be approximated by ĥ2 1/ĥ2, where ĥj is an estimator of h j, for j = 1, 2. To estimate h j, construct a sample W (1) 1,..., W (1) n from f W (1) and a sample W (2) 1,..., W n (2) from f W (2), by taking W (1) i = W i + U (1) i and W (2) i = W (1) i + U (2) i for j = 1, 2, U (j) 1,..., U n (j) where, is a sample of independent observations generated from f U, independently of the W i s. Let f W () f W, f () ( θ) f( θ) f U, f (1) ( θ) = f () ( θ) f U, f (2) ( θ) = f (1) ( θ) f U, and, for j =, 1, 2, let r (j) = f W (j)/f (j) ( θ). By definition, we have, for j =, 1, h 4 µ 2 h j+1 = argmin 2(K) R r h 4 (j)f (j) ( θ) } + 1 ϕ K (t) 2πnh ϕ U (t/h) and to estimate h j+1 it suffices to estimate r (j) (x) and f (j)( θ). We estimate the latter by f (j) ( ˆθ), and to estimate the former, we take ˆr (j) (x) = ˇf W (j)(x)/f (j) (x ˆθ)}, where ˇf W () and ˇf W (1) denote standard error-free kernel estimators of f W and f W (1) computed from the data W 1,..., W n and W (1) 1,..., W n (1), respectively, and using, for example, a normal reference bandwidth for estimating second derivatives of densities; see Sheather and Jones (1991) and the references therein. In practice, to compute 2 dt,

14 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 14 R r (j) f (j)( θ) } we truncate the integral to the interval [W.5, (j) W.95], (j) where W α () and W α (1) denote the 1αth percentile of, respectively, the W i s and the W (1) s. Again, here, when computing ˆr (j)(x) = ˇf (x) ˇfW W (j) f (j) (x θ) (j)(x)f(j) (x θ) f(j) 2 (x θ) 2 ˇf W (j)(x)f(j) (x θ) + (x θ) f 2 (j) 2 ˇf W (j)(x)f(j) (x θ)}2 f(j) 3 (x θ), we need to use a ridge at each denominator. We implemented this procedure in the Laplace case only, and we used the same ridge as the one described in section C.1. Note that estimated bandwidth depends on the SIMEX sample W (j) 1..., W n (j). To reduce the effect of the random sampling step, as in Delaigle and Hall (28) we generate B sets of SIMEX samples of size n (we took B = 3), and take h j+1 to minimise the average of the corresponding B estimated AMISE values. C.3 NR bandwidth for ˆf dec with the sinc kernel The mean integrated squared error (MISE) of ˆf dec, computed with the sinc kernel, is given by i MISE(h) = 1 2πn 1/h 1/h ϕ U (t) 2 dt 1 + n 1 2π 1/h 1/h ϕ X (t) 2 dt + f 2 X, (C.1) and to find the bandwidth that minimises the MISE, we need to find the roots of MISE (h) = 1 πnh 2 ϕ U(1/h) 2 + n + 1 nπh 2 ϕ W (1/h) 2 ϕ U (1/h) 2. Equivalently, this bandwidth is a solution of 1 + (n + 1) ϕ W (1/h) 2 =. The NR rule assumes that X N(ˆµ, ˆσ 2 ), where ˆµ = W and ˆσ 2 = W and Var(W ) Var(U), with Var(W ) denoting the empirical mean and variance of the W i s. This amounts to estimating ϕ W (1/h) by ˆϕ W (1/h) = ϕ U (1/h) exp(iˆµ/h) exp(.5ˆσ 2 /h 2 ). Then the bandwidth is estimated by ĥnr, the solution of 1 + (n + 1) ˆϕ W (1/h) 2 =.

15 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 15 D Additional simulation results Tables D.1 and D.3 report the integrated squared bias (ISB), computed from 1 samples, of the estimators ˆf dec, ˆfbco and ˆf wgt for the various cases considered in our simulations. These tables show that, as could be expected, ˆfbco is less biased than ˆf dec. ˆfwgt also benefits from a bias reduction in the Laplace error case, and also in the normal case for densities (i) and (ii). However, for the other densities, in the normal error case, the ISB of ˆfwgt is slightly larger than that of ˆfdec. This is because θ is more difficult to estimate in the normal error case than in the Laplace error case. As a result, in the normal error case, the parametric component of ˆf wgt is less good, and the weight ŵ of ˆf wgt, which also depends on θ, is smaller. See Table D.2, where we show the median and interquartile range of ŵ, when estimating densities (i) to (iii). Table D.1: 1 3 ISB of 1 estimators of densities (i) to (iii) in the Laplace and normal error cases, when NSR = 1% or 25% and n = 2 or 7, using the estimators ˆf dec, ˆf bco ( ; ˆθ MD ) and ˆf wgt ( ; ˆθ ML ). Laplace Normal Density (i) Density (ii) Density (iii) Density (i) Density (ii) Density (iii) n n NSR = 1% NSR = 1% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML ) NSR = 25% NSR = 25% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML )

16 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 16 Table D.2: Median (IQR) of 1 values of ŵ when estimating densities (i) to (iii) in the Laplace and normal error cases, with NSR = 1% or 25% and n = 2 or 7. Density (i) Density (ii) Density (iii) n ˆθ ML Lap, NSR = 1%.7 (.7).74 (.7).71 (.9).73 (.7).68 (.9).72 (.1) Lap, NSR = 25%.73 (.7).76 (.5).73 (.9).78 (.7).71 (.8).73 (.8) Norm, NSR = 1%.29 (.9).28 (.12).29 (.9).28 (.9).14 (.7).5 (.2) Norm, NSR = 25%.33 (.1).33 (.9).3 (.9).33 (.12).24 (.13).17 (.9) Table D.3: 1 3 ISB of 1 estimators of densities (iv) to (vi) in the normal error case, when NSR = 1% or 25% and n = 2 or 7, using the estimators ˆf dec, ˆf bco ( ; ˆθ MD ) and ˆf wgt ( ; ˆθ ML ). Density (iv) Density (v) Density (vi) Density (iv) Density (v) Density (vi) n n NSR = 1% NSR = 25% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML ) References Cook, J.R. and Stefanski, L.A. (1994). Simulation-extrapolation estimation in parametric measurement error models. J. Amer. Statist. Assoc., Hall, P. and Heyde, C. (198). Martingale Limit Theory and its Application. Academic Press. Sheather, S. J. and Jones, M. C. (1991). A reliable data-based bandwidth selection method for kernel density estimation. J. Royal Statist. Society Ser. B., Stefanski, L. and Cook, J.R. (1995). Simulation-extrapolation: The measurement error jackknife. J. Amer. Statist. Assoc

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

arxiv: v1 [stat.me] 17 Jan 2008

arxiv: v1 [stat.me] 17 Jan 2008 Some thoughts on the asymptotics of the deconvolution kernel density estimator arxiv:0801.2600v1 [stat.me] 17 Jan 08 Bert van Es Korteweg-de Vries Instituut voor Wiskunde Universiteit van Amsterdam Plantage

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

A Novel Nonparametric Density Estimator

A Novel Nonparametric Density Estimator A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

Smooth simultaneous confidence bands for cumulative distribution functions

Smooth simultaneous confidence bands for cumulative distribution functions Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang

More information

Nonparametric confidence intervals. for receiver operating characteristic curves

Nonparametric confidence intervals. for receiver operating characteristic curves Nonparametric confidence intervals for receiver operating characteristic curves Peter G. Hall 1, Rob J. Hyndman 2, and Yanan Fan 3 5 December 2003 Abstract: We study methods for constructing confidence

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential. Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample

More information

VARIANCE REDUCTION BY SMOOTHING REVISITED BRIAN J E R S K Y (POMONA)

VARIANCE REDUCTION BY SMOOTHING REVISITED BRIAN J E R S K Y (POMONA) PROBABILITY AND MATHEMATICAL STATISTICS Vol. 33, Fasc. 1 2013), pp. 79 92 VARIANCE REDUCTION BY SMOOTHING REVISITED BY ANDRZEJ S. KO Z E K SYDNEY) AND BRIAN J E R S K Y POMONA) Abstract. Smoothing is a

More information

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence

More information

Bandwith selection based on a special choice of the kernel

Bandwith selection based on a special choice of the kernel Bandwith selection based on a special choice of the kernel Thomas Oksavik Master of Science in Physics and Mathematics Submission date: June 2007 Supervisor: Nikolai Ushakov, MATH Norwegian University

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram

More information

On variable bandwidth kernel density estimation

On variable bandwidth kernel density estimation JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced

More information

Quantitative Economics for the Evaluation of the European Policy. Dipartimento di Economia e Management

Quantitative Economics for the Evaluation of the European Policy. Dipartimento di Economia e Management Quantitative Economics for the Evaluation of the European Policy Dipartimento di Economia e Management Irene Brunetti 1 Davide Fiaschi 2 Angela Parenti 3 9 ottobre 2015 1 ireneb@ec.unipi.it. 2 davide.fiaschi@unipi.it.

More information

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(

More information

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1

36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1 36. Multisample U-statistics jointly distributed U-statistics Lehmann 6.1 In this topic, we generalize the idea of U-statistics in two different directions. First, we consider single U-statistics for situations

More information

TIKHONOV S REGULARIZATION TO DECONVOLUTION PROBLEM

TIKHONOV S REGULARIZATION TO DECONVOLUTION PROBLEM TIKHONOV S REGULARIZATION TO DECONVOLUTION PROBLEM Dang Duc Trong a, Cao Xuan Phuong b, Truong Trung Tuyen c, and Dinh Ngoc Thanh a a Faculty of Mathematics and Computer Science, Ho Chi Minh City National

More information

Log-Density Estimation with Application to Approximate Likelihood Inference

Log-Density Estimation with Application to Approximate Likelihood Inference Log-Density Estimation with Application to Approximate Likelihood Inference Martin Hazelton 1 Institute of Fundamental Sciences Massey University 19 November 2015 1 Email: m.hazelton@massey.ac.nz WWPMS,

More information

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,

More information

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,

More information

Continuous Distributions

Continuous Distributions Chapter 3 Continuous Distributions 3.1 Continuous-Type Data In Chapter 2, we discuss random variables whose space S contains a countable number of outcomes (i.e. of discrete type). In Chapter 3, we study

More information

Adaptive Kernel Estimation of The Hazard Rate Function

Adaptive Kernel Estimation of The Hazard Rate Function Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized

More information

Kernel Density Estimation

Kernel Density Estimation Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Partial Differential Equations

Partial Differential Equations Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,

More information

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 = Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values

More information

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta

Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density

More information

Kernel density estimation of reliability with applications to extreme value distribution

Kernel density estimation of reliability with applications to extreme value distribution University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School 2008 Kernel density estimation of reliability with applications to extreme value distribution Branko Miladinovic

More information

Integral approximation by kernel smoothing

Integral approximation by kernel smoothing Integral approximation by kernel smoothing François Portier Université catholique de Louvain - ISBA August, 29 2014 In collaboration with Bernard Delyon Topic of the talk: Given ϕ : R d R, estimation of

More information

Density Estimation with Replicate Heteroscedastic Measurements

Density Estimation with Replicate Heteroscedastic Measurements Ann Inst Stat Math (011) 63:81 99 DOI 10.1007/s10463-009-00-x Density Estimation with Replicate Heteroscedastic Measurements Julie McIntyre Leonard A. Stefanski Received: 9 January 008 / Revised: 4 September

More information

Random fraction of a biased sample

Random fraction of a biased sample Random fraction of a biased sample old models and a new one Statistics Seminar Salatiga Geurt Jongbloed TU Delft & EURANDOM work in progress with Kimberly McGarrity and Jilt Sietsma September 2, 212 1

More information

Density Deconvolution for Generalized Skew-Symmetric Distributions

Density Deconvolution for Generalized Skew-Symmetric Distributions Density Deconvolution for Generalized Skew-Symmetric Distributions Cornelis J. Potgieter Department of Statistical Science, Southern Methodist University, Dallas, TX arxiv:1706.01507v1 [stat.me] 5 Jun

More information

ON SOME TWO-STEP DENSITY ESTIMATION METHOD

ON SOME TWO-STEP DENSITY ESTIMATION METHOD UNIVESITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XLIII 2005 ON SOME TWO-STEP DENSITY ESTIMATION METHOD by Jolanta Jarnicka Abstract. We introduce a new two-step kernel density estimation method,

More information

ON THE CHOICE OF TEST STATISTIC FOR CONDITIONAL MOMENT INEQUALITES. Timothy B. Armstrong. October 2014 Revised July 2017

ON THE CHOICE OF TEST STATISTIC FOR CONDITIONAL MOMENT INEQUALITES. Timothy B. Armstrong. October 2014 Revised July 2017 ON THE CHOICE OF TEST STATISTIC FOR CONDITIONAL MOMENT INEQUALITES By Timothy B. Armstrong October 2014 Revised July 2017 COWLES FOUNDATION DISCUSSION PAPER NO. 1960R2 COWLES FOUNDATION FOR RESEARCH IN

More information

the convolution of f and g) given by

the convolution of f and g) given by 09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

ON DECONVOLUTION WITH REPEATED MEASUREMENTS

ON DECONVOLUTION WITH REPEATED MEASUREMENTS ON DECONVOLUTION WITH REPEATED MEASUREMENTS Aurore Delaigle 1,2 Peter Hall 2,3 Alexander Meister 2,4 ABSTRACT. In a large class of statistical inverse problems it is necessary to suppose that the transformation

More information

Nonparametric Density Estimation (Multidimension)

Nonparametric Density Estimation (Multidimension) Nonparametric Density Estimation (Multidimension) Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann February 19, 2007 Setup One-dimensional

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Estimation of the Bivariate and Marginal Distributions with Censored Data

Estimation of the Bivariate and Marginal Distributions with Censored Data Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new

More information

A Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem

A Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem A Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem Aurore Delaigle, Jianqing Fan, and Raymond J. Carroll Abstract: Local polynomial estimators are popular techniques for nonparametric

More information

ζ (s) = s 1 s {u} [u] ζ (s) = s 0 u 1+sdu, {u} Note how the integral runs from 0 and not 1.

ζ (s) = s 1 s {u} [u] ζ (s) = s 0 u 1+sdu, {u} Note how the integral runs from 0 and not 1. Problem Sheet 3. From Theorem 3. we have ζ (s) = + s s {u} u+sdu, (45) valid for Res > 0. i) Deduce that for Res >. [u] ζ (s) = s u +sdu ote the integral contains [u] in place of {u}. ii) Deduce that for

More information

Properties of Principal Component Methods for Functional and Longitudinal Data Analysis 1

Properties of Principal Component Methods for Functional and Longitudinal Data Analysis 1 Properties of Principal Component Methods for Functional and Longitudinal Data Analysis 1 Peter Hall 2 Hans-Georg Müller 3,4 Jane-Ling Wang 3,5 Abstract The use of principal components methods to analyse

More information

EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018

EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018 EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018 While these notes are under construction, I expect there will be many typos. The main reference for this is volume 1 of Hörmander, The analysis of liner

More information

UNIVERSITÄT POTSDAM Institut für Mathematik

UNIVERSITÄT POTSDAM Institut für Mathematik UNIVERSITÄT POTSDAM Institut für Mathematik Testing the Acceleration Function in Life Time Models Hannelore Liero Matthias Liero Mathematische Statistik und Wahrscheinlichkeitstheorie Universität Potsdam

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

The Central Limit Theorem: More of the Story

The Central Limit Theorem: More of the Story The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit

More information

INTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES

INTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES INTRODUCTION TO REAL ANALYSIS II MATH 433 BLECHER NOTES. As in earlier classnotes. As in earlier classnotes (Fourier series) 3. Fourier series (continued) (NOTE: UNDERGRADS IN THE CLASS ARE NOT RESPONSIBLE

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Extending clustered point process-based rainfall models to a non-stationary climate

Extending clustered point process-based rainfall models to a non-stationary climate Extending clustered point process-based rainfall models to a non-stationary climate Jo Kaczmarska 1, 2 Valerie Isham 2 Paul Northrop 2 1 Risk Management Solutions 2 Department of Statistical Science, University

More information

1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample

More information

Limiting Distributions

Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results

More information

MATH 205C: STATIONARY PHASE LEMMA

MATH 205C: STATIONARY PHASE LEMMA MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

Stochastic Processes

Stochastic Processes Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54 Summary from measure theory De nition (X, A) is a measurable space

More information

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote Real Variables, Fall 4 Problem set 4 Solution suggestions Exercise. Let f be of bounded variation on [a, b]. Show that for each c (a, b), lim x c f(x) and lim x c f(x) exist. Prove that a monotone function

More information

Efficient Regressions via Optimally Combining Quantile Information

Efficient Regressions via Optimally Combining Quantile Information Efficient Regressions via Optimally Combining Quantile Information Zhibiao Zhao Penn State University Zhijie Xiao Boston College September 29, 2011 Abstract We study efficient estimation of regression

More information

Bootstrap of residual processes in regression: to smooth or not to smooth?

Bootstrap of residual processes in regression: to smooth or not to smooth? Bootstrap of residual processes in regression: to smooth or not to smooth? arxiv:1712.02685v1 [math.st] 7 Dec 2017 Natalie Neumeyer Ingrid Van Keilegom December 8, 2017 Abstract In this paper we consider

More information

A Note on Tail Behaviour of Distributions. the max domain of attraction of the Frechét / Weibull law under power normalization

A Note on Tail Behaviour of Distributions. the max domain of attraction of the Frechét / Weibull law under power normalization ProbStat Forum, Volume 03, January 2010, Pages 01-10 ISSN 0974-3235 A Note on Tail Behaviour of Distributions in the Max Domain of Attraction of the Frechét/ Weibull Law under Power Normalization S.Ravi

More information

arxiv: v2 [stat.me] 3 Dec 2016

arxiv: v2 [stat.me] 3 Dec 2016 Accelerated Nonparametric Maximum Likelihood Density Deconvolution Using Bernstein Polynomial arxiv:1601.0642v2 [stat.me] Dec 2016 Abstract Zhong Guan Department of Mathematical Sciences Indiana University

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

6 Lecture 6b: the Euler Maclaurin formula

6 Lecture 6b: the Euler Maclaurin formula Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Fall 217 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 217 March 26, 218 6 Lecture 6b: the Euler Maclaurin formula

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

MATH 220 solution to homework 4

MATH 220 solution to homework 4 MATH 22 solution to homework 4 Problem. Define v(t, x) = u(t, x + bt), then v t (t, x) a(x + u bt) 2 (t, x) =, t >, x R, x2 v(, x) = f(x). It suffices to show that v(t, x) F = max y R f(y). We consider

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

ON THE ESTIMATION OF EXTREME TAIL PROBABILITIES. By Peter Hall and Ishay Weissman Australian National University and Technion

ON THE ESTIMATION OF EXTREME TAIL PROBABILITIES. By Peter Hall and Ishay Weissman Australian National University and Technion The Annals of Statistics 1997, Vol. 25, No. 3, 1311 1326 ON THE ESTIMATION OF EXTREME TAIL PROBABILITIES By Peter Hall and Ishay Weissman Australian National University and Technion Applications of extreme

More information

Estimation of a quadratic regression functional using the sinc kernel

Estimation of a quadratic regression functional using the sinc kernel Estimation of a quadratic regression functional using the sinc kernel Nicolai Bissantz Hajo Holzmann Institute for Mathematical Stochastics, Georg-August-University Göttingen, Maschmühlenweg 8 10, D-37073

More information

Summer 2017 MATH Solution to Exercise 5

Summer 2017 MATH Solution to Exercise 5 Summer 07 MATH00 Solution to Exercise 5. Find the partial derivatives of the following functions: (a (xy 5z/( + x, (b x/ x + y, (c arctan y/x, (d log((t + 3 + ts, (e sin(xy z 3, (f x α, x = (x,, x n. (a

More information

A Gentle Introduction to Stein s Method for Normal Approximation I

A Gentle Introduction to Stein s Method for Normal Approximation I A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein

More information

Conditional Distributions

Conditional Distributions Conditional Distributions The goal is to provide a general definition of the conditional distribution of Y given X, when (X, Y ) are jointly distributed. Let F be a distribution function on R. Let G(,

More information

Sobolev Spaces. Chapter 10

Sobolev Spaces. Chapter 10 Chapter 1 Sobolev Spaces We now define spaces H 1,p (R n ), known as Sobolev spaces. For u to belong to H 1,p (R n ), we require that u L p (R n ) and that u have weak derivatives of first order in L p

More information

Solutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) =

Solutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) = Solutions to Exam, Math 56 The function f(x) e x + x 3 + x is one-to-one (there is no need to check this) What is (f ) ( + e )? Solution Because f(x) is one-to-one, we know the inverse function exists

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

x 2 y = 1 2. Problem 2. Compute the Taylor series (at the base point 0) for the function 1 (1 x) 3.

x 2 y = 1 2. Problem 2. Compute the Taylor series (at the base point 0) for the function 1 (1 x) 3. MATH 8.0 - FINAL EXAM - SOME REVIEW PROBLEMS WITH SOLUTIONS 8.0 Calculus, Fall 207 Professor: Jared Speck Problem. Consider the following curve in the plane: x 2 y = 2. Let a be a number. The portion of

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

Asymptotic Distributions for the Nelson-Aalen and Kaplan-Meier estimators and for test statistics.

Asymptotic Distributions for the Nelson-Aalen and Kaplan-Meier estimators and for test statistics. Asymptotic Distributions for the Nelson-Aalen and Kaplan-Meier estimators and for test statistics. Dragi Anevski Mathematical Sciences und University November 25, 21 1 Asymptotic distributions for statistical

More information

Fourier Series. (Com S 477/577 Notes) Yan-Bin Jia. Nov 29, 2016

Fourier Series. (Com S 477/577 Notes) Yan-Bin Jia. Nov 29, 2016 Fourier Series (Com S 477/577 otes) Yan-Bin Jia ov 9, 016 1 Introduction Many functions in nature are periodic, that is, f(x+τ) = f(x), for some fixed τ, which is called the period of f. Though function

More information

Discussion Paper No. 28

Discussion Paper No. 28 Discussion Paper No. 28 Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on the Circle Yasuhito Tsuruta Masahiko Sagae Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on

More information

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Closest Moment Estimation under General Conditions

Closest Moment Estimation under General Conditions Closest Moment Estimation under General Conditions Chirok Han Victoria University of Wellington New Zealand Robert de Jong Ohio State University U.S.A October, 2003 Abstract This paper considers Closest

More information

Asymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs

Asymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs Asymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs M. Huebner S. Lototsky B.L. Rozovskii In: Yu. M. Kabanov, B. L. Rozovskii, and A. N. Shiryaev editors, Statistics

More information

Statistica Sinica Preprint No: SS

Statistica Sinica Preprint No: SS Statistica Sinica Preprint No: SS-017-0013 Title A Bootstrap Method for Constructing Pointwise and Uniform Confidence Bands for Conditional Quantile Functions Manuscript ID SS-017-0013 URL http://wwwstatsinicaedutw/statistica/

More information

Week 9 The Central Limit Theorem and Estimation Concepts

Week 9 The Central Limit Theorem and Estimation Concepts Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population

More information

Anomalous transport of particles in Plasma physics

Anomalous transport of particles in Plasma physics Anomalous transport of particles in Plasma physics L. Cesbron a, A. Mellet b,1, K. Trivisa b, a École Normale Supérieure de Cachan Campus de Ker Lann 35170 Bruz rance. b Department of Mathematics, University

More information

Technical Supplement for: The Triangular Model with Random Coefficients

Technical Supplement for: The Triangular Model with Random Coefficients Technical Supplement for: The Triangular Model with andom Coefficients Stefan Hoderlein Boston College Hajo Holzmann Marburg March 9, 2016 Alexander Meister ostock Abstract We provide additional technical

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

Introduction to Machine Learning. Lecture 2

Introduction to Machine Learning. Lecture 2 Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for

More information

EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I. M. Bloznelis. April Introduction

EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I. M. Bloznelis. April Introduction EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I M. Bloznelis April 2000 Abstract. For symmetric asymptotically linear statistics based on simple random samples, we construct the one-term

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information