SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1
|
|
- Jason Gaines
- 5 years ago
- Views:
Transcription
1 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 B Technical details B.1 Variance of ˆf dec in the ersmooth case We derive the variance in the case where ϕ K (t) = ( 1 t 2) κ I( 1 t 1), i.e. we take K as in (A3), and assume that ϕ U (t) = a(t) exp ( t α), where r, α > and a denotes a symmetric, real-valued function satisfying a() = 1 and a(t) ξ t α 1 as t, (B.1) with < α 1 < and ξ >. That is, we put γ = 1, α 2 = α 1 and d = d 1 in (3.2). The more general setting at (3.2) obtains similarly. Assume too that the distribution of which f W is the density has finite variance: w 2 f W (w) dw <. (B.2) Recall the definition of K U at (2.2). In this notation, it is well known that the asymptotic variance of ˆf dec is given by Var ˆfdec (x) } = n 1 K 2 U f W (x) n 1 E ˆf dec (x) } 2. Theorem 3. If (B.1) (B.2) hold then, for each x, K 2 U f W (x) as h. 2 κ ξ 1 Γ(κ + 1) h (κ+1)α+α 1 exp ( h α) } 2 2π α κ+1 cos(x y)} 2 f W (y) dy (B.3) Proof. It is notationally convenient to put b = a 1. Note that 2π K U (x) = = h 1 cos(tx) ( 1 t 2) κ b(t/h) exp ( t/h α ) dt 1/h cos(hux) 1 (hu) 2} κ b(u) exp ( u α ) du = h I 1 (x; h), (B.4)
2 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 2 say. Let ϵ = ϵ(h) decrease to zero slowly as h, at a rate that we shall address shortly, and observe that, by (B.1), 1/h I 1(x; h) cos(hux) 1 (hu) 2} κ ( ) b(u) exp u α du (1 ϵ)/h (1 ϵ)/h C } 1 1 (hu) 2 κ (1 + u ) α 1 exp ( u α) du (1 ϵ)/h C 2 (1 + u ) α 1 exp ( u α) du C 3 h α 1 exp [ (1 ϵ)/h} α], (B.5) where, here and below, C 1, C 2,... denote generic positive constants not depending on h or x. Note too that 1/h (1 ϵ)/h cos(hux) 1 (hu) 2} κ b(u) exp ( u α ) du cos(x) 1/h C 3 ϵ x (1 ϵ)/h 1/h (1 ϵ)/h } 1 (hu) 2 κ ( ) b(u) exp u α du 1 (hu) 2 } κ (1 + u ) α 1 exp ( u α) du. (B.6) Since α > then, writing o u (1) for a generic function of u that satisfies u ϵ/h α o u (1) = o(1), and taking ϵ to decrease to zero so slowly that ϵ/h α, we have: 1/h 1 (hu) 2 } κ b(u) exp ( u α ) du (B.7) (1 ϵ)/h = ϵ/h 2 hv (hv) 2 } κ b ( h 1 v ) exp ( h 1 v α) dv ϵ/h = (2h) κ v ( κ 1 1 hv) κ ( 2 b h 1 v ) exp ( h 1 v α) dv ξ 1 (2h) κ h α 1 = ξ 1 (2h) κ h α 1 ϵ/h ϵ/h v κ exp ( h 1 v α) dv v κ exp ( h α 1 hv α) dv
3 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 3 ϵ/h α = 2 κ ξ 1 h (κ+1)α+α 1 1 = 2 κ ξ 1 h (κ+1)α+α 1 1 ϵ/h α u κ exp ( h α 1 h α u α) du u κ exp h α (1 α h α u) + o u (1) } du = 2 κ ξ 1 h (κ+1)α+α1 1 exp ( h α) ϵ/h α u κ exp α u + o u (1) } du 2 κ ξ 1 h (κ+1)α+α1 1 exp ( h α) u κ exp( α u) du (B.8) = 2κ ξ 1 Γ(κ + 1) α κ+1 h (κ+1)α+α 1 1 exp ( h α). (B.9) Combining (B.6) and (B.9) we deduce that 1/h cos(hux) 1 (hu) 2} κ ( ) b(u) exp u α du (1 ϵ)/h cos(x) + o unif (1)} 2κ ξ 1 Γ(κ + 1) α κ+1 C 4 ϵ x h (κ+1)α+α 1 1 exp ( h α), h (κ+1)α+α1 1 exp ( h α) (B.1) where, here and below, o unif (1) and O unif (1) denote quantities that depend on x and equal o(1) and O(1), respectively, uniformly in < x <, as h. Together, (B.5) and (B.1) imply that I 1 (x; h) = cos(x) 2κ ξ 1 Γ(κ + 1) α κ+1 h (κ+1)α+α 1 1 exp ( h α) + o unif (1) + O unif (1) ϵ x } h (κ+1)α+α 1 1 exp ( h α) + O unif (1) h α 1 exp [ (1 ϵ)/h} α]. (B.11) Together, (B.4) and (B.11) imply that, if ϵ decreases to zero slowly as h, KU 2 f W (x) = KU(x 2 y) f W (y) dy = h2 I (2π) 1(x 2 y; h) f 2 W (y) dy 2 κ ξ 1 Γ(κ + 1) = h (κ+1)α+α 1 exp ( h α) } 2 2π α κ+1 cos(x y)} 2 f W (y) dy + o[ h (κ+1)α+α 1 exp ( h α)} ] 2. (B.12) (Here we have used (B.2).) This result is equivalent to (B.3).
4 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 4 B.2 Main steps in derivation of convergence rate of ˆf rat (see (3.13)) in the case n 1/2 η Theorem 4 below demonstrates that, under the assumption that ˆθ = θ 1 + O p (n 1/2 ), the asymptotic bias of ˆfrat (x ˆθ) equals that of ˆfrat (x θ 1 ), and the error about the mean of ˆf rat (x ˆθ) equals that of ˆf dec (x) plus negligible terms. Write E fx for expectation when the distribution of the data W j has density f U f X ; and, for any random variable R with finite mean, let (1 E fx ) R denote R E fx (R). Theorem 4. If (3.1), (3.6) and (B1) (B6) in section 3.5 hold, then, for each ϵ >, ˆf rat (x ˆθ) (1 E fx ) ˆf dec (x) E fx ˆfrat (x θ 1 ) } = O p n ( ϵ nh 2α 1) } 1/2 + n 1/2 + h r s+1. (B.13) Moreover, the remainder term on the right-hand side is of the stated order uniformly in densities f X for which (3.1), (B1) (B6) in section 3.5, and (3.6) hold, in the sense that lim lim C n f X F > C P fx [ ˆf rat (x ˆθ) (1 E fx ) ˆf dec (x) E fx ˆfrat (x θ 1 ) } n ϵ ( nh 2α 1) 1/2 + n 1/2 + h r s+1 } ] =. (B.14) Our next result describes the bias of ˆf rat ( θ 1 ). Let s(ψ) be as in (3.15). Theorem 5. If (3.1), and (B1) (B6) in section 3.5 hold, then θ Θ h β s(ψ) } 1 EfX ˆfrat (x θ 1 )} f X (x) = O(1) as n. (B.15) Let Var fx denote the variance operator when the data W j have density f U f X. If (3.1), (B1), (B2) and (B3) hold, then standard deconvolution results imply that f X F Var fx ˆfdec (x) } = O (nh 2α+1 ) 1}, (B.16)
5 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 5 where the order is exact when f W (x) is nonzero. Hence, the remainder on the righthand side of (B.13) is negligibly small relative to the term (1 E fx ) ˆf dec (x) on the left-hand side, for any x such that f W (x) is nonzero. Combining (B.14), (B.15) and (B.16) we deduce that, under the conditions of Theorem 5, if the term in h r s+1 in (B.14) can be ignored then [ lim lim P fx ˆfrat (x) f X (x) (nh > C ) }] 2α+1 1/2 + h β s(ψ) =. C n θ Θ (B.17) Under the assumptions for Theorem 5, if ψ represents a sequence of functions, and we take h min1, (nη 2 ) 1/(2α+2β+1) }, then the minimiser, with respect to h, of the term within braces (i.e. the coefficient of C) on the left-hand side of (B.17) is of size (η 2α+1 /n β ) 1/(2α+2β+1) in the case n 1/2 η, in which instance h is asymptotic to a constant multiple of (nη 2 ) 1/(2α+2β+1). The term in h r s+1 in (B.14) can be ignored if h r s+1 = O(η 2α+1 /n β ) 1/(2α+2β+1) }, or equivalently, since h is asymptotic to a constant multiple of (nη 2 ) 1/(2α+2β+1), if n 1 = O ( η (2α+2r 2s+3)/(r s+1 β)) (B.18) (provided that r > s 1 + β). For example, if η = O(n t ) where t (, 1 2 ) and 2 t α + β (1 2 t) (r s + 1) t, then (B.18) holds, in which case h r s+1 can be ignored in (B.2) and therefore property (3.13) follows from (B.17). B.3 Proof of Theorem 4 Observe that 2π f(x θ) L x (w θ, h) =f(x θ) =f(x θ) e it(x w) = ϕ U (t) dt e itw ϕ K x( θ,h)(t) dt ϕ U (t) e itw ϕ U (t) dt e it(x hu) K(u) f(x hu θ) du e ihtu f(x θ) K(u) f(x hu θ) du
6 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 6 = r k= h k e it(x w) k! g k(x θ) ϕ U (t) dt + h r+1 e it(x w) ϕ U (t) dt Since γ h,x has s derivatives then s integrations by parts give: e ihtu ω(hu, x θ) u r+1 K(u) du = (iht) s Therefore, e it(x w) ϕ U (t) dt e ihtu ω(hu, x θ) u r+1 K(u) du 1 dt γ h,x (u θ) du + ϕ U (t) 1 e ihtu u k K(u) du e ihtu ω(hu, x θ) u r+1 K(u) du. t >1 e ihtu γ (s) h,x (u θ) du. Combining this bound with (B.19), and noting (B6), we deduce that: r f(x θ) L h k g k (x θ) x(w θ, h) <w< θ Θ 2πk! k= 1 B 4 h r+1 dt ϕ U (t) + 1 (B.19) dt γ (s) ht s h,x ϕ U (t) (u θ) du. e it(x w) ϕ U (t) dt t >1 dt ht s ϕ U (t) e ihtu u k K(u) du }, (B.2) uniformly in < h C 1, where B 4 > is a constant. Property (3.1) implies that the first integral on the right-hand side of (B.2) is finite, and (3.1) and (B6) imply that the second integral is bounded above by t >1 dt ht s ϕ U (t) C (1 + t ) α 1 dt = O ( h s). t >1 ht s These results, (B.2) and the property e ihtu u k K(u) du = ( i) k ϕ (k) K ( ht), entail: f(x θ) L x(w θ, h) <w< θ Θ r ( ih) k g k (x θ) e it(w x) 2πk! ϕ U (t) ϕ K (k) (ht) dt = O( h r s+1). (B.21) k=
7 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 7 Note that (B.21), and (B.22) and (B.24) below, involve only the densities f( θ) and f U, which we take to be fixed; these formulae do not involve f X, which can vary with n. Let s k = ( 1) k/2 if k is even, and s k = ( 1) (k 1)/2 if k is odd. Since the distribution of U is symmetric then ϕ U is symmetric, and so, defining cos(tu) ϕ U (t) ϕ K (k) (ht) dt if k is even ψ k (u, h) = sin(tu) ϕ U (t) ϕ K (k) (ht) dt if k is odd we have: <w< θ Θ Hence, defining f(x θ) L x(w θ, h) 1 2π r k= s k h k k! Ψ k (x) = 1 n g k (x θ) ψ k (w x, h) = O( h r s+1). n ψ k (W j x, h), j=1 (B.22) (B.23) we have: for a constant B 5 >, not depending on h, n or the data W 1,..., W n, ˆf rat (x θ) 1 r s k h k g k (x θ) 2π k! Ψ k (x) B 5 h r s+1, (B.24) θ Θ k= where the right-hand side denotes a deterministic quantity of the stated size. Define and note that, by (B2), L k (u) = L k (u) B 6 h α <u< exp(itu) ϕ U (t/h) ϕ K (k) (t) dt (1 + t ) α ϕk (k) (t) dt B7 h α, (B.25) where B 6, B 7 > are constants not depending on h or n. Using (3.1), (B1), (B2) and (B.25) we deduce that, for each integer m 2, h m E ψ k (W x, h) m} E [ L k (W x)/h} m] L k m 2 E [ L k (W x)/h} 2]
8 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 8 =h L k m 2 L k (u) 2 f W (x + hu) du B 8 h ( B 7 h α) m 2 L k (u) 2 du =B 9 (m) h 1 (m 2)α (2π) 1 ϕ (k) K (t) ϕ U (t/h) B 1 (m) h 1 (m 2)α (1 + t/h ) α ϕ K (k) (t) } 2 dt =O ( h 1 mα), (B.26) 2 dt where the bounds apply uniformly in x I, and B 8 = f W, B 9 (m) = B m 2 7 B 8, and B l (m) or B l, for l 1, denote constants not depending on h or n. (The second identity in the string at (B.26) follows using Parseval s identity. Note that if (B1) holds then f X (x) is bounded uniformly in x and n, from which it follows that the same is true for f W (x).) Therefore, E ψ k (W x, h) m} = O ( h 1 m(α+1)). (B.27) Let 1 j r, and recall the definition of Ψ k at (B.23). Replacing m by either 2 or 2m in the bound at (B.27), we have for each m 1, using Rosenthal s identity (see e.g. Hall and Heyde (198), p. 23): E (1 E) Ψj (x) ( [ 2m B 11 (m) n 1 E ψ k (W x, h) 2}] m + n (2m 1) E ψ k (W x, h) 2m}) (nh 2α+1 B 12 (m) ) m ( + h α nh α+1) } (2m 1) B 13 (m) ( nh 2α+1) m, uniformly in x I, where the last inequality is valid whenever nh 1. Hence, if I(n) is a set of at most n B values x I, then, for each ϵ >, P (n) (1 E) Ψj (x) ( > n ϵ nh 2α+1) } 1/2
9 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 9 n B (1 P E) Ψj (x) ( > n ϵ nh 2α+1) } 1/2 (n) n B n ϵ ( nh 2α+1) 1/2 } 2m E (1 E) Ψj (x) 2m B 14 (m) n B 2mϵ. (B.28) Observe too that if 1 k r and x, x I, then, using (3.1), (B2) and (B3), Ψk (x) Ψ k (x ) 1 n n j=1 C 11 n 1 x x ψk (W j x, h) ψ k (W j x, h) n j=1 t (1 + t ) α ϕk (k) (ht) dt B 15 x x h (α+2) B 16 x x n (α+2)/(2α+1). The last inequality follows from the fact that, by (B3), nh 2α+1 is bounded away from. Therefore, if I(n) represents a grid in I with edge width n 1 (α+2)/(2α+1), and if for each x I we define x to be the point in I(n) nearest to x, then and therefore Ψk (x) Ψ k (x ) B17 n 1 (α+2)/(2α+1) n (α+2)/(2α+1) = B 17 n 1, (1 E) Ψk (x) Ψ k (x ) } 2 B 17 n 1. Hence the following version of (B.28), with I(n) there replaced by I, holds: for each ϵ >, P (1 E) Ψ j (x) > n ( ϵ nh 2α+1) } 1/2. Combining (B.24) and (B.29) we deduce that, for each ϵ >, θ Θ (B.29) ˆf rat (x θ) g (x θ) Ψ (x) 1 r s k h k g k (x θ) E Ψ k (x)} 2π k! k=1 = O p n ( ϵ nh 2α 1) } 1/2 + h r s+1. (B.3) Note too that e it(w x) E ϕ U (t) } e ϕ (k) itx ϕ X (t) ϕ U (t) K (ht) dt = ϕ U (t) ϕ K (k) (ht) dt
10 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 1 = e itx ϕ X (t) ϕ K (k) (ht) dt = 2π i k f X (x hu) u k K(u) du, which implies that E Ψk (x) } = 2π s k f X (x hu) u k K(u) du. (B.31) Furthermore, g (x θ) 1. Hence, by (B.3), ˆf rat (x θ) Ψ r (x) θ Θ where ϵ > is arbitrary. k=1 h k k! g k(x θ) f X (x hu) u k K(u) du = O p n ϵ ( nh 2α 1) 1/2 + h r s+1 }, (B.32) Assumption (B5) implies that g k (x θ), and its first derivative with respect to θ, are bounded uniformly in x I and θ Θ. Using this result, (B5) and (3.6) we deduce that: max gk (x ˆθ) g k (x θ 1 ) ( ) = Op n 1/2. 1 k r Conditions (B1) and (B2) imply that f X (x hu) u k K(u) du is bounded uniformly in x and h, for k =,..., r. Combining these results with (B.32), and recalling that g (x θ 1 ) 1, we deduce that, for each ϵ >, ˆf rat (x ˆθ) (1 E) Ψ r (x) h k k! g k(x θ 1 ) f X (x hu) u k K(u) du k= = O p n ( ϵ nh 2α 1) } 1/2 + n 1/2 + h r s+1. (B.33) Result (B.22) implies that f(x θ 1) EL x (W θ 1, h)} r k= h k k! g k(x θ 1 ) f X (x hu) u k K(u) du} = O( h r s+1). (B.34) Since Ψ = ˆf dec where ˆf dec is as defined at (2.6), and E ˆfrat (x θ 1 ) } = f(x θ 1 ) EL x (W θ 1, h)},
11 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 11 where ˆf rat (x θ) is as defined at (2.3), then (B.33) and (B.34) together imply (B.13). The results mentioned in the second and third sentences of this paragraph, together with (B.24), (B.29), (B.31) and (B.34), similarly imply (B.14). B.4 Proof of Theorem 5 Noting that f X = f( θ 1 ) + ψ, and that, by construction, EL x (W θ 1, h)} = EK x (W θ 1, h)}, we have: EL x (W θ 1, h)} = K x (u θ 1, h) du 1 e itu ϕ X (t) dt 2π = 1 ( x u ) fx (u) K h h f(u θ 1 ) du = K(u) f(x hu θ 1) + ψ(x hu) du f(x hu θ 1 ) ψ(x hu) =1 + K(u) f(x hu θ 1 ) du. Therefore, in notation introduced in section 3.5, E ˆfrat (x θ 1 )} f X (x) = = K(u) f(x θ 1 ) f(x hu θ 1 ) ψ(x hu) du ψ(x) K(u) ψ(x hu) ψ(x)} du + h r+1 γ h,x (u θ 1 ) ψ(x hu) du + r k=1 h k k! g k(x θ 1 ) u k K(u) ψ(x hu) du (B.35) =O h β s(ψ) }, (B.36) uniformly in θ Θ and x I. This is equivalent to (B.15). The last identity in (B.36) follows on using the moment condition u j K(u) du =, for j = 1,..., β 1.
12 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 12 C Details of implementation C.1 Using a ridge when computing ˆf rat If U has a Laplace distribution with scale parameter λ, so that ϕ U (t) = (1 + λ 2 t 2 ) 1, then if K and f( θ) are twice differentiable, we can write L x (u θ, h) = 1 2π e itu (1 + λ 2 t 2 ) ϕ Kx (t θ, h)dt = K x (u θ, h) λ 2 K x(u θ, h). More generally, if U has the distribution of an N-fold Laplace convolution, then L x is a linear combination of derivatives of K x of the form K x (k) (W j θ, h) = k K(x u)/h} = u k hf(u θ) u=wj k+1 l=1 g l (W j x, h, θ) f l (W j θ), for some functions g l which are sums and products of positive powers of K(x W j )/h}, f(w j θ) and their derivatives. In practice, the denominators f l (W j θ) are often too close to zero for some W j s, which makes the estimator work rather poorly. To avoid this problem, we can use a ridge parameter in the denominators. We tried several approaches to ridging in the particular Laplace error case, and found that the following approach performed well in practice: replace f l ( θ) by maxf l ( θ), δ} where δ > is a ridge parameter. In our numerical work, we used this approach with δ =.4 f( θ). C.2 Details of SIMEX bandwidth for ˆf rat It follows from our theoretical results that, under regularity conditions, the asymptotic mean integrated squared error (AMISE) of ˆf rat is equal to AMISE ˆfrat ( θ) } = h 4 µ 2 2(K)Rr f( θ)}/4 + (2πnh) 1 ϕ K (t) 2 ϕ U (t/h) 2 dt, where µ 2 (K) = x 2 K(x) dx, r = f X /f( θ), and we used the notation R(f) = f 2. We suggest choosing h for ˆf rat by minimising a SIMEX estimator of the AMISE;
13 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 13 see Cook and Stefanski (1994) and Stefanski and Cook (1995) for an introduction to SIMEX. The idea of SIMEX methods is that, in some way, the relation between data from f W and f X can be mimicked by that between data from f W (1) f W f U and f W (2) f W (1) f U. Now, quantities related to f W are easy to estimate from the data, since we have a sample of W i s and can generate data from f W (1) be exploited to estimate unknown quantities related to f X. and f W (2). This can Let h, h 1 and h 2 be the bandwidths that minimise, respectively, AMISE ˆfrat ( θ) }, AMISE ˆfrat,W ( θ) } and AMISE ˆfrat,W (1)( θ) }, where ˆf rat,w and ˆf rat,w (1) denote the ratio estimators of, respectively, f W and f W (1) computed from a sample of size n having a density, respectively f W (1) and f W (2). Then, extending the SIMEX-based bandwidth of Delaigle and Hall (28) to our problem, h /h 1 can be mimicked by h 1 /h 2, and hence h can be approximated by ĥ2 1/ĥ2, where ĥj is an estimator of h j, for j = 1, 2. To estimate h j, construct a sample W (1) 1,..., W (1) n from f W (1) and a sample W (2) 1,..., W n (2) from f W (2), by taking W (1) i = W i + U (1) i and W (2) i = W (1) i + U (2) i for j = 1, 2, U (j) 1,..., U n (j) where, is a sample of independent observations generated from f U, independently of the W i s. Let f W () f W, f () ( θ) f( θ) f U, f (1) ( θ) = f () ( θ) f U, f (2) ( θ) = f (1) ( θ) f U, and, for j =, 1, 2, let r (j) = f W (j)/f (j) ( θ). By definition, we have, for j =, 1, h 4 µ 2 h j+1 = argmin 2(K) R r h 4 (j)f (j) ( θ) } + 1 ϕ K (t) 2πnh ϕ U (t/h) and to estimate h j+1 it suffices to estimate r (j) (x) and f (j)( θ). We estimate the latter by f (j) ( ˆθ), and to estimate the former, we take ˆr (j) (x) = ˇf W (j)(x)/f (j) (x ˆθ)}, where ˇf W () and ˇf W (1) denote standard error-free kernel estimators of f W and f W (1) computed from the data W 1,..., W n and W (1) 1,..., W n (1), respectively, and using, for example, a normal reference bandwidth for estimating second derivatives of densities; see Sheather and Jones (1991) and the references therein. In practice, to compute 2 dt,
14 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 14 R r (j) f (j)( θ) } we truncate the integral to the interval [W.5, (j) W.95], (j) where W α () and W α (1) denote the 1αth percentile of, respectively, the W i s and the W (1) s. Again, here, when computing ˆr (j)(x) = ˇf (x) ˇfW W (j) f (j) (x θ) (j)(x)f(j) (x θ) f(j) 2 (x θ) 2 ˇf W (j)(x)f(j) (x θ) + (x θ) f 2 (j) 2 ˇf W (j)(x)f(j) (x θ)}2 f(j) 3 (x θ), we need to use a ridge at each denominator. We implemented this procedure in the Laplace case only, and we used the same ridge as the one described in section C.1. Note that estimated bandwidth depends on the SIMEX sample W (j) 1..., W n (j). To reduce the effect of the random sampling step, as in Delaigle and Hall (28) we generate B sets of SIMEX samples of size n (we took B = 3), and take h j+1 to minimise the average of the corresponding B estimated AMISE values. C.3 NR bandwidth for ˆf dec with the sinc kernel The mean integrated squared error (MISE) of ˆf dec, computed with the sinc kernel, is given by i MISE(h) = 1 2πn 1/h 1/h ϕ U (t) 2 dt 1 + n 1 2π 1/h 1/h ϕ X (t) 2 dt + f 2 X, (C.1) and to find the bandwidth that minimises the MISE, we need to find the roots of MISE (h) = 1 πnh 2 ϕ U(1/h) 2 + n + 1 nπh 2 ϕ W (1/h) 2 ϕ U (1/h) 2. Equivalently, this bandwidth is a solution of 1 + (n + 1) ϕ W (1/h) 2 =. The NR rule assumes that X N(ˆµ, ˆσ 2 ), where ˆµ = W and ˆσ 2 = W and Var(W ) Var(U), with Var(W ) denoting the empirical mean and variance of the W i s. This amounts to estimating ϕ W (1/h) by ˆϕ W (1/h) = ϕ U (1/h) exp(iˆµ/h) exp(.5ˆσ 2 /h 2 ). Then the bandwidth is estimated by ĥnr, the solution of 1 + (n + 1) ˆϕ W (1/h) 2 =.
15 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 15 D Additional simulation results Tables D.1 and D.3 report the integrated squared bias (ISB), computed from 1 samples, of the estimators ˆf dec, ˆfbco and ˆf wgt for the various cases considered in our simulations. These tables show that, as could be expected, ˆfbco is less biased than ˆf dec. ˆfwgt also benefits from a bias reduction in the Laplace error case, and also in the normal case for densities (i) and (ii). However, for the other densities, in the normal error case, the ISB of ˆfwgt is slightly larger than that of ˆfdec. This is because θ is more difficult to estimate in the normal error case than in the Laplace error case. As a result, in the normal error case, the parametric component of ˆf wgt is less good, and the weight ŵ of ˆf wgt, which also depends on θ, is smaller. See Table D.2, where we show the median and interquartile range of ŵ, when estimating densities (i) to (iii). Table D.1: 1 3 ISB of 1 estimators of densities (i) to (iii) in the Laplace and normal error cases, when NSR = 1% or 25% and n = 2 or 7, using the estimators ˆf dec, ˆf bco ( ; ˆθ MD ) and ˆf wgt ( ; ˆθ ML ). Laplace Normal Density (i) Density (ii) Density (iii) Density (i) Density (ii) Density (iii) n n NSR = 1% NSR = 1% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML ) NSR = 25% NSR = 25% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML )
16 SUPPLEMENTARY MATERIAL FOR PUBLICATION ONLINE 16 Table D.2: Median (IQR) of 1 values of ŵ when estimating densities (i) to (iii) in the Laplace and normal error cases, with NSR = 1% or 25% and n = 2 or 7. Density (i) Density (ii) Density (iii) n ˆθ ML Lap, NSR = 1%.7 (.7).74 (.7).71 (.9).73 (.7).68 (.9).72 (.1) Lap, NSR = 25%.73 (.7).76 (.5).73 (.9).78 (.7).71 (.8).73 (.8) Norm, NSR = 1%.29 (.9).28 (.12).29 (.9).28 (.9).14 (.7).5 (.2) Norm, NSR = 25%.33 (.1).33 (.9).3 (.9).33 (.12).24 (.13).17 (.9) Table D.3: 1 3 ISB of 1 estimators of densities (iv) to (vi) in the normal error case, when NSR = 1% or 25% and n = 2 or 7, using the estimators ˆf dec, ˆf bco ( ; ˆθ MD ) and ˆf wgt ( ; ˆθ ML ). Density (iv) Density (v) Density (vi) Density (iv) Density (v) Density (vi) n n NSR = 1% NSR = 25% ˆf dec ˆf bco ( ; ˆθ MD ) ˆf wgt ( ; ˆθ ML ) References Cook, J.R. and Stefanski, L.A. (1994). Simulation-extrapolation estimation in parametric measurement error models. J. Amer. Statist. Assoc., Hall, P. and Heyde, C. (198). Martingale Limit Theory and its Application. Academic Press. Sheather, S. J. and Jones, M. C. (1991). A reliable data-based bandwidth selection method for kernel density estimation. J. Royal Statist. Society Ser. B., Stefanski, L. and Cook, J.R. (1995). Simulation-extrapolation: The measurement error jackknife. J. Amer. Statist. Assoc
Goodness-of-fit tests for the cure rate in a mixture cure model
Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,
More informationarxiv: v1 [stat.me] 17 Jan 2008
Some thoughts on the asymptotics of the deconvolution kernel density estimator arxiv:0801.2600v1 [stat.me] 17 Jan 08 Bert van Es Korteweg-de Vries Instituut voor Wiskunde Universiteit van Amsterdam Plantage
More informationOn robust and efficient estimation of the center of. Symmetry.
On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract
More informationLocal Polynomial Regression
VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based
More informationDensity estimators for the convolution of discrete and continuous random variables
Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract
More informationA Novel Nonparametric Density Estimator
A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with
More informationO Combining cross-validation and plug-in methods - for kernel density bandwidth selection O
O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem
More informationSmooth simultaneous confidence bands for cumulative distribution functions
Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang
More informationNonparametric confidence intervals. for receiver operating characteristic curves
Nonparametric confidence intervals for receiver operating characteristic curves Peter G. Hall 1, Rob J. Hyndman 2, and Yanan Fan 3 5 December 2003 Abstract: We study methods for constructing confidence
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend
More informationTime Series and Forecasting Lecture 4 NonLinear Time Series
Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationVARIANCE REDUCTION BY SMOOTHING REVISITED BRIAN J E R S K Y (POMONA)
PROBABILITY AND MATHEMATICAL STATISTICS Vol. 33, Fasc. 1 2013), pp. 79 92 VARIANCE REDUCTION BY SMOOTHING REVISITED BY ANDRZEJ S. KO Z E K SYDNEY) AND BRIAN J E R S K Y POMONA) Abstract. Smoothing is a
More informationPointwise convergence rates and central limit theorems for kernel density estimators in linear processes
Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence
More informationBandwith selection based on a special choice of the kernel
Bandwith selection based on a special choice of the kernel Thomas Oksavik Master of Science in Physics and Mathematics Submission date: June 2007 Supervisor: Nikolai Ushakov, MATH Norwegian University
More information12 - Nonparametric Density Estimation
ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6
More informationECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd
ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram
More informationOn variable bandwidth kernel density estimation
JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced
More informationQuantitative Economics for the Evaluation of the European Policy. Dipartimento di Economia e Management
Quantitative Economics for the Evaluation of the European Policy Dipartimento di Economia e Management Irene Brunetti 1 Davide Fiaschi 2 Angela Parenti 3 9 ottobre 2015 1 ireneb@ec.unipi.it. 2 davide.fiaschi@unipi.it.
More informationNonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction
Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(
More information36. Multisample U-statistics and jointly distributed U-statistics Lehmann 6.1
36. Multisample U-statistics jointly distributed U-statistics Lehmann 6.1 In this topic, we generalize the idea of U-statistics in two different directions. First, we consider single U-statistics for situations
More informationTIKHONOV S REGULARIZATION TO DECONVOLUTION PROBLEM
TIKHONOV S REGULARIZATION TO DECONVOLUTION PROBLEM Dang Duc Trong a, Cao Xuan Phuong b, Truong Trung Tuyen c, and Dinh Ngoc Thanh a a Faculty of Mathematics and Computer Science, Ho Chi Minh City National
More informationLog-Density Estimation with Application to Approximate Likelihood Inference
Log-Density Estimation with Application to Approximate Likelihood Inference Martin Hazelton 1 Institute of Fundamental Sciences Massey University 19 November 2015 1 Email: m.hazelton@massey.ac.nz WWPMS,
More informationConvergence rates in weighted L 1 spaces of kernel density estimators for linear processes
Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,
More information1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).
Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,
More informationContinuous Distributions
Chapter 3 Continuous Distributions 3.1 Continuous-Type Data In Chapter 2, we discuss random variables whose space S contains a countable number of outcomes (i.e. of discrete type). In Chapter 3, we study
More informationAdaptive Kernel Estimation of The Hazard Rate Function
Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized
More informationKernel Density Estimation
Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationPartial Differential Equations
Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,
More information1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =
Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values
More informationBoundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta
Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density
More informationKernel density estimation of reliability with applications to extreme value distribution
University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School 2008 Kernel density estimation of reliability with applications to extreme value distribution Branko Miladinovic
More informationIntegral approximation by kernel smoothing
Integral approximation by kernel smoothing François Portier Université catholique de Louvain - ISBA August, 29 2014 In collaboration with Bernard Delyon Topic of the talk: Given ϕ : R d R, estimation of
More informationDensity Estimation with Replicate Heteroscedastic Measurements
Ann Inst Stat Math (011) 63:81 99 DOI 10.1007/s10463-009-00-x Density Estimation with Replicate Heteroscedastic Measurements Julie McIntyre Leonard A. Stefanski Received: 9 January 008 / Revised: 4 September
More informationRandom fraction of a biased sample
Random fraction of a biased sample old models and a new one Statistics Seminar Salatiga Geurt Jongbloed TU Delft & EURANDOM work in progress with Kimberly McGarrity and Jilt Sietsma September 2, 212 1
More informationDensity Deconvolution for Generalized Skew-Symmetric Distributions
Density Deconvolution for Generalized Skew-Symmetric Distributions Cornelis J. Potgieter Department of Statistical Science, Southern Methodist University, Dallas, TX arxiv:1706.01507v1 [stat.me] 5 Jun
More informationON SOME TWO-STEP DENSITY ESTIMATION METHOD
UNIVESITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XLIII 2005 ON SOME TWO-STEP DENSITY ESTIMATION METHOD by Jolanta Jarnicka Abstract. We introduce a new two-step kernel density estimation method,
More informationON THE CHOICE OF TEST STATISTIC FOR CONDITIONAL MOMENT INEQUALITES. Timothy B. Armstrong. October 2014 Revised July 2017
ON THE CHOICE OF TEST STATISTIC FOR CONDITIONAL MOMENT INEQUALITES By Timothy B. Armstrong October 2014 Revised July 2017 COWLES FOUNDATION DISCUSSION PAPER NO. 1960R2 COWLES FOUNDATION FOR RESEARCH IN
More informationthe convolution of f and g) given by
09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that
More informationMinimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.
Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of
More informationON DECONVOLUTION WITH REPEATED MEASUREMENTS
ON DECONVOLUTION WITH REPEATED MEASUREMENTS Aurore Delaigle 1,2 Peter Hall 2,3 Alexander Meister 2,4 ABSTRACT. In a large class of statistical inverse problems it is necessary to suppose that the transformation
More informationNonparametric Density Estimation (Multidimension)
Nonparametric Density Estimation (Multidimension) Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann February 19, 2007 Setup One-dimensional
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationEstimation of the Bivariate and Marginal Distributions with Censored Data
Estimation of the Bivariate and Marginal Distributions with Censored Data Michael Akritas and Ingrid Van Keilegom Penn State University and Eindhoven University of Technology May 22, 2 Abstract Two new
More informationA Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem
A Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem Aurore Delaigle, Jianqing Fan, and Raymond J. Carroll Abstract: Local polynomial estimators are popular techniques for nonparametric
More informationζ (s) = s 1 s {u} [u] ζ (s) = s 0 u 1+sdu, {u} Note how the integral runs from 0 and not 1.
Problem Sheet 3. From Theorem 3. we have ζ (s) = + s s {u} u+sdu, (45) valid for Res > 0. i) Deduce that for Res >. [u] ζ (s) = s u +sdu ote the integral contains [u] in place of {u}. ii) Deduce that for
More informationProperties of Principal Component Methods for Functional and Longitudinal Data Analysis 1
Properties of Principal Component Methods for Functional and Longitudinal Data Analysis 1 Peter Hall 2 Hans-Georg Müller 3,4 Jane-Ling Wang 3,5 Abstract The use of principal components methods to analyse
More informationEXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018
EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018 While these notes are under construction, I expect there will be many typos. The main reference for this is volume 1 of Hörmander, The analysis of liner
More informationUNIVERSITÄT POTSDAM Institut für Mathematik
UNIVERSITÄT POTSDAM Institut für Mathematik Testing the Acceleration Function in Life Time Models Hannelore Liero Matthias Liero Mathematische Statistik und Wahrscheinlichkeitstheorie Universität Potsdam
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationThe Central Limit Theorem: More of the Story
The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit
More informationINTRODUCTION TO REAL ANALYSIS II MATH 4332 BLECHER NOTES
INTRODUCTION TO REAL ANALYSIS II MATH 433 BLECHER NOTES. As in earlier classnotes. As in earlier classnotes (Fourier series) 3. Fourier series (continued) (NOTE: UNDERGRADS IN THE CLASS ARE NOT RESPONSIBLE
More informationEcon 582 Nonparametric Regression
Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume
More informationExtending clustered point process-based rainfall models to a non-stationary climate
Extending clustered point process-based rainfall models to a non-stationary climate Jo Kaczmarska 1, 2 Valerie Isham 2 Paul Northrop 2 1 Risk Management Solutions 2 Department of Statistical Science, University
More information1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationMATH 205C: STATIONARY PHASE LEMMA
MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)
More informationNonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix
Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract
More informationAdditive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535
Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June
More informationStochastic Processes
Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54 Summary from measure theory De nition (X, A) is a measurable space
More informationh(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote
Real Variables, Fall 4 Problem set 4 Solution suggestions Exercise. Let f be of bounded variation on [a, b]. Show that for each c (a, b), lim x c f(x) and lim x c f(x) exist. Prove that a monotone function
More informationEfficient Regressions via Optimally Combining Quantile Information
Efficient Regressions via Optimally Combining Quantile Information Zhibiao Zhao Penn State University Zhijie Xiao Boston College September 29, 2011 Abstract We study efficient estimation of regression
More informationBootstrap of residual processes in regression: to smooth or not to smooth?
Bootstrap of residual processes in regression: to smooth or not to smooth? arxiv:1712.02685v1 [math.st] 7 Dec 2017 Natalie Neumeyer Ingrid Van Keilegom December 8, 2017 Abstract In this paper we consider
More informationA Note on Tail Behaviour of Distributions. the max domain of attraction of the Frechét / Weibull law under power normalization
ProbStat Forum, Volume 03, January 2010, Pages 01-10 ISSN 0974-3235 A Note on Tail Behaviour of Distributions in the Max Domain of Attraction of the Frechét/ Weibull Law under Power Normalization S.Ravi
More informationarxiv: v2 [stat.me] 3 Dec 2016
Accelerated Nonparametric Maximum Likelihood Density Deconvolution Using Bernstein Polynomial arxiv:1601.0642v2 [stat.me] Dec 2016 Abstract Zhong Guan Department of Mathematical Sciences Indiana University
More informationSome Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model
Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;
More information6 Lecture 6b: the Euler Maclaurin formula
Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Fall 217 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 217 March 26, 218 6 Lecture 6b: the Euler Maclaurin formula
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationMATH 220 solution to homework 4
MATH 22 solution to homework 4 Problem. Define v(t, x) = u(t, x + bt), then v t (t, x) a(x + u bt) 2 (t, x) =, t >, x R, x2 v(, x) = f(x). It suffices to show that v(t, x) F = max y R f(y). We consider
More informationMonte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan
Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis
More informationON THE ESTIMATION OF EXTREME TAIL PROBABILITIES. By Peter Hall and Ishay Weissman Australian National University and Technion
The Annals of Statistics 1997, Vol. 25, No. 3, 1311 1326 ON THE ESTIMATION OF EXTREME TAIL PROBABILITIES By Peter Hall and Ishay Weissman Australian National University and Technion Applications of extreme
More informationEstimation of a quadratic regression functional using the sinc kernel
Estimation of a quadratic regression functional using the sinc kernel Nicolai Bissantz Hajo Holzmann Institute for Mathematical Stochastics, Georg-August-University Göttingen, Maschmühlenweg 8 10, D-37073
More informationSummer 2017 MATH Solution to Exercise 5
Summer 07 MATH00 Solution to Exercise 5. Find the partial derivatives of the following functions: (a (xy 5z/( + x, (b x/ x + y, (c arctan y/x, (d log((t + 3 + ts, (e sin(xy z 3, (f x α, x = (x,, x n. (a
More informationA Gentle Introduction to Stein s Method for Normal Approximation I
A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein
More informationConditional Distributions
Conditional Distributions The goal is to provide a general definition of the conditional distribution of Y given X, when (X, Y ) are jointly distributed. Let F be a distribution function on R. Let G(,
More informationSobolev Spaces. Chapter 10
Chapter 1 Sobolev Spaces We now define spaces H 1,p (R n ), known as Sobolev spaces. For u to belong to H 1,p (R n ), we require that u L p (R n ) and that u have weak derivatives of first order in L p
More informationSolutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) =
Solutions to Exam, Math 56 The function f(x) e x + x 3 + x is one-to-one (there is no need to check this) What is (f ) ( + e )? Solution Because f(x) is one-to-one, we know the inverse function exists
More informationModelling Non-linear and Non-stationary Time Series
Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September
More informationx 2 y = 1 2. Problem 2. Compute the Taylor series (at the base point 0) for the function 1 (1 x) 3.
MATH 8.0 - FINAL EXAM - SOME REVIEW PROBLEMS WITH SOLUTIONS 8.0 Calculus, Fall 207 Professor: Jared Speck Problem. Consider the following curve in the plane: x 2 y = 2. Let a be a number. The portion of
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationAsymptotic Distributions for the Nelson-Aalen and Kaplan-Meier estimators and for test statistics.
Asymptotic Distributions for the Nelson-Aalen and Kaplan-Meier estimators and for test statistics. Dragi Anevski Mathematical Sciences und University November 25, 21 1 Asymptotic distributions for statistical
More informationFourier Series. (Com S 477/577 Notes) Yan-Bin Jia. Nov 29, 2016
Fourier Series (Com S 477/577 otes) Yan-Bin Jia ov 9, 016 1 Introduction Many functions in nature are periodic, that is, f(x+τ) = f(x), for some fixed τ, which is called the period of f. Though function
More informationDiscussion Paper No. 28
Discussion Paper No. 28 Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on the Circle Yasuhito Tsuruta Masahiko Sagae Asymptotic Property of Wrapped Cauchy Kernel Density Estimation on
More informationIEOR 165 Lecture 7 1 Bias-Variance Tradeoff
IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationClosest Moment Estimation under General Conditions
Closest Moment Estimation under General Conditions Chirok Han Victoria University of Wellington New Zealand Robert de Jong Ohio State University U.S.A October, 2003 Abstract This paper considers Closest
More informationAsymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs
Asymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs M. Huebner S. Lototsky B.L. Rozovskii In: Yu. M. Kabanov, B. L. Rozovskii, and A. N. Shiryaev editors, Statistics
More informationStatistica Sinica Preprint No: SS
Statistica Sinica Preprint No: SS-017-0013 Title A Bootstrap Method for Constructing Pointwise and Uniform Confidence Bands for Conditional Quantile Functions Manuscript ID SS-017-0013 URL http://wwwstatsinicaedutw/statistica/
More informationWeek 9 The Central Limit Theorem and Estimation Concepts
Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population
More informationAnomalous transport of particles in Plasma physics
Anomalous transport of particles in Plasma physics L. Cesbron a, A. Mellet b,1, K. Trivisa b, a École Normale Supérieure de Cachan Campus de Ker Lann 35170 Bruz rance. b Department of Mathematics, University
More informationTechnical Supplement for: The Triangular Model with Random Coefficients
Technical Supplement for: The Triangular Model with andom Coefficients Stefan Hoderlein Boston College Hajo Holzmann Marburg March 9, 2016 Alexander Meister ostock Abstract We provide additional technical
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationIntroduction to Machine Learning. Lecture 2
Introduction to Machine Learning Lecturer: Eran Halperin Lecture 2 Fall Semester Scribe: Yishay Mansour Some of the material was not presented in class (and is marked with a side line) and is given for
More informationEMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I. M. Bloznelis. April Introduction
EMPIRICAL EDGEWORTH EXPANSION FOR FINITE POPULATION STATISTICS. I M. Bloznelis April 2000 Abstract. For symmetric asymptotically linear statistics based on simple random samples, we construct the one-term
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationWeb Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.
Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More information