Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Size: px
Start display at page:

Download "Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model"

Transcription

1 Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004; Fan and Huang, 2005) is a very useful semiparametric model which models the key covariates linearly and models the rest of the covariates nonparametrically. Suppose that we are given a random sample {(x i, z i, U i, Y i ); 1 i n}, Y i is the response variable and (x i, z i, U i ) are covariates. The varying-coefficient partially linear model assumes the following conditional linear structure: Y = x T α(u) + z T β + ϵ, (1.1) x = (X 1,..., X p ) T, α(u) = {α 1 (u),..., α p (u)} T, β = (β 1,..., β q ) T, and E(ϵ x, z, U) = 0. Here, we allow the dependence of ϵ on (x, z, U). In this article, we assume U is a scalar. The results can be extended for multivariate U. However, the extension may not be useful in practice due to the curse of dimensionality. Note that, if β = 0, (1.1) becomes a standard varying-coefficient model. If α j ( ) = 0, j = 2,..., p and X 1 = 1, then (1.1) becomes the standard partial linear model (Wahba, 1984; Green and Silverman, 1994). If α j ( ) = 0, j = 2,..., p, X 1 = 1 and β = 0, then (1.1) becomes the nonparametric regression model. In this article, we employ the local linear techniques for estimating α(u 0 ). Other nonparametric smoothers, such as splines, can be also used. We propose to get an initial consistent estimate of α(u 0 ) by minimizing the following loss function, l 1 (a, b, β) = ρ [ Y i { a T x i + b T x i (U i u) + z T i β }] K h (U i U 0 ), (1.2) 1

2 Note that the traditional least squares criterion corresponds to ρ(t) = t 2. In this article, we will mainly consider a robust criterion ρ( ). The median regression corresponds to L 1 loss. The Huber function (Huber 1981) is one commonly used robust loss function, ψ c (t) = ρ (t) = max{ c, min(c, t)} and the tuning constant c regulates the amount of robustness. Huber (1981) recommends using c = in practice. This choice produces a relative efficiency of approximately 95% when the error density is normal. Another possibility for ψ( ) is Tukey s bisquare function ψ c (t) = t{1 (t/c) 2 } 2 +, which weighs the tail contribution of t by a biweight function. In the parametric robustness literature, the use of c = 4.685, which produces 95% efficiency, is recommended. Let {â (0), ˆb (0), ˆβ (0) } be the minimizer of (1.4). Then ˆα (0) (U 0 ) = â (0). We will prove that α (0) (U 0 ) and ˆβ (0) are nh consistent. Note that the global parameters β was estimated locally here. To improve the efficiency, β should be estimated globally. By replacing α(u) by ˆα (0) (u) in (1.1), we estimate β by minimizing l 2 (β) = ρ [ Y i { x T i ˆα (0)(U i ) + z T i β }]. (1.3) Denote by ˆβ (1) the minimizer of β. The root n consistency of ˆβ (1) will be proved in the next section. Note that ˆα (0) (u) obtained in (1.2) needs to account for the uncertainty of β since we estimate α(u) and β locally simultaneously. Therefore, we can further improve the efficiency of the estimate of α(u 0 ) by minimizing the loss function l 3 (a, b) = [ }] ρ Y i {a T x i + b T x i (U i u) + z i ˆβ(1) K h (U i U 0 ), (1.4) ˆβ (1) is obtained in (1.3). Let {â (1), ˆb (1) } be the minimizer of (1.4). Then 2

3 ˆα (1) (U 0 ) = â (1). Since ˆβ (1) has root n consistency, ˆα (1) (U 0 ) has the same efficiency as if β were known. Therefore, the one step backfitting estimate ˆα (1) (U 0 ) will be more efficient than ˆα (0) (U 0 ). The formal result will be provided in the next section. After replacing α(u) by ˆα (1) (u) in (1.1), we can then estimate β by minimizing l 4 (β) = ρ [ Y i { x T i ˆα (1)(U i ) + z T i β }]. (1.5) Denote by ˆβ (2) the minimizer of (1.5). Since ˆα (1) (u) is more efficient than ˆα (0) (u), it will be our interest to check whether ˆβ (2) is more efficient than ˆβ (1). In addition, since the efficiency of ˆα (1) (u) cannot be improved further more, it is natural to think that any further iteration will not improve the efficiency of ˆβ (2). We will prove in the next section that the above two arguments are not necessary true and further demonstrate that the second arguments are in fact wrong in many common situations. 2. Computation In this section, we will consider how to minimize (1.2) (1.5). In general, let s consider how to minimize l(θ) = ρ ( Y i x T i θ ) K h (U i U 0 ). (2.1) Note that the estimate ˆθ is the solution of 0 = ρ ( Y i x T i θ ) x i K h (U i U 0 ) = W (r i )(Y i x T i θ)x i K h (U i U 0 ), (2.2) W (r i ) = ρ (r i ) /r i and r i = Y i x T i θ. If W (r i ) is given, (2.2) is a weighted least squares estimation and have explicit solution. Therefore, we can iterate between estimating W (r i ) and θ. Given the initial 3

4 estimate θ (0), at (k + 1)th step, θ (k+1) = ( W (r (k) i )K h (U i U 0 )x i x T i ) 1 W (r (k) i )K h (U i U 0 )x i Y i, r (k) i = Y i x T i θ (k). The above computation method can be applied to minimize any objective function of (1.2) (1.5). Note that the recommended c values used in ρ( ) introduced in Section 1 assumes the data has unit variance 1. Therefore, in practice, we need to account for the scale of the data. For example, for Huber s ψ function, we can use c = 1.345σ, σ is the standard deviation of ϵ. In practice, we can estimate σ by median absolute deviation (MAD) of the initial estimated residual ˆϵ i s. ˆσ = Median( ˆε i Median(ˆε i ) )/0.675, ˆϵ i can be calculated by using L 1 loss (median regression) for ρ( ) (R might have package to calculate median regression or we can use the computation method introduced in this section to compute median regression). 3. Main Asymptotic results 3.1. Initial estimate using varying coefficient model To establish the asymptotic properties of local maximizer { β, ã, b} of (1.4), let θ = (a T, hb T, β T ) T and x i = {x T i, x T i (U i U 0 )/h, z T i } T. Then, l 1 (θ) = K h (U i U 0 )ρ(y i x T i θ). (3.1) 4

5 Denote by f(u) the marginal density of u. Let γ j (U) = E{ρ (ϵ)(x j x T, 0 1 p, X j z T ) U}, A(X, Z) = xx T 0 p p xz T 0 p p µ 2 xx T 0 p q zx T 0 q p zz T and Ã(X, z) = ν 0 xx T 0 p p ν 0 xz T 0 p p ν 2 xx T 0 p q ν 0 zx T 0 q p ν 0 zz T, (3.2) µ l = u l K(u) du, ν l = u l K 2 (u) du and 0 a b is a by b matrix with all elements 0. Let A(U 0 ) = E{A(x, z)ρ (ϵ) U = U 0 } and Ã(U 0) = E{ρ (ϵ) 2 Ã(x, z) U = U 0 }. Theorem 3.1. Supposing ˆθ (0) is the minimizer of (3.1). Then, as n, h 0, and nh, we have a) θ θ =n 1 {f(u 0 )A(U 0 )} 1 ρ (Y i x T i θ) x i K h (U i U 0 ) + O p (a n ), (3.3) j=1 a n = h 2 + nh log 1/2 (1/h) and O p (a n ) holds uniformly for all U 0 in the support of U. b) nh { θ θ bias( θ) } d N {0, f(u 0 ) 1 A(U 0 ) 1 Ã(U 0 )A(U 0 ) 1 }, bias( θ) = 1 2 µ 2h 2 p j=1 α j (U 0 )A(U 0 ) 1 γ j (U 0 ) + o p (h 2 ). (3.4) 3.2 One step estimate One step of β: 5

6 Theorem 3.2. Suppose ˆβ (1) is the minimizer of (1.3). As n, nh 4 0, and nh 2 / log(1/h), n(ˆβ(1) β) d N(0, Σ 1 ), Σ 1 = B 1 var { ρ (ϵ)z E(ρ (ϵ)zx T U)ψ(x, Y, z, U) } B 1. B = E { ρ (ϵ)zz T } (3.5) and ψ(x j, Y j, z j, U j ) is a p 1 vector, in which the elements are taken from the first p entries of A(U j ) 1 {ρ (ϵ j ) x j }. One step of α( ): Let ξ = {α(u 0 ) T, hα (U 0 ) T } T, ˇx i = {x T i, x T i (U i U 0 )/h} T. Then the first step estimate ξ (1) = { ˆα (1) (U 0 ) T, h ˆα (1)(U 0 ) T } T is the minimizer of l 3 (ξ) = {ρ(y i ˇx T i ξ z T ˆβ i (1) )}. Let (U) = E{ρ (ϵ)xx T U}, (U) = E{ρ (ϵ) 2 xx T U}, η j (U) = E[ρ (ϵ)xx j U], and H = µ 2 (U 0 ), H = ν ν 2 (U 0 ), (3.6) is the Kronecker product. 6

7 Theorem 3.3. As n, h 0, and nh, [ nh ˆα(1) (U 0 ) α(u 0 ) bias{ ˆα (1) (U 0 )} ] { d N 0, ν 0 f(u 0 ) 1 1 (U 0 ) (U } 0 ) 1 (U 0 ), bias{ ˆα (1) (U 0 )} = 1 2 µ 2h 2 1 (U 0 ) p j=1 α j (U 0 )η j (U 0 ) + o p (h 2 ). (3.7) 3.3 Two step estimate Let ω (x j, Y j, z j, U j ) = E { ρ (ϵ)zx T U = Uj } 1 (U j )x j ρ (ϵ j ), (3.8) and M = E [E { ρ (ϵ)zx } { T U E ρ (ϵ)xx } T 1 { U E ρ (ϵ)zx } ] T T U. (3.9) Theorem 3.4. Suppose ˆβ (2) is the minimizer of (1.5). n(ˆβ(2) β) d N(0, Σ 2 ), Σ 2 = B 1 var [ ρ (ϵ i )z i + ω (x i, Y i, z i, U i ) MB 1 { ρ (ϵ i )z i + ω(x i, Y i, z i, U i )} ] B Full iteration Let ˆβ full be the estimate by full iteration of backing fitting algorithm. 7

8 Theorem 3.5. As n, nh 4 0, and nh 2 / log(1/h), n(ˆβfull β) d N(0, Σ full ), Σ full = (M + B) 1 var [ ρ (ϵ i )z i + ω (x i, Y i, z i, U i )] (M + B) 1, ω and M are defined in (3.8) and (3.9). 3.5 Partially linear models It is very difficult to compare Σ 1, Σ 2, and Σ full for general varying coefficient partially linear model due to their complicated expressions. In this section, similar to Carroll et al. (1997), we use a simpler partially linear model to compare different proposed estimators. Example: Considering the partially linear model (i.e., x = 1), we assume ρ(t) = t 2 /2, ϵ is independent of U and z with var(ϵ) = σ 2, and E(z) = 0 (see, Carroll et al. (1997); Opsomer and Ruppert (1999)). Then B = E(zz T ) = var(z) M = E [ E(z U)E(z T U) ] = var{e(z U)} = var(q(u)) ω (x i, Y i, z i, U i ) = E(z U i )ϵ i ω(x i, Y i, z i, U i ) = E(z U i )e T 1 A 1 (1, z T i ) T ϵ i. q(u) = E(z U), e 1 is the unit vector with 1 in the first position and 0 8

9 else. A = 1 E(z T U i ) E(z U i ) E(zz T U i ) 1. For one step estimate, note that since cov { z, E(z U)e T 1 A 1 (1, z T i ) T } = 0, we have Σ 1 = B 1 var {ρ (ϵ)z ω(x, Y, z, U)} B 1 = B 1 σ 2 var { z E(z U)e T 1 A 1 (1, z T i ) T } B 1 = σ 2 [ var(z) 1 + var(z) 1 E { q(u)q(u) T e T 1 A 1 e 1 } var(z) 1 ], e T 1 A 1 e 1 = 1 1 q(u) T E(zz T U) 1 q(u) = 1 + q(u)t var(z U) 1 q(u). For two step estimator, note that cov { ρ (ϵ i )z + ω (x i, Y i, z i, U i ), ρ (ϵ i )z i + ω(x i, Y i, z i, U i )} =σ 2 var(z) σ 2 var{e(z U)} + σ 2 var{e(z U} = σ 2 var(z). 9

10 Therefore, Σ 2 =σ 2 B ( 1 E {var(z U)} 2MB 1 var(z) +MB {[ 1 var(z) + E { q(u)q(u) T a 11 (U) }]} B 1 M ) B 1 = σ 2 B 1 [var(z) + var{q(u)}] B 1 + B 1 MΣ 1 MB 1 = σ [ 2 var(z) 1 + var(z) 1 var{q(u)}var(z) 1] + var(z) 1 var{q(u)}σ 1 var{q(u)}var(z) 1. Note that M + B = E {var(z U)} and var { ρ (ϵ i )z i + ω (x i, Y i, z i, U i )} = var { ϵz i + E(z U i )ϵ i } (3.10) = σ 2 E {var(z U)} (3.11) So Σ full = E {var(z U)} 1 σ 2 When z is independent of U, q(u) = 0 and E(var(z U)) = var(z). Therefore, Σ 1 = Σ 2 = Σ full = σ 2 var(z) 1, i.e., the one step estimator is as efficient as the fully iterated estimator. Carroll et al. (1997) derived a similar result for partial linear models. If we only assume that q(u) = E(z U) is independent of var(z U) (for example, var(z U) does not depend on U). Note that var(z) = var{q(u)} + E{var(z U)}, var{q(u)} = E{q(U)q(U) T },E{var(z U) 1/2 } E{var(z U)} 1/2, and E { q(u)q(u) T var(z U) 1 q(u)q(u) T } =var{q(u)}e{var(z U)} 1 var{q(u)} + cov [ q(u)q(u) T var(z U)} 1/2]. 10

11 Then Σ 1 =σ [ 2 var(z) 1 + var(z) 1 var {q(u)} var(z) 1 (3.12) +var(z) 1 E { q(u)q(u) T var(z U) 1 q(u)q(u) } T var(z) 1] σ [ 2 var(z) 1 + var(z) 1 var {q(u)} var(z) 1 +var(z) 1 var{q(u)}e{var(z U)} 1 var{q(u)}var(z) 1] =σ ( 2 var(z) 1 + var(z) 1 [var(z) E{var(z U)}] var(z) 1 +var(z) 1 [var(z) E{var(z U)}] E{var(z U)} 1 [var(z) E{var(z U)}] var(z) 1) = σ 2 E{var(z U)} 1 = Σ full. (3.13) The above equality only holds when q(u)q(u) T var(z U) 1 q(u)q(u) T does not depend on U, i.e., is a constant. For two step estimator, since Σ 1 σ 2 E{var(z U)} 1, we have Σ 2 σ 2 [ var(z) 1 + var(z) 1 var{q(u)}var(z) 1 +var(z) 1 var{q(u)}e{var(z U)} 1 var{q(u)}var(z) 1] = σ 2 E{var(z U)} 1 = Σ full. The above equality only holds when q(u)q(u) T var(z U) 1 q(u)q(u) T is a constant. Therefore, the full iteration is at least as efficient as the one step and two step estimators. In addition, σ 2 var{q(u)}σ 1 var{q(u)} = var{q(u)} [ var(z) 1 + var(z) 1 E { q(u)q(u) T } var(z) 1 +var(z) 1 E { q(u)q(u) T var(z U) 1 q(u)q(u) T } var(z) 1] var{q(u)} 11

12 and for any positive definite matrix C, we have C var{q(u)}var(z) 1 Cvar(z) 1 var{q(u)} 0 noting var(z) var{q(u)} and var(z)c 1 var(z) var{q(u)}c 1 var{q(u)}. Based on (3.13), we have σ 2 var(z)(σ 1 Σ 2 )var(z) =E { q(u)q(u) T var(z U) 1 q(u)q(u) } T σ 2 var{q(u)}σ 1 var{q(u)} var{q(u)}e{var(z U)} 1 var{q(u)} var{q(u)}var(z) 1 var{q(u)} var{q(u)}var(z) 1 var {q(u)} var(z) 1 var{q(u)} var{q(u)}var(z) 1 var{q(u)}e{var(z U)} 1 var{q(u)}var(z) 1 var{q(u)} = var{q(u)} [ E{var(z U)} 1 E{var(z U)} 1] var{q(u)} = 0. Therefore, Σ 1 Σ 2, i.e., the second step estimator is at least as efficient as the one step estimator. In summary, if z and U are independent, all three estimators have the same efficiency. Therefore, the one step estimator is enough. However, for general situation, if q(u) = E(z U) is independent of var(z U), we only have Σ 1 Σ 2 Σ full, although they have the same convergence rate. The above results are only proved for partial linear model. We conjecture that the above results also holds for varying coefficient partially linear model but the formal generalization needs further research. Appendix The following technical conditions are imposed in this section. They are not the weakest possible conditions, but they are imposed to facilitate the proofs 12

13 (A1) α(u) has continuous second derivative on its support D. (A2) f(u) has continuous first derivative. (A3) x and z have a bounded support. (A4) S(u) and T (u) are continuous and T (u) > 0. (A5) K( ) is a symmetric (about 0) probability density with compact support [ 1, 1]. (A6) E{ρ (ϵ)} = 0 and E{ρ (ϵ)} > 0. (A7) E{ρ (ϵ)} 3 <. Proof of Theorem 3.1: Supposing ˆθ (0) is the minimizer of (3.1). Let ˆθ (0) = γ 1 n (ˆθ (0) θ), γ n = (nh) 1/2 and θ = {α(u 0 ), hα (U 0 ), β} is the true value. Then ˆθ (0) minimizes l 1(θ ) = h Based on Taylor expansion, we have [ K h (U i U 0 ) ρ(y i x T i (θ + θ / ] nh)) ρ(y i x T i θ) l 1(θ ) = V T nθ θ T W n θ {1 + o p (1)}, Note that V n = θ l 1(θ) = hγ n W n = 2 θ θ l 1(θ) = hγ 2 n ρ (Y i x T i θ) x i K h (U i U 0 ) ρ (Y i x T i θ) x i x T i K h (U i U 0 ) W n = E { K h (U U 0 )ρ (ϵ) x x T } (1 + o p (1)) = f(u 0 )A(U 0 ) + o p (1). 13

14 Therefore, we have l 1(θ ) = V T nθ θ T f(u 0 )A(U 0 )θ + o p (1). By applying the convexity Lemma (Pollard, 1991), we obtain ˆθ (0) = {f(u 0 )A(U 0 )} 1 V n + o p (1). (3.14) Note that [ nh E(V n ) = 2 E K h (U U 0 )ρ (ϵ) nh p = 2 µ 2h 2 f(u 0 ) j=1 p j=1 α j (U 0 )(U U 0 ) 2 X j x α j (U 0 )γ j (U 0 )(1 + o p (1)), cov(v n ) = E [ K 2 h(u U 0 )ρ (ϵ) 2 x x T ] (1 + o p (1)) = 1 nh f(u 0)Ã(U 0)(1 + o p (1)). ] (1 + o p (1)) Hence, based on the central limit theorem by checking the Liapounov s condition, we have nh { θ θ bias( θ) } d N {0, f(u 0 ) 1 A(U 0 ) 1 Ã(U 0 )A(U 0 ) 1 }, bias( θ) is defined in (3.4). Based on Lemma A.1 and the result (A.6) of Carroll et al. (1997), we also have a 14

15 stronger representation of (3.14) ˆθ (0) (U) θ(u) =n 1 {f(u)a(u)} 1 ρ (Y i x T i θ) x i K h (U i U) + o p (a n1 ). (3.15) holds uniformly over all u D, a n1 = γ n h 2 + γ 2 n log(1/h). Proof of Theorem 3.2: Denote ˆβ (1) = n(ˆβ (1) β), β is the true value. Then ˆβ (1) minimizes l 2(β ) = { ρ(yi x T i ˆα (0)(U i ) z T i (β + β / n)) ρ(y i x T i ˆα (0)(U i ) z T i β) } (3.16). By a Taylor expansion and some calculation, l 2(β ) = A n β β T B n β + o p (1), (3.17) A n = n 1/2 ρ (Y i x T i ˆα (0)(U i ) z T i β)z i, For B n, it can be shown that B n = n 1 ρ (Y i x T i ˆα (0)(U i ) z T i β)z i z T i. B n = E { ρ (ϵ)z i z T i } = B + op (1). 15

16 Then, we have l n (β ) = A n β 1 2 β T B β + o p (1). (3.18) Next, we expand A n as A n = 1 n ρ (Y i x T i α(u i ) z T i β)z i + T n1 + O p (a 1n ). a 1n = n 1/2 ˆα (0) α 2 and T n1 = 1 n ρ (Y i x T i α(u i ) z T i β)z i x T i { ˆα (0) (U i ) α(u i )}. (3.19) Let ψ(x j, Y j, z j, U j ) be a p 1 vector, in which the elements are taken from the first p entries of A 1 (U j ) {ρ (Y j x T j α(u j ) z T j β) x i }. Based on the result (3.15), by condition nh 2 / log(1/h), we have O p (n 1/2 a n1 ) = o p (1). Since x T j θ(u i ) x T j α(u j ) z T j β = O{(U i U j ) 2 }, we have T n1 = n 3/2 ρ (ϵ i )z i x T i f 1 (U i )ψ(x j, Y j, z j, U j )K h (U i U j ) + O p (n 1/2 h 2 ) j=1 = T n2 + O p (n 1/2 h 2 ). It can be shown, by calculating the second moment, that T n2 T n3 p 0, (3.20) 16

17 T n3 = n 1/2 n j=1 ω(x j, Y j, z j, U j ), with ω(x j, Y j, z j, U j ) = E { ρ (ϵ)zx T U = Uj } ψ(xj, Y j, z j, U j ). By condition nh 4 0, we know A n = n 1/2 { ρ (Y i x T i α(u i ) z T i β)z i + ω(x i, Y i, z i, U i ) } + o p (1). By (3.18) and quadratic approximation lemma, ˆβ (1) = B 1 A n + o p (1). (3.21) Note that E {ρ (ϵ) x, z, U} = 0 and E {ω(x, z, U)} = 0. So based on the central limit theorem, we have n(ˆβ(1) β) d N(0, B 1 ΣB 1 ), Σ = var { ρ (Y x T α(u) z T β)z ω(x, Y, z, U) }. Proof of Theorem 3.3: We have nh{ ξ ξ} = f(u0 ) 1 H 1 ˆ n + o p (1), (3.22) 17

18 h ˆ n = n ρ (Y i ˇX T i ξ z T ˆβ i (1) ) ˇX i K h (U i U 0 ). It can be calculated that h ˆ n = n ρ (Y i ˇX T i ξ z T i β) ˇX i K h (U i U 0 ) + D n + o p (1), D n = h n ρ (Y i ˇX T i ξ z T i β) ˇX i z T i (ˆβ (1) β)k h (U i U 0 ) Since n(ˆβ (1) β) = O p (1), it can be shown that D n = O p (h). Hence nh{ ξ ξ} = f(u0 ) 1 H 1 n + o p (1), h n = n We can show that ρ (Y i ˇX T i ξ z T i β) ˇX i K h (U i U 0 ). var( n ) f(u 0 ) H and E( n ) nh 2 µ 2h 2 f(u 0 ) p j=1 α j (U 0 )(1, 0) T η j (U 0 ), 18

19 So, the asymptotic bias of the resulting estimator ᾰ(u 0 ) is bias{ᾰ(u 0 )} = 1 2 µ 2h 2 1 (U 0 ) p j=1 α j (U 0 )η j (U 0 ) + o p (h 2 1), and the asymptotic variance of ᾰ(u 0 ) is cov{ᾰ(u 0 )} = ν 0 nhf(u 0 ) 1 (U 0 ) (U 0 ) 1 (U 0 )(1 + o p (1)). Based on the central limit theorem, the asymptotic normality nh [ᾰ(u0 ) α(u 0 ) bias{ᾰ(u 0 )}] { d N 0, ν 0 f(u 0 ) 1 1 (U 0 ) (U } 0 ) 1 (U 0 ), bias(ˆθ) is defined in (3.7), can also be proved under some regularity conditions. Proof of Theorem 3.4: Based on the one step result (3.22), ( ˆα (0) (U i ) α(u i )) n 1 {f(u i )} 1 1 (U i ) ρ (Y j ˇX T j ξ z T j β)x j K h (U j U i ) j=1 (3.23) n 1 {f(u i )} 1 1 (U i )Dn, (3.24) D n = ρ (Y j ˇX T j ξ z T j β)x j z T j (ˆβ (1) β)k h (U j U i ) j=1 19

20 From (3.19), we have T n1 =n 3/2 j=1 n 3/2 j=1 ρ (ϵ i )z i x T i 1 (U i )ρ (ϵ j )x j K h (U i U j ) f(u i ) =T n2 T n3 n(ˆβ(1) β) ρ (ϵ i )z i x T i 1 (U i )ρ (ϵ j )x j z T j f(u i ) (ˆβ (1) β)k h (U i U j ) + O p (n 1/2 h 2 ) It can be shown, by calculating the second moment, that T n2 T n4 p 0, (3.25) T n4 = n 1/2 n j=1 ω (x j, Y j, z j, U j ), with ω (x j, Y j, z j, U j ) = E { ρ (ϵ)zx T U = Uj } 1 (U j )x j ρ (ϵ j ) In addition, we have T n3 = E [E { ρ (ϵ)zx } T U 1 (U)E { ρ (ϵ)zx } ] T T U (1 + o p (1)) = M(1 + o p (1)) So n(ˆβ β) = B 1 A n + o p (1), (3.26) 20

21 B = E{ρ (ϵ)zz T }. A n = n 1/2 { ρ (ϵ i )z i + ω (x i, Y i, z i, U i )} M n(ˆβ (1) β), (3.27) and So, n(ˆβ(1) β) = B 1 n 1/2 { ρ (ϵ i )z i + ω(x i, Y i, z i, U i )} n(ˆβ β) d N(0, Σ 2 ), Σ 2 = B 1 var [ ρ (ϵ i )z i + ω (x i, Y i, z i, U i ) MB 1 { ρ (ϵ i )z i + ω(x i, Y i, z i, U i )} ] B 1. Proof of Theorem 3.5: At convergence, (ˆβ (1) β) = (ˆβ β), then (I + B 1 M) n(ˆβ β) = B 1 n 1/2 { ρ (ϵ i )z i + ω (x i, Y i, z i, U i )} So n(ˆβfull β) d N(0, Σ full ), Σ full = (M + B) 1 var [ ρ (ϵ i )z i + ω (x i, Y i, z i, U i )] (M + B) 1 References Carroll, R. J., Fan, J., Gijbels, I. and Wand, M. P. (1997). Generalized Partially Linear Single-Index Models. Journal of American and Statisical Association. 92, 21

22 Fan, J. and Huang, T. (2005). Profile Likelihood Inferences on Semiparametric Varying-Coefficient Partially Linear Models. Bernoulli, 11, Green, P. J. and Silverman, B. W. (1994). Nonparametric Regression and Generalized Linear Models: A Roughness Penalty Approach. London: Chapman and Hall. Opsomer, J. D. and Ruppert, D. (1999) A root-n consistent backfitting estimator for semiparametric additive modeling. Journal of Computational and Graphical Statistics, 8, Wahba, G. (1984). Partial Spline Models for Semiparametric Estimation of Functions of Several Variables. In Statistical Analysis of Time Series, Proceedings of the Japan U.S. Joint Seminar, Tokyo, Institute of Statistical Mathematics, Tokyo. Xia, Y., Zhang, W., and Tong, H. (2004). Efficient Estimation for Semivaryingcoefficient Models. Biometrika, 91, Zhang, W., Lee, S. Y., and Song, X. (2002). Local Polynomial Fitting in Semivarying Coefficent Models. Journal of Multivariate Analysis, 82,

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University Integrated Likelihood Estimation in Semiparametric Regression Models Thomas A. Severini Department of Statistics Northwestern University Joint work with Heping He, University of York Introduction Let Y

More information

Robust Variable Selection Through MAVE

Robust Variable Selection Through MAVE Robust Variable Selection Through MAVE Weixin Yao and Qin Wang Abstract Dimension reduction and variable selection play important roles in high dimensional data analysis. Wang and Yin (2008) proposed sparse

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

Function of Longitudinal Data

Function of Longitudinal Data New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li Abstract This paper develops a new estimation of nonparametric regression functions for

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,

More information

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data ew Local Estimation Procedure for onparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li The Pennsylvania State University Technical Report Series #0-03 College of Health and Human

More information

Local Modal Regression

Local Modal Regression Local Modal Regression Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Bruce G. Lindsay and Runze Li Department of Statistics, The Pennsylvania

More information

Regularization in Cox Frailty Models

Regularization in Cox Frailty Models Regularization in Cox Frailty Models Andreas Groll 1, Trevor Hastie 2, Gerhard Tutz 3 1 Ludwig-Maximilians-Universität Munich, Department of Mathematics, Theresienstraße 39, 80333 Munich, Germany 2 University

More information

An Empirical Characteristic Function Approach to Selecting a Transformation to Normality

An Empirical Characteristic Function Approach to Selecting a Transformation to Normality Communications for Statistical Applications and Methods 014, Vol. 1, No. 3, 13 4 DOI: http://dx.doi.org/10.5351/csam.014.1.3.13 ISSN 87-7843 An Empirical Characteristic Function Approach to Selecting a

More information

Efficiency of Profile/Partial Likelihood in the Cox Model

Efficiency of Profile/Partial Likelihood in the Cox Model Efficiency of Profile/Partial Likelihood in the Cox Model Yuichi Hirose School of Mathematics, Statistics and Operations Research, Victoria University of Wellington, New Zealand Summary. This paper shows

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

Mixture of Regression Models with Varying Mixing Proportions: A Semiparametric Approach

Mixture of Regression Models with Varying Mixing Proportions: A Semiparametric Approach Mixture of Regression Models with Varying Mixing Proportions: A Semiparametric Approach Mian Huang and Weixin Yao Abstract In this paper, we study a class of semiparametric mixtures of regression models,

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

On the Robust Modal Local Polynomial Regression

On the Robust Modal Local Polynomial Regression International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression

More information

The EM Algorithm for the Finite Mixture of Exponential Distribution Models

The EM Algorithm for the Finite Mixture of Exponential Distribution Models Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 2, 57-64 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.312133 The EM Algorithm for the Finite Mixture of Exponential Distribution

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Econometrics Workshop UNC

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Estimation of Treatment Effects under Essential Heterogeneity

Estimation of Treatment Effects under Essential Heterogeneity Estimation of Treatment Effects under Essential Heterogeneity James Heckman University of Chicago and American Bar Foundation Sergio Urzua University of Chicago Edward Vytlacil Columbia University March

More information

SEMI-LINEAR LINEAR INDEX MODEL WHEN THE LINEAR COVARIATES AND INDICES ARE INDEPENDENT

SEMI-LINEAR LINEAR INDEX MODEL WHEN THE LINEAR COVARIATES AND INDICES ARE INDEPENDENT SEMI-LINEAR LINEAR INDEX MODEL WHEN THE LINEAR COVARIATES AND INDICES ARE INDEPENDENT By Yun Sam Chong, Jane-Ling Wang and Lixing Zhu Summary Wecker Associate, University of California at Davis, and The

More information

9. Robust regression

9. Robust regression 9. Robust regression Least squares regression........................................................ 2 Problems with LS regression..................................................... 3 Robust regression............................................................

More information

Semi-parametric estimation of non-stationary Pickands functions

Semi-parametric estimation of non-stationary Pickands functions Semi-parametric estimation of non-stationary Pickands functions Linda Mhalla 1 Joint work with: Valérie Chavez-Demoulin 2 and Philippe Naveau 3 1 Geneva School of Economics and Management, University of

More information

On a Nonparametric Notion of Residual and its Applications

On a Nonparametric Notion of Residual and its Applications On a Nonparametric Notion of Residual and its Applications Bodhisattva Sen and Gábor Székely arxiv:1409.3886v1 [stat.me] 12 Sep 2014 Columbia University and National Science Foundation September 16, 2014

More information

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives TR-No. 14-06, Hiroshima Statistical Research Group, 1 11 Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives Mariko Yamamura 1, Keisuke Fukui

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA Kasun Rathnayake ; A/Prof Jun Ma Department of Statistics Faculty of Science and Engineering Macquarie University

More information

Nonparametric Small Area Estimation Using Penalized Spline Regression

Nonparametric Small Area Estimation Using Penalized Spline Regression Nonparametric Small Area Estimation Using Penalized Spline Regression 0verview Spline-based nonparametric regression Nonparametric small area estimation Prediction mean squared error Bootstrapping small

More information

Final Exam. Economics 835: Econometrics. Fall 2010

Final Exam. Economics 835: Econometrics. Fall 2010 Final Exam Economics 835: Econometrics Fall 2010 Please answer the question I ask - no more and no less - and remember that the correct answer is often short and simple. 1 Some short questions a) For each

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

LOCAL LINEAR REGRESSION FOR GENERALIZED LINEAR MODELS WITH MISSING DATA

LOCAL LINEAR REGRESSION FOR GENERALIZED LINEAR MODELS WITH MISSING DATA The Annals of Statistics 1998, Vol. 26, No. 3, 1028 1050 LOCAL LINEAR REGRESSION FOR GENERALIZED LINEAR MODELS WITH MISSING DATA By C. Y. Wang, 1 Suojin Wang, 2 Roberto G. Gutierrez and R. J. Carroll 3

More information

Robust variable selection through MAVE

Robust variable selection through MAVE This is the author s final, peer-reviewed manuscript as accepted for publication. The publisher-formatted version may be available through the publisher s web site or your institution s library. Robust

More information

A note on profile likelihood for exponential tilt mixture models

A note on profile likelihood for exponential tilt mixture models Biometrika (2009), 96, 1,pp. 229 236 C 2009 Biometrika Trust Printed in Great Britain doi: 10.1093/biomet/asn059 Advance Access publication 22 January 2009 A note on profile likelihood for exponential

More information

A Resampling Method on Pivotal Estimating Functions

A Resampling Method on Pivotal Estimating Functions A Resampling Method on Pivotal Estimating Functions Kun Nie Biostat 277,Winter 2004 March 17, 2004 Outline Introduction A General Resampling Method Examples - Quantile Regression -Rank Regression -Simulation

More information

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Statistica Sinica 18(2008), 515-534 DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Kani Chen 1, Jianqing Fan 2 and Zhezhen Jin 3 1 Hong Kong University of Science and Technology,

More information

The BLP Method of Demand Curve Estimation in Industrial Organization

The BLP Method of Demand Curve Estimation in Industrial Organization The BLP Method of Demand Curve Estimation in Industrial Organization 9 March 2006 Eric Rasmusen 1 IDEAS USED 1. Instrumental variables. We use instruments to correct for the endogeneity of prices, the

More information

Final Exam November 24, Problem-1: Consider random walk with drift plus a linear time trend: ( t

Final Exam November 24, Problem-1: Consider random walk with drift plus a linear time trend: ( t Problem-1: Consider random walk with drift plus a linear time trend: y t = c + y t 1 + δ t + ϵ t, (1) where {ϵ t } is white noise with E[ϵ 2 t ] = σ 2 >, and y is a non-stochastic initial value. (a) Show

More information

Regression, Ridge Regression, Lasso

Regression, Ridge Regression, Lasso Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University

SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES. Liping Zhu, Mian Huang, & Runze Li. The Pennsylvania State University SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES Liping Zhu, Mian Huang, & Runze Li The Pennsylvania State University Technical Report Series #10-104 College of Health and Human Development

More information

Functional Latent Feature Models. With Single-Index Interaction

Functional Latent Feature Models. With Single-Index Interaction Generalized With Single-Index Interaction Department of Statistics Center for Statistical Bioinformatics Institute for Applied Mathematics and Computational Science Texas A&M University Naisyin Wang and

More information

Cointegration Lecture I: Introduction

Cointegration Lecture I: Introduction 1 Cointegration Lecture I: Introduction Julia Giese Nuffield College julia.giese@economics.ox.ac.uk Hilary Term 2008 2 Outline Introduction Estimation of unrestricted VAR Non-stationarity Deterministic

More information

Lecture 6: Discrete Choice: Qualitative Response

Lecture 6: Discrete Choice: Qualitative Response Lecture 6: Instructor: Department of Economics Stanford University 2011 Types of Discrete Choice Models Univariate Models Binary: Linear; Probit; Logit; Arctan, etc. Multinomial: Logit; Nested Logit; GEV;

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Likelihood Ratio Tests. that Certain Variance Components Are Zero. Ciprian M. Crainiceanu. Department of Statistical Science

Likelihood Ratio Tests. that Certain Variance Components Are Zero. Ciprian M. Crainiceanu. Department of Statistical Science 1 Likelihood Ratio Tests that Certain Variance Components Are Zero Ciprian M. Crainiceanu Department of Statistical Science www.people.cornell.edu/pages/cmc59 Work done jointly with David Ruppert, School

More information

Estimation in Semiparametric Single Index Panel Data Models with Fixed Effects

Estimation in Semiparametric Single Index Panel Data Models with Fixed Effects Estimation in Semiparametric Single Index Panel Data Models with Fixed Effects By Degui Li, Jiti Gao and Jia Chen School of Economics, The University of Adelaide, Adelaide, Australia Abstract In this paper,

More information

SIMULTANEOUS CONFIDENCE BANDS AND HYPOTHESIS TESTING FOR SINGLE-INDEX MODELS

SIMULTANEOUS CONFIDENCE BANDS AND HYPOTHESIS TESTING FOR SINGLE-INDEX MODELS Statistica Sinica 24 (2014), 937-955 doi:http://dx.doi.org/10.5705/ss.2012.127 SIMULTANEOUS CONFIDENCE BANDS AND HYPOTHESIS TESTING FOR SINGLE-INDEX MODELS Gaorong Li 1, Heng Peng 2, Kai Dong 2 and Tiejun

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation

Preface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation Preface Nonparametric econometrics has become one of the most important sub-fields in modern econometrics. The primary goal of this lecture note is to introduce various nonparametric and semiparametric

More information

On robust and efficient estimation of the center of. Symmetry.

On robust and efficient estimation of the center of. Symmetry. On robust and efficient estimation of the center of symmetry Howard D. Bondell Department of Statistics, North Carolina State University Raleigh, NC 27695-8203, U.S.A (email: bondell@stat.ncsu.edu) Abstract

More information

Testing Restrictions and Comparing Models

Testing Restrictions and Comparing Models Econ. 513, Time Series Econometrics Fall 00 Chris Sims Testing Restrictions and Comparing Models 1. THE PROBLEM We consider here the problem of comparing two parametric models for the data X, defined by

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

Modification and Improvement of Empirical Likelihood for Missing Response Problem

Modification and Improvement of Empirical Likelihood for Missing Response Problem UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu

More information

Web-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach

Web-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach 1 Web-based Supplementary Material for Dependence Calibration in Conditional Copulas: A Nonparametric Approach Elif F. Acar, Radu V. Craiu, and Fang Yao Web Appendix A: Technical Details The score and

More information

NONPARAMETRIC MIXTURE OF REGRESSION MODELS

NONPARAMETRIC MIXTURE OF REGRESSION MODELS NONPARAMETRIC MIXTURE OF REGRESSION MODELS Huang, M., & Li, R. The Pennsylvania State University Technical Report Series #09-93 College of Health and Human Development The Pennsylvania State University

More information

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001)

Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Presented by Yang Zhao March 5, 2010 1 / 36 Outlines 2 / 36 Motivation

More information

Asymptotic standard errors of MLE

Asymptotic standard errors of MLE Asymptotic standard errors of MLE Suppose, in the previous example of Carbon and Nitrogen in soil data, that we get the parameter estimates For maximum likelihood estimation, we can use Hessian matrix

More information

SIMULTANEOUS CONFIDENCE INTERVALS FOR SEMIPARAMETRIC LOGISTICS REGRESSION AND CONFIDENCE REGIONS FOR THE MULTI-DIMENSIONAL EFFECTIVE DOSE

SIMULTANEOUS CONFIDENCE INTERVALS FOR SEMIPARAMETRIC LOGISTICS REGRESSION AND CONFIDENCE REGIONS FOR THE MULTI-DIMENSIONAL EFFECTIVE DOSE Statistica Sinica 20 (2010), 637-659 SIMULTANEOUS CONFIDENCE INTERVALS FOR SEMIPARAMETRIC LOGISTICS REGRESSION AND CONFIDENCE REGIONS FOR THE MULTI-DIMENSIONAL EFFECTIVE DOSE Jialiang Li 1, Chunming Zhang

More information

A PARADOXICAL EFFECT OF NUISANCE PARAMETERS ON EFFICIENCY OF ESTIMATORS

A PARADOXICAL EFFECT OF NUISANCE PARAMETERS ON EFFICIENCY OF ESTIMATORS J. Japan Statist. Soc. Vol. 34 No. 1 2004 75 86 A PARADOXICAL EFFECT OF NUISANCE PARAMETERS ON EFFICIENCY OF ESTIMATORS Masayuki Henmi* This paper is concerned with parameter estimation in the presence

More information

Inference for High Dimensional Robust Regression

Inference for High Dimensional Robust Regression Department of Statistics UC Berkeley Stanford-Berkeley Joint Colloquium, 2015 Table of Contents 1 Background 2 Main Results 3 OLS: A Motivating Example Table of Contents 1 Background 2 Main Results 3 OLS:

More information

Clustering as a Design Problem

Clustering as a Design Problem Clustering as a Design Problem Alberto Abadie, Susan Athey, Guido Imbens, & Jeffrey Wooldridge Harvard-MIT Econometrics Seminar Cambridge, February 4, 2016 Adjusting standard errors for clustering is common

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Nonconcave Penalized Likelihood with A Diverging Number of Parameters Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized

More information

Robust estimators for additive models using backfitting

Robust estimators for additive models using backfitting Robust estimators for additive models using backfitting Graciela Boente Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires and CONICET, Argentina Alejandra Martínez Facultad de Ciencias

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Quantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation

Quantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation Quantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation Maria Ponomareva University of Western Ontario May 8, 2011 Abstract This paper proposes a moments-based

More information

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL

TESTING SERIAL CORRELATION IN SEMIPARAMETRIC VARYING COEFFICIENT PARTIALLY LINEAR ERRORS-IN-VARIABLES MODEL Jrl Syst Sci & Complexity (2009) 22: 483 494 TESTIG SERIAL CORRELATIO I SEMIPARAMETRIC VARYIG COEFFICIET PARTIALLY LIEAR ERRORS-I-VARIABLES MODEL Xuemei HU Feng LIU Zhizhong WAG Received: 19 September

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, N.C.,

More information

Issues on quantile autoregression

Issues on quantile autoregression Issues on quantile autoregression Jianqing Fan and Yingying Fan We congratulate Koenker and Xiao on their interesting and important contribution to the quantile autoregression (QAR). The paper provides

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

KANSAS STATE UNIVERSITY

KANSAS STATE UNIVERSITY ROBUST MIXTURES OF REGRESSION MODELS by XIUQIN BAI M.S., Kansas State University, USA, 2010 AN ABSTRACT OF A DISSERTATION submitted in partial fulfillment of the requirements for the degree DOCTOR OF PHILOSOPHY

More information

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING

SEMIPARAMETRIC ESTIMATION OF CONDITIONAL HETEROSCEDASTICITY VIA SINGLE-INDEX MODELING Statistica Sinica 3 (013), 135-155 doi:http://dx.doi.org/10.5705/ss.01.075 SEMIPARAMERIC ESIMAION OF CONDIIONAL HEEROSCEDASICIY VIA SINGLE-INDEX MODELING Liping Zhu, Yuexiao Dong and Runze Li Shanghai

More information

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St.

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St. Regression Graphics R. D. Cook Department of Applied Statistics University of Minnesota St. Paul, MN 55108 Abstract This article, which is based on an Interface tutorial, presents an overview of regression

More information

1 Appendix A: Matrix Algebra

1 Appendix A: Matrix Algebra Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Calibration Estimation for Semiparametric Copula Models under Missing Data

Calibration Estimation for Semiparametric Copula Models under Missing Data Calibration Estimation for Semiparametric Copula Models under Missing Data Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Economics and Economic Growth Centre

More information

Adaptive Piecewise Polynomial Estimation via Trend Filtering

Adaptive Piecewise Polynomial Estimation via Trend Filtering Adaptive Piecewise Polynomial Estimation via Trend Filtering Liubo Li, ShanShan Tu The Ohio State University li.2201@osu.edu, tu.162@osu.edu October 1, 2015 Liubo Li, ShanShan Tu (OSU) Trend Filtering

More information

This paper is not to be removed from the Examination Halls

This paper is not to be removed from the Examination Halls ~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Robust estimation, efficiency, and Lasso debiasing

Robust estimation, efficiency, and Lasso debiasing Robust estimation, efficiency, and Lasso debiasing Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics WHOA-PSI workshop Washington University in St. Louis Aug 12, 2017 Po-Ling

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

DavidAllanBurt. Dissertation submitted to the Faculty of the. Virginia Polytechnic Institute and State University

DavidAllanBurt. Dissertation submitted to the Faculty of the. Virginia Polytechnic Institute and State University Bandwidth Selection Concerns for Jump Point Discontinuity Preservation in the Regression Setting Using M-smoothers and the Extension to Hypothesis Testing DavidAllanBurt Dissertation submitted to the Faculty

More information

Short Questions (Do two out of three) 15 points each

Short Questions (Do two out of three) 15 points each Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Topic 12 Overview of Estimation

Topic 12 Overview of Estimation Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Spatial Process Estimates as Smoothers: A Review

Spatial Process Estimates as Smoothers: A Review Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

On prediction and density estimation Peter McCullagh University of Chicago December 2004

On prediction and density estimation Peter McCullagh University of Chicago December 2004 On prediction and density estimation Peter McCullagh University of Chicago December 2004 Summary Having observed the initial segment of a random sequence, subsequent values may be predicted by calculating

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information