A note on profile likelihood for exponential tilt mixture models

Size: px
Start display at page:

Download "A note on profile likelihood for exponential tilt mixture models"

Transcription

1 Biometrika (2009), 96, 1,pp C 2009 Biometrika Trust Printed in Great Britain doi: /biomet/asn059 Advance Access publication 22 January 2009 A note on profile likelihood for exponential tilt mixture models BY Z. TAN Department of Statistics, Rutgers University, Piscataway, New Jersey 08854, U.S.A. ztan@stat.rutgers.edu SUMMARY Suppose that independent observations are drawn from multiple distributions, each of which is a mixture of two component distributions such that their log density ratio satisfies a linear model with a slope parameter and an intercept parameter. Inference for such models has been studied using empirical likelihood, and mixed results have been obtained. The profile empirical likelihood of the slope and intercept has an irregularity at the null hypothesis so that the two component distributions are equal. We derive a profile empirical likelihood and maximum likelihood estimator of the slope alone, and obtain the usual asymptotic properties for the estimator and the likelihood ratio statistic regardless of the null. Furthermore, we show the maximum likelihood estimator of the slope and intercept jointly is consistent and asymptotically normal regardless of the null. At the null, the joint maximum likelihood estimator falls along a straight line through the origin with perfect correlation asymptotically to the first order. Some key words: Empirical likelihood; Exponential tilt; Likelihood ratio statistic; Maximum likelihood; Mixture model; Profile likelihood. 1. INTRODUCTION Suppose that a sample of n j independent observations, {x j1,...,x jnj }, is drawn from each of m( 2) distributions P j ( j 1,...,m), and each P j is a mixture of two component distributions, G 0 and G 1, related in an exponential tilt form: dp j λ j dg 0 + (1 λ j )dg 1, where 0 λ j 1 are known mixture proportions, λ j λ k ( j k), and dg 1 exp(β 0 + β 1 x)dg 0, where G 0 is an unknown baseline distribution, β 1 is an unknown parameter, and β 0 log{ exp(β 1 x)dg 0 }. Note that β 1 0 implies that β 0 0 for any G 0.The jth sample, {x j1,...,x jnj }, is obtained from dp j {λ j + (1 λ j )exp(β 0 + β 1 x)} dg 0. (1) The special case of model (1) where m 2, λ 1 1, and λ 2 0 is known as the two-sample exponential tilt model (Qin, 1998), first studied by Anderson (1972) and Prentice & Pyke (1979) in the context of case-control studies. Inference for β (β 1,β 0 ) T and G 0 in model (1) has been studied using empirical likelihood (Owen, 2001), and mixed results have been obtained. Qin (1999) derived a profile empirical likelihood and the resulting maximum likelihood estimator of β in the case where m 3, λ 1 1, λ 2 0, and 0 <λ 3 < 1 may be unknown. However, Zou et al. (2002) pointed out that the profile empirical likelihood of β has an irregularity at the null hypothesis that G 1 G 0, i.e., β 0, and the asymptotic arguments in Qin (1999) for the maximum likelihood estimator of β are not applicable there. Zou & Fine (2002) claimed that

2 230 Z. TAN the maximum likelihood estimator of β is not consistent when β 0. To address this, Zou et al. (2002) proposed a partial profile empirical likelihood, and showed that the maximum partial likelihood estimator of β is consistent and asymptotically normal whether or not β 0, and the profile loglikelihood ratio for testing β 0 has a chi-squared distribution with one degree of freedom asymptotically. A limitation of this method is that, when β 0, the maximum partial likelihood estimator of β always has asymptotic variance no smaller than that of the maximum likelihood estimator in Qin (1999). See Zou & Fine (2002) for a further discussion on the construction of the maximum partial likelihood estimator and the efficiency comparisons of several related estimators. In this note we first derive a profile empirical likelihood and the resulting maximum likelihood estimator of β 1 alone, and show that the profile empirical likelihood of β 1 does not have an irregularity at β 1 0, in contrast to that of β. In fact, under mild regularity conditions, the profile empirical likelihood and the maximum likelihood estimator of β 1 behave like usual profile likelihoods and maximum likelihood estimators with the following properties, whether or not β 1 0: the maximum likelihood estimator is consistent and asymptotically normal; the inverse curvature of the profile loglikelihood at its maximum is a consistent estimator of the asymptotic variance of the maximum likelihood estimator; and the loglikelihood ratio statistic has a chi-squared distribution with one degree of freedom asymptotically. Second, we derive in our approach the maximum likelihood estimator of β, which agrees with that in Qin (1999), and show that it is consistent and asymptotically normal, and always has asymptotic variance no greater than that of the maximum partial likelihood estimator of β in Zou et al. (2002), whether or not β 0. A subtle point is that when β 0, the maximum likelihood estimator of β falls along a straight line through the origin with perfect correlation asymptotically to the first order. The proofs of the main results are provided in the Appendix. The loglikelihood of (β 1, G 0 )is 2. PROFILE LIKELIHOOD AND MAXIMUM LIKELIHOOD ESTIMATOR l(β 1, G 0 ) [log{λ j + (1 λ j )exp(β 0 + β 1 x ji )}+log G 0 ({x ji })], where β 0 log{ exp(β 1 x)dg 0 } is a function of (β 1, G 0 ). The profile loglikelihood of β 1 is defined for each fixed β 1 by maximizing l(β 1, G 0 ) over all possible probability distributions G 0, that is, pl(β 1 ) max G0 l(β 1, G 0 ). The maximum likelihood estimator of β 1 is defined by maximizing the profile loglikelihood, that is, ˆβ 1 arg max pl(β 1 ). We provide a formula for computing the profile loglikelihood pl(β 1 ) in Proposition 1 (i). The formula involves the function κ(β,α 0 ) log { } λ j + (1 λ j )exp(β 0 + β 1 x ji ) n log n, α 0 + (1 α 0 )exp(β 0 + β 1 x ji ) where β (β 1,β 0 ) T and 0 α 0 1 are free arguments. This function has a duality relationship with the loglikelihood l(β 1, G 0 ) as shown in Proposition 1 (ii). PROPOSITION 1. (i) For a fixed β 1, the loglikelihood l(β 1, G 0 ) achieves the maximum at Ĝ 0 Ĝ 0 (β 1 ) supported on {x ji : i 1,...,n j, j 1,...,m} with Ĝ({x ji }) n 1 ˆα 0 + (1 ˆα 0 )exp(ˆβ 0 + β 1 x ji ), where n m j1 n j, and ˆβ 0 ˆβ 0 (β 1 ) and ˆα 0 ˆα 0 (β 1 ) satisfy α 0 1 n λ j λ j + (1 λ j )exp(β 0 + β 1 x ji ), (2)

3 Miscellanea n 1 1 n 1 α 0 + (1 α 0 )exp(β 0 + β 1 x ji ), (3) exp(β 0 + β 1 x ji ) α 0 + (1 α 0 )exp(β 0 + β 1 x ji ). (4) The profile loglikelihood is pl(β 1 ) κ{β 1, ˆβ 0 (β 1 ), ˆα 0 (β 1 )}. (ii) Furthermore, for any G 0 and any (β 0,α 0 ) satisfying 0 α 0 1, max l(β 1, G 0 ) min max κ(β,α 0 ), G 0 α 0 β 0 where the equality holds when G 0 Ĝ 0 (β 1 ) and (β 0,α 0 ) {ˆβ 0 (β 1 ), ˆα 0 (β 1 )}. The profile loglikelihood pl(β 1 ) is a function of β 1 alone, with ˆβ 0 (β 1 ) and ˆα 0 (β 1 ) implicitly defined by equations (2) (4). Note that (3) and (4) are equivalent to each other, and both are equivalent to the equation 0 κ/ α 0, and that (2), (3) and (4) jointly are equivalent to 0 κ/ β 0 and 0 κ/ α 0 jointly. See the Appendix for the gradient and Hessian of κ(β,α 0 ). By implicit differentiation, the gradient and Hessian of pl(β 1 )are pl κ β 1 β, (5) 1 γ ˆγ (β1 ) { 2 pl 2 ( κ β1 2 β1 2 2 κ 2 ) 1 } κ 2 κ, (6) β 1 γ T γ γ T γ β 1 γ ˆγ (β1 ) where κ(β,α 0 ) is treated as κ(β 1,γ) with γ (β 0,α 0 ) T, and ˆγ (β 1 ) {ˆβ 0 (β 1 ), ˆα 0 (β 1 )} T. These formulas can be used in the Newton Raphson algorithm to numerically maximize pl(β 1 ) and find the maximum likelihood estimator ˆβ 1. In asymptotic considerations, assume that n j /n tends to a constant 0 <ρ j < 1asn ( j 1,...,m). Let β be the true value of β, α0 m j1 ρ jλ j, and γ (β0,α 0. The following lemma )T provides basic calculations used in Proposition 2, due to the law of large numbers and the central limit theorem. LEMMA 1. Suppose that β and α 0 are evaluated at the true values β and α 0. (i) As n, in probability, where S 11 m j1 S 12 S T 21 s 22 and z (x, 1) T. 1 n ( 2 κ β β T 2 κ β α 0 2 κ α 0 β T 2 κ α 2 0 ) ( U S11 S 12 S 21 s 22 λ j (1 λ j )exp(β T z)zz T α ρ j λ j + (1 λ j )exp(β T z) dg (1 α0 )exp(β T z)zz T α0 + (1 α 0 )exp(β T z) dg 0, exp(β T z)z α 0 + (1 α 0 )exp(β T z) dg 0, {1 exp(β T z)} 2 α 0 + (1 α 0 )exp(β T z) dg 0, ),

4 232 Z. TAN (ii) As n,n 1/2 ( κ/ β T, κ/ α 0 ) T converges in distribution to a multivariate normal variable with mean 0 and variance matrix ( ) V S11 δs 12 S 21 δs 12 s 22 δs 21 s 22 s 22 δs22 2, where δ m j1 ρ jλ 2 j α 2 0. We show in Proposition 2 (i) that (5) and (6) satisfy the second Bartlett identity: the variance of the gradient equals the expectation of the negative Hessian for a loglikelihood function, evaluated at the true value of the parameter. This leads directly to the negative Hessian representation for the asymptotic variance of the maximum likelihood estimator, and the chi-squared convergence of the loglikelihood ratio statistic in (ii) and (iii). PROPOSITION 2. Rewrite U and V in Lemma 1 by the partition (β 1,γ) instead of (β,α 0 ): ( ) ( ) ( ) U σ11 10, V U 0 S12 S12 δ (S S 21 2s 22 s 21, s 22 ). 22 Assume that U and 00 are nonsingular. (i) Equations (2) (4) for β 1 β1 admit a solution ˆγ (β 1 ) that converges to γ in probability, 1 pl n 1/2 β N (0, V ), 1 β1 β 1 in distribution, and 1 n 2 pl U, β1 β 1 β 2 1 in probability, where U V σ (ii) Assume that x 2 dg 0 < and x 2 dg 1 <. Then 0 pl/ β 1 admits a solution ˆβ 1 that converges in probability to β1, and n1/2 ( ˆβ 1 β1 ) converges in distribution to N (0, V ). (iii) Furthermore, 2{pl(β1 ) pl( ˆβ 1 )} converges in distribution to a chi-squared variable with one degree of freedom. COROLLARY 1. Under the same conditions in Proposition 2 (ii), n 1/2 { ˆβ 1 β1, ˆγ ( ˆβ 1 ) γ } converges in distribution to a multivariate normal variable with mean zero and variance matrix ( ) S s 22, (7) δ where s 22 and S 11 are defined by the partition of U 1 : ( U 1 S 11 S 12 S 21 s 22 Several interesting conclusions can be drawn. First, the maximum likelihood estimator of β is given by ˆβ {ˆβ 1, ˆβ 0 ( ˆβ 1 )} T, and it agrees with that derived in Qin (1999). In fact, { ˆβ 1, ˆγ ( ˆβ 1 )} satisfies 0 κ/ β 1 and 0 κ/ γ, which are equivalent to equations (2 2) and (2 5) in Qin (1999). By Corollary 1, n 1/2 ( ˆβ β ) converges in distribution to a multivariate normal variable with mean 0 and variance S 11, which equals S S 1 11 S 12s 22 S 21 S Second, ˆβ is asymptotically at least as efficient as ˆβ P, the maximum profile likelihood estimator of β in Zou et al. (2002), which is defined as a maximizer to κ(β, α 0 ) with α 0 n j1 (n j/n)λ j.asshown by Zou et al. (2002), n 1/2 ( ˆβ P β ) converges in distribution to a multivariate normal variable with mean zero and variance S 1 11 δs 1 11 S 12S 21 S 1 11, which is no smaller than S11 because δ s 22 by the positive semidefiniteness of (7) in Corollary 1. ).

5 Miscellanea 233 Third, when β 0, ˆβ falls along the line {β :(μ, 1)β 0} asymptotically to the order of n 1/2, where μ x dg 0.Letσ 2 x 2 dg 0 μ 2. The asymptotic variance (7) becomes 1/(δσ 2 ) μ/(δσ 2 ) 0 μ/(δσ 2 ) μ 2 /(δσ 2 ) 0. (8) As a result, ˆα 0 ( ˆβ 1 ) α0 + o p(n 1/2 ), and (μ, 1) ˆβ o p (n 1/2 ). Note that s 22 δ in this case. The asymptotic variance of n 1/2 ( ˆβ P β ) is identical to that of n 1/2 ( ˆβ β ), and therefore ˆβ P falls along the line {β :(μ, 1)β 0} in a similar manner as ˆβ. In the special case where m 2, λ 1 1, and λ 2 0, corresponding to the two-sample exponential tilt model (Qin, 1998), the assumption that U is nonsingular is violated for Proposition 2, because the 3 3 matrix δs xx δs x1 s x1 U δs 1x δs 11 s 11 s 1x s 11 δ 1 s 11 is singular, where s uv exp ( β0 + β 1 x) uv α0 + ( ) ( 1 α0 exp β 0 + β1 x) dg 0 for (u,v) (x, x), (x, 1), or (1,1). This reduction occurs because equation (2) implies that ˆα 0 (β 1 ) n 1 /n, independent of β 1. Nevertheless, similar conclusions to Proposition 2 and Corollary 1 can be obtained. The asymptotic variance of n 1/2 ( ˆβ 1 β1 )isu 1, where U δ(s xx sx1 2 /s 11). The asymptotic variance of n 1/2 ( ˆβ β )is { (sxx ) 1 ( ) } δ 1 s x1 0 (0, 1), s 1x s 11 1 which was previously obtained in Prentice & Pyke (1979) and Qin & Zhang (1997). When β 0, this matrix equals the first 2 2 principal submatrix of (8), and therefore ˆβ falls along the line {β :(μ, 1)β 0} asymptotically to the order of n 1/2. 3. DISCUSSION For model (1), the profile empirical likelihood of β 1 is well behaved with the usual properties of profile likelihoods whether or not β 0, in contrast to that of β, which has an irregularity when β 0. The reason for this difference is that β 0 is a function of (β 1, G 0 ) such that β 0 0 whenever β 1 0. The profile empirical likelihood of β is not defined on the line {(0,β 0 ) T : β 0 0}, which intersects each nonempty neighbourhood of β when β 0. Therefore, the profile empirical likelihood of β does not have a proper Hessian at β when β 0. The maximum likelihood estimator ˆβ is consistent and n 1/2 ( ˆβ β ) is asymptotically normal whether or not β 0. This estimator can be derived in two ways, either by combining ˆβ 1 and ˆβ 0 ( ˆβ 1 ) from the profile empirical likelihood of β 1 in Section 2 or by maximizing the profile empirical likelihood of β in Qin (1999). However, these two derivations are not equally useful for studying the asymptotic behavior of ˆβ. Qin (1999) aimed to show that the profile empirical likelihood of β is well-defined and has a maximizer in a small neighbourhood of β, but the first property fails when β 0. In contrast, such arguments can be properly applied to the profile empirical likelihood of β 1 to obtain Proposition 2 and Corollary 1. Alternatively in our proofs, we focus on the estimating equations and apply the theory of M-estimators. An interesting point is that when β 0, ˆβ falls along the line {β :(μ, 1)β 0} with perfect correlation asymptotically to the first order. As a result, ˆβ lies away from the line {(0,β 0 ) T : β 0 0}, on which the profile empirical likelihood of β is not defined. This separation allows ˆβ, obtained as a maximizer to the

6 234 Z. TAN profile empirical likelihood of β, to still behave asymptotically in the usual manner except with perfect correlation. Furthermore, the restriction of ˆβ to a straight line suggests that when β 0, the loglikelihood ratio statistic for β converges in distribution to a chi-squared variable with one degree of freedom. Formally, this result follows from Proposition 2 (iii), because the profile empirical likelihood of β at ˆβ, orβ 0, is identical to that of β 1 at ˆβ 1,orβ 1 0. ACKNOWLEDGEMENT The author thanks an associate editor and two referees for helpful comments. This research was supported by the U.S. National Science Foundation. APPENDIX Technical details Proof of Proposition 1. (i) If G 0 ({x}) > 0 for some point x outside the sample {x ji : i 1,...,n j, j 1,...,m}, then l(β 1, G 0 ) < l(β 1, G 0 ), where G is a probability distribution such that G 0 ({x}) 0 and G 0 ({x }) G 0 ({x })/{1 G 0 ({x})} for x x. Therefore, we restrict G 0 to probability distributions supported on the sample, and write w ji G 0 ({x ji }). For a fixed β 1, we maximize the loglikelihood log{λ j + (1 λ j )exp(β 0 + β 1 x ji )}+log w ji, over w ji (i 1,...,n j, j 1,...,m) and β 0 subject to the normalization conditions 1 w ji, 1 exp(β 0 + β 1 x ji )w ji. By introducing Lagrange multipliers nα 0 and nα 1 and setting the derivatives with respect to w ji and β 0 to 0, we obtain 0 1 w ji nα 0 nα 1 exp(β 0 + β 1 x ji ), α 1 1 n (1 λ j )exp(β 0 + β 1 x ji ) λ j + (1 λ j )exp(β 0 + β 1 x ji ). Multiplying the left-hand equation here by w ji and summing over the sample yields α 0 + α 1 1. Therefore, w ji s are of the claimed form, equation (2) results from the right-hand equation, and (3) and (4) correspond to the normalization conditions. (ii) The claimed inequality follows because for any distribution G 0 supported on the sample, any 0 α 0 1, and β 0 log{ exp(β 1 x)dg 0 },wehavel(β 1, G 0 ) κ(β,α 0 ). In fact, by Jensen s inequality, n n j l(β 1, G 0 ) κ(β,α 0 ) log G 0 ({x ji }) n{α 0 + (1 α 0 )exp(β 0 + β 1 x ji )} n n j n log G 0 ({x ji }){α 0 + (1 α 0 )exp(β 0 + β 1 x ji )} 0.

7 Miscellanea 235 Gradient and Hessian of κ(β,α 0 ) To find the Gradient and Hessian of κ(β,α 0 ), write z ji (x ji, 1) T (i 1,...,n j, j 1,...,m). By taking derivatives, we find 1 κ 1 n α 0 n 1 κ n β 1 n By taking second derivatives, we find 1 n 1 2 κ n α0 2 2 κ β β T 1 n 1 n 1 2 κ 1 n β α 0 n 1 + exp(β T z ji ) α 0 + (1 α 0 )exp(β T z ji ), { (1 λ j )exp(β T z ji )z ji λ j + (1 λ j )exp(β T z ji ) {(1 exp(β T z ji )} 2 {α 0 + (1 α 0 )exp(β T z ji )} 2, (1 α } 0)exp(β T z ji )z ji, α 0 + (1 α 0 )exp(β T z ji ) [ λ j (1 λ j )exp(β T z ji )z ji z T ji {λ j + (1 λ j )exp(β T z ji )} α ] 0(1 α 0 )exp(β T z ji )z ji z T ji, 2 {α 0 + (1 α 0 )exp(β T z ji )} 2 exp(β T z ji )z ji {α 0 + (1 α 0 )exp(β T z ji )} 2. Proof of Proposition 2. (i) Fix β 1 β1. Equations (2) (4) are equivalent to 0 κ/ γ β 1 β.the 1 individual terms in κ/ γ and 2 κ/( γ γ T ) are uniformly bounded by constants for γ in a neighbourhood of γ. By the asymptotic theory of M-estimators, the equation 0 κ/ γ β1 β admits a solution ˆγ (β 1 1 ) γ + O p (n 1/2 ). More specifically, ˆγ ( ( β1 ) γ 2 ) 1 κ κ γ γ T γ + o p (n 1/2 ). β1 β 1,γ γ By a Taylor expansion of ( pl/ β 1 )(β1 ), given by (5), with ˆγ (β 1 ) around γ, we find { 1 pl 1 ( κ 2 κ 2 ) 1 } κ κ + o n β 1 n β 1 β 1 γ T γ γ T γ p (n 1/2 ). β1 β 1,γ γ β1 β 1 By Lemma 1, as n, n 1/2 ( pl/ β 1 )(β1 ) converges in distribution to a multivariate normal variable with mean zero and variance matrix V ( 1, ( ) V ) σ , where the simplification is due to (1, )(S 12, s 22 ) T 0. Next, by Lemma 1 and (6), n 1 ( 2 pl/ β1 2)(β 1 ) converges in probability to U σ It follows immediately that U V. (ii) Note that ˆβ 1 satisfies 0 pl/ β 1 if and only if { ˆβ 1, ˆγ ( ˆβ 1 )} satisfies 0 κ/ β 1 and 0 κ/ γ. The individual terms in κ/ β 1 and κ/ γ and the second-order derivatives are uniformly bounded by quadratic functions of observations from G 0 and G 1 for (β 1,γ) in a neighbourhood of (β1,γ ). By the asymptotic theory of M-estimators, there exists a solution { ˆβ 1, ˆγ ( ˆβ 1 )}(β1,γ ) + O p (n 1/2 ). By a Taylor expansion of ( pl/ β 1 )(β1 ), we obtain ( ) 1 ˆβ 1 β1 2 pl pl β + o p (n 1/2 ), 1 β1 β 1 which together with (i) implies that n 1/2 ( ˆβ 1 β1 ) converges in distribution to N (0, V ). β 2 1

8 236 Z. TAN (iii) By a second-order Taylor expansion of pl(β1 ), we obtain pl( ˆβ 1 ) pl(β1 ) 1 2 pl ( 2 β1 2 ˆβ 1 β ) op (1), β1 β 1 which together with (i) implies that 2{pl(β1 ) pl( ˆβ 1 )} converges in distribution to a chi-squared distribution with one degree of freedom, as n. Proof of Corollary 1. The proof of Proposition 2 (ii) shows that n 1/2 { ˆβ 1 β1, ˆγ ( ˆβ 1 ) γ } converges in distribution to a multivariate normal variable with mean 0 and variance U 1 V U 1. The matrix U 1 V U 1 reduces to the claimed form, because U 1 (S 12, s 22 ) T (0, 0, 1) T and hence ( ) ( ) U 1 0 S12 U s 22 S 21 s 22. REFERENCES ANDERSON, J. A. (1972). Separate sample logistic discrimination. Biometrika 59, OWEN, A. (2001). Empirical Likelihood. New York: Chapman & Hall. PRENTICE, R. L.& PYKE, R. (1979). Logistic disease incidence models and case-control studies. Biometrika 66, QIN, J. (1998). Inferences for case-control and semiparametric two-sample density ratio models. Biometrika 85, QIN, J. (1999). Empirical likelihood ratio based confidence intervals for mixture proportions. Ann. Statist. 27, QIN, J.& ZHANG, B. (1997). A goodness-of-fit test for logistic regression models based on case-control data. Biometrika 84, ZOU, F.&FINE, J. P. (2002). A note on a partial empirical likelihood. Biometrika 89, ZOU, F., FINE, J. P.& YANDELL, B. S. (2002). On empirical likelihood for a semiparametric mixture model. Biometrika 89, [Received March Revised September 2008]

Nuisance parameter elimination for proportional likelihood ratio models with nonignorable missingness and random truncation

Nuisance parameter elimination for proportional likelihood ratio models with nonignorable missingness and random truncation Biometrika Advance Access published October 24, 202 Biometrika (202), pp. 8 C 202 Biometrika rust Printed in Great Britain doi: 0.093/biomet/ass056 Nuisance parameter elimination for proportional likelihood

More information

University of California, Berkeley

University of California, Berkeley University of California, Berkeley U.C. Berkeley Division of Biostatistics Working Paper Series Year 24 Paper 153 A Note on Empirical Likelihood Inference of Residual Life Regression Ying Qing Chen Yichuan

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

Sample size calculations for logistic and Poisson regression models

Sample size calculations for logistic and Poisson regression models Biometrika (2), 88, 4, pp. 93 99 2 Biometrika Trust Printed in Great Britain Sample size calculations for logistic and Poisson regression models BY GWOWEN SHIEH Department of Management Science, National

More information

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Takeshi Emura and Hisayuki Tsukuma Abstract For testing the regression parameter in multivariate

More information

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2

More information

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests

Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University

More information

Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho

Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho M. de Carvalho and F. J. Marques Abstract We discuss jackknife Euclidean likelihood-based inference methods, with a special focus on the

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH

FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH Jian-Jian Ren 1 and Mai Zhou 2 University of Central Florida and University of Kentucky Abstract: For the regression parameter

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

Lecture 5: LDA and Logistic Regression

Lecture 5: LDA and Logistic Regression Lecture 5: and Logistic Regression Hao Helen Zhang Hao Helen Zhang Lecture 5: and Logistic Regression 1 / 39 Outline Linear Classification Methods Two Popular Linear Models for Classification Linear Discriminant

More information

Minimum distance estimation for the logistic regression model

Minimum distance estimation for the logistic regression model Biometrika (2005), 92, 3, pp. 724 731 2005 Biometrika Trust Printed in Great Britain Minimum distance estimation for the logistic regression model BY HOWARD D. BONDELL Department of Statistics, Rutgers

More information

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12 Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following

More information

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model 1 Linear Regression 2 Linear Regression In this lecture we will study a particular type of regression model: the linear regression model We will first consider the case of the model with one predictor

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

Full likelihood inferences in the Cox model: an empirical likelihood approach

Full likelihood inferences in the Cox model: an empirical likelihood approach Ann Inst Stat Math 2011) 63:1005 1018 DOI 10.1007/s10463-010-0272-y Full likelihood inferences in the Cox model: an empirical likelihood approach Jian-Jian Ren Mai Zhou Received: 22 September 2008 / Revised:

More information

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

Modification and Improvement of Empirical Likelihood for Missing Response Problem

Modification and Improvement of Empirical Likelihood for Missing Response Problem UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation

More information

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note

More information

11 Survival Analysis and Empirical Likelihood

11 Survival Analysis and Empirical Likelihood 11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with

More information

Adjusted Empirical Likelihood for Long-memory Time Series Models

Adjusted Empirical Likelihood for Long-memory Time Series Models Adjusted Empirical Likelihood for Long-memory Time Series Models arxiv:1604.06170v1 [stat.me] 21 Apr 2016 Ramadha D. Piyadi Gamage, Wei Ning and Arjun K. Gupta Department of Mathematics and Statistics

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Semiparametric Regression

Semiparametric Regression Semiparametric Regression Patrick Breheny October 22 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Introduction Over the past few weeks, we ve introduced a variety of regression models under

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

Modeling Real Estate Data using Quantile Regression

Modeling Real Estate Data using Quantile Regression Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices

More information

Classical regularity conditions

Classical regularity conditions Chapter 3 Classical regularity conditions Preliminary draft. Please do not distribute. The results from classical asymptotic theory typically require assumptions of pointwise differentiability of a criterion

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

Lecture 14: Introduction to Poisson Regression

Lecture 14: Introduction to Poisson Regression Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu 8 May 2007 1 / 52 Overview Modelling counts Contingency tables Poisson regression models 2 / 52 Modelling counts I Why

More information

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview Modelling counts I Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu Why count data? Number of traffic accidents per day Mortality counts in a given neighborhood, per week

More information

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X. Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may

More information

8. Hypothesis Testing

8. Hypothesis Testing FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements

More information

STAT 6350 Analysis of Lifetime Data. Failure-time Regression Analysis

STAT 6350 Analysis of Lifetime Data. Failure-time Regression Analysis STAT 6350 Analysis of Lifetime Data Failure-time Regression Analysis Explanatory Variables for Failure Times Usually explanatory variables explain/predict why some units fail quickly and some units survive

More information

UNIVERSITY OF CALIFORNIA, SAN DIEGO

UNIVERSITY OF CALIFORNIA, SAN DIEGO UNIVERSITY OF CALIFORNIA, SAN DIEGO Estimation of the primary hazard ratio in the presence of a secondary covariate with non-proportional hazards An undergraduate honors thesis submitted to the Department

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model

Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;

More information

Chap 2. Linear Classifiers (FTH, ) Yongdai Kim Seoul National University

Chap 2. Linear Classifiers (FTH, ) Yongdai Kim Seoul National University Chap 2. Linear Classifiers (FTH, 4.1-4.4) Yongdai Kim Seoul National University Linear methods for classification 1. Linear classifiers For simplicity, we only consider two-class classification problems

More information

Classification. Chapter Introduction. 6.2 The Bayes classifier

Classification. Chapter Introduction. 6.2 The Bayes classifier Chapter 6 Classification 6.1 Introduction Often encountered in applications is the situation where the response variable Y takes values in a finite set of labels. For example, the response Y could encode

More information

Statistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio

Statistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio \"( Statistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio When the True Parameter Is on the Boundary of the Parameter Space by Ziding Feng 1 and Charles E. McCulloch2

More information

arxiv: v2 [stat.me] 8 Jun 2016

arxiv: v2 [stat.me] 8 Jun 2016 Orthogonality of the Mean and Error Distribution in Generalized Linear Models 1 BY ALAN HUANG 2 and PAUL J. RATHOUZ 3 University of Technology Sydney and University of Wisconsin Madison 4th August, 2013

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Two sample T 2 test 1 Two sample T 2 test 2 Analogous to the univariate context, we

More information

Size Distortion and Modi cation of Classical Vuong Tests

Size Distortion and Modi cation of Classical Vuong Tests Size Distortion and Modi cation of Classical Vuong Tests Xiaoxia Shi University of Wisconsin at Madison March 2011 X. Shi (UW-Mdsn) H 0 : LR = 0 IUPUI 1 / 30 Vuong Test (Vuong, 1989) Data fx i g n i=1.

More information

High-dimensional asymptotic expansions for the distributions of canonical correlations

High-dimensional asymptotic expansions for the distributions of canonical correlations Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

STAT5044: Regression and Anova

STAT5044: Regression and Anova STAT5044: Regression and Anova Inyoung Kim 1 / 15 Outline 1 Fitting GLMs 2 / 15 Fitting GLMS We study how to find the maxlimum likelihood estimator ˆβ of GLM parameters The likelihood equaions are usually

More information

Biometrika Trust. Biometrika Trust is collaborating with JSTOR to digitize, preserve and extend access to Biometrika.

Biometrika Trust. Biometrika Trust is collaborating with JSTOR to digitize, preserve and extend access to Biometrika. Biometrika Trust An Improved Bonferroni Procedure for Multiple Tests of Significance Author(s): R. J. Simes Source: Biometrika, Vol. 73, No. 3 (Dec., 1986), pp. 751-754 Published by: Biometrika Trust Stable

More information

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING DATE AND TIME: August 30, 2018, 14.00 19.00 RESPONSIBLE TEACHER: Niklas Wahlström NUMBER OF PROBLEMS: 5 AIDING MATERIAL: Calculator, mathematical

More information

Survival Analysis for Case-Cohort Studies

Survival Analysis for Case-Cohort Studies Survival Analysis for ase-ohort Studies Petr Klášterecký Dept. of Probability and Mathematical Statistics, Faculty of Mathematics and Physics, harles University, Prague, zech Republic e-mail: petr.klasterecky@matfyz.cz

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

Lecture 16 Solving GLMs via IRWLS

Lecture 16 Solving GLMs via IRWLS Lecture 16 Solving GLMs via IRWLS 09 November 2015 Taylor B. Arnold Yale Statistics STAT 312/612 Notes problem set 5 posted; due next class problem set 6, November 18th Goals for today fixed PCA example

More information

STT 843 Key to Homework 1 Spring 2018

STT 843 Key to Homework 1 Spring 2018 STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ

More information

On consistency of Kendall s tau under censoring

On consistency of Kendall s tau under censoring Biometria (28), 95, 4,pp. 997 11 C 28 Biometria Trust Printed in Great Britain doi: 1.193/biomet/asn37 Advance Access publication 17 September 28 On consistency of Kendall s tau under censoring BY DAVID

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

An Introduction to Multivariate Statistical Analysis

An Introduction to Multivariate Statistical Analysis An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents

More information

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Nonconcave Penalized Likelihood with A Diverging Number of Parameters Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized

More information

Multivariate Analysis and Likelihood Inference

Multivariate Analysis and Likelihood Inference Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density

More information

Scatter plot of data from the study. Linear Regression

Scatter plot of data from the study. Linear Regression 1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25

More information

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives

More information

Likelihood-Based Methods

Likelihood-Based Methods Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

QED. Queen s Economics Department Working Paper No Hypothesis Testing for Arbitrary Bounds. Jeffrey Penney Queen s University

QED. Queen s Economics Department Working Paper No Hypothesis Testing for Arbitrary Bounds. Jeffrey Penney Queen s University QED Queen s Economics Department Working Paper No. 1319 Hypothesis Testing for Arbitrary Bounds Jeffrey Penney Queen s University Department of Economics Queen s University 94 University Avenue Kingston,

More information

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood Mai Zhou Yifan Yang Received:

More information

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department

More information

Binary choice 3.3 Maximum likelihood estimation

Binary choice 3.3 Maximum likelihood estimation Binary choice 3.3 Maximum likelihood estimation Michel Bierlaire Output of the estimation We explain here the various outputs from the maximum likelihood estimation procedure. Solution of the maximum likelihood

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Primal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing

Primal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing Primal-dual Covariate Balance and Minimal Double Robustness via (Joint work with Daniel Percival) Department of Statistics, Stanford University JSM, August 9, 2015 Outline 1 2 3 1/18 Setting Rubin s causal

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

Likelihood Ratio Test in High-Dimensional Logistic Regression Is Asymptotically a Rescaled Chi-Square

Likelihood Ratio Test in High-Dimensional Logistic Regression Is Asymptotically a Rescaled Chi-Square Likelihood Ratio Test in High-Dimensional Logistic Regression Is Asymptotically a Rescaled Chi-Square Yuxin Chen Electrical Engineering, Princeton University Coauthors Pragya Sur Stanford Statistics Emmanuel

More information

i=1 h n (ˆθ n ) = 0. (2)

i=1 h n (ˆθ n ) = 0. (2) Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating

More information

Survival Analysis Math 434 Fall 2011

Survival Analysis Math 434 Fall 2011 Survival Analysis Math 434 Fall 2011 Part IV: Chap. 8,9.2,9.3,11: Semiparametric Proportional Hazards Regression Jimin Ding Math Dept. www.math.wustl.edu/ jmding/math434/fall09/index.html Basic Model Setup

More information

MAS3301 / MAS8311 Biostatistics Part II: Survival

MAS3301 / MAS8311 Biostatistics Part II: Survival MAS3301 / MAS8311 Biostatistics Part II: Survival M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2009-10 1 13 The Cox proportional hazards model 13.1 Introduction In the

More information

September Math Course: First Order Derivative

September Math Course: First Order Derivative September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which

More information

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

Lecture 8 CORRELATION AND LINEAR REGRESSION

Lecture 8 CORRELATION AND LINEAR REGRESSION Announcements CBA5 open in exam mode - deadline midnight Friday! Question 2 on this week s exercises is a prize question. The first good attempt handed in to me by 12 midday this Friday will merit a prize...

More information

Scatter plot of data from the study. Linear Regression

Scatter plot of data from the study. Linear Regression 1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25

More information

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course. Name of the course Statistical methods and data analysis Audience The course is intended for students of the first or second year of the Graduate School in Materials Engineering. The aim of the course

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method

446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method 446 CHAP. 8 NUMERICAL OPTIMIZATION Newton's Search for a Minimum of f(xy) Newton s Method The quadratic approximation method of Section 8.1 generated a sequence of seconddegree Lagrange polynomials. It

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

Managing Uncertainty

Managing Uncertainty Managing Uncertainty Bayesian Linear Regression and Kalman Filter December 4, 2017 Objectives The goal of this lab is multiple: 1. First it is a reminder of some central elementary notions of Bayesian

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Other Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model

Other Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model Other Survival Models (1) Non-PH models We briefly discussed the non-proportional hazards (non-ph) model λ(t Z) = λ 0 (t) exp{β(t) Z}, where β(t) can be estimated by: piecewise constants (recall how);

More information

You can compute the maximum likelihood estimate for the correlation

You can compute the maximum likelihood estimate for the correlation Stat 50 Solutions Comments on Assignment Spring 005. (a) _ 37.6 X = 6.5 5.8 97.84 Σ = 9.70 4.9 9.70 75.05 7.80 4.9 7.80 4.96 (b) 08.7 0 S = Σ = 03 9 6.58 03 305.6 30.89 6.58 30.89 5.5 (c) You can compute

More information

Optimization Methods. Lecture 19: Line Searches and Newton s Method

Optimization Methods. Lecture 19: Line Searches and Newton s Method 15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions

More information

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo PROCEDURES CONTROLLING THE k-fdr USING BIVARIATE DISTRIBUTIONS OF THE NULL p-values Sanat K. Sarkar and Wenge Guo Temple University and National Institute of Environmental Health Sciences Abstract: Procedures

More information

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University

More information

University of California San Diego and Stanford University and

University of California San Diego and Stanford University and First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent

More information