A note on profile likelihood for exponential tilt mixture models
|
|
- Natalie Davis
- 5 years ago
- Views:
Transcription
1 Biometrika (2009), 96, 1,pp C 2009 Biometrika Trust Printed in Great Britain doi: /biomet/asn059 Advance Access publication 22 January 2009 A note on profile likelihood for exponential tilt mixture models BY Z. TAN Department of Statistics, Rutgers University, Piscataway, New Jersey 08854, U.S.A. ztan@stat.rutgers.edu SUMMARY Suppose that independent observations are drawn from multiple distributions, each of which is a mixture of two component distributions such that their log density ratio satisfies a linear model with a slope parameter and an intercept parameter. Inference for such models has been studied using empirical likelihood, and mixed results have been obtained. The profile empirical likelihood of the slope and intercept has an irregularity at the null hypothesis so that the two component distributions are equal. We derive a profile empirical likelihood and maximum likelihood estimator of the slope alone, and obtain the usual asymptotic properties for the estimator and the likelihood ratio statistic regardless of the null. Furthermore, we show the maximum likelihood estimator of the slope and intercept jointly is consistent and asymptotically normal regardless of the null. At the null, the joint maximum likelihood estimator falls along a straight line through the origin with perfect correlation asymptotically to the first order. Some key words: Empirical likelihood; Exponential tilt; Likelihood ratio statistic; Maximum likelihood; Mixture model; Profile likelihood. 1. INTRODUCTION Suppose that a sample of n j independent observations, {x j1,...,x jnj }, is drawn from each of m( 2) distributions P j ( j 1,...,m), and each P j is a mixture of two component distributions, G 0 and G 1, related in an exponential tilt form: dp j λ j dg 0 + (1 λ j )dg 1, where 0 λ j 1 are known mixture proportions, λ j λ k ( j k), and dg 1 exp(β 0 + β 1 x)dg 0, where G 0 is an unknown baseline distribution, β 1 is an unknown parameter, and β 0 log{ exp(β 1 x)dg 0 }. Note that β 1 0 implies that β 0 0 for any G 0.The jth sample, {x j1,...,x jnj }, is obtained from dp j {λ j + (1 λ j )exp(β 0 + β 1 x)} dg 0. (1) The special case of model (1) where m 2, λ 1 1, and λ 2 0 is known as the two-sample exponential tilt model (Qin, 1998), first studied by Anderson (1972) and Prentice & Pyke (1979) in the context of case-control studies. Inference for β (β 1,β 0 ) T and G 0 in model (1) has been studied using empirical likelihood (Owen, 2001), and mixed results have been obtained. Qin (1999) derived a profile empirical likelihood and the resulting maximum likelihood estimator of β in the case where m 3, λ 1 1, λ 2 0, and 0 <λ 3 < 1 may be unknown. However, Zou et al. (2002) pointed out that the profile empirical likelihood of β has an irregularity at the null hypothesis that G 1 G 0, i.e., β 0, and the asymptotic arguments in Qin (1999) for the maximum likelihood estimator of β are not applicable there. Zou & Fine (2002) claimed that
2 230 Z. TAN the maximum likelihood estimator of β is not consistent when β 0. To address this, Zou et al. (2002) proposed a partial profile empirical likelihood, and showed that the maximum partial likelihood estimator of β is consistent and asymptotically normal whether or not β 0, and the profile loglikelihood ratio for testing β 0 has a chi-squared distribution with one degree of freedom asymptotically. A limitation of this method is that, when β 0, the maximum partial likelihood estimator of β always has asymptotic variance no smaller than that of the maximum likelihood estimator in Qin (1999). See Zou & Fine (2002) for a further discussion on the construction of the maximum partial likelihood estimator and the efficiency comparisons of several related estimators. In this note we first derive a profile empirical likelihood and the resulting maximum likelihood estimator of β 1 alone, and show that the profile empirical likelihood of β 1 does not have an irregularity at β 1 0, in contrast to that of β. In fact, under mild regularity conditions, the profile empirical likelihood and the maximum likelihood estimator of β 1 behave like usual profile likelihoods and maximum likelihood estimators with the following properties, whether or not β 1 0: the maximum likelihood estimator is consistent and asymptotically normal; the inverse curvature of the profile loglikelihood at its maximum is a consistent estimator of the asymptotic variance of the maximum likelihood estimator; and the loglikelihood ratio statistic has a chi-squared distribution with one degree of freedom asymptotically. Second, we derive in our approach the maximum likelihood estimator of β, which agrees with that in Qin (1999), and show that it is consistent and asymptotically normal, and always has asymptotic variance no greater than that of the maximum partial likelihood estimator of β in Zou et al. (2002), whether or not β 0. A subtle point is that when β 0, the maximum likelihood estimator of β falls along a straight line through the origin with perfect correlation asymptotically to the first order. The proofs of the main results are provided in the Appendix. The loglikelihood of (β 1, G 0 )is 2. PROFILE LIKELIHOOD AND MAXIMUM LIKELIHOOD ESTIMATOR l(β 1, G 0 ) [log{λ j + (1 λ j )exp(β 0 + β 1 x ji )}+log G 0 ({x ji })], where β 0 log{ exp(β 1 x)dg 0 } is a function of (β 1, G 0 ). The profile loglikelihood of β 1 is defined for each fixed β 1 by maximizing l(β 1, G 0 ) over all possible probability distributions G 0, that is, pl(β 1 ) max G0 l(β 1, G 0 ). The maximum likelihood estimator of β 1 is defined by maximizing the profile loglikelihood, that is, ˆβ 1 arg max pl(β 1 ). We provide a formula for computing the profile loglikelihood pl(β 1 ) in Proposition 1 (i). The formula involves the function κ(β,α 0 ) log { } λ j + (1 λ j )exp(β 0 + β 1 x ji ) n log n, α 0 + (1 α 0 )exp(β 0 + β 1 x ji ) where β (β 1,β 0 ) T and 0 α 0 1 are free arguments. This function has a duality relationship with the loglikelihood l(β 1, G 0 ) as shown in Proposition 1 (ii). PROPOSITION 1. (i) For a fixed β 1, the loglikelihood l(β 1, G 0 ) achieves the maximum at Ĝ 0 Ĝ 0 (β 1 ) supported on {x ji : i 1,...,n j, j 1,...,m} with Ĝ({x ji }) n 1 ˆα 0 + (1 ˆα 0 )exp(ˆβ 0 + β 1 x ji ), where n m j1 n j, and ˆβ 0 ˆβ 0 (β 1 ) and ˆα 0 ˆα 0 (β 1 ) satisfy α 0 1 n λ j λ j + (1 λ j )exp(β 0 + β 1 x ji ), (2)
3 Miscellanea n 1 1 n 1 α 0 + (1 α 0 )exp(β 0 + β 1 x ji ), (3) exp(β 0 + β 1 x ji ) α 0 + (1 α 0 )exp(β 0 + β 1 x ji ). (4) The profile loglikelihood is pl(β 1 ) κ{β 1, ˆβ 0 (β 1 ), ˆα 0 (β 1 )}. (ii) Furthermore, for any G 0 and any (β 0,α 0 ) satisfying 0 α 0 1, max l(β 1, G 0 ) min max κ(β,α 0 ), G 0 α 0 β 0 where the equality holds when G 0 Ĝ 0 (β 1 ) and (β 0,α 0 ) {ˆβ 0 (β 1 ), ˆα 0 (β 1 )}. The profile loglikelihood pl(β 1 ) is a function of β 1 alone, with ˆβ 0 (β 1 ) and ˆα 0 (β 1 ) implicitly defined by equations (2) (4). Note that (3) and (4) are equivalent to each other, and both are equivalent to the equation 0 κ/ α 0, and that (2), (3) and (4) jointly are equivalent to 0 κ/ β 0 and 0 κ/ α 0 jointly. See the Appendix for the gradient and Hessian of κ(β,α 0 ). By implicit differentiation, the gradient and Hessian of pl(β 1 )are pl κ β 1 β, (5) 1 γ ˆγ (β1 ) { 2 pl 2 ( κ β1 2 β1 2 2 κ 2 ) 1 } κ 2 κ, (6) β 1 γ T γ γ T γ β 1 γ ˆγ (β1 ) where κ(β,α 0 ) is treated as κ(β 1,γ) with γ (β 0,α 0 ) T, and ˆγ (β 1 ) {ˆβ 0 (β 1 ), ˆα 0 (β 1 )} T. These formulas can be used in the Newton Raphson algorithm to numerically maximize pl(β 1 ) and find the maximum likelihood estimator ˆβ 1. In asymptotic considerations, assume that n j /n tends to a constant 0 <ρ j < 1asn ( j 1,...,m). Let β be the true value of β, α0 m j1 ρ jλ j, and γ (β0,α 0. The following lemma )T provides basic calculations used in Proposition 2, due to the law of large numbers and the central limit theorem. LEMMA 1. Suppose that β and α 0 are evaluated at the true values β and α 0. (i) As n, in probability, where S 11 m j1 S 12 S T 21 s 22 and z (x, 1) T. 1 n ( 2 κ β β T 2 κ β α 0 2 κ α 0 β T 2 κ α 2 0 ) ( U S11 S 12 S 21 s 22 λ j (1 λ j )exp(β T z)zz T α ρ j λ j + (1 λ j )exp(β T z) dg (1 α0 )exp(β T z)zz T α0 + (1 α 0 )exp(β T z) dg 0, exp(β T z)z α 0 + (1 α 0 )exp(β T z) dg 0, {1 exp(β T z)} 2 α 0 + (1 α 0 )exp(β T z) dg 0, ),
4 232 Z. TAN (ii) As n,n 1/2 ( κ/ β T, κ/ α 0 ) T converges in distribution to a multivariate normal variable with mean 0 and variance matrix ( ) V S11 δs 12 S 21 δs 12 s 22 δs 21 s 22 s 22 δs22 2, where δ m j1 ρ jλ 2 j α 2 0. We show in Proposition 2 (i) that (5) and (6) satisfy the second Bartlett identity: the variance of the gradient equals the expectation of the negative Hessian for a loglikelihood function, evaluated at the true value of the parameter. This leads directly to the negative Hessian representation for the asymptotic variance of the maximum likelihood estimator, and the chi-squared convergence of the loglikelihood ratio statistic in (ii) and (iii). PROPOSITION 2. Rewrite U and V in Lemma 1 by the partition (β 1,γ) instead of (β,α 0 ): ( ) ( ) ( ) U σ11 10, V U 0 S12 S12 δ (S S 21 2s 22 s 21, s 22 ). 22 Assume that U and 00 are nonsingular. (i) Equations (2) (4) for β 1 β1 admit a solution ˆγ (β 1 ) that converges to γ in probability, 1 pl n 1/2 β N (0, V ), 1 β1 β 1 in distribution, and 1 n 2 pl U, β1 β 1 β 2 1 in probability, where U V σ (ii) Assume that x 2 dg 0 < and x 2 dg 1 <. Then 0 pl/ β 1 admits a solution ˆβ 1 that converges in probability to β1, and n1/2 ( ˆβ 1 β1 ) converges in distribution to N (0, V ). (iii) Furthermore, 2{pl(β1 ) pl( ˆβ 1 )} converges in distribution to a chi-squared variable with one degree of freedom. COROLLARY 1. Under the same conditions in Proposition 2 (ii), n 1/2 { ˆβ 1 β1, ˆγ ( ˆβ 1 ) γ } converges in distribution to a multivariate normal variable with mean zero and variance matrix ( ) S s 22, (7) δ where s 22 and S 11 are defined by the partition of U 1 : ( U 1 S 11 S 12 S 21 s 22 Several interesting conclusions can be drawn. First, the maximum likelihood estimator of β is given by ˆβ {ˆβ 1, ˆβ 0 ( ˆβ 1 )} T, and it agrees with that derived in Qin (1999). In fact, { ˆβ 1, ˆγ ( ˆβ 1 )} satisfies 0 κ/ β 1 and 0 κ/ γ, which are equivalent to equations (2 2) and (2 5) in Qin (1999). By Corollary 1, n 1/2 ( ˆβ β ) converges in distribution to a multivariate normal variable with mean 0 and variance S 11, which equals S S 1 11 S 12s 22 S 21 S Second, ˆβ is asymptotically at least as efficient as ˆβ P, the maximum profile likelihood estimator of β in Zou et al. (2002), which is defined as a maximizer to κ(β, α 0 ) with α 0 n j1 (n j/n)λ j.asshown by Zou et al. (2002), n 1/2 ( ˆβ P β ) converges in distribution to a multivariate normal variable with mean zero and variance S 1 11 δs 1 11 S 12S 21 S 1 11, which is no smaller than S11 because δ s 22 by the positive semidefiniteness of (7) in Corollary 1. ).
5 Miscellanea 233 Third, when β 0, ˆβ falls along the line {β :(μ, 1)β 0} asymptotically to the order of n 1/2, where μ x dg 0.Letσ 2 x 2 dg 0 μ 2. The asymptotic variance (7) becomes 1/(δσ 2 ) μ/(δσ 2 ) 0 μ/(δσ 2 ) μ 2 /(δσ 2 ) 0. (8) As a result, ˆα 0 ( ˆβ 1 ) α0 + o p(n 1/2 ), and (μ, 1) ˆβ o p (n 1/2 ). Note that s 22 δ in this case. The asymptotic variance of n 1/2 ( ˆβ P β ) is identical to that of n 1/2 ( ˆβ β ), and therefore ˆβ P falls along the line {β :(μ, 1)β 0} in a similar manner as ˆβ. In the special case where m 2, λ 1 1, and λ 2 0, corresponding to the two-sample exponential tilt model (Qin, 1998), the assumption that U is nonsingular is violated for Proposition 2, because the 3 3 matrix δs xx δs x1 s x1 U δs 1x δs 11 s 11 s 1x s 11 δ 1 s 11 is singular, where s uv exp ( β0 + β 1 x) uv α0 + ( ) ( 1 α0 exp β 0 + β1 x) dg 0 for (u,v) (x, x), (x, 1), or (1,1). This reduction occurs because equation (2) implies that ˆα 0 (β 1 ) n 1 /n, independent of β 1. Nevertheless, similar conclusions to Proposition 2 and Corollary 1 can be obtained. The asymptotic variance of n 1/2 ( ˆβ 1 β1 )isu 1, where U δ(s xx sx1 2 /s 11). The asymptotic variance of n 1/2 ( ˆβ β )is { (sxx ) 1 ( ) } δ 1 s x1 0 (0, 1), s 1x s 11 1 which was previously obtained in Prentice & Pyke (1979) and Qin & Zhang (1997). When β 0, this matrix equals the first 2 2 principal submatrix of (8), and therefore ˆβ falls along the line {β :(μ, 1)β 0} asymptotically to the order of n 1/2. 3. DISCUSSION For model (1), the profile empirical likelihood of β 1 is well behaved with the usual properties of profile likelihoods whether or not β 0, in contrast to that of β, which has an irregularity when β 0. The reason for this difference is that β 0 is a function of (β 1, G 0 ) such that β 0 0 whenever β 1 0. The profile empirical likelihood of β is not defined on the line {(0,β 0 ) T : β 0 0}, which intersects each nonempty neighbourhood of β when β 0. Therefore, the profile empirical likelihood of β does not have a proper Hessian at β when β 0. The maximum likelihood estimator ˆβ is consistent and n 1/2 ( ˆβ β ) is asymptotically normal whether or not β 0. This estimator can be derived in two ways, either by combining ˆβ 1 and ˆβ 0 ( ˆβ 1 ) from the profile empirical likelihood of β 1 in Section 2 or by maximizing the profile empirical likelihood of β in Qin (1999). However, these two derivations are not equally useful for studying the asymptotic behavior of ˆβ. Qin (1999) aimed to show that the profile empirical likelihood of β is well-defined and has a maximizer in a small neighbourhood of β, but the first property fails when β 0. In contrast, such arguments can be properly applied to the profile empirical likelihood of β 1 to obtain Proposition 2 and Corollary 1. Alternatively in our proofs, we focus on the estimating equations and apply the theory of M-estimators. An interesting point is that when β 0, ˆβ falls along the line {β :(μ, 1)β 0} with perfect correlation asymptotically to the first order. As a result, ˆβ lies away from the line {(0,β 0 ) T : β 0 0}, on which the profile empirical likelihood of β is not defined. This separation allows ˆβ, obtained as a maximizer to the
6 234 Z. TAN profile empirical likelihood of β, to still behave asymptotically in the usual manner except with perfect correlation. Furthermore, the restriction of ˆβ to a straight line suggests that when β 0, the loglikelihood ratio statistic for β converges in distribution to a chi-squared variable with one degree of freedom. Formally, this result follows from Proposition 2 (iii), because the profile empirical likelihood of β at ˆβ, orβ 0, is identical to that of β 1 at ˆβ 1,orβ 1 0. ACKNOWLEDGEMENT The author thanks an associate editor and two referees for helpful comments. This research was supported by the U.S. National Science Foundation. APPENDIX Technical details Proof of Proposition 1. (i) If G 0 ({x}) > 0 for some point x outside the sample {x ji : i 1,...,n j, j 1,...,m}, then l(β 1, G 0 ) < l(β 1, G 0 ), where G is a probability distribution such that G 0 ({x}) 0 and G 0 ({x }) G 0 ({x })/{1 G 0 ({x})} for x x. Therefore, we restrict G 0 to probability distributions supported on the sample, and write w ji G 0 ({x ji }). For a fixed β 1, we maximize the loglikelihood log{λ j + (1 λ j )exp(β 0 + β 1 x ji )}+log w ji, over w ji (i 1,...,n j, j 1,...,m) and β 0 subject to the normalization conditions 1 w ji, 1 exp(β 0 + β 1 x ji )w ji. By introducing Lagrange multipliers nα 0 and nα 1 and setting the derivatives with respect to w ji and β 0 to 0, we obtain 0 1 w ji nα 0 nα 1 exp(β 0 + β 1 x ji ), α 1 1 n (1 λ j )exp(β 0 + β 1 x ji ) λ j + (1 λ j )exp(β 0 + β 1 x ji ). Multiplying the left-hand equation here by w ji and summing over the sample yields α 0 + α 1 1. Therefore, w ji s are of the claimed form, equation (2) results from the right-hand equation, and (3) and (4) correspond to the normalization conditions. (ii) The claimed inequality follows because for any distribution G 0 supported on the sample, any 0 α 0 1, and β 0 log{ exp(β 1 x)dg 0 },wehavel(β 1, G 0 ) κ(β,α 0 ). In fact, by Jensen s inequality, n n j l(β 1, G 0 ) κ(β,α 0 ) log G 0 ({x ji }) n{α 0 + (1 α 0 )exp(β 0 + β 1 x ji )} n n j n log G 0 ({x ji }){α 0 + (1 α 0 )exp(β 0 + β 1 x ji )} 0.
7 Miscellanea 235 Gradient and Hessian of κ(β,α 0 ) To find the Gradient and Hessian of κ(β,α 0 ), write z ji (x ji, 1) T (i 1,...,n j, j 1,...,m). By taking derivatives, we find 1 κ 1 n α 0 n 1 κ n β 1 n By taking second derivatives, we find 1 n 1 2 κ n α0 2 2 κ β β T 1 n 1 n 1 2 κ 1 n β α 0 n 1 + exp(β T z ji ) α 0 + (1 α 0 )exp(β T z ji ), { (1 λ j )exp(β T z ji )z ji λ j + (1 λ j )exp(β T z ji ) {(1 exp(β T z ji )} 2 {α 0 + (1 α 0 )exp(β T z ji )} 2, (1 α } 0)exp(β T z ji )z ji, α 0 + (1 α 0 )exp(β T z ji ) [ λ j (1 λ j )exp(β T z ji )z ji z T ji {λ j + (1 λ j )exp(β T z ji )} α ] 0(1 α 0 )exp(β T z ji )z ji z T ji, 2 {α 0 + (1 α 0 )exp(β T z ji )} 2 exp(β T z ji )z ji {α 0 + (1 α 0 )exp(β T z ji )} 2. Proof of Proposition 2. (i) Fix β 1 β1. Equations (2) (4) are equivalent to 0 κ/ γ β 1 β.the 1 individual terms in κ/ γ and 2 κ/( γ γ T ) are uniformly bounded by constants for γ in a neighbourhood of γ. By the asymptotic theory of M-estimators, the equation 0 κ/ γ β1 β admits a solution ˆγ (β 1 1 ) γ + O p (n 1/2 ). More specifically, ˆγ ( ( β1 ) γ 2 ) 1 κ κ γ γ T γ + o p (n 1/2 ). β1 β 1,γ γ By a Taylor expansion of ( pl/ β 1 )(β1 ), given by (5), with ˆγ (β 1 ) around γ, we find { 1 pl 1 ( κ 2 κ 2 ) 1 } κ κ + o n β 1 n β 1 β 1 γ T γ γ T γ p (n 1/2 ). β1 β 1,γ γ β1 β 1 By Lemma 1, as n, n 1/2 ( pl/ β 1 )(β1 ) converges in distribution to a multivariate normal variable with mean zero and variance matrix V ( 1, ( ) V ) σ , where the simplification is due to (1, )(S 12, s 22 ) T 0. Next, by Lemma 1 and (6), n 1 ( 2 pl/ β1 2)(β 1 ) converges in probability to U σ It follows immediately that U V. (ii) Note that ˆβ 1 satisfies 0 pl/ β 1 if and only if { ˆβ 1, ˆγ ( ˆβ 1 )} satisfies 0 κ/ β 1 and 0 κ/ γ. The individual terms in κ/ β 1 and κ/ γ and the second-order derivatives are uniformly bounded by quadratic functions of observations from G 0 and G 1 for (β 1,γ) in a neighbourhood of (β1,γ ). By the asymptotic theory of M-estimators, there exists a solution { ˆβ 1, ˆγ ( ˆβ 1 )}(β1,γ ) + O p (n 1/2 ). By a Taylor expansion of ( pl/ β 1 )(β1 ), we obtain ( ) 1 ˆβ 1 β1 2 pl pl β + o p (n 1/2 ), 1 β1 β 1 which together with (i) implies that n 1/2 ( ˆβ 1 β1 ) converges in distribution to N (0, V ). β 2 1
8 236 Z. TAN (iii) By a second-order Taylor expansion of pl(β1 ), we obtain pl( ˆβ 1 ) pl(β1 ) 1 2 pl ( 2 β1 2 ˆβ 1 β ) op (1), β1 β 1 which together with (i) implies that 2{pl(β1 ) pl( ˆβ 1 )} converges in distribution to a chi-squared distribution with one degree of freedom, as n. Proof of Corollary 1. The proof of Proposition 2 (ii) shows that n 1/2 { ˆβ 1 β1, ˆγ ( ˆβ 1 ) γ } converges in distribution to a multivariate normal variable with mean 0 and variance U 1 V U 1. The matrix U 1 V U 1 reduces to the claimed form, because U 1 (S 12, s 22 ) T (0, 0, 1) T and hence ( ) ( ) U 1 0 S12 U s 22 S 21 s 22. REFERENCES ANDERSON, J. A. (1972). Separate sample logistic discrimination. Biometrika 59, OWEN, A. (2001). Empirical Likelihood. New York: Chapman & Hall. PRENTICE, R. L.& PYKE, R. (1979). Logistic disease incidence models and case-control studies. Biometrika 66, QIN, J. (1998). Inferences for case-control and semiparametric two-sample density ratio models. Biometrika 85, QIN, J. (1999). Empirical likelihood ratio based confidence intervals for mixture proportions. Ann. Statist. 27, QIN, J.& ZHANG, B. (1997). A goodness-of-fit test for logistic regression models based on case-control data. Biometrika 84, ZOU, F.&FINE, J. P. (2002). A note on a partial empirical likelihood. Biometrika 89, ZOU, F., FINE, J. P.& YANDELL, B. S. (2002). On empirical likelihood for a semiparametric mixture model. Biometrika 89, [Received March Revised September 2008]
Nuisance parameter elimination for proportional likelihood ratio models with nonignorable missingness and random truncation
Biometrika Advance Access published October 24, 202 Biometrika (202), pp. 8 C 202 Biometrika rust Printed in Great Britain doi: 0.093/biomet/ass056 Nuisance parameter elimination for proportional likelihood
More informationUniversity of California, Berkeley
University of California, Berkeley U.C. Berkeley Division of Biostatistics Working Paper Series Year 24 Paper 153 A Note on Empirical Likelihood Inference of Residual Life Regression Ying Qing Chen Yichuan
More informationFULL LIKELIHOOD INFERENCES IN THE COX MODEL
October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach
More informationSample size calculations for logistic and Poisson regression models
Biometrika (2), 88, 4, pp. 93 99 2 Biometrika Trust Printed in Great Britain Sample size calculations for logistic and Poisson regression models BY GWOWEN SHIEH Department of Management Science, National
More informationHypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations
Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations Takeshi Emura and Hisayuki Tsukuma Abstract For testing the regression parameter in multivariate
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationSize and Shape of Confidence Regions from Extended Empirical Likelihood Tests
Biometrika (2014),,, pp. 1 13 C 2014 Biometrika Trust Printed in Great Britain Size and Shape of Confidence Regions from Extended Empirical Likelihood Tests BY M. ZHOU Department of Statistics, University
More informationJackknife Euclidean Likelihood-Based Inference for Spearman s Rho
Jackknife Euclidean Likelihood-Based Inference for Spearman s Rho M. de Carvalho and F. J. Marques Abstract We discuss jackknife Euclidean likelihood-based inference methods, with a special focus on the
More informationFULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH
FULL LIKELIHOOD INFERENCES IN THE COX MODEL: AN EMPIRICAL LIKELIHOOD APPROACH Jian-Jian Ren 1 and Mai Zhou 2 University of Central Florida and University of Kentucky Abstract: For the regression parameter
More information(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )
Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative
More informationLecture 5: LDA and Logistic Regression
Lecture 5: and Logistic Regression Hao Helen Zhang Hao Helen Zhang Lecture 5: and Logistic Regression 1 / 39 Outline Linear Classification Methods Two Popular Linear Models for Classification Linear Discriminant
More informationMinimum distance estimation for the logistic regression model
Biometrika (2005), 92, 3, pp. 724 731 2005 Biometrika Trust Printed in Great Britain Minimum distance estimation for the logistic regression model BY HOWARD D. BONDELL Department of Statistics, Rutgers
More informationPartitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12
Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following
More informationLinear Regression. In this lecture we will study a particular type of regression model: the linear regression model
1 Linear Regression 2 Linear Regression In this lecture we will study a particular type of regression model: the linear regression model We will first consider the case of the model with one predictor
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationFull likelihood inferences in the Cox model: an empirical likelihood approach
Ann Inst Stat Math 2011) 63:1005 1018 DOI 10.1007/s10463-010-0272-y Full likelihood inferences in the Cox model: an empirical likelihood approach Jian-Jian Ren Mai Zhou Received: 22 September 2008 / Revised:
More informationPairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion
Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationModification and Improvement of Empirical Likelihood for Missing Response Problem
UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu
More informationDiscussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon
Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics
More informationChapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27
Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation
More informationA Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints
Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationAdjusted Empirical Likelihood for Long-memory Time Series Models
Adjusted Empirical Likelihood for Long-memory Time Series Models arxiv:1604.06170v1 [stat.me] 21 Apr 2016 Ramadha D. Piyadi Gamage, Wei Ning and Arjun K. Gupta Department of Mathematics and Statistics
More informationLarge Sample Properties of Estimators in the Classical Linear Regression Model
Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in
More informationSTAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015
STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis
More informationSemiparametric Regression
Semiparametric Regression Patrick Breheny October 22 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Introduction Over the past few weeks, we ve introduced a variety of regression models under
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationModeling Real Estate Data using Quantile Regression
Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices
More informationClassical regularity conditions
Chapter 3 Classical regularity conditions Preliminary draft. Please do not distribute. The results from classical asymptotic theory typically require assumptions of pointwise differentiability of a criterion
More information1 Directional Derivatives and Differentiability
Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationLecture 14: Introduction to Poisson Regression
Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu 8 May 2007 1 / 52 Overview Modelling counts Contingency tables Poisson regression models 2 / 52 Modelling counts I Why
More informationModelling counts. Lecture 14: Introduction to Poisson Regression. Overview
Modelling counts I Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu Why count data? Number of traffic accidents per day Mortality counts in a given neighborhood, per week
More informationOptimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.
Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may
More information8. Hypothesis Testing
FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements
More informationSTAT 6350 Analysis of Lifetime Data. Failure-time Regression Analysis
STAT 6350 Analysis of Lifetime Data Failure-time Regression Analysis Explanatory Variables for Failure Times Usually explanatory variables explain/predict why some units fail quickly and some units survive
More informationUNIVERSITY OF CALIFORNIA, SAN DIEGO
UNIVERSITY OF CALIFORNIA, SAN DIEGO Estimation of the primary hazard ratio in the presence of a secondary covariate with non-proportional hazards An undergraduate honors thesis submitted to the Department
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationSome Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model
Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;
More informationChap 2. Linear Classifiers (FTH, ) Yongdai Kim Seoul National University
Chap 2. Linear Classifiers (FTH, 4.1-4.4) Yongdai Kim Seoul National University Linear methods for classification 1. Linear classifiers For simplicity, we only consider two-class classification problems
More informationClassification. Chapter Introduction. 6.2 The Bayes classifier
Chapter 6 Classification 6.1 Introduction Often encountered in applications is the situation where the response variable Y takes values in a finite set of labels. For example, the response Y could encode
More informationStatistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio
\"( Statistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio When the True Parameter Is on the Boundary of the Parameter Space by Ziding Feng 1 and Charles E. McCulloch2
More informationarxiv: v2 [stat.me] 8 Jun 2016
Orthogonality of the Mean and Error Distribution in Generalized Linear Models 1 BY ALAN HUANG 2 and PAUL J. RATHOUZ 3 University of Technology Sydney and University of Wisconsin Madison 4th August, 2013
More informationConstrained Optimization
1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange
More informationInference For High Dimensional M-estimates. Fixed Design Results
: Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Two sample T 2 test 1 Two sample T 2 test 2 Analogous to the univariate context, we
More informationSize Distortion and Modi cation of Classical Vuong Tests
Size Distortion and Modi cation of Classical Vuong Tests Xiaoxia Shi University of Wisconsin at Madison March 2011 X. Shi (UW-Mdsn) H 0 : LR = 0 IUPUI 1 / 30 Vuong Test (Vuong, 1989) Data fx i g n i=1.
More informationHigh-dimensional asymptotic expansions for the distributions of canonical correlations
Journal of Multivariate Analysis 100 2009) 231 242 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva High-dimensional asymptotic
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More informationSTAT5044: Regression and Anova
STAT5044: Regression and Anova Inyoung Kim 1 / 15 Outline 1 Fitting GLMs 2 / 15 Fitting GLMS We study how to find the maxlimum likelihood estimator ˆβ of GLM parameters The likelihood equaions are usually
More informationBiometrika Trust. Biometrika Trust is collaborating with JSTOR to digitize, preserve and extend access to Biometrika.
Biometrika Trust An Improved Bonferroni Procedure for Multiple Tests of Significance Author(s): R. J. Simes Source: Biometrika, Vol. 73, No. 3 (Dec., 1986), pp. 751-754 Published by: Biometrika Trust Stable
More informationEXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING
EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING DATE AND TIME: August 30, 2018, 14.00 19.00 RESPONSIBLE TEACHER: Niklas Wahlström NUMBER OF PROBLEMS: 5 AIDING MATERIAL: Calculator, mathematical
More informationSurvival Analysis for Case-Cohort Studies
Survival Analysis for ase-ohort Studies Petr Klášterecký Dept. of Probability and Mathematical Statistics, Faculty of Mathematics and Physics, harles University, Prague, zech Republic e-mail: petr.klasterecky@matfyz.cz
More informationAdvanced Quantitative Methods: maximum likelihood
Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y
More informationOutline of GLMs. Definitions
Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density
More informationLecture 16 Solving GLMs via IRWLS
Lecture 16 Solving GLMs via IRWLS 09 November 2015 Taylor B. Arnold Yale Statistics STAT 312/612 Notes problem set 5 posted; due next class problem set 6, November 18th Goals for today fixed PCA example
More informationSTT 843 Key to Homework 1 Spring 2018
STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ
More informationOn consistency of Kendall s tau under censoring
Biometria (28), 95, 4,pp. 997 11 C 28 Biometria Trust Printed in Great Britain doi: 1.193/biomet/asn37 Advance Access publication 17 September 28 On consistency of Kendall s tau under censoring BY DAVID
More informationGeneralized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses
Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationAn Introduction to Multivariate Statistical Analysis
An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents
More informationNonconcave Penalized Likelihood with A Diverging Number of Parameters
Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationScatter plot of data from the study. Linear Regression
1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25
More informationRestricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model
Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives
More informationLikelihood-Based Methods
Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)
More informationPOLI 8501 Introduction to Maximum Likelihood Estimation
POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,
More informationQED. Queen s Economics Department Working Paper No Hypothesis Testing for Arbitrary Bounds. Jeffrey Penney Queen s University
QED Queen s Economics Department Working Paper No. 1319 Hypothesis Testing for Arbitrary Bounds Jeffrey Penney Queen s University Department of Economics Queen s University 94 University Avenue Kingston,
More informationA Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood
Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints and Its Application to Empirical Likelihood Mai Zhou Yifan Yang Received:
More informationHigh Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data
High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department
More informationBinary choice 3.3 Maximum likelihood estimation
Binary choice 3.3 Maximum likelihood estimation Michel Bierlaire Output of the estimation We explain here the various outputs from the maximum likelihood estimation procedure. Solution of the maximum likelihood
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationPrimal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing
Primal-dual Covariate Balance and Minimal Double Robustness via (Joint work with Daniel Percival) Department of Statistics, Stanford University JSM, August 9, 2015 Outline 1 2 3 1/18 Setting Rubin s causal
More informationAsymptotic inference for a nonstationary double ar(1) model
Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk
More informationInferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data
Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone
More informationLikelihood Ratio Test in High-Dimensional Logistic Regression Is Asymptotically a Rescaled Chi-Square
Likelihood Ratio Test in High-Dimensional Logistic Regression Is Asymptotically a Rescaled Chi-Square Yuxin Chen Electrical Engineering, Princeton University Coauthors Pragya Sur Stanford Statistics Emmanuel
More informationi=1 h n (ˆθ n ) = 0. (2)
Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating
More informationSurvival Analysis Math 434 Fall 2011
Survival Analysis Math 434 Fall 2011 Part IV: Chap. 8,9.2,9.3,11: Semiparametric Proportional Hazards Regression Jimin Ding Math Dept. www.math.wustl.edu/ jmding/math434/fall09/index.html Basic Model Setup
More informationMAS3301 / MAS8311 Biostatistics Part II: Survival
MAS3301 / MAS8311 Biostatistics Part II: Survival M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2009-10 1 13 The Cox proportional hazards model 13.1 Introduction In the
More informationSeptember Math Course: First Order Derivative
September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which
More informationT.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS
ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationThe Multivariate Gaussian Distribution [DRAFT]
The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,
More informationLecture 8 CORRELATION AND LINEAR REGRESSION
Announcements CBA5 open in exam mode - deadline midnight Friday! Question 2 on this week s exercises is a prize question. The first good attempt handed in to me by 12 midday this Friday will merit a prize...
More informationScatter plot of data from the study. Linear Regression
1 2 Linear Regression Scatter plot of data from the study. Consider a study to relate birthweight to the estriol level of pregnant women. The data is below. i Weight (g / 100) i Weight (g / 100) 1 7 25
More information* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.
Name of the course Statistical methods and data analysis Audience The course is intended for students of the first or second year of the Graduate School in Materials Engineering. The aim of the course
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More information446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method
446 CHAP. 8 NUMERICAL OPTIMIZATION Newton's Search for a Minimum of f(xy) Newton s Method The quadratic approximation method of Section 8.1 generated a sequence of seconddegree Lagrange polynomials. It
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationManaging Uncertainty
Managing Uncertainty Bayesian Linear Regression and Kalman Filter December 4, 2017 Objectives The goal of this lab is multiple: 1. First it is a reminder of some central elementary notions of Bayesian
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationOther Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model
Other Survival Models (1) Non-PH models We briefly discussed the non-proportional hazards (non-ph) model λ(t Z) = λ 0 (t) exp{β(t) Z}, where β(t) can be estimated by: piecewise constants (recall how);
More informationYou can compute the maximum likelihood estimate for the correlation
Stat 50 Solutions Comments on Assignment Spring 005. (a) _ 37.6 X = 6.5 5.8 97.84 Σ = 9.70 4.9 9.70 75.05 7.80 4.9 7.80 4.96 (b) 08.7 0 S = Σ = 03 9 6.58 03 305.6 30.89 6.58 30.89 5.5 (c) You can compute
More informationOptimization Methods. Lecture 19: Line Searches and Newton s Method
15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions
More informationPROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo
PROCEDURES CONTROLLING THE k-fdr USING BIVARIATE DISTRIBUTIONS OF THE NULL p-values Sanat K. Sarkar and Wenge Guo Temple University and National Institute of Environmental Health Sciences Abstract: Procedures
More informationAsymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University
More informationUniversity of California San Diego and Stanford University and
First International Workshop on Functional and Operatorial Statistics. Toulouse, June 19-21, 2008 K-sample Subsampling Dimitris N. olitis andjoseph.romano University of California San Diego and Stanford
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More information