Figure 36: Respiratory infection versus time for the first 49 children.
|
|
- Maurice Ford
- 5 years ago
- Views:
Transcription
1 y BINARY DATA MODELS We devote an entire chapter to binary data since such data are challenging, both in terms of modeling the dependence, and parameter interpretation. We again consider mixed effects models approaches (with inference from likelihood or Bayes perspectives), and GEE. Motivating Data - Indonesian Child Health DHLZ describe a dataset in which the response is the absence/presence of respiratory illness in 275 children, with mutiple measurements being collected over time quarterly measurements were taken for up to six consecutive quarters. Age also recorded. Question of interest: Is the prevalence of respiratory infection higher amongst children who suffer from xerophthalmia, an ocular manifestation of chronic vitamin A deficiency? Figure 36 shows the infection indicator versus time for the first 49 children. 329 id id id id id id id time Figure 36: Respiratory infection versus time for the first 49 children. 330
2 xero id id id id id id id time Figure 37: Xerophthalmia versus time for the first 49 children. 331 Exploratory Summaries With binary responses and covariates, plots are not so informative - instead look at tables. > table(n) # no of responses per child (ie 22 children had 1),1200 in total > xtabs(~xerocross+ycross) # cross-sectional comparison of respiratory ycross # event (ycross) and xerophthalmia (xerocross) xerocross 0 1 # at first time point > 234/(9*31) [1] > summary(glm(cbind(ycross,1-ycross) ~ xerocross, family="binomial")) Estimate Std. Error z value Pr(> z ) (Intercept) <2e-16 *** xerocross > exp(-.1759) [1]
3 Now add in age (potential confounder). > modglm <- glm(cbind(ycross,1-ycross) ~ agecross+xerocross, family="binomial") > summary(modglm) # Cross-sectional analyses Coefficients: Estimate Std. Error z value Pr(> z ) (Intercept) <2e-16 *** agecross * xerocross Null deviance: on 274 degrees of freedom Residual deviance: on 272 degrees of freedom > modglmq <- glm(cbind(ycross,1-ycross) ~ agecross+xerocross,family=quasibinomial()) > summary(modglmq) Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) <2e-16 *** agecross * xerocross (Dispersion parameter for quasibinomial family taken to be ) Null deviance: on 274 degrees of freedom Residual deviance: on 272 degrees of freedom 333 Overview of Modeling Clustered Binary Data For a single binary variable, Y, all moments are determined by p = E[Y ]. Specifically E[Y r ] = p for r 1. Consider n observations from a single individual Y = (Y 1,..., Y n ) T. The saturated model has 2 n 1 parameters (in contrast to n means, n variances and n(n 1)/2 correlations for the saturated normal model). There is no obvious or natural manner by which multivariate binary data can be specified. We begin with a mixed effects formulation that induces correlations. 334
4 Mixed Effects Models: Conjugate Model gives n Pr(Y = y) = y Marginal moments: Y p Binomial(n, p) Γ(a + b) Γ(a)Γ(b) E[Y ] = ne[p] = n a a + b, p Beta(a, b) Γ(a + y)γ(b + n y), y = 0,1,..., n. Γ(a + b + n) + b + n var(y ) = ne[p](1 E[p])a a + b + 1 Limitations: likelihood is complex, extension to covariates increases complexity, and no tractable way to allow random effect slopes. 335 Mixed Effects Models: Logistic Model A GLMM for binary data takes the binomial exponential family, with the canonical link being logistic. We have Stage 1: Y ij p ij ind Binomial(n ij, p ij ) with «pij log = x ij β + z ij b i 1 p ij Stage 2: b i D iid N(0, D). Marginal moments are not available in closed form. 336
5 Likelihood Inference Likelihood inference proceeds in the usual asymptotic fashion, following integration of the random effects. A Bayesian analysis adds priors for the fixed effects and variance components. Model checking is very difficult for binary data > library(lme4) > memmod1 <- lmer(cbind(y,1-y) ~ age+xero+(1 id),family=binomial,method="pql") > summary(memmod1) Random effects: Groups Name Variance Std.Dev. id (Intercept) Fixed effects: Estimate Std. Error z value Pr(> z ) (Intercept) < 2.2e-16 *** age ** xero memmod2 <- lmer(cbind(y,1-y) ~ age+xero+(1 id),family=binomial,method="laplace") > summary(memmod2) Random effects: Groups Name Variance Std.Dev. id (Intercept) Fixed effects: Estimate Std. Error z value Pr(> z ) (Intercept) < 2e-16 *** age * xero
6 Likelihoods for Multivariate Binary Data Log-Linear Model We have 2 n 1 distinct probabilities, but we wish to consider formulations that allow more parsimonious descriptions as a function of covariates. One choice is the log-linear model: 0 1 Pr(Y = y) = c(θ) X θ (1) j y j + X θ (2) j 1 j y j1 y j θ (n) n y 1...y n A, j j 1 <j 2 with 2 n 1 parameters θ = (θ (1) 1,..., θ(1) n, θ (2) and where c(θ) is the normalizing constant. 12,..., θ(2) n 1n,..., θ(n) 12...n )T, This formulation allows calculation of cell probabilities (but is less useful for describing Pr(Y = y) as a function of x). Note that we have 2 n 1 parameters and we have two aims: reduce this number, and introduce a regression model. 339 Example: n = 2. We have Pr(Y 1 = y 1, Y 2 = y 2 ) = c(θ) exp θ (1) 1 y 1 + θ (1) 2 y 2 + θ (2) 12 y 1y 2, where θ = (θ (1) 1, θ(1) 2, θ(2) 12 )T and c(θ) 1 = 1X 1X y 1 =0 y 2 =0 exp θ (1) 1 y 1 + θ (1) 2 y 2 + θ (2) 12 y 1y 2 y 1 y 2 Pr(Y 1 = y 1, Y 2 = y 2 ) 0 0 c(θ) 1 0 c(θ) exp(θ (1) 1 ) 0 1 c(θ) exp(θ (1) 2 ) 1 1 c(θ) exp(θ (1) 1 + θ (1) 2 + θ (2) 12 ) 340
7 Hence we have (conditional) interpretations: exp(θ (1) 1 ) = Pr(Y 1 = 1, Y 2 = 0) Pr(Y 1 = 0, Y 2 = 0) = Pr(Y 1 = 1 Y 2 = 0) Pr(Y 1 = 0 Y 2 = 0) the odds of an event at trial 1, given no event at trial 2; exp(θ (1) 2 ) = Pr(Y 1 = 0, Y 2 = 1) Pr(Y 1 = 0, Y 2 = 0) = Pr(Y 2 = 1 Y 1 = 0) Pr(Y 2 = 0 Y 1 = 0) the odds of an event at trial 2, given an event at trial 1; exp(θ (12) 12 ) = Pr(Y 1 = 1, Y 2 = 1)Pr(Y 1 = 0, Y 2 = 0) Pr(Y 1 = 1, Y 2 = 0)Pr(Y 1 = 0, Y 2 = 1) = Pr(Y 2 = 1 Y 1 = 1)/ Pr(Y 2 = 0 Y 1 = 1) Pr(Y 2 = 1 Y 1 = 0)/ Pr(Y 2 = 0 Y 1 = 0) the ratio of odds of an event at trial 2 given an event at trial 1, divided by the odds of an event at trial 2 given no event at trial 1. Hence if this parameter is larger than 1 we have positive dependence. 341 Quadratic Exponential Log-Linear Model We describe three approaches to modeling binary data: conditional odds ratios, correlations, marginal odds ratios. Zhao and Prentice (1990) consider the log-linear model with third and higher-order terms set to zero, so that 0 1 Pr(Y = y) = c(θ) X j θ (1) j y j + X θ (2) jk y jy k A. j<k For this model Interpretation: Pr(Y j = 1 Y k = y k, Y l = 0, l j, k) Pr(Y j = 0 Y k = y k, Y l = 0, l j, k) = exp(θ(1) j + θ (2) jk y k). exp(θ (1) j ) is the odds of a success, given all other responses are zero. exp(θ (2) jk ) is the odds ratio describing the association between Y j and Y k, given all other responses are fixed and equal to zero. 342
8 Issues: 1. Suppose we now wish to model θ as a function of x. Example: Y respiratory infection, x mother s smoking (no/yes). Then we could let the parameters θ depend on x, i.e. θ(x). But the difference between θ (1) j (x = 1) and θ (1) j (x = 0) represent the effect of smoking on the conditional probability of respiratory infection at visit j, given that there was no infection at any other visits. Since smoking is likely to influence the response at all visits a part of the smoking effect will be conditioned away. Difficult to interpret, and we would rather model the marginal probability. 2. The interpretation of the θ parameters depends on the number of responses n particularly a problem in a longitudinal setting with different n i. 343 A Related Model Fitzmaurice and Laird (1993) parameterize in terms of µ j = Pr(Y j = 1) (the marginal mean) and the second and higher-order parameters, θ (2) 12,..., θ(2) n 1n,..., θ(n) 12...n. If we let θ 1 = (θ (1) 1,..., θ(1) n ) and θ 2 = (θ (12) 12, θ(2) 13,..., θ(2) n 1n,..., θ(n) 1...n ) and w = (y 1 y 2, y 1 y 3,..., y n 1n, y 123,..., y 1 y 2...y n ), then the loglinear model can be written as Pr(Y = y) = c(θ 1, θ 2 )exp(y T θ 1 + w T θ 2 ) The parameters of interest are µ = (µ 1,..., µ n ) = µ(θ 1, θ 2 ) so that we can reparameterize from (θ 1, θ 2 ) to (µ, θ 2 ). GEE used for inference (with a logistic mean model, for example). The limitiation of this parameterizatio is that θ 2 as a conditional odds depends on the number of other responses. So most useful for cases in which we have the same number of observations per cluster. If the association between responses is of interest then not so easy to interpret. 344
9 Bahadur Representation Another approach to parameterizing a multivariate binary model was proposed by Bahadur (1961) who used marginal means, as well as second-order moments specified in terms of correlations. Let Then we can write 0 R j = Y j µ j [µ j (1 µ j )] 1/2 ρ jk = corr(y j, Y k ) = E[R j R k ] ρ jkl = E[R j R k R l ] ρ 1,...,n = E[R 1...R n ] Pr(Y = y) + X ρ jk r j r k + X j<k j<k<l ny j=1 µ y j j (1 µ j) 1 y j 1 ρ jkl r j r k r l ρ 1,...,n r 1 r 2...r n A Appealing because we have the marginal means µ j and nuisance parameters. 345 Limitations: Unfortunately, the correlations are constrained in complicated ways by the marginal means. Example: consider measurements on a single individual, Y 1 and Y 2, with means µ 1 and µ 2. We have and corr(y 1, Y 2 ) = Pr(Y 1 = 1, Y 2 = 1) µ 1 µ 2 {µ 1 (1 µ 1 )µ 2 (1 µ 2 )} 1/2 max(0, µ 1 + µ 2 1) Pr(Y 1 = 1, Y 2 = 1) min(µ 1, µ 2 ), which implies complex constraints on the correlation. For example, if µ 1 = and µ 2 = then 0 corr(y 1, Y 2 )
10 Marginal Odds Ratios An alternative is to parameterize in terms of the marginal means and the marginal odds ratios defined by γ jk = Pr(Y j = 1, Y k = 1) Pr(Y j = 0, Y k = 0) Pr(Y j = 1, Y k = 0)Pr(Y j = 0, Y k = 1) = Pr(Y j = 1 Y k = 1)/Pr(Y j = 0 Y k = 1) Pr(Y j = 1 Y k = 0)/Pr(Y j = 0 Y k = 0) which is the odds that the j-th observation is a 1, given the k-th observation is a 1, divided by the odds that the j-th observation is a 1, given the k-th observation is a 0. Hence if γ jk > 1 we have positive dependence between outcomes j and k. It is then possible to obtain the joint distribution in terms of the means µ, where µ j = Pr(Y j = 1) the odds ratios γ = (γ 12,..., γ n 1,n ) and contrasts of odds ratios 347 We need to find E[Y j Y k ] = µ jk, so that we can write down the likelihood function, or an estimating function. For the case of n = 2 we have and so γ 12 = Pr(Y 1 = 1, Y 2 = 1) Pr(Y 1 = 0, Y 2 = 0) Pr(Y 1 = 1, Y 2 = 0)Pr(Y 1 = 0, Y 2 = 1) = µ 12(1 µ 1 µ 2 + µ 12 ) (µ 1 µ 12 )(µ 2 µ 12 ), µ 2 12(γ 12 1) + µ 12 b + γ 12 µ 1 µ 2 = 0, where b = (µ 1 + µ 2 )(1 γ 12 ) 1, to give µ 12 = b ± p b 2 4(γ 12 1)µ 1 µ 2. 2(γ 12 1) Y Y µ 1 1 µ 12 µ 1 1 µ 2 µ 2 348
11 Limitations In a longitudinal setting (we add an i subscript to denote individuals), finding the µ ijk terms is computationally complex (see Qaqish and Ivanova, 2006, Biometrika). Large numbers of nuisance odds ratios if n i s are large assumptions such as γ ijk = γ for all i, j, k may be made. Another possibility is to take log γ ijk = α 0 + α 1 t ij t ik 1, so that the degree of association is inversely proportional to the time between observations. 349 Modeling Multivariate Binary Data Using GEE For a marginal Bernoulli outcome we have Pr(Y ij = y ij x ij ) = µ y ij ij (1 µ ij) 1 y ij = exp(y ij θ ij log{1 + e θ ij }), where θ ij = log(µ ij /(1 µ ij ), an exponential family representation. For independent responses we therefore have the likelihood mx Xn i mx Xn i mx Xn i Pr(Y x) = exp 4 y ij θ ij log{1 + e θ ij } 5 = exp 4 i=1 j=1 i=1 j=1 To find the MLEs we consider the score equation: G(β) = l β = mx Xn i i=1 j=1 l ij θ ij θ ij β = mx Xn i x ij (y ij µ ij ) = i=1 j=1 i=1 j=1 l ij 3 5. mx x T i (y i µ i ). So GEE with working independence can be implemented with standard software, though we need to fix-up the standard errors via sandwich estimation. i=1 350
12 Non-independence GEE Assuming working correlation matrices: R i (α) and estimating equation G(β) = where W i = 1/2 i R i (α) 1/2 i. mx i=1 D T i W 1 i (y i µ i ), Here α are parameters that we need a consistent estimator of (Newey 1990, shows that the choice of estimator for α has no effect on the asymptotic efficiency). Define a set of n i (n i 1)/2 empirical covariances R ijk = (Y ij µ ij )(Y ik µ ik ) We can then define a set of moment-based estimating equations to obtain estimates of α. 351 First Extension to GEE Rather than have a method of moments estimator for α, Prentice (1988) proposed using a second set of estimating equations for α. In the context of data with var(y ij ) = v(µ ij ) recall: G 1 (β, α) = G 2 (β, α) = mx i=1 mx i=1 where R ij = {Y ij µ ij }, to give data D T i W 1 i (Y i µ i ) E T i H 1 i (T i Σ i ) T T i = (R i1r i2,..., R ini 1R ini, R 2 i1,..., R2 in i ), Σ i (α) = E[T i ] is a model for the correlations and variances of the standardized residuals, E i = Σ i α, and H i = cov(t i ) is the working covariance model. The vector T i has n i (n i 1)/2 + n i elements in general the working covariance model H i is in general complex. 352
13 In the context of binary data Prentice (1988) the variances are determined by the mean and so the last n i terms of T i are dropped; he also suggests taking a diagonal working covariance model, H i, i.e. ignoring the covariances. The theoetical variances ae given by var(r ij R ik ) = 1+(1 2p ij )(1 2p ik ){p ij (1 p ij )p ik (1 p ik )} 1/2 Σ(α) ijk Σ(α) 2 ijk which depend on the assumed correlation model Σ these may be taken as the diagonal elements of H i. 353 Application of GEE Extension to Marginal Odds Model We have the marginal mean model logit E[Y ij X ij ] = βx ij. Suppose we specify a model for the associations in terms of the marginal log odds ratios: j ff Pr(Yij = 1, Y ik = 1) Pr(Y ij = 0, Y ik = 0) α ijk = log. Pr(Y ij = 1, Y ik = 0) Pr(Y ij = 0, Y ik = 1) These are nuisance parameters, but how do we estimate them? Carey et al. (1992) suggest the following approach fo estimating β and α. Let µ ij = Pr(Y ij = 1) µ ik = Pr(Y ik = 1) µ ijk = Pr(Y ij = 1, Y ik = 1) 354
14 It is easy to show that Pr(Y ij = 1 Y ik = y ik ) Pr(Y ij = 0 Y ik = y ik ) = exp(y ik α ijk ) Pr(Y ij = 1, Y ik = 0) Pr(Y ij = 0, Y ik = 0) «µ ij µ ijk = exp(y ik α ijk ) 1 µ ij µ ik + µ ijk which can be written as a logistic regression model in terms of conditional probabilities: «Pr(Yij = 1 Y ik = y ik ) logit E[Y ij Y ik ] = log Pr(Y ij = 0 Y ik = y ik ) «µ ij µ ijk = y ik α ijk + log 1 µ ij µ ik + µ ijk where the term on the right is a known offset (the µ s are a function of β only). Suppose for simplicity that α ijk = α then given current estimates of β, α, we can fit a logistic regression model by regressing Y ij on Y ik for 1 j < k n i, to estimate α this can then be used within the working correlation model. Carey et al. (1992) refer to this method as alternating logistic regressions. 355 Indonesian Children s Health Example > summary(geese(y ~ xero+age,corstr="independence",id=id,family="binomial")) Mean Model: Mean Link: logit Variance to Mean Relation: binomial Coefficients: estimate san.se wald p (Intercept) e+00 age e-07 xero e-02 Scale Model: Scale Link: identity Estimated Scale Parameters: estimate san.se wald p (Intercept) Correlation Model: Correlation Structure: independence Number of clusters: 275 Maximum cluster size: 6 356
15 > summary(geese(y ~ age+xero, corstr="exchangeable", id=id,family="binomial")) estimate san.se wald p (Intercept) e+00 age e-06 xero e-01 Estimated Scale Parameters: estimate san.se wald p (Intercept) Estimated Correlation Parameters: estimate san.se wald p alpha > summary(geese(y ~ age+xero, corstr="ar1", id=id,family="binomial")) Coefficients: estimate san.se wald p (Intercept) e+00 age e-07 xero e-01 Estimated Scale Parameters: estimate san.se wald p (Intercept) Estimated Correlation Parameters: estimate san.se wald p alpha Conditional Likelihood: Binary Longitudinal Data Recall that conditional likelihood is a technique for eliminating nuisance parameters, here what we have previously modeled as random effects. Consider individual i with binary observations y i1,..., y ini and assume the random intercepts model Y ij γ i, β Bernoulli(p ij ), where «pij log = x ij β + γ i 1 p ij where γ i = x i β + b i and x ij (a slight change from our usual notation), are those covariates which change within an individual. 358
16 We have Pr(y i1,..., y ini γ i, β) = = = n i Y exp (γ i y ij + x ij βy ij ) 1 + exp (γ j=1 i + x ij β) P ni exp γ i j=1 y ij + P n i j=1 x ijy ij β Q ni j=1 [1 + exp (γ i + x ij β)] exp (γ i t 2i + t 1i β) Q ni j=1 [1 + exp (γ i + x ij β)] = exp (γ it 2i + t 1i β) k(γ i, β) = p(t 1i, t 2i γ i, β) where t 1i = k(γ i, β) = n i X X x ij y ij, t 2i = j=1 n i Y j=1 n i j=1 y ij [1 + exp (γ i + x ij β)]. 359 We have where L c (β) = p(t 2i γ i, β) = my p(t 1i t 2i, β) = i=1 P ni y i+ l=1 exp my i=1 p(t 1i, t 2i γ i, β) p(t 2i γ i, β) γ i P ni j=1 y ij + P n i k=1 x iky l ik β k(γ i, β), where the summation is over the ` n i y i+ ways of choosing yi+ ones out of n i, and y l i = (yl i1,..., yl in i ), l = 1,..., ` n i y i+ is the collection of these ways. Hence L c (β) = = my i=1 my i=1 P ni y i+ P ni exp γ i j=1 y ij + P n i j=1 x ijy ij β l=1 exp P ni y i+ P ni γ i j=1 y ij + P n i k=1 x ikyik l β exp Pni j=1 x ijy ij β l=1 exp `P n i k=1 x ikyik l β 360
17 Notes Can be computationally expensive to evaluate likelihood if n i is large, e.g. if n i = 20 and y i+ = 10, ` n i y i+ = 184,756. There is no contribution to the conditional likelihood from individuals: With n i = 1. With y i+ = 0 or y i+ = n i. For those covariates with x i1 =... = x ini = x i. The conditional likelihood estimates β s that are associated with within-individual covariates. If a covariate only varies between individuals, then it cannot be estimated using conditional likelihood. For covariates that vary both between and within individuals, only the within-individual contrasts are used. The similarity to Cox s partial likelihood may be exploited to carry out computation. We have not made a distributional assumption for the γ i s! 361 Examples: If n i = 3 and y i = (0,0, 1) so that y i+ = 1 then y 1 i = (1, 0,0), y2 i = (0, 1,0), y3 i = (0,0, 1), and the contribution to the conditional likelihood is exp(x i3 β) exp(x i1 β) + exp(x i2 β) + exp(x i3 β). If n i = 3 and y i = (1,0, 1) so that y i+ = 2 then y 1 i = (1, 1,0), y2 i = (1, 0,1), y3 i = (0,1, 1), and the contribution to the conditional likelihood is exp(x i1 β + x i3 β) exp(x i1 β + x i2 β) + exp(x i1 β + x i3 β) + exp(x i2 β + x i3 β). 362
,..., θ(2),..., θ(n)
Likelihoods for Multivariate Binary Data Log-Linear Model We have 2 n 1 distinct probabilities, but we wish to consider formulations that allow more parsimonious descriptions as a function of covariates.
More informationModeling the scale parameter ϕ A note on modeling correlation of binary responses Using marginal odds ratios to model association for binary responses
Outline Marginal model Examples of marginal model GEE1 Augmented GEE GEE1.5 GEE2 Modeling the scale parameter ϕ A note on modeling correlation of binary responses Using marginal odds ratios to model association
More informationGeneralized Linear Models. Last time: Background & motivation for moving beyond linear
Generalized Linear Models Last time: Background & motivation for moving beyond linear regression - non-normal/non-linear cases, binary, categorical data Today s class: 1. Examples of count and ordered
More informationReview. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 770: Categorical Data Analysis
Review Timothy Hanson Department of Statistics, University of South Carolina Stat 770: Categorical Data Analysis 1 / 22 Chapter 1: background Nominal, ordinal, interval data. Distributions: Poisson, binomial,
More informationPart IV: Marginal models
Part IV: Marginal models 246 BIO 245, Spring 2018 Indonesian Children s Health Study (ICHS) Study conducted to determine the effects of vitamin A deficiency in preschool children Data on K=275 children
More informationLecture 3.1 Basic Logistic LDA
y Lecture.1 Basic Logistic LDA 0.2.4.6.8 1 Outline Quick Refresher on Ordinary Logistic Regression and Stata Women s employment example Cross-Over Trial LDA Example -100-50 0 50 100 -- Longitudinal Data
More informationGeneralized linear models for binary data. A better graphical exploratory data analysis. The simple linear logistic regression model
Stat 3302 (Spring 2017) Peter F. Craigmile Simple linear logistic regression (part 1) [Dobson and Barnett, 2008, Sections 7.1 7.3] Generalized linear models for binary data Beetles dose-response example
More informationNATIONAL UNIVERSITY OF SINGAPORE EXAMINATION (SOLUTIONS) ST3241 Categorical Data Analysis. (Semester II: )
NATIONAL UNIVERSITY OF SINGAPORE EXAMINATION (SOLUTIONS) Categorical Data Analysis (Semester II: 2010 2011) April/May, 2011 Time Allowed : 2 Hours Matriculation No: Seat No: Grade Table Question 1 2 3
More informationLOGISTIC REGRESSION Joseph M. Hilbe
LOGISTIC REGRESSION Joseph M. Hilbe Arizona State University Logistic regression is the most common method used to model binary response data. When the response is binary, it typically takes the form of
More information36-720: The Rasch Model
36-720: The Rasch Model Brian Junker October 15, 2007 Multivariate Binary Response Data Rasch Model Rasch Marginal Likelihood as a GLMM Rasch Marginal Likelihood as a Log-Linear Model Example For more
More informationHigh-Throughput Sequencing Course
High-Throughput Sequencing Course DESeq Model for RNA-Seq Biostatistics and Bioinformatics Summer 2017 Outline Review: Standard linear regression model (e.g., to model gene expression as function of an
More informationSTAT 526 Advanced Statistical Methodology
STAT 526 Advanced Statistical Methodology Fall 2017 Lecture Note 10 Analyzing Clustered/Repeated Categorical Data 0-0 Outline Clustered/Repeated Categorical Data Generalized Linear Mixed Models Generalized
More informationLinear Mixed Models. One-way layout REML. Likelihood. Another perspective. Relationship to classical ideas. Drawbacks.
Linear Mixed Models One-way layout Y = Xβ + Zb + ɛ where X and Z are specified design matrices, β is a vector of fixed effect coefficients, b and ɛ are random, mean zero, Gaussian if needed. Usually think
More informationRepeated ordinal measurements: a generalised estimating equation approach
Repeated ordinal measurements: a generalised estimating equation approach David Clayton MRC Biostatistics Unit 5, Shaftesbury Road Cambridge CB2 2BW April 7, 1992 Abstract Cumulative logit and related
More informationBayesian Multivariate Logistic Regression
Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of
More informationModule 6 Case Studies in Longitudinal Data Analysis
Module 6 Case Studies in Longitudinal Data Analysis Benjamin French, PhD Radiation Effects Research Foundation SISCR 2018 July 24, 2018 Learning objectives This module will focus on the design of longitudinal
More informationLinear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52
Statistics for Applications Chapter 10: Generalized Linear Models (GLMs) 1/52 Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52 Components of a linear model The two
More informationSample size calculations for logistic and Poisson regression models
Biometrika (2), 88, 4, pp. 93 99 2 Biometrika Trust Printed in Great Britain Sample size calculations for logistic and Poisson regression models BY GWOWEN SHIEH Department of Management Science, National
More informationA Generalized Linear Model for Binomial Response Data. Copyright c 2017 Dan Nettleton (Iowa State University) Statistics / 46
A Generalized Linear Model for Binomial Response Data Copyright c 2017 Dan Nettleton (Iowa State University) Statistics 510 1 / 46 Now suppose that instead of a Bernoulli response, we have a binomial response
More informationContrasting Marginal and Mixed Effects Models Recall: two approaches to handling dependence in Generalized Linear Models:
Contrasting Marginal and Mixed Effects Models Recall: two approaches to handling dependence in Generalized Linear Models: Marginal models: based on the consequences of dependence on estimating model parameters.
More informationLinear Regression Models P8111
Linear Regression Models P8111 Lecture 25 Jeff Goldsmith April 26, 2016 1 of 37 Today s Lecture Logistic regression / GLMs Model framework Interpretation Estimation 2 of 37 Linear regression Course started
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science
UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator
More informationGeneralized linear models
Generalized linear models Douglas Bates November 01, 2010 Contents 1 Definition 1 2 Links 2 3 Estimating parameters 5 4 Example 6 5 Model building 8 6 Conclusions 8 7 Summary 9 1 Generalized Linear Models
More informationSemiparametric Generalized Linear Models
Semiparametric Generalized Linear Models North American Stata Users Group Meeting Chicago, Illinois Paul Rathouz Department of Health Studies University of Chicago prathouz@uchicago.edu Liping Gao MS Student
More informationST3241 Categorical Data Analysis I Generalized Linear Models. Introduction and Some Examples
ST3241 Categorical Data Analysis I Generalized Linear Models Introduction and Some Examples 1 Introduction We have discussed methods for analyzing associations in two-way and three-way tables. Now we will
More informationStat 5102 Final Exam May 14, 2015
Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions
More informationMultinomial Regression Models
Multinomial Regression Models Objectives: Multinomial distribution and likelihood Ordinal data: Cumulative link models (POM). Ordinal data: Continuation models (CRM). 84 Heagerty, Bio/Stat 571 Models for
More informationGeneralized Estimating Equations (gee) for glm type data
Generalized Estimating Equations (gee) for glm type data Søren Højsgaard mailto:sorenh@agrsci.dk Biometry Research Unit Danish Institute of Agricultural Sciences January 23, 2006 Printed: January 23, 2006
More informationClassification. Chapter Introduction. 6.2 The Bayes classifier
Chapter 6 Classification 6.1 Introduction Often encountered in applications is the situation where the response variable Y takes values in a finite set of labels. For example, the response Y could encode
More informationA measure of partial association for generalized estimating equations
A measure of partial association for generalized estimating equations Sundar Natarajan, 1 Stuart Lipsitz, 2 Michael Parzen 3 and Stephen Lipshultz 4 1 Department of Medicine, New York University School
More informationModeling Longitudinal Count Data with Excess Zeros and Time-Dependent Covariates: Application to Drug Use
Modeling Longitudinal Count Data with Excess Zeros and : Application to Drug Use University of Northern Colorado November 17, 2014 Presentation Outline I and Data Issues II Correlated Count Regression
More informationGeneralized linear models
Generalized linear models Søren Højsgaard Department of Mathematical Sciences Aalborg University, Denmark October 29, 202 Contents Densities for generalized linear models. Mean and variance...............................
More informationPart 8: GLMs and Hierarchical LMs and GLMs
Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course
More informationLogistic Regression. James H. Steiger. Department of Psychology and Human Development Vanderbilt University
Logistic Regression James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) Logistic Regression 1 / 38 Logistic Regression 1 Introduction
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationCharles E. McCulloch Biometrics Unit and Statistics Center Cornell University
A SURVEY OF VARIANCE COMPONENTS ESTIMATION FROM BINARY DATA by Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University BU-1211-M May 1993 ABSTRACT The basic problem of variance components
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More informationSTA216: Generalized Linear Models. Lecture 1. Review and Introduction
STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general
More informationStat 579: Generalized Linear Models and Extensions
Stat 579: Generalized Linear Models and Extensions Linear Mixed Models for Longitudinal Data Yan Lu April, 2018, week 15 1 / 38 Data structure t1 t2 tn i 1st subject y 11 y 12 y 1n1 Experimental 2nd subject
More informationPrevious lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing.
Previous lecture P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Interaction Outline: Definition of interaction Additive versus multiplicative
More informationBiostatistics Advanced Methods in Biostatistics IV
Biostatistics 140.754 Advanced Methods in Biostatistics IV Jeffrey Leek Assistant Professor Department of Biostatistics jleek@jhsph.edu 1 / 35 Tip + Paper Tip Meet with seminar speakers. When you go on
More informationNon-maximum likelihood estimation and statistical inference for linear and nonlinear mixed models
Optimum Design for Mixed Effects Non-Linear and generalized Linear Models Cambridge, August 9-12, 2011 Non-maximum likelihood estimation and statistical inference for linear and nonlinear mixed models
More informationCovariance Models (*) X i : (n i p) design matrix for fixed effects β : (p 1) regression coefficient for fixed effects
Covariance Models (*) Mixed Models Laird & Ware (1982) Y i = X i β + Z i b i + e i Y i : (n i 1) response vector X i : (n i p) design matrix for fixed effects β : (p 1) regression coefficient for fixed
More informationUsing Estimating Equations for Spatially Correlated A
Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship
More informationGEE for Longitudinal Data - Chapter 8
GEE for Longitudinal Data - Chapter 8 GEE: generalized estimating equations (Liang & Zeger, 1986; Zeger & Liang, 1986) extension of GLM to longitudinal data analysis using quasi-likelihood estimation method
More informationMixed models in R using the lme4 package Part 5: Generalized linear mixed models
Mixed models in R using the lme4 package Part 5: Generalized linear mixed models Douglas Bates Madison January 11, 2011 Contents 1 Definition 1 2 Links 2 3 Example 7 4 Model building 9 5 Conclusions 14
More informationChapter 22: Log-linear regression for Poisson counts
Chapter 22: Log-linear regression for Poisson counts Exposure to ionizing radiation is recognized as a cancer risk. In the United States, EPA sets guidelines specifying upper limits on the amount of exposure
More information1/15. Over or under dispersion Problem
1/15 Over or under dispersion Problem 2/15 Example 1: dogs and owners data set In the dogs and owners example, we had some concerns about the dependence among the measurements from each individual. Let
More informationLongitudinal Modeling with Logistic Regression
Newsom 1 Longitudinal Modeling with Logistic Regression Longitudinal designs involve repeated measurements of the same individuals over time There are two general classes of analyses that correspond to
More informationSample Size and Power Considerations for Longitudinal Studies
Sample Size and Power Considerations for Longitudinal Studies Outline Quantities required to determine the sample size in longitudinal studies Review of type I error, type II error, and power For continuous
More informationSTA 216: GENERALIZED LINEAR MODELS. Lecture 1. Review and Introduction. Much of statistics is based on the assumption that random
STA 216: GENERALIZED LINEAR MODELS Lecture 1. Review and Introduction Much of statistics is based on the assumption that random variables are continuous & normally distributed. Normal linear regression
More informationLatent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent
Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary
More informationVarious Issues in Fitting Contingency Tables
Various Issues in Fitting Contingency Tables Statistics 149 Spring 2006 Copyright 2006 by Mark E. Irwin Complete Tables with Zero Entries In contingency tables, it is possible to have zero entries in a
More informationˆπ(x) = exp(ˆα + ˆβ T x) 1 + exp(ˆα + ˆβ T.
Exam 3 Review Suppose that X i = x =(x 1,, x k ) T is observed and that Y i X i = x i independent Binomial(n i,π(x i )) for i =1,, N where ˆπ(x) = exp(ˆα + ˆβ T x) 1 + exp(ˆα + ˆβ T x) This is called the
More informationResearch Projects. Hanxiang Peng. March 4, Department of Mathematical Sciences Indiana University-Purdue University at Indianapolis
Hanxiang Department of Mathematical Sciences Indiana University-Purdue University at Indianapolis March 4, 2009 Outline Project I: Free Knot Spline Cox Model Project I: Free Knot Spline Cox Model Consider
More informationGeneralized Linear Models. Kurt Hornik
Generalized Linear Models Kurt Hornik Motivation Assuming normality, the linear model y = Xβ + e has y = β + ε, ε N(0, σ 2 ) such that y N(μ, σ 2 ), E(y ) = μ = β. Various generalizations, including general
More information1. Hypothesis testing through analysis of deviance. 3. Model & variable selection - stepwise aproaches
Sta 216, Lecture 4 Last Time: Logistic regression example, existence/uniqueness of MLEs Today s Class: 1. Hypothesis testing through analysis of deviance 2. Standard errors & confidence intervals 3. Model
More informationExam Applied Statistical Regression. Good Luck!
Dr. M. Dettling Summer 2011 Exam Applied Statistical Regression Approved: Tables: Note: Any written material, calculator (without communication facility). Attached. All tests have to be done at the 5%-level.
More informationStatistics: A review. Why statistics?
Statistics: A review Why statistics? What statistical concepts should we know? Why statistics? To summarize, to explore, to look for relations, to predict What kinds of data exist? Nominal, Ordinal, Interval
More informationPoisson regression: Further topics
Poisson regression: Further topics April 21 Overdispersion One of the defining characteristics of Poisson regression is its lack of a scale parameter: E(Y ) = Var(Y ), and no parameter is available to
More informationMixed models in R using the lme4 package Part 5: Generalized linear mixed models
Mixed models in R using the lme4 package Part 5: Generalized linear mixed models Douglas Bates 2011-03-16 Contents 1 Generalized Linear Mixed Models Generalized Linear Mixed Models When using linear mixed
More informationTopic 12 Overview of Estimation
Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the
More informationLecture 5: LDA and Logistic Regression
Lecture 5: and Logistic Regression Hao Helen Zhang Hao Helen Zhang Lecture 5: and Logistic Regression 1 / 39 Outline Linear Classification Methods Two Popular Linear Models for Classification Linear Discriminant
More informationApplied Economics. Regression with a Binary Dependent Variable. Department of Economics Universidad Carlos III de Madrid
Applied Economics Regression with a Binary Dependent Variable Department of Economics Universidad Carlos III de Madrid See Stock and Watson (chapter 11) 1 / 28 Binary Dependent Variables: What is Different?
More informationRobust covariance estimator for small-sample adjustment in the generalized estimating equations: A simulation study
Science Journal of Applied Mathematics and Statistics 2014; 2(1): 20-25 Published online February 20, 2014 (http://www.sciencepublishinggroup.com/j/sjams) doi: 10.11648/j.sjams.20140201.13 Robust covariance
More informationFREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE
FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE Donald A. Pierce Oregon State Univ (Emeritus), RERF Hiroshima (Retired), Oregon Health Sciences Univ (Adjunct) Ruggero Bellio Univ of Udine For Perugia
More informationOverdispersion Workshop in generalized linear models Uppsala, June 11-12, Outline. Overdispersion
Biostokastikum Overdispersion is not uncommon in practice. In fact, some would maintain that overdispersion is the norm in practice and nominal dispersion the exception McCullagh and Nelder (1989) Overdispersion
More informationIntroduction to mtm: An R Package for Marginalized Transition Models
Introduction to mtm: An R Package for Marginalized Transition Models Bryan A. Comstock and Patrick J. Heagerty Department of Biostatistics University of Washington 1 Introduction Marginalized transition
More informationBayesian SAE using Complex Survey Data Lecture 4A: Hierarchical Spatial Bayes Modeling
Bayesian SAE using Complex Survey Data Lecture 4A: Hierarchical Spatial Bayes Modeling Jon Wakefield Departments of Statistics and Biostatistics University of Washington 1 / 37 Lecture Content Motivation
More informationSCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models
SCHOOL OF MATHEMATICS AND STATISTICS Linear and Generalised Linear Models Autumn Semester 2017 18 2 hours Attempt all the questions. The allocation of marks is shown in brackets. RESTRICTED OPEN BOOK EXAMINATION
More informationLogistic Regressions. Stat 430
Logistic Regressions Stat 430 Final Project Final Project is, again, team based You will decide on a project - only constraint is: you are supposed to use techniques for a solution that are related to
More informationSTAT 526 Spring Final Exam. Thursday May 5, 2011
STAT 526 Spring 2011 Final Exam Thursday May 5, 2011 Time: 2 hours Name (please print): Show all your work and calculations. Partial credit will be given for work that is partially correct. Points will
More informationGeneralized Linear Models 1
Generalized Linear Models 1 STA 2101/442: Fall 2012 1 See last slide for copyright information. 1 / 24 Suggested Reading: Davison s Statistical models Exponential families of distributions Sec. 5.2 Chapter
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationPoisson regression 1/15
Poisson regression 1/15 2/15 Counts data Examples of counts data: Number of hospitalizations over a period of time Number of passengers in a bus station Blood cells number in a blood sample Number of typos
More informationSTA6938-Logistic Regression Model
Dr. Ying Zhang STA6938-Logistic Regression Model Topic 2-Multiple Logistic Regression Model Outlines:. Model Fitting 2. Statistical Inference for Multiple Logistic Regression Model 3. Interpretation of
More information9 Generalized Linear Models
9 Generalized Linear Models The Generalized Linear Model (GLM) is a model which has been built to include a wide range of different models you already know, e.g. ANOVA and multiple linear regression models
More informationSimulating Longer Vectors of Correlated Binary Random Variables via Multinomial Sampling
Simulating Longer Vectors of Correlated Binary Random Variables via Multinomial Sampling J. Shults a a Department of Biostatistics, University of Pennsylvania, PA 19104, USA (v4.0 released January 2015)
More informationLinear Regression. Data Model. β, σ 2. Process Model. ,V β. ,s 2. s 1. Parameter Model
Regression: Part II Linear Regression y~n X, 2 X Y Data Model β, σ 2 Process Model Β 0,V β s 1,s 2 Parameter Model Assumptions of Linear Model Homoskedasticity No error in X variables Error in Y variables
More informationGeneralized Quasi-likelihood versus Hierarchical Likelihood Inferences in Generalized Linear Mixed Models for Count Data
Sankhyā : The Indian Journal of Statistics 2009, Volume 71-B, Part 1, pp. 55-78 c 2009, Indian Statistical Institute Generalized Quasi-likelihood versus Hierarchical Likelihood Inferences in Generalized
More informationSingle-level Models for Binary Responses
Single-level Models for Binary Responses Distribution of Binary Data y i response for individual i (i = 1,..., n), coded 0 or 1 Denote by r the number in the sample with y = 1 Mean and variance E(y) =
More informationNATIONAL UNIVERSITY OF SINGAPORE EXAMINATION. ST3241 Categorical Data Analysis. (Semester II: ) April/May, 2011 Time Allowed : 2 Hours
NATIONAL UNIVERSITY OF SINGAPORE EXAMINATION Categorical Data Analysis (Semester II: 2010 2011) April/May, 2011 Time Allowed : 2 Hours Matriculation No: Seat No: Grade Table Question 1 2 3 4 5 6 Full marks
More informationModels for Clustered Data
Models for Clustered Data Edps/Psych/Soc 589 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2019 Outline Notation NELS88 data Fixed Effects ANOVA
More informationEfficiency of generalized estimating equations for binary responses
J. R. Statist. Soc. B (2004) 66, Part 4, pp. 851 860 Efficiency of generalized estimating equations for binary responses N. Rao Chaganty Old Dominion University, Norfolk, USA and Harry Joe University of
More informationStatistical Methods III Statistics 212. Problem Set 2 - Answer Key
Statistical Methods III Statistics 212 Problem Set 2 - Answer Key 1. (Analysis to be turned in and discussed on Tuesday, April 24th) The data for this problem are taken from long-term followup of 1423
More informationModels for Clustered Data
Models for Clustered Data Edps/Psych/Stat 587 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline Notation NELS88 data Fixed Effects ANOVA
More informationBinary Response: Logistic Regression. STAT 526 Professor Olga Vitek
Binary Response: Logistic Regression STAT 526 Professor Olga Vitek March 29, 2011 4 Model Specification and Interpretation 4-1 Probability Distribution of a Binary Outcome Y In many situations, the response
More informationI r j Binom(m j, p j ) I L(, ; y) / exp{ y j + (x j y j ) m j log(1 + e + x j. I (, y) / L(, ; y) (, )
Today I Bayesian analysis of logistic regression I Generalized linear mixed models I CD on fixed and random effects I HW 2 due February 28 I Case Studies SSC 2014 Toronto I March/April: Semi-parametric
More informationGeneralized Quasi-likelihood (GQL) Inference* by Brajendra C. Sutradhar Memorial University address:
Generalized Quasi-likelihood (GQL) Inference* by Brajendra C. Sutradhar Memorial University Email address: bsutradh@mun.ca QL Estimation for Independent Data. For i = 1,...,K, let Y i denote the response
More informationLogistic Regression - problem 6.14
Logistic Regression - problem 6.14 Let x 1, x 2,, x m be given values of an input variable x and let Y 1,, Y m be independent binomial random variables whose distributions depend on the corresponding values
More informationLecture 12: Effect modification, and confounding in logistic regression
Lecture 12: Effect modification, and confounding in logistic regression Ani Manichaikul amanicha@jhsph.edu 4 May 2007 Today Categorical predictor create dummy variables just like for linear regression
More informationConditional Inference Functions for Mixed-Effects Models with Unspecified Random-Effects Distribution
Conditional Inference Functions for Mixed-Effects Models with Unspecified Random-Effects Distribution Peng WANG, Guei-feng TSAI and Annie QU 1 Abstract In longitudinal studies, mixed-effects models are
More informationConfidence Intervals for the Odds Ratio in Logistic Regression with Two Binary X s
Chapter 866 Confidence Intervals for the Odds Ratio in Logistic Regression with Two Binary X s Introduction Logistic regression expresses the relationship between a binary response variable and one or
More informationTime-Dependent Covariates. Outline Introduction Data Example: the MSCM Study Stochastic Covariates Lagged Covariates Time-Dependent Confounders
Outline Introduction Data Example: the MSCM Study Stochastic Covariates Lagged Covariates Time-Dependent Confounders 1 Introduction One advantage of a longitudinal study: to observe the temporal order
More information22s:152 Applied Linear Regression. Example: Study on lead levels in children. Ch. 14 (sec. 1) and Ch. 15 (sec. 1 & 4): Logistic Regression
22s:52 Applied Linear Regression Ch. 4 (sec. and Ch. 5 (sec. & 4: Logistic Regression Logistic Regression When the response variable is a binary variable, such as 0 or live or die fail or succeed then
More informationCovariance modelling for longitudinal randomised controlled trials
Covariance modelling for longitudinal randomised controlled trials G. MacKenzie 1,2 1 Centre of Biostatistics, University of Limerick, Ireland. www.staff.ul.ie/mackenzieg 2 CREST, ENSAI, Rennes, France.
More informationAdministration. Homework 1 on web page, due Feb 11 NSERC summer undergraduate award applications due Feb 5 Some helpful books
STA 44/04 Jan 6, 00 / 5 Administration Homework on web page, due Feb NSERC summer undergraduate award applications due Feb 5 Some helpful books STA 44/04 Jan 6, 00... administration / 5 STA 44/04 Jan 6,
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationConfidence Intervals for the Odds Ratio in Logistic Regression with One Binary X
Chapter 864 Confidence Intervals for the Odds Ratio in Logistic Regression with One Binary X Introduction Logistic regression expresses the relationship between a binary response variable and one or more
More information