Methods for Handling Missing Data

Size: px
Start display at page:

Download "Methods for Handling Missing Data"

Transcription

1 Methods for Handling Missing Data Joseph Hogan Brown University MDEpiNet Conference Workshop October 22, 2018 Hogan (MDEpiNet) Missing Data October 22, / 160

2 Course Overview I 1 Introduction and Background Introduce case studies Missing data mechanisms Review and critique of commonly-used methods 2 Case Study 1: Growth Hormone Study Analysis using mixture models Setting up sensitivity analysis Inference about treatment effects Hogan (MDEpiNet) Missing Data October 22, / 160

3 Course Overview II 3 Case Study 2: Smoking cessation study Exploratory analysis for long sequence of binary data Analysis via IPW methods under MAR Comparative analysis via GEE, ML, LOCF Hogan (MDEpiNet) Missing Data October 22, / 160

4 INTRODUCTION AND BACKGROUND Hogan (MDEpiNet) Missing Data October 22, / 160

5 Study 1: Growth Hormone Study NIH-funded trial to study effect of rhgh for increasing muscle strength in elderly About 240 patients randomized to one of 4 arms: Placebo rhgh Exercise + Placebo (EP) Exercise + rhgh (EG) Primary outcome Quadriceps strength, in ft-lbs of torque Measured at baseline, 6 months, 12 months Our analysis Mean quad strength at 12 months Compare EP and EG arms only, for illustration Hogan (MDEpiNet) Missing Data October 22, / 160

6 Summary Statistics: Growth Hormone Study Month Treatment k n k EP (32) (52) 86 (51) (24) 81 (25) 73 (21) All (26) 82 (26) 73 (21) EG (26) (15) 68 (26) (24) 90 (32) 88 (32) All (25) 87 (32) 88 (32) Hogan (MDEpiNet) Missing Data October 22, / 160

7 Questions to be addressed What is the mean quad strength at 12 months, among all individuals who initiated therapy? What is the treatment effect at 12 months, among all individuals who initiated therapy? These are questions about the full data, or the data we intended to observed but did not. The main difficulty is that a significant proportion of the data are missing. Hogan (MDEpiNet) Missing Data October 22, / 160

8 Study 2: Smoking Cessation Study NIH-funded study to reduce smoking among sedentary women Roughly 300 individuals randomized to two arms: Supervised exercise vs. Wellness education program Primary outcome Weekly smoking status over 12 weeks Treatment comparison Smoking rate at week 12 following baseline Analysis issues Binary outcomes Mean has some structure as a function of time Large number of repeated measures Hogan (MDEpiNet) Missing Data October 22, / 160

9 Smoking Cessation Study: Summaries Hogan (MDEpiNet) Missing Data October 22, / 160

10 Smoking Cessation Study: Summaries Hogan (MDEpiNet) Missing Data October 22, / 160

11 Basics of inference with incomplete data Formulate the precise question you want to answer Define the quantity you want to estimate Ascertain what information is available in the data And... what information is unavailable Apply a statistical method to estimate quantity of interest Apply statistical principles to quantify uncertainty Sampling variability Uncertainty due to missing data or untestable assumptions Hogan (MDEpiNet) Missing Data October 22, / 160

12 Course objectives Develop an understanding of Mechanisms that lead to missing data Biases missing data may cause Methods of addressing missing data Use examples to illustrate the methods, and provide understanding of how they work Hogan (MDEpiNet) Missing Data October 22, / 160

13 A word about the data examples They are relatively simple in nature (stylized) They are designed to promote understanding of the methods The idea is that when you apply the methods on more complex problems, you will have a feel for how and why they work (and how and why they don t) In real life, data analysis problems can be much harder than the ones we are using You will have to do further research to implement these methods on complex datasets Hogan (MDEpiNet) Missing Data October 22, / 160

14 Some reasons for missing data Refusal to respond Drop out of a study (patient decision) Removal from a study (researcher / doctor decision) Death Administrative reasons (funding, etc) Hogan (MDEpiNet) Missing Data October 22, / 160

15 Defining the estimation target First, some notation needed Y = outcome variable (e.g., CD4 count) R = response indicator { 1 if Y observed = 0 if Y missing X = covariates of direct interest V = auxiliary covariates (available, not of direct interest) Hogan (MDEpiNet) Missing Data October 22, / 160

16 Defining the estimation target with incomplete data Possible targets of estimation Full-data parameter: Mean outcome among all individuals intended to be in the sample, whether or not they are observed µ = E(Y ) Observed-data parameter: Mean response among all individuals whose outcome was observed µ 1 = E(Y R = 1) Hogan (MDEpiNet) Missing Data October 22, / 160

17 Defining the estimation target with incomplete data Full data parameter: Regression parameters among all individuals intended to be in the sample E(Y X ) = X T β Observed-data parameter: Regression parameters among individuals with observed outcome E(Y X, R = 1) = X T β 1 Important questions to ask: When does µ = µ 1? When does β = β 1? Hogan (MDEpiNet) Missing Data October 22, / 160

18 Illustration using univariate mean Consider univariate sample Data: n units targeted m units respond (m < n) Y 1,..., Y m observed; Y m+1,..., Y n missing R 1 = R 2 = = R m = 1 and R m+1 = = R n = 0 X 1,..., X n is a baseline covariate, observed on everyone Target of inference: µ = E(Y ) (full-data parameter) Hogan (MDEpiNet) Missing Data October 22, / 160

19 Data excerpt from Growth Hormone trial V Y R [1,] 35.0 NA 0 [2,] [3,] [4,] [5,] 68.6 NA 0 [6,] [7,] 39.0 NA 0 [8,] 52.0 NA 0 [9,] [10,] [11,] V = baseline quad strength Y = quad strength at one year Hogan (MDEpiNet) Missing Data October 22, / 160

20 What can be estimated? What we can estimate: µ 1 = E(Y R = 1) What we cannot estimate (without making assumptions) µ = E(Y ) Main reason we cannot estimate these is because µ = E(Y ) = E(Y R = 1)P(R = 1) + E(Y R = 0)P(R = 0) Hogan (MDEpiNet) Missing Data October 22, / 160

21 The need for assumptions to estimate full-data parameters Cannot estimate parameters for parts of the data that are missing Hence need assumptions about the missing data These are called missing data mechanisms Under most circumstances, these assumptions cannot be tested This motivates the need to: State the assumptions unambiguously so others can critique them Carry out sensitivity analysis wherever possible Hogan (MDEpiNet) Missing Data October 22, / 160

22 Missing data mechanisms Classification of association between R and Y MCAR Missing completely at random MAR Missing at random MNAR Missing not at random These are sometimes defined conditionally on covariates X or, in the case of repeated measures, on the data history up to a specific time point. More later. Hogan (MDEpiNet) Missing Data October 22, / 160

23 Missing data mechanisms Joint distribution of two random variables MDM for univariate samples MDM for multivariate sample, where interest is in regression Have model covariates only Have model covariates and auxiliary information Hogan (MDEpiNet) Missing Data October 22, / 160

24 Statistical independence The notation X Y means that the random variable X is independent of the random variable Y. Implications of independence Joint distribution can be factored f (x, y) = f (x)f (y) Conditional distributions and expectations f (x y) = f (x) E(X Y ) = E(X ) i.e., knowing Y does not influence the distribution or expectation of X Hogan (MDEpiNet) Missing Data October 22, / 160

25 Joint distribution of two random variables To characterize inference from incomplete data, always sampling at least two variables, Y and R Usually we are interested in some aspect of f (y), such as the mean or median. Denote this as θ. For example θ = E(Y ) = y f (y) dy But to carry out inference, we need to make assumptions about the joint distribution of Y and R, denoted by f (y, r) Hogan (MDEpiNet) Missing Data October 22, / 160

26 Joint distribution of two random variables The joint distribution can be decomposed into conditional distributions as follows Mixture factorization f (y, r) = f (r) f (y r) Selection factorization f (y, r) = f (y) f (r y) We will focus on the selection factorization for now. Hogan (MDEpiNet) Missing Data October 22, / 160

27 Selection factorization The selection factorization describes the joint distribution in terms of The distribution or model for the variable of interest, f (y) The distribution of the response indicators and their dependence on y, written as f (r y). Missing data mechanism Selection mechanism This allows us to characterize different types of missing data mechanisms formally. Hogan (MDEpiNet) Missing Data October 22, / 160

28 MDM for univariate sampling Missing values of Y are missing completely at random (MCAR) if R Y, or equivalently if f (r y) = f (r) For univariate samples, this is also classified as missing at random (MAR). More on this distinction later. Hogan (MDEpiNet) Missing Data October 22, / 160

29 MDM for univariate sampling Missing values of Y are missing not at random (MNAR) if there exists at least one value of y such that f (r y) f (r) Or in words, if the probability of response is systematically higher/lower for particular values of y. Hogan (MDEpiNet) Missing Data October 22, / 160

30 MDM for univariate sampling Under MAR, methods applied to the observed data only will generally yield valid inferences about the population. Estimates will be consistent Standard errors may be larger than if you had the full data Under MNAR, methods applied to observed data only generally will not yield valid inferences Hogan (MDEpiNet) Missing Data October 22, / 160

31 MDM for multivariate sampling regression Consider the setting where we are interested in the regression of Y on X. Let µ(x ) = E(Y X ). Assume there are no other covariates available The model we have in mind is g{µ(x )} = β 0 + β 1 X Here, the function g tells you the type of regression model you are fitting (linear, logistic, etc.) The full data are (Y 1, X 1, R 1 ), (Y 2, X 2, R 2 ),..., (Y n, X n, R n ) Hogan (MDEpiNet) Missing Data October 22, / 160

32 MDM for multivariate sampling regression In cases like this, we can define missing data mechanisms relative to the objective of inference. The Y s are missing at random if Y R X In words, the MDM is a random deletion mechanism within distinct levels of X Another way to write this: f (r y, x) = f (r x) The deletion mechanism depends on X, but within levels of X it does not depend on Y Hogan (MDEpiNet) Missing Data October 22, / 160

33 Examples of MAR in regression Let Y denote blood pressure, X denote gender (1 = F, 0 = M). The regression model of interest is E(Y X ) = β 0 + β 1 X, so that β 0 = mean BP among men, β 1 = mean difference. Let s assume men have higher BP on average. Randomly delete BP for 20% of men and 40% of women. R does depend on Y, but only through X Men have higher BP Men less likely to be deleted those with higher BP less likely to be deleted Within levels of X, deletion mechanism is completely random. Hogan (MDEpiNet) Missing Data October 22, / 160

34 MAR in regression some practical issues Revisit the MAR condition. If R Y X, this also means f (y x, r) = f (y x), or f (y x, r = 1) = f (y x, r = 0) The relationship between Y and X is the same whether R = 1 or R = 0. Consequence is that a regression fit to those with R = 1 gives valid estimates of regression parameters. (Standard errors will be higher relative to having all the data.) Hogan (MDEpiNet) Missing Data October 22, / 160

35 MAR in regression some practical issues Under MAR, the inferences are still valid even if The X distribution is different between those with missing and observed Y s Question you have to ask to (subjectively) assess MAR: Is the missing data mechanism a random deletion of Y s among people who have the same X values? Equivalent formulation of this question: Is the relationship between X and Y the same among those with missing and observed Y values? Hogan (MDEpiNet) Missing Data October 22, / 160

36 MAR for regression when auxiliary variables are available In some cases we have information on more than just the X variables In a clinical trial we may be interested in E(Y X ) when X is a treatment group, but we have collected lots of baseline covariates V. In a longitudinal study, we may be interested in the mean outcome at the last measurement time, but we have accumulated information on the outcome at previous measurement times. When auxiliary information is available, we can use it in some cases to make MAR more plausible. Here MAR has a slightly different formulation. Hogan (MDEpiNet) Missing Data October 22, / 160

37 MAR with auxiliary covariates The relationship of interest is g{µ(x )} = β 0 + β 1 X The full data are (Y 1, X 1, V 1, R 1 ), (Y 2, X 2, V 2, R 2 ),..., (Y n, X n, V n, R n ) where Y is observed when R = 1 and is missing when R = 0. Hogan (MDEpiNet) Missing Data October 22, / 160

38 MAR with auxiliary covariates Values of Y are missing at random (MAR) if Y R (X, V ) Two equivalent ways to write this are: f (r x, v, y) = f (r x, v) f (y x, v, r = 1) = f (y x, v, r = 0) The first says that within distinct levels defined by (X, V ), missingness in Y is a random deletion mechanism The second says that the relationship between Y and (X, V ) is the same whether Y is missing or not Hogan (MDEpiNet) Missing Data October 22, / 160

39 MAR with auxiliaries example Return to our BP example, but now assume V denotes income level. Recall, we are interested in the coefficient β 1 from E(Y X ) = β 0 + β 1 X and not the coefficient α 1 from E(Y X, V ) = α 0 + α 1 X + α 2 V Hogan (MDEpiNet) Missing Data October 22, / 160

40 Missing data mechanisms for longitudinal data Need to define some notation for longitudinal data Y j = value of Y at time j R j = 1 if Y j observed, 0 otherwise Y j = (Y 1, Y 2,..., Y j ) = outcome history up to time j X j = covariate history up to time j H j = (X j, Y j 1 ) Allows us to define MCAR, MAR, MNAR for longitudinal data Hogan (MDEpiNet) Missing Data October 22, / 160

41 Missing data mechanisms for longitudinal data MAR Missing at random (MAR) If interest is in marginal means such as E(Y j ), MAR means R j Y j (R j 1 = 1, H j ) Interpretation: Among those in follow up at time j, missingness is independent of the outcome Yj conditional on the previously observed Y s. Missingness does not depend on present or future Y s, given the past. Hogan (MDEpiNet) Missing Data October 22, / 160

42 Missing data mechanisms for longitudinal data MAR Implications 1 Selection mechanism [R j R j 1 = 1, H J ] = [R j R j 1 = 1, H j ] Can model selection probability as a function of observed past 2 Imputation mechanism [Y j R j = 0, R j 1 = 1, H j ] = [Y j R j = 1, R j 1 = 1, H j ] Can impute missing Yj using a model of the observed Y j Critical: Must correctly specify observed-data model Hogan (MDEpiNet) Missing Data October 22, / 160

43 LOCF We can characterize LOCF in this framework It is an imputation mechanism Missing Yj is equal to the most recently observed value of Y Missing value filled in with probability one (no variance) Formally where j = max k<j {R k = 1} [Y j R j = 0, H j ] = Y j with probability one, Not an MAR mechanism in general Conditional distribution of missing Yj not equal to that for observed Y j Hogan (MDEpiNet) Missing Data October 22, / 160

44 Random effects and parametric models Assume a joint distribution for the repeated measures Model applies to the full data, hence cannot be checked Example: Multivariate normal where µ J 1 = E(Y ) and Σ J J = var(y ) (Y 1,..., Y J ) T N(µ, Σ), Special case: Random effects model Particular way of structuring mean and variance Hogan (MDEpiNet) Missing Data October 22, / 160

45 Random effects and parametric models When do these models yield valid inference? Most parametric models valid if MAR holds All parts of the model are correctly specified These models have an implied distribution for the conditionals [Y j Y 1,..., Y j 1 ] Under MAR, the implied distribution applies to those with complete and incomplete data, but... Parametric assumptions cannot be checked empirically Hogan (MDEpiNet) Missing Data October 22, / 160

46 GEE I Assume a mean and variance structure for the repeated measures Not necessarily a full parametric model Assumed variance structure is the working covariance With complete data Inferences are most efficient when covariance correctly specified Correct inference about time-specific means even if covariance mis-specified Reason: all information about time-specific means is already observed Hogan (MDEpiNet) Missing Data October 22, / 160

47 GEE II With incomplete data Information about time-specific means relies on imputation of missing observations These imputations come from the conditional distribution [Y j Y 1,..., Y j 1 ] The form of the conditional distribution depends on the working covariance Implication: Correct inference about time-specific means only when both mean and covariance are correctly specified Can get different treatment effects with different working covariances Hogan (MDEpiNet) Missing Data October 22, / 160

48 Dependence of estimates on working covariance From Hogan et al., 2004 Statistics in Medicine Hogan (MDEpiNet) Missing Data October 22, / 160

49 Structure of case studies 1 Introduce modeling approach 2 Relate modeling approach to missing data hierarchy 3 Illustrate on simple cases 4 Include a treatment comparison 5 Discussion of key points from case study Hogan (MDEpiNet) Missing Data October 22, / 160

50 CASE STUDY I: MIXTURE MODEL ANALYSIS OF GROWTH HORMONE TRIAL Hogan (MDEpiNet) Missing Data October 22, / 160

51 Outline of analysis Objective: Compare EG to EP at month 12 Variable: Y3 Estimation of E(Y 3 ) for EG arm only Ignoring baseline covariates Using information from baseline covariate Y1 MAR and MNAR (sensitivity analysis) Treatment comparisons Expand to longitudinal case MAR using regression imputation MNAR sensitivity analysis Hogan (MDEpiNet) Missing Data October 22, / 160

52 Estimate E(Y ) from univariate sample Y3 R [1,] NA 0 [2,] [3,] [4,] [5,] NA 0 [6,] [7,] NA 0 [8,] NA 0 [9,] [10,] [11,] [12,] [13,] NA 0 [14,] Hogan (MDEpiNet) Missing Data October 22, / 160

53 Estimating E(Y ) from univariate sample Model: E(Y R = 1) = µ 1 E(Y R = 0) = µ 0 (not identifiable) Target of estimation E(Y ) = µ 1 P(R = 1) + µ 0 P(R = 0) Question: what to assume about µ 0? In a sense, we are going to impute a value for µ 0, or impute values of the missing Y s that will lead to an estimate of µ 0. Hogan (MDEpiNet) Missing Data October 22, / 160

54 Parameterizing departures from MAR Target of estimation: E µ0 (Y ) = µ 1 P(R = 1) + µ 0 P(R = 0) = µ 1 + (µ 0 µ 1 ) P(R = 0) Suggests sensitivity parameter (µ 0 ) = µ 0 µ 1. Leads to E (Y ) = µ 1 + P(R = 0) Features of this format: Centered at MAR ( = 0) cannot be estimated from observed data Can vary for sensitivity analysis Allows Bayesian approach by placing prior on Hogan (MDEpiNet) Missing Data October 22, / 160

55 Estimation under MNAR Recall model E (Y ) = µ 1 + P(R = 0) Estimate known quantities n 1 = i R i n 0 = i (1 R i) P(R = 0) = n 0 /(n 1 + n 0 ) µ 1 = (1/n 1 ) i Y i Have one unknown quantity = µ 0 µ 1 Hogan (MDEpiNet) Missing Data October 22, / 160

56 Estimation under MAR Plug into model Interpretation Under MAR ( = 0), Ê (Y ) = µ 1 + (µ 0 µ 1 ) }{{} P(R = 0) Ê (Y ) = µ 1 Estimator is the observed-data mean = µ 1 + P(R = 0) Under MNAR ( 0), Shift observed-dta mean by P(R = 0) Shift is proportional to fraction of missing observations Hogan (MDEpiNet) Missing Data October 22, / 160

57 y3.delta delta Y [R=1] = 88.3 P(R = 0) = 0.42 Hogan (MDEpiNet) Missing Data October 22, / 160

58 Using information from baseline covariates Y1 Y3 R [1,] 35.0 NA 0 [2,] [3,] [4,] [5,] 68.6 NA 0 [6,] [7,] 39.0 NA 0 [8,] 52.0 NA 0 [9,] [10,] [11,] [12,] [13,] 63.4 NA 0 [14,] Hogan (MDEpiNet) Missing Data October 22, / 160

59 General model for Y Objective: Inference for E(Y 3 ) Model General form [Y 3 Y 1, R = 1] F 1 (y 3 y 1 ) [Y 3 Y 1, R = 0] F 0 (y 3 y 1 ) The general form encompasses all possible models that can be assumed for the observed and missing values of Y 3 The model F 0 cannot be estimated from data Hogan (MDEpiNet) Missing Data October 22, / 160

60 The model under MAR and MNAR Under MAR, F 0 = F 1. Suggests following strategy Fit model for F1 using observed data; call it F 1 Use this model to impute missing values of Y3 Under MNAR, F 0 F 1. Suggests following strategy Parameterize a model so that F0 is related to F 1 through a sensitivity parameter Generically write this as F0 = F 1 Use fitted version of F 1 to impute missing Y 3 Hogan (MDEpiNet) Missing Data October 22, / 160

61 Regression parameterization of F 1 and F 0 Take the case of MAR first MAR implies [Y 3 Y 1, R = 1] = [Y 3 Y 1, R = 0] Assume regression model for [Y X, R = 1] E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 Assume a model for [Y X, R = 0] of similar form E(Y 3 Y 1, R = 0) = α 0 + β 0 Y 1 Cannot estimate parameters from observed data Hogan (MDEpiNet) Missing Data October 22, / 160

62 Regression parameterization of F 1 and F 0 Recall model Link the models. One way to do this: E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 E(Y 3 Y 1, R = 0) = α 0 + β 0 Y 1 β 0 = β 1 + β α 0 = α 1 + α Under MAR: α = β = 0 Hogan (MDEpiNet) Missing Data October 22, / 160

63 Caveats to using this (or any!) approach Recall model E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 E(Y 3 Y 1, R = 0) = α 0 + β 0 Y 1 More general version of missing-data model E(Y 3 Y 1, R = 0) = g(y 1 ; θ) Do we know the form of g? Do we know the value of θ? Do we know that Y 1 is sufficient to predict Y 3? We are assuming we know all of these things. Hogan (MDEpiNet) Missing Data October 22, / 160

64 Estimation of E(Y 3 ) under MAR 1 Fit the model E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 obtain α 1, β 1. 2 For those with R = 0, impute predicted value via Ŷ 3i = Ê(Y 3 Y 1i, R i = 0) = α 1 + β 1 Y 1i 3 Estimate overall mean as the mixture Ê(Y 3 ) = (1/n) i R i Y 3i + (1 R i )Ŷ3i Hogan (MDEpiNet) Missing Data October 22, / 160

65 Regression imputation under MAR y3[r == 1] y1[r == 1] Hogan (MDEpiNet) Missing Data October 22, / 160

66 Sample means and imputed means under MAR n r Y 1 Y 3 R = R = MAR Hogan (MDEpiNet) Missing Data October 22, / 160

67 Some intuition behind this (simple) estimator Could base estimate purely on regression model E(Y 3 ) = E Y1,R {E(Y 3 Y 1, R)} = E R [ EY1 R {E(Y 3 Y 1, R)} ] = E R [α 1 + β 1 E(Y 1 R)] = α 1 + β 1 E(Y 1 ) Plug in the estimators for each term May be more efficient when regression model is correct Hogan (MDEpiNet) Missing Data October 22, / 160

68 Some more details... If we don t want to use the regression model for Y 3, can write where the term in red is Hence E(Y 3 ) = E(Y 3 R = 1)P(R = 1) + E(Y 3 R = 0)P(R = 0), E(Y 3 R = 0) = E Y1 R=0 {E(Y 3 Y 1, R = 0)} Ê(Y 3 ) = Y [R=1] 3 = E Y1 R=0(α 1 + β 1 Y 1 R = 0) = α 1 + β 1 E(Y 1 R = 0) ( ˆP(R = 1) + α 1 + β ) 1 Y [R=0] 1 ˆP(R = 0) Hogan (MDEpiNet) Missing Data October 22, / 160

69 Inference and treatment comparisons SE and CI: bootstrap Draw bootstrap sample Carry out imputation procedure Repeat for lots of bootstrap samples (say B) Base SE and CI on the B bootstrapped estimators Why not multiple imputation? Estimators are linear Bootstrap takes care of missing data uncertainty here Treatment comparisons coming later Hogan (MDEpiNet) Missing Data October 22, / 160

70 Estimation of E(Y 3 ) under MNAR Recall model E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 E(Y 3 Y 1, R = 0) = α 0 + β 0 Y 1 More general version of model E(Y 3 Y 1, R = 1) = g 1 (Y 1 ; θ 1 ) E(Y 3 Y 1, R = 0) = g 0 (Y 1 ; θ 1, ) = h{g 1 (Y 1 ; θ 1 ), } The h function relates missing-data and observed-data models. User needs to specify form of h The parameter should not be estimable from observed data Vary in a sensitivity analysis Hogan (MDEpiNet) Missing Data October 22, / 160

71 Regression-based specification under MNAR Specify observed-data model E(Y 3 Y 1, R = 1) = g 1 (Y 1, θ 1 ) = α 1 + β 1 Y 1 Specify missing-data model E(Y 3 Y 1, R = 0) = h{g 1 (Y 1, θ 1 ), } = + g 1 (Y 1, θ 1 ) = + (α 1 + β 1 Y 1 ) Many other choices are possible Here, add a constant to the MAR imputation Have MAR when = 0, and MNAR otherwise Hogan (MDEpiNet) Missing Data October 22, / 160

72 Estimation of E(Y 3 ) under MNAR 1 Fit the model E(Y 3 Y 1, R = 1) = α 1 + β 1 Y 1 obtain α 1, β 1. 2 For those with R = 0, impute predicted value via Ŷ 3i = Ê(Y 3 Y 1i, R i = 0) = + α 1 + β 1 Y 1i 3 Estimate overall mean as the mixture Ê(Y 3 ) = (1/n) R i Y 3i + (1 R i )Ŷ3i i ( = Y [R=1] 3 ˆP(R = 1) + + α 1 + β ) 1 Y [R=0] 1 ˆP(R = 0) Hogan (MDEpiNet) Missing Data October 22, / 160

73 Sensitivity analysis based on varying What should be the anchor point? Usually appropriate to anchor analysis at MAR Examine effect of MNAR by varying away from 0 How to select a range for? Will always be specific to application. Ensure that range is appropriate to context (see upcoming example) Can use data-driven range for, e.g. based on SD Reporting final inferences Stress test approach Inverted sensitivity analysis find values of that would change substantive conclusions Average over plausible values Hogan (MDEpiNet) Missing Data October 22, / 160

74 Calibrating How should the range and scale of be chosen? Direction > 0 dropouts have higher mean < 0 dropouts have lower mean Range and scale Residual variation in outcome quantified by SD of regression error [Y 3 Y 1, R = 1] = α 1 + β 1 Y 1 + e var(e) = σ 2 Suggests scaling in units of σ Will illustrate in longitudinal case Hogan (MDEpiNet) Missing Data October 22, / 160

75 Moving to longitudinal setting Set-up for single treatment arm Illustrate ideas with analysis of GH data Compare treatments Illustrate sensitivity analysis under MNAR Discuss how to report results Hogan (MDEpiNet) Missing Data October 22, / 160

76 Longitudinal case: notation Assume missing data pattern is monotone K = dropout time = j R j E k (Y j ) = E(Y j K = k) When j > k, cannot estimate E k (Y j ) from data Hogan (MDEpiNet) Missing Data October 22, / 160

77 Longitudinal model with J = 3: Set up E k (Y 1 ) = E(Y 1 K = k) identified for k = 1, 2, 3 For other means, we have j = 2 j = 3 K = 1 E 1 (Y 2 Y 1 ) E 1 (Y 3 Y 1, Y 2 ) K = 2 E 2 (Y 2 Y 1 ) E 2 (Y 3 Y 1, Y 2 ) K = 3 E 3 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Y 2 ) Components in red cannot be estimated. Need assumptions Hogan (MDEpiNet) Missing Data October 22, / 160

78 Longitudinal model with J = 3: MAR j = 2 j = 3 K = 1 ωe 2 (Y 2 Y 1 ) + (1 ω)e 3 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Ŷ2) K = 2 E 2 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Y 2 ) K = 3 E 3 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Y 2 ) Here, ω is a weight such that 0 ω 1 Hogan (MDEpiNet) Missing Data October 22, / 160

79 Longitudinal model with J = 3: MNAR j = 2 j = 3 K = 1 ωe 2 (Y 2 Y 1 ) + (1 ω)e 3 (Y 2 Y 1 ) + 1 E 3 (Y 3 Y 1, Ŷ2) + 2 K = 2 E 2 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Y 2 ) + 3 K = 3 E 3 (Y 2 Y 1 ) E 3 (Y 3 Y 1, Y 2 ) Hogan (MDEpiNet) Missing Data October 22, / 160

80 Procedure with several longitudinal measures Start by imputing those with missing data at j = 2 1 Fit model for E(Y 2 Y 1, R 2 = 1) obtain α (2), β (2) 1 E(Y 2 Y 1, R 2 = 1) = α (2) + β (2) 1 Y 1 This is a model that combines those with K = 2 and K = 3 2 Impute missing Y 2 as before Ŷ 2i = + α (2) + β (2) 1 Y 1i Hogan (MDEpiNet) Missing Data October 22, / 160

81 Procedure with several longitudinal measures Now impute those with missing data at j = 3 1 Fit model for E(Y 3 Y 1, Y 2, R 3 = 1) obtain α (3), E(Y 3 Y 1, Y 2, R 3 = 1) = α (3) + β (3) 1 Y 1 + β (3) 2 Y 2 (3) (3) β 1, β 2 2 Impute missing Y 3 as follows: For those with Y1, Y 2 observed, Ŷ 3i = + α (3) + (3) β 1 Y (3) 1i + β 2 Y 2i For those with only Y1 observed, Ŷ 3i = + α (3) + β (3) 1 Y 1i + (3) β 2 Ŷ 2i Hogan (MDEpiNet) Missing Data October 22, / 160

82 Side note Recall imputation for those with only Y 1 observed: Ŷ 3i = + α (3) + β (3) 1 Y 1i + (3) β 2 Ŷ 2i This is really just using information from the observed Y 1 because Ŷ 2i = + α (2) + β (2) 1 Y 1i Hence the imputation is from the (linear) model of E(Y 3 Y 1 ) that is implied by the other imputation models. Hogan (MDEpiNet) Missing Data October 22, / 160

83 Calibration of At each time point j, is actually a multiplier of the residual SD for the observed data regression [Y j Y 1,..., Y j 1, R = 1] For example, the imputation model at j = 3 is actually Ŷ 3i = σ 3 + α (3) + where σ 2 3 = var(y 3 Y 1, Y 2, R 3 = 1) (3) β 1 Y (3) 1i + β 2 Y 2i Generally will suppress this for clarity Hogan (MDEpiNet) Missing Data October 22, / 160

84 Procedure with several longitudinal measures Final step: compute estimate of E(Y 3 ) Ê (Y 3 ) = (1/n) i R 3i Y 3i + (1 R 3i )Ŷ3i( ) Based on the imputations, this turns out to be a weighted average of Y [K=3] 3, Y [K=2] 2, and Y [K=1] 1. Weights depend on dropout rates at each time coefficients in imputation models sensitivity parameter(s) Hogan (MDEpiNet) Missing Data October 22, / 160

85 Analysis components Fitted regression models at each time point Can check validity of imputation Contour plots: vary separately by treatment Treatment effect estimates p-values Summary table Treatment effect, SE, p-value These are computed using bootstrap Hogan (MDEpiNet) Missing Data October 22, / 160

86 Hogan (MDEpiNet) Missing Data October 22, / 160

87 Hogan (MDEpiNet) Missing Data October 22, / 160

88 Hogan (MDEpiNet) Missing Data October 22, / 160

89 Hogan (MDEpiNet) Missing Data October 22, / 160

90 Case Study 2: Chronic Schizophrenia Major breakthroughs have been made in the treatment of patients with psychotic symptoms. However, side effects associated with some medications have limited their usefulness. RIS-INT-3 (Marder and Meibach, 1994, Chouinard et al., 1993) was a multi-center study designed to assess the effectiveness and adverse experiences of four fixed doses of risperidone compared to haliperidol and placebo in the treatment of chronic schizophrenia. Hogan (MDEpiNet) Missing Data October 22, / 160

91 RIS-INT-3 Patients were required to have a PANSS (Positive and Negative Syndrome Scale) score between 60 and 120. Prior to randomization, one-week washout phase (all anti-psychotic medications discontinued). If acute psychotic symptoms occurred, patients randomized to a double-blind treatment phase, schedule to last 8 weeks. Patients randomized to one of 6 treatment groups: risperidone 2, 6, 10 or 16 mg, haliperidol 20 mg, or placebo. Dose titration occurred during the first week of the double-blind phase. Hogan (MDEpiNet) Missing Data October 22, / 160

92 RSIP-INT-3 Patients scheduled for 5 post-baseline assessments at weeks 1,2,4,6, and 8 of the double-blind phase. Primary efficacy variable: PANSS score Patients who did not respond to treatment and discontinued therapy or those who completed the study were eligible to receive risperidone in an open-label extension study. 521 patients randomized to receive placebo (n = 88), haliperidol 20 mg (n = 87), risperidone 2mg (n = 87), risperidone 6mg (n = 86), risperidone 10 mg (n = 86), or risperidone 16 mg (n = 87). Hogan (MDEpiNet) Missing Data October 22, / 160

93 Dropout and withdrawal Only 49% of patients completed the 8 week treatment period. The most common reason for discontinuation was insufficient response. Other main reasons included: adverse events, uncooperativeness, and withdrawal of consent. Hogan (MDEpiNet) Missing Data October 22, / 160

94 Dropout and Withdrawal Placebo Haliperidol Risp 2mg Risp 6mg Risp 10mg Risp 16 mg (n = 88) (n = 87) (n = 87) (n = 86) (n = 86) (n = 87) Completed 27 31% 36 41% 36 41% 53 62% 48 56% 54 62% Withdrawn 61 69% 51 59% 51 59% 33 38% 38 44% 33 38% Lack of Efficacy 51 58% 36 41% 41 47% 12 14% 25 29% 18 21% Other 10 11% 15 17% 10 11% 21 24% 13 15% 15 17% Hogan (MDEpiNet) Missing Data October 22, / 160

95 Central Question What is the difference in the mean PANSS scores at week 8 between risperidone at a specified dose level vs. placebo in the counterfactual world in which all patients were followed to that week? Hogan (MDEpiNet) Missing Data October 22, / 160

96 Sample means and imputed means under MAR N INS HOST EPS Y 0 µ R Placebo R = ?? R = Risperidone R = ?? R = Hogan (MDEpiNet) Missing Data October 22, / 160

97 Sample means and imputed means under MAR N INS HOST EPS Y 0 µ R Placebo R = R = Risperidone R = R = Hogan (MDEpiNet) Missing Data October 22, / 160

98 Regression imputation under MAR Placebo Hogan (MDEpiNet) Missing Data October 22, / 160

99 Regression imputation under MAR Risperidone Hogan (MDEpiNet) Missing Data October 22, / 160

100 Sample means by dropout time: Aggregated data PANSS Visit Hogan (MDEpiNet) Missing Data October 22, / 160

101 Extrapolation of means when ν = 0 (MAR) PANSS Visit Hogan (MDEpiNet) Missing Data October 22, / 160

102 Extrapolation of means when ν = 1 PANSS Visit Hogan (MDEpiNet) Missing Data October 22, / 160

103 Extrapolation of means when ν = +1 PANSS Visit Hogan (MDEpiNet) Missing Data October 22, / 160

104 Summary of means Placebo Risperidone E(Y 5 R 5 = 1) E(Y 5 R 5 = 0) ν = ν = ν = E(Y 5 ) ν = ν = ν = Hogan (MDEpiNet) Missing Data October 22, / 160

105 Summary Full-data mean parameterized as a mixture of observed- and missing-data means Implemented using imputation We used regression imputation, but this is only one possibility Regression models need not be linear More complex models may require more complex imputation procedures Key features of this approach Missing data distribution indexed by sensitivity parameter that cannot be estimated from data Separates testable from untestable assumptions Easy to assess effect of departures from MAR Hogan (MDEpiNet) Missing Data October 22, / 160

106 Summary Model parameterization Need to limit number of s to make inferences manageable Need sensible scale and range for s Scope of sensitivity analysis should be specified as part of trial protocol to avoid reliance on post-hoc analyses Inference about treatment effects Sensitivity analysis provides range of conclusions Can use as a stress-test : under what MNAR scenario would our conclusions change? Can also use Bayesian formulations that average results over a prior for sensitivity parameter (Daniels & Hogan, 2008) Hogan (MDEpiNet) Missing Data October 22, / 160

107 CASE STUDY II: INVERSE PROBABILITY WEIGHTING METHODS Hogan (MDEpiNet) Missing Data October 22, / 160

108 Inverse Probability Weighting General idea Consider estimating E(Y ) from a sample of data If we had all the data, we would use the sample mean Ê(Y ) = (1/n)(Y 1 + Y Y n ) Problem: Only some of the Y s are observed The observed Y s may not be a random draw from the full sample Hogan (MDEpiNet) Missing Data October 22, / 160

109 Inverse Probability Weighting General idea The solution: use a weighted mean Probability of being observed: π i = P(R i = 1) Issues: Weighted mean Ê IPW (Y ) = i R iy i /π i i R i/π i Probability of being observed may depend on individual characteristics X May also depend on the actual (but unobserved) outcome Y Hogan (MDEpiNet) Missing Data October 22, / 160

110 Inverse Probability Weighting under MAR Recall MAR: Y R X MAR implies [R X, Y ] [R X ] IPW theory Define π(x i ) = P(R = 1 X i ) Assume MAR Assume π(x i ) > 0 for all i The weighted estimator is a consistent estimate of E(Y ) Ê IPW (Y ) = i R iy i /π(x i ) i R i/π(x i ) Hogan (MDEpiNet) Missing Data October 22, / 160

111 Inverse Probability Weighting under MAR Remains true when π(x i ) is replaced by a consistent estimator π(x i ). To estimate π(x i ), must specify a model such as logit π(x i ) = X T i β In this case, π(x i ) = exp(xi T β) 1 + exp(xi T β) Can use other models as well but the model must yield consistent estimates of π(x i ) for IPW to give valid estimator. Hogan (MDEpiNet) Missing Data October 22, / 160

112 Example: Estimate E(Y 3 ) from GH Data Treat Y 1 as the covariate MAR assumption is [R Y 3, Y 1 ] [R Y 1 ] Assume logit model for selection logit π(y 1 ) = γ 0 + γ 1 Y 1 Compute weighted estimator as above Hogan (MDEpiNet) Missing Data October 22, / 160

113 Fitted model π(y 1 ) prob y1 Hogan (MDEpiNet) Missing Data October 22, / 160

114 Relative weights: Plot of Y 3 vs 1/ π(y 1 ) wt[r == 1] y3[r == 1] Hogan (MDEpiNet) Missing Data October 22, / 160

115 Compare estimators Complete cases: 88 Imputation under MAR: 79 IPW under MAR: 80 Hogan (MDEpiNet) Missing Data October 22, / 160

116 IPW Longitudinal case To illustrate, we assume monotone missingness (as in the GH trial) Target of inference: E(Y 3 ). Weighted estimator is Ê IPW (Y 3 ) = How to construct the response probabilities π 3i? i R 3iY 3i /π 3i i R 3i/π 3i Hogan (MDEpiNet) Missing Data October 22, / 160

117 Longitudinal case no covariates R 3 = 1 is equivalent to the joint event R 1 = 1, R 2 = 1, R 3 = 1 Hence P(R 3 = 1) = P(R 1 = 1, R 2 = 1, R 3 = 1) = P(R 3 = 1 R 2 = 1, R 1 = 1) P(R 2 = 1 R 1 = 1) P(R 1 = 1) = φ 3 φ 2 φ 1 Notation φ j = P(R j = 1 R 1 = R 2 = = R j 1 = 1) Hogan (MDEpiNet) Missing Data October 22, / 160

118 Longitudinal case with covariates Recall longitudinal version of MAR [R j R j 1 = 1, H J ] = [R j R j 1 = 1, H j ] In words, missingness at j depends only on the observable past history of X and Y. Implication: can use observable history in models of φ j Example: logit φ j (H j ) = X T i1β + X T ij γ + θy i,j 1 Hogan (MDEpiNet) Missing Data October 22, / 160

119 Procedure for inference about E(Y J ) 1 Formulate and fit models for φ j (H ij ) = P(R j = 1 R i,j 1 = 1, H ij ) 2 Compute estimated value of π J (H ij ) = P(R J = 1 H ij ) as π J (H ij ) = φ 1 (H i1 ) φ 2 (H i2 ) φ J (H ij ) 3 Compute weighted mean Ê IPW (Y J ) = i R ijy ij / π J (H ij ) i R ij/ π J (H ij ) Hogan (MDEpiNet) Missing Data October 22, / 160

120 Inverse Probability Weighting under MAR Practical issues to consider Stability of weights Weights near zero lead to bias and inefficiency Need to check histogram Can use stabilized weights Fit of weight models No guarantees here Can use lack of fit diagnostics to weed out poor-fitting models Selection of weight models Poses a more serious problem with respect to final inferences Not good to pick weight model that gives lowest p-value! Pre-specify weight covariates that are related to missingness and outcome Hogan (MDEpiNet) Missing Data October 22, / 160

121 Analysis of Smoking Cessation Data via IPW Specify outcome model Select baseline covariates Specify and fit weight model Fit weighted longitudinal regression for treatment comparison Hogan (MDEpiNet) Missing Data October 22, / 160

122 Outcome model Outcome and treatment Y j = quit status (1 if yes, 0 if no) Z = 1 if exercise, 0 if wellness θ j = P(Y j = 1) Model constant quit rate up to week 4 separate treatment quit rates after week 4, but constant over time logit θ j = γ 0 1(j 4) + (γ 1 + βz) 1(j > 4) β = treatment log odds ratio Hogan (MDEpiNet) Missing Data October 22, / 160

123 Hogan (MDEpiNet) Missing Data October 22, / 160

124 Exploratory analysis Ascertain whether previous Y s belong in weight model This shows strong correlation between R j and Y j 1 Hogan (MDEpiNet) Missing Data October 22, / 160

125 Exploratory analysis Hogan (MDEpiNet) Missing Data October 22, / 160

126 Covariates for selection model Hogan (MDEpiNet) Missing Data October 22, / 160

127 Check weight distribution at each time Hogan (MDEpiNet) Missing Data October 22, / 160

128 Summary of results Treatment effect relatively robust in odds ratio scale Arm-specific cessation rates very different Tx effect not robust in other scales Hogan (MDEpiNet) Missing Data October 22, / 160

129 Hogan (MDEpiNet) Missing Data October 22, / 160

130 Multiple Imputation Hogan (MDEpiNet) Missing Data October 22, / 160

131 Overview Imputing missing data from a parametric model Build around an example using GH data Goal 1: Estimate E(Y3 ) in GH data Goal 2: Estimate treatment effect in CTQ data Process of imputation Model specification Drawing imputed values from the model Combining observed and imputed information Standard error estimation Sensitivity analysis (CTQ data) Hogan (MDEpiNet) Missing Data October 22, / 160

132 Recall excerpt from GH data id tx Y1 Y2 Y3 R Hogan (MDEpiNet) Missing Data October 22, / 160

133 The strategy behind multiple imputation Setting: Full data for an individual is (Y, R, X, V ) Objective: Interested in some feature of f (y) or f (y x) Assumptions: 1 MAR, in that f (y x, v, r = 0) = f (y x, v, r = 1) 2 Model for f (y x, v, r = 1) has known form Hogan (MDEpiNet) Missing Data October 22, / 160

134 The strategy behind multiple imputation 1 Fit a model for f (y x, v, r = 1); if it is a parametric model, this means estimating the parameters α in the model f (y x, v, r = 1, α) 2 For each person having R = 0, take a draw of Y X, V from the fitted model. That means, for person i having R i = 0, plug in their values of X i and V i, and draw a value of Y i from the fitted model. Example coming soon. 3 Do this several times for each individual, so that each person has multiple draws of Y i. Can call these Now have K filled-in datasets. Ŷ (1) i, Ŷ (2) i,..., Ŷ (K) i Hogan (MDEpiNet) Missing Data October 22, / 160

135 The strategy behind multiple imputation 4 Perform the analysis you would have carried out had the data been complete. If you are interested in the parameter θ, this gives you K parameter estimates, θ (1), θ (2),..., θ (K) Hogan (MDEpiNet) Missing Data October 22, / 160

136 The strategy behind multiple imputatation 5 Now need an estimate and standard error The estimate is the sample mean θ = (1/K) K θ (j) j=1 The (estimate of) variance of θ combines the between- and within-imputation variance, var( θ) = 1 K 1 K ( θ (j) θ) 2 + j=1 1 K K var( θ (j) ), j=1 where var( θ (j) ) = { } 2 s.e.( θ (j) ) Hogan (MDEpiNet) Missing Data October 22, / 160

137 Analysis 1: Use MI to estimate E(Y 3 ) As with single imputation, we treat Y 1 as an auxiliary variable to use in the imputation model Missing data assumption: MAR f (y 3 y 1, r = 1) = f (y 3 y 1, r = 0) Specify parametric imputation model f (y 3 y 1, r = 1) Y 3 = β 0 + β 1 Y 1 + e e N(0, σ 2 ) This implies that the model f (y3 y 1, r = 1) is a normal distribution with mean and variance E(Y 3 Y 1 ) = β 0 + β 1 Y 1 var(y 3 Y 1 ) = σ 2 Hogan (MDEpiNet) Missing Data October 22, / 160

138 Applying MI to the GH data 1 Fit a model for f (y 3 y 1, r 3 = 1) fitted.model = lm(y3 ~ Y1, subset = (R3==1)) beta.hat = fitted.model$coefficients sigma.sq = anova.lm(fitted.model)$ Mean Sq [2] sigma = sqrt(sigma.sq) For the GH data, this means fitting the regression model specified above. The estimated parameters are β 0 = 15.6 β 1 = 0.19 σ = 20.9 Hogan (MDEpiNet) Missing Data October 22, / 160

139 Y1 R3 Y3 Y3.mean [1,] NA [2,] [3,] [4,] [5,] NA [6,] [7,] NA Hogan (MDEpiNet) Missing Data October 22, / 160 Applying MI to the GH data 2 For each person having R = 0, take a draw of Y 3 Y 1 from the fitted model. That means, for person i having R i = 0, plug in their value of Y 1i, and draw a value of Y 3 from the fitted model. So we draw imputed values of the missing Y 3 from X.matrix = cbind( rep(1,length(y1)), Y1) Y3.mean = X.matrix %*% beta.hat Y 3 N( Y 1i, )

140 Applying MI to GH data 3 Do this several times for each individual, so that each person has multiple draws of Y 3. Can call these Now have K filled-in datasets. Ŷ (1) # Draw 10 imputations of Y3 K = 10 n = length(y3) Y3.imp = matrix(0, nrow=n, ncol=k) 3i, Ŷ (2) 3i,..., Ŷ (K) 3i for (j in 1:K) { # this line imputes a value for each person Y3.imp[,j] = rnorm(n=n, mean=y3.mean, sd = sigma ) # this line replaces the values with observed Y3 where Y3 is observed Y3.imp[R3==1,] = Y3[R3==1] } Hogan (MDEpiNet) Missing Data October 22, / 160

141 Excerpt from imputed data Y1 R3 Y3 Y3.mean === FIRST 4 IMPUTATIONS === [1,] NA [2,] [3,] [4,] [5,] NA [6,] [7,] NA Hogan (MDEpiNet) Missing Data October 22, / 160

142 Applying MI to GH data 4 Perform the analysis you would have carried out had the data been complete. If you are interested in the parameter θ, this gives you K parameter estimates, θ (1), θ (2),..., θ (K) Hogan (MDEpiNet) Missing Data October 22, / 160

143 Applying MI to GH data # Step 4: Calculate E(Y3) for each replicated dataset # We will do this on the matrix Y3.imp, calculating column means and s.e. Y3.bar.imp = apply(y3.imp, 2, mean) Y3.sd.imp = apply(y3.imp, 2, sd) Y3.se.imp = Y3.sd / sqrt(n) Hogan (MDEpiNet) Missing Data October 22, / 160

144 Applying MI to GH data > # Here are the means and s.e. from each imputed dataset > cbind(y3.bar.imp, Y3.se.imp) Y3.bar.imp Y3.se.imp [1,] [2,] [3,] [4,] [5,] [6,] [7,] [8,] [9,] [10,] Hogan (MDEpiNet) Missing Data October 22, / 160

145 Apply MI to GH data 5 The (estimate of) variance of θ combines the between- and within-imputation variance, var( θ) = 1 K 1 K ( θ (j) θ) K j=1 K var( θ (j) ), j=1 # Step 5: Calculate overall mean and SE theta.hat = mean(y3.bar.imp) var.w = mean( Y3.se.imp^2 ) var.b = (1 + 1/K) * sd( Y3.bar.imp )^2 se.theta.hat = sqrt(var.w + var.b) # missing information miss.info = var.b / (var.w + var.b) c(theta.hat, se.theta.hat, miss.info) [1] Hogan (MDEpiNet) Missing Data October 22, / 160

146 Comparing results from three methods Method Ê(Y 3 ) s.e. Observed data only IPW Regression imputation with bootstrap s.e. 6.8 Multiple imputation Hogan (MDEpiNet) Missing Data October 22, / 160

147 Example 2: CTQ Data Goal of analysis: treatment comparison between wellness and exercise Available data: Y = indicator of quit status at week 12 (1 if yes, 0 if no) X = treatment indicator V = auxiliary covariates measured at baseline R = indicator of whether Y is observed Model of interest logit{pr(y = 1 X )} = β 0 + β 1 X Hogan (MDEpiNet) Missing Data October 22, / 160

148 Example 2: CTQ Data Auxiliary covariates V = (F, W ): F = Fagerstrom index of nicotine dependence (1-10) W = weight at baseline Recall: model of interest does not involve F, W logit{pr(y = 1 X )} = β 0 + β 1 X Hogan (MDEpiNet) Missing Data October 22, / 160

149 CTQ Data Imputation model under MAR logit{p(y = 1 X, F, W )} = α 0 + α 1 X + α 2 F + α 3 W S i = α 0 + α 1 X i + α 2 F i + α 3 W i φ i = exp(s i ) 1 + exp(s i ) Y i Ber(φ i ) Imputation involves drawing multiple values of Y from the appropriate Bernoulli distribution. Hogan (MDEpiNet) Missing Data October 22, / 160

150 Apply MI to CTQ Data Data excerpt > ctq id week y z weight j r fs basewt NA 1 NA NA 1 NA Hogan (MDEpiNet) Missing Data October 22, / 160

151 Apply MI to CTQ Data Fit treatment model to observed data only logit{p(y = 1 X )} = β 0 + β 1 X > # fit regression model of smoking to z with observed data > model.0 = glm(y ~ z, family=binomial(link="logit") ) > summary(model.0) Call: glm(formula = y ~ z, family = binomial(link = "logit")) Coefficients: Estimate Std. Error z value Pr(> z ) (Intercept) *** z Null deviance: on 186 degrees of freedom (90 observations deleted due to missingness) Hogan (MDEpiNet) Missing Data October 22, / 160

152 Apply MI to CTQ Data 2 Fit imputation model logit{p(y = 1 X, F, W )} = α 0 + α 1 X + α 2 F + α 3 W = S > imp.model = glm(y ~ z + basewt + fs, family = binomial) > summary(imp.model) glm(formula = y ~ z + basewt + fs, family = binomial) Estimate Std. Error z value Pr(> z ) (Intercept) z basewt * fs *** Hogan (MDEpiNet) Missing Data October 22, / 160

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A. Linero and M. Daniels UF, UT-Austin SRC 2014, Galveston, TX 1 Background 2 Working model

More information

Whether to use MMRM as primary estimand.

Whether to use MMRM as primary estimand. Whether to use MMRM as primary estimand. James Roger London School of Hygiene & Tropical Medicine, London. PSI/EFSPI European Statistical Meeting on Estimands. Stevenage, UK: 28 September 2015. 1 / 38

More information

Global Sensitivity Analysis for Repeated Measures Studies with Informative Drop-out: A Semi-Parametric Approach

Global Sensitivity Analysis for Repeated Measures Studies with Informative Drop-out: A Semi-Parametric Approach Global for Repeated Measures Studies with Informative Drop-out: A Semi-Parametric Approach Daniel Aidan McDermott Ivan Diaz Johns Hopkins University Ibrahim Turkoz Janssen Research and Development September

More information

Statistical Methods. Missing Data snijders/sm.htm. Tom A.B. Snijders. November, University of Oxford 1 / 23

Statistical Methods. Missing Data  snijders/sm.htm. Tom A.B. Snijders. November, University of Oxford 1 / 23 1 / 23 Statistical Methods Missing Data http://www.stats.ox.ac.uk/ snijders/sm.htm Tom A.B. Snijders University of Oxford November, 2011 2 / 23 Literature: Joseph L. Schafer and John W. Graham, Missing

More information

Bayesian methods for missing data: part 1. Key Concepts. Nicky Best and Alexina Mason. Imperial College London

Bayesian methods for missing data: part 1. Key Concepts. Nicky Best and Alexina Mason. Imperial College London Bayesian methods for missing data: part 1 Key Concepts Nicky Best and Alexina Mason Imperial College London BAYES 2013, May 21-23, Erasmus University Rotterdam Missing Data: Part 1 BAYES2013 1 / 68 Outline

More information

7 Sensitivity Analysis

7 Sensitivity Analysis 7 Sensitivity Analysis A recurrent theme underlying methodology for analysis in the presence of missing data is the need to make assumptions that cannot be verified based on the observed data. If the assumption

More information

Extending causal inferences from a randomized trial to a target population

Extending causal inferences from a randomized trial to a target population Extending causal inferences from a randomized trial to a target population Issa Dahabreh Center for Evidence Synthesis in Health, Brown University issa dahabreh@brown.edu January 16, 2019 Issa Dahabreh

More information

MISSING or INCOMPLETE DATA

MISSING or INCOMPLETE DATA MISSING or INCOMPLETE DATA A (fairly) complete review of basic practice Don McLeish and Cyntha Struthers University of Waterloo Dec 5, 2015 Structure of the Workshop Session 1 Common methods for dealing

More information

Clinical Trials. Olli Saarela. September 18, Dalla Lana School of Public Health University of Toronto.

Clinical Trials. Olli Saarela. September 18, Dalla Lana School of Public Health University of Toronto. Introduction to Dalla Lana School of Public Health University of Toronto olli.saarela@utoronto.ca September 18, 2014 38-1 : a review 38-2 Evidence Ideal: to advance the knowledge-base of clinical medicine,

More information

Some methods for handling missing values in outcome variables. Roderick J. Little

Some methods for handling missing values in outcome variables. Roderick J. Little Some methods for handling missing values in outcome variables Roderick J. Little Missing data principles Likelihood methods Outline ML, Bayes, Multiple Imputation (MI) Robust MAR methods Predictive mean

More information

A Sampling of IMPACT Research:

A Sampling of IMPACT Research: A Sampling of IMPACT Research: Methods for Analysis with Dropout and Identifying Optimal Treatment Regimes Marie Davidian Department of Statistics North Carolina State University http://www.stat.ncsu.edu/

More information

2 Naïve Methods. 2.1 Complete or available case analysis

2 Naïve Methods. 2.1 Complete or available case analysis 2 Naïve Methods Before discussing methods for taking account of missingness when the missingness pattern can be assumed to be MAR in the next three chapters, we review some simple methods for handling

More information

Global Sensitivity Analysis for Repeated Measures Studies with. Informative Drop-out: A Fully Parametric Approach

Global Sensitivity Analysis for Repeated Measures Studies with. Informative Drop-out: A Fully Parametric Approach Global Sensitivity Analysis for Repeated Measures Studies with Informative Drop-out: A Fully Parametric Approach Daniel Scharfstein (dscharf@jhsph.edu) Aidan McDermott (amcdermo@jhsph.edu) Department of

More information

Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs

Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs Michael J. Daniels and Chenguang Wang Jan. 18, 2009 First, we would like to thank Joe and Geert for a carefully

More information

Missing Data in Longitudinal Studies: Mixed-effects Pattern-Mixture and Selection Models

Missing Data in Longitudinal Studies: Mixed-effects Pattern-Mixture and Selection Models Missing Data in Longitudinal Studies: Mixed-effects Pattern-Mixture and Selection Models Hedeker D & Gibbons RD (1997). Application of random-effects pattern-mixture models for missing data in longitudinal

More information

A Generalized Linear Model for Binomial Response Data. Copyright c 2017 Dan Nettleton (Iowa State University) Statistics / 46

A Generalized Linear Model for Binomial Response Data. Copyright c 2017 Dan Nettleton (Iowa State University) Statistics / 46 A Generalized Linear Model for Binomial Response Data Copyright c 2017 Dan Nettleton (Iowa State University) Statistics 510 1 / 46 Now suppose that instead of a Bernoulli response, we have a binomial response

More information

De-mystifying random effects models

De-mystifying random effects models De-mystifying random effects models Peter J Diggle Lecture 4, Leahurst, October 2012 Linear regression input variable x factor, covariate, explanatory variable,... output variable y response, end-point,

More information

Analysing longitudinal data when the visit times are informative

Analysing longitudinal data when the visit times are informative Analysing longitudinal data when the visit times are informative Eleanor Pullenayegum, PhD Scientist, Hospital for Sick Children Associate Professor, University of Toronto eleanor.pullenayegum@sickkids.ca

More information

TUTORIAL IN BIOSTATISTICS Handling drop-out in longitudinal studies

TUTORIAL IN BIOSTATISTICS Handling drop-out in longitudinal studies STATISTICS IN MEDICINE Statist. Med. 2004; 23:1455 1497 (DOI: 10.1002/sim.1728) TUTORIAL IN BIOSTATISTICS Handling drop-out in longitudinal studies Joseph W. Hogan 1; ;, Jason Roy 2; and Christina Korkontzelou

More information

Linear Regression Models P8111

Linear Regression Models P8111 Linear Regression Models P8111 Lecture 25 Jeff Goldsmith April 26, 2016 1 of 37 Today s Lecture Logistic regression / GLMs Model framework Interpretation Estimation 2 of 37 Linear regression Course started

More information

36 th Annual Meeting Preconference Workshop P3 Handout

36 th Annual Meeting Preconference Workshop P3 Handout 36 th Annual Meeting Preconference Workshop P3 Handout Global Sensitivity Analysis of Randomized Trials with Missing Data Daniel Scharfstein Johns Hopkins University dscharf@jhu.edu May 17, 2015 1 / 75

More information

Basics of Modern Missing Data Analysis

Basics of Modern Missing Data Analysis Basics of Modern Missing Data Analysis Kyle M. Lang Center for Research Methods and Data Analysis University of Kansas March 8, 2013 Topics to be Covered An introduction to the missing data problem Missing

More information

Classification. Chapter Introduction. 6.2 The Bayes classifier

Classification. Chapter Introduction. 6.2 The Bayes classifier Chapter 6 Classification 6.1 Introduction Often encountered in applications is the situation where the response variable Y takes values in a finite set of labels. For example, the response Y could encode

More information

Time-Invariant Predictors in Longitudinal Models

Time-Invariant Predictors in Longitudinal Models Time-Invariant Predictors in Longitudinal Models Today s Class (or 3): Summary of steps in building unconditional models for time What happens to missing predictors Effects of time-invariant predictors

More information

Personalized Treatment Selection Based on Randomized Clinical Trials. Tianxi Cai Department of Biostatistics Harvard School of Public Health

Personalized Treatment Selection Based on Randomized Clinical Trials. Tianxi Cai Department of Biostatistics Harvard School of Public Health Personalized Treatment Selection Based on Randomized Clinical Trials Tianxi Cai Department of Biostatistics Harvard School of Public Health Outline Motivation A systematic approach to separating subpopulations

More information

Shared Random Parameter Models for Informative Missing Data

Shared Random Parameter Models for Informative Missing Data Shared Random Parameter Models for Informative Missing Data Dean Follmann NIAID NIAID/NIH p. A General Set-up Longitudinal data (Y ij,x ij,r ij ) Y ij = outcome for person i on visit j R ij = 1 if observed

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

BIOSTATISTICAL METHODS

BIOSTATISTICAL METHODS BIOSTATISTICAL METHODS FOR TRANSLATIONAL & CLINICAL RESEARCH Cross-over Designs #: DESIGNING CLINICAL RESEARCH The subtraction of measurements from the same subject will mostly cancel or minimize effects

More information

Estimating and Using Propensity Score in Presence of Missing Background Data. An Application to Assess the Impact of Childbearing on Wellbeing

Estimating and Using Propensity Score in Presence of Missing Background Data. An Application to Assess the Impact of Childbearing on Wellbeing Estimating and Using Propensity Score in Presence of Missing Background Data. An Application to Assess the Impact of Childbearing on Wellbeing Alessandra Mattei Dipartimento di Statistica G. Parenti Università

More information

Binomial Model. Lecture 10: Introduction to Logistic Regression. Logistic Regression. Binomial Distribution. n independent trials

Binomial Model. Lecture 10: Introduction to Logistic Regression. Logistic Regression. Binomial Distribution. n independent trials Lecture : Introduction to Logistic Regression Ani Manichaikul amanicha@jhsph.edu 2 May 27 Binomial Model n independent trials (e.g., coin tosses) p = probability of success on each trial (e.g., p =! =

More information

Case Study in the Use of Bayesian Hierarchical Modeling and Simulation for Design and Analysis of a Clinical Trial

Case Study in the Use of Bayesian Hierarchical Modeling and Simulation for Design and Analysis of a Clinical Trial Case Study in the Use of Bayesian Hierarchical Modeling and Simulation for Design and Analysis of a Clinical Trial William R. Gillespie Pharsight Corporation Cary, North Carolina, USA PAGE 2003 Verona,

More information

Lecture 10: Introduction to Logistic Regression

Lecture 10: Introduction to Logistic Regression Lecture 10: Introduction to Logistic Regression Ani Manichaikul amanicha@jhsph.edu 2 May 2007 Logistic Regression Regression for a response variable that follows a binomial distribution Recall the binomial

More information

ST 790, Homework 1 Spring 2017

ST 790, Homework 1 Spring 2017 ST 790, Homework 1 Spring 2017 1. In EXAMPLE 1 of Chapter 1 of the notes, it is shown at the bottom of page 22 that the complete case estimator for the mean µ of an outcome Y given in (1.18) under MNAR

More information

Generalized Linear Models. Last time: Background & motivation for moving beyond linear

Generalized Linear Models. Last time: Background & motivation for moving beyond linear Generalized Linear Models Last time: Background & motivation for moving beyond linear regression - non-normal/non-linear cases, binary, categorical data Today s class: 1. Examples of count and ordered

More information

Time-Invariant Predictors in Longitudinal Models

Time-Invariant Predictors in Longitudinal Models Time-Invariant Predictors in Longitudinal Models Topics: What happens to missing predictors Effects of time-invariant predictors Fixed vs. systematically varying vs. random effects Model building strategies

More information

Bootstrapping Sensitivity Analysis

Bootstrapping Sensitivity Analysis Bootstrapping Sensitivity Analysis Qingyuan Zhao Department of Statistics, The Wharton School University of Pennsylvania May 23, 2018 @ ACIC Based on: Qingyuan Zhao, Dylan S. Small, and Bhaswar B. Bhattacharya.

More information

A Flexible Bayesian Approach to Monotone Missing. Data in Longitudinal Studies with Nonignorable. Missingness with Application to an Acute

A Flexible Bayesian Approach to Monotone Missing. Data in Longitudinal Studies with Nonignorable. Missingness with Application to an Acute A Flexible Bayesian Approach to Monotone Missing Data in Longitudinal Studies with Nonignorable Missingness with Application to an Acute Schizophrenia Clinical Trial Antonio R. Linero, Michael J. Daniels

More information

IP WEIGHTING AND MARGINAL STRUCTURAL MODELS (CHAPTER 12) BIOS IPW and MSM

IP WEIGHTING AND MARGINAL STRUCTURAL MODELS (CHAPTER 12) BIOS IPW and MSM IP WEIGHTING AND MARGINAL STRUCTURAL MODELS (CHAPTER 12) BIOS 776 1 12 IPW and MSM IP weighting and marginal structural models ( 12) Outline 12.1 The causal question 12.2 Estimating IP weights via modeling

More information

Combining Non-probability and Probability Survey Samples Through Mass Imputation

Combining Non-probability and Probability Survey Samples Through Mass Imputation Combining Non-probability and Probability Survey Samples Through Mass Imputation Jae-Kwang Kim 1 Iowa State University & KAIST October 27, 2018 1 Joint work with Seho Park, Yilin Chen, and Changbao Wu

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Figure 36: Respiratory infection versus time for the first 49 children.

Figure 36: Respiratory infection versus time for the first 49 children. y BINARY DATA MODELS We devote an entire chapter to binary data since such data are challenging, both in terms of modeling the dependence, and parameter interpretation. We again consider mixed effects

More information

Analyzing Pilot Studies with Missing Observations

Analyzing Pilot Studies with Missing Observations Analyzing Pilot Studies with Missing Observations Monnie McGee mmcgee@smu.edu. Department of Statistical Science Southern Methodist University, Dallas, Texas Co-authored with N. Bergasa (SUNY Downstate

More information

Estimating the Mean Response of Treatment Duration Regimes in an Observational Study. Anastasios A. Tsiatis.

Estimating the Mean Response of Treatment Duration Regimes in an Observational Study. Anastasios A. Tsiatis. Estimating the Mean Response of Treatment Duration Regimes in an Observational Study Anastasios A. Tsiatis http://www.stat.ncsu.edu/ tsiatis/ Introduction to Dynamic Treatment Regimes 1 Outline Description

More information

A weighted simulation-based estimator for incomplete longitudinal data models

A weighted simulation-based estimator for incomplete longitudinal data models To appear in Statistics and Probability Letters, 113 (2016), 16-22. doi 10.1016/j.spl.2016.02.004 A weighted simulation-based estimator for incomplete longitudinal data models Daniel H. Li 1 and Liqun

More information

Monte Carlo Studies. The response in a Monte Carlo study is a random variable.

Monte Carlo Studies. The response in a Monte Carlo study is a random variable. Monte Carlo Studies The response in a Monte Carlo study is a random variable. The response in a Monte Carlo study has a variance that comes from the variance of the stochastic elements in the data-generating

More information

Global Sensitivity Analysis of Randomized Trials with Missing Data

Global Sensitivity Analysis of Randomized Trials with Missing Data Global Sensitivity Analysis of Randomized Trials with Missing Data FDA Shortcourse Daniel Scharfstein Johns Hopkins University dscharf@jhu.edu May 8, 2017 1 / 149 Funding Acknowledgments FDA PCORI 2 /

More information

Analysis of Incomplete Non-Normal Longitudinal Lipid Data

Analysis of Incomplete Non-Normal Longitudinal Lipid Data Analysis of Incomplete Non-Normal Longitudinal Lipid Data Jiajun Liu*, Devan V. Mehrotra, Xiaoming Li, and Kaifeng Lu 2 Merck Research Laboratories, PA/NJ 2 Forrest Laboratories, NY *jiajun_liu@merck.com

More information

Introduction to mtm: An R Package for Marginalized Transition Models

Introduction to mtm: An R Package for Marginalized Transition Models Introduction to mtm: An R Package for Marginalized Transition Models Bryan A. Comstock and Patrick J. Heagerty Department of Biostatistics University of Washington 1 Introduction Marginalized transition

More information

Lecture 14: Introduction to Poisson Regression

Lecture 14: Introduction to Poisson Regression Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu 8 May 2007 1 / 52 Overview Modelling counts Contingency tables Poisson regression models 2 / 52 Modelling counts I Why

More information

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview

Modelling counts. Lecture 14: Introduction to Poisson Regression. Overview Modelling counts I Lecture 14: Introduction to Poisson Regression Ani Manichaikul amanicha@jhsph.edu Why count data? Number of traffic accidents per day Mortality counts in a given neighborhood, per week

More information

Mixed-Effects Pattern-Mixture Models for Incomplete Longitudinal Data. Don Hedeker University of Illinois at Chicago

Mixed-Effects Pattern-Mixture Models for Incomplete Longitudinal Data. Don Hedeker University of Illinois at Chicago Mixed-Effects Pattern-Mixture Models for Incomplete Longitudinal Data Don Hedeker University of Illinois at Chicago This work was supported by National Institute of Mental Health Contract N44MH32056. 1

More information

Advanced Quantitative Methods: limited dependent variables

Advanced Quantitative Methods: limited dependent variables Advanced Quantitative Methods: Limited Dependent Variables I University College Dublin 2 April 2013 1 2 3 4 5 Outline Model Measurement levels 1 2 3 4 5 Components Model Measurement levels Two components

More information

Potential Outcomes Model (POM)

Potential Outcomes Model (POM) Potential Outcomes Model (POM) Relationship Between Counterfactual States Causality Empirical Strategies in Labor Economics, Angrist Krueger (1999): The most challenging empirical questions in economics

More information

Methods for inferring short- and long-term effects of exposures on outcomes, using longitudinal data on both measures

Methods for inferring short- and long-term effects of exposures on outcomes, using longitudinal data on both measures Methods for inferring short- and long-term effects of exposures on outcomes, using longitudinal data on both measures Ruth Keogh, Stijn Vansteelandt, Rhian Daniel Department of Medical Statistics London

More information

Hypothesis testing, part 2. With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal

Hypothesis testing, part 2. With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal Hypothesis testing, part 2 With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal 1 CATEGORICAL IV, NUMERIC DV 2 Independent samples, one IV # Conditions Normal/Parametric Non-parametric

More information

Longitudinal analysis of ordinal data

Longitudinal analysis of ordinal data Longitudinal analysis of ordinal data A report on the external research project with ULg Anne-Françoise Donneau, Murielle Mauer June 30 th 2009 Generalized Estimating Equations (Liang and Zeger, 1986)

More information

Integrated approaches for analysis of cluster randomised trials

Integrated approaches for analysis of cluster randomised trials Integrated approaches for analysis of cluster randomised trials Invited Session 4.1 - Recent developments in CRTs Joint work with L. Turner, F. Li, J. Gallis and D. Murray Mélanie PRAGUE - SCT 2017 - Liverpool

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

Combining multiple observational data sources to estimate causal eects

Combining multiple observational data sources to estimate causal eects Department of Statistics, North Carolina State University Combining multiple observational data sources to estimate causal eects Shu Yang* syang24@ncsuedu Joint work with Peng Ding UC Berkeley May 23,

More information

Introduction to Statistical Analysis

Introduction to Statistical Analysis Introduction to Statistical Analysis Changyu Shen Richard A. and Susan F. Smith Center for Outcomes Research in Cardiology Beth Israel Deaconess Medical Center Harvard Medical School Objectives Descriptive

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Causal Inference with General Treatment Regimes: Generalizing the Propensity Score

Causal Inference with General Treatment Regimes: Generalizing the Propensity Score Causal Inference with General Treatment Regimes: Generalizing the Propensity Score David van Dyk Department of Statistics, University of California, Irvine vandyk@stat.harvard.edu Joint work with Kosuke

More information

Lecture Outline. Biost 518 Applied Biostatistics II. Choice of Model for Analysis. Choice of Model. Choice of Model. Lecture 10: Multiple Regression:

Lecture Outline. Biost 518 Applied Biostatistics II. Choice of Model for Analysis. Choice of Model. Choice of Model. Lecture 10: Multiple Regression: Biost 518 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture utline Choice of Model Alternative Models Effect of data driven selection of

More information

MISSING or INCOMPLETE DATA

MISSING or INCOMPLETE DATA MISSING or INCOMPLETE DATA A (fairly) complete review of basic practice Don McLeish and Cyntha Struthers University of Waterloo Dec 5, 2015 Structure of the Workshop Session 1 Common methods for dealing

More information

Modelling Dropouts by Conditional Distribution, a Copula-Based Approach

Modelling Dropouts by Conditional Distribution, a Copula-Based Approach The 8th Tartu Conference on MULTIVARIATE STATISTICS, The 6th Conference on MULTIVARIATE DISTRIBUTIONS with Fixed Marginals Modelling Dropouts by Conditional Distribution, a Copula-Based Approach Ene Käärik

More information

Global Sensitivity Analysis of Randomized Trials with Missing Data

Global Sensitivity Analysis of Randomized Trials with Missing Data Global Sensitivity Analysis of Randomized Trials with Missing Data ASA Webinar Daniel Scharfstein Johns Hopkins University dscharf@jhu.edu September 20, 2016 1 / 163 Funding Acknowledgments FDA PCORI 2

More information

Analysis of Longitudinal Data. Patrick J. Heagerty PhD Department of Biostatistics University of Washington

Analysis of Longitudinal Data. Patrick J. Heagerty PhD Department of Biostatistics University of Washington Analysis of Longitudinal Data Patrick J Heagerty PhD Department of Biostatistics University of Washington Auckland 8 Session One Outline Examples of longitudinal data Scientific motivation Opportunities

More information

Chapter 11. Correlation and Regression

Chapter 11. Correlation and Regression Chapter 11. Correlation and Regression The word correlation is used in everyday life to denote some form of association. We might say that we have noticed a correlation between foggy days and attacks of

More information

Time Invariant Predictors in Longitudinal Models

Time Invariant Predictors in Longitudinal Models Time Invariant Predictors in Longitudinal Models Longitudinal Data Analysis Workshop Section 9 University of Georgia: Institute for Interdisciplinary Research in Education and Human Development Section

More information

A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data

A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data Alexina Mason, Sylvia Richardson and Nicky Best Department of Epidemiology and Biostatistics, Imperial College

More information

6. Fractional Imputation in Survey Sampling

6. Fractional Imputation in Survey Sampling 6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Statistics Part A Comprehensive Exam Methodology Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Statistics Part A Comprehensive Exam Methodology Paper Student Name: ID: McGill University Faculty of Science Department of Mathematics and Statistics Statistics Part A Comprehensive Exam Methodology Paper Date: Friday, May 13, 2016 Time: 13:00 17:00 Instructions

More information

(3) Review of Probability. ST440/540: Applied Bayesian Statistics

(3) Review of Probability. ST440/540: Applied Bayesian Statistics Review of probability The crux of Bayesian statistics is to compute the posterior distribution, i.e., the uncertainty distribution of the parameters (θ) after observing the data (Y) This is the conditional

More information

multilevel modeling: concepts, applications and interpretations

multilevel modeling: concepts, applications and interpretations multilevel modeling: concepts, applications and interpretations lynne c. messer 27 october 2010 warning social and reproductive / perinatal epidemiologist concepts why context matters multilevel models

More information

Summary and discussion of The central role of the propensity score in observational studies for causal effects

Summary and discussion of The central role of the propensity score in observational studies for causal effects Summary and discussion of The central role of the propensity score in observational studies for causal effects Statistics Journal Club, 36-825 Jessica Chemali and Michael Vespe 1 Summary 1.1 Background

More information

Lecture 3.1 Basic Logistic LDA

Lecture 3.1 Basic Logistic LDA y Lecture.1 Basic Logistic LDA 0.2.4.6.8 1 Outline Quick Refresher on Ordinary Logistic Regression and Stata Women s employment example Cross-Over Trial LDA Example -100-50 0 50 100 -- Longitudinal Data

More information

Causal Inference in Observational Studies with Non-Binary Treatments. David A. van Dyk

Causal Inference in Observational Studies with Non-Binary Treatments. David A. van Dyk Causal Inference in Observational Studies with Non-Binary reatments Statistics Section, Imperial College London Joint work with Shandong Zhao and Kosuke Imai Cass Business School, October 2013 Outline

More information

Chapter 4. Parametric Approach. 4.1 Introduction

Chapter 4. Parametric Approach. 4.1 Introduction Chapter 4 Parametric Approach 4.1 Introduction The missing data problem is already a classical problem that has not been yet solved satisfactorily. This problem includes those situations where the dependent

More information

7/28/15. Review Homework. Overview. Lecture 6: Logistic Regression Analysis

7/28/15. Review Homework. Overview. Lecture 6: Logistic Regression Analysis Lecture 6: Logistic Regression Analysis Christopher S. Hollenbeak, PhD Jane R. Schubart, PhD The Outcomes Research Toolbox Review Homework 2 Overview Logistic regression model conceptually Logistic regression

More information

ECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam

ECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam ECLT 5810 Linear Regression and Logistic Regression for Classification Prof. Wai Lam Linear Regression Models Least Squares Input vectors is an attribute / feature / predictor (independent variable) The

More information

Latent Variable Model for Weight Gain Prevention Data with Informative Intermittent Missingness

Latent Variable Model for Weight Gain Prevention Data with Informative Intermittent Missingness Journal of Modern Applied Statistical Methods Volume 15 Issue 2 Article 36 11-1-2016 Latent Variable Model for Weight Gain Prevention Data with Informative Intermittent Missingness Li Qin Yale University,

More information

Alexina Mason. Department of Epidemiology and Biostatistics Imperial College, London. 16 February 2010

Alexina Mason. Department of Epidemiology and Biostatistics Imperial College, London. 16 February 2010 Strategy for modelling non-random missing data mechanisms in longitudinal studies using Bayesian methods: application to income data from the Millennium Cohort Study Alexina Mason Department of Epidemiology

More information

Logistic Regression - problem 6.14

Logistic Regression - problem 6.14 Logistic Regression - problem 6.14 Let x 1, x 2,, x m be given values of an input variable x and let Y 1,, Y m be independent binomial random variables whose distributions depend on the corresponding values

More information

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M.

CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M. CRE METHODS FOR UNBALANCED PANELS Correlated Random Effects Panel Data Models IZA Summer School in Labor Economics May 13-19, 2013 Jeffrey M. Wooldridge Michigan State University 1. Introduction 2. Linear

More information

Lab 3: Two levels Poisson models (taken from Multilevel and Longitudinal Modeling Using Stata, p )

Lab 3: Two levels Poisson models (taken from Multilevel and Longitudinal Modeling Using Stata, p ) Lab 3: Two levels Poisson models (taken from Multilevel and Longitudinal Modeling Using Stata, p. 376-390) BIO656 2009 Goal: To see if a major health-care reform which took place in 1997 in Germany was

More information

E(Y ij b i ) = f(x ijβ i ), (13.1) β i = A i β + B i b i. (13.2)

E(Y ij b i ) = f(x ijβ i ), (13.1) β i = A i β + B i b i. (13.2) 1 Advanced topics 1.1 Introduction In this chapter, we conclude with brief overviews of several advanced topics. Each of these topics could realistically be the subject of an entire course! 1. Generalized

More information

Structural Nested Mean Models for Assessing Time-Varying Effect Moderation. Daniel Almirall

Structural Nested Mean Models for Assessing Time-Varying Effect Moderation. Daniel Almirall 1 Structural Nested Mean Models for Assessing Time-Varying Effect Moderation Daniel Almirall Center for Health Services Research, Durham VAMC & Dept. of Biostatistics, Duke University Medical Joint work

More information

9. Linear Regression and Correlation

9. Linear Regression and Correlation 9. Linear Regression and Correlation Data: y a quantitative response variable x a quantitative explanatory variable (Chap. 8: Recall that both variables were categorical) For example, y = annual income,

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model EPSY 905: Multivariate Analysis Lecture 1 20 January 2016 EPSY 905: Lecture 1 -

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

,..., θ(2),..., θ(n)

,..., θ(2),..., θ(n) Likelihoods for Multivariate Binary Data Log-Linear Model We have 2 n 1 distinct probabilities, but we wish to consider formulations that allow more parsimonious descriptions as a function of covariates.

More information

BIOS 2083: Linear Models

BIOS 2083: Linear Models BIOS 2083: Linear Models Abdus S Wahed September 2, 2009 Chapter 0 2 Chapter 1 Introduction to linear models 1.1 Linear Models: Definition and Examples Example 1.1.1. Estimating the mean of a N(μ, σ 2

More information

Models for binary data

Models for binary data Faculty of Health Sciences Models for binary data Analysis of repeated measurements 2015 Julie Lyng Forman & Lene Theil Skovgaard Department of Biostatistics, University of Copenhagen 1 / 63 Program for

More information

Comparison of methods for repeated measures binary data with missing values. Farhood Mohammadi. A thesis submitted in partial fulfillment of the

Comparison of methods for repeated measures binary data with missing values. Farhood Mohammadi. A thesis submitted in partial fulfillment of the Comparison of methods for repeated measures binary data with missing values by Farhood Mohammadi A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Biostatistics

More information

Discussion on Fygenson (2007, Statistica Sinica): a DS Perspective

Discussion on Fygenson (2007, Statistica Sinica): a DS Perspective 1 Discussion on Fygenson (2007, Statistica Sinica): a DS Perspective Chuanhai Liu Purdue University 1. Introduction In statistical analysis, it is important to discuss both uncertainty due to model choice

More information

Exam Applied Statistical Regression. Good Luck!

Exam Applied Statistical Regression. Good Luck! Dr. M. Dettling Summer 2011 Exam Applied Statistical Regression Approved: Tables: Note: Any written material, calculator (without communication facility). Attached. All tests have to be done at the 5%-level.

More information

Today. HW 1: due February 4, pm. Aspects of Design CD Chapter 2. Continue with Chapter 2 of ELM. In the News:

Today. HW 1: due February 4, pm. Aspects of Design CD Chapter 2. Continue with Chapter 2 of ELM. In the News: Today HW 1: due February 4, 11.59 pm. Aspects of Design CD Chapter 2 Continue with Chapter 2 of ELM In the News: STA 2201: Applied Statistics II January 14, 2015 1/35 Recap: data on proportions data: y

More information

An Empirical Comparison of Multiple Imputation Approaches for Treating Missing Data in Observational Studies

An Empirical Comparison of Multiple Imputation Approaches for Treating Missing Data in Observational Studies Paper 177-2015 An Empirical Comparison of Multiple Imputation Approaches for Treating Missing Data in Observational Studies Yan Wang, Seang-Hwane Joo, Patricia Rodríguez de Gil, Jeffrey D. Kromrey, Rheta

More information

Probability and Information Theory. Sargur N. Srihari

Probability and Information Theory. Sargur N. Srihari Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal

More information

STA 4504/5503 Sample Exam 1 Spring 2011 Categorical Data Analysis. 1. Indicate whether each of the following is true (T) or false (F).

STA 4504/5503 Sample Exam 1 Spring 2011 Categorical Data Analysis. 1. Indicate whether each of the following is true (T) or false (F). STA 4504/5503 Sample Exam 1 Spring 2011 Categorical Data Analysis 1. Indicate whether each of the following is true (T) or false (F). (a) T In 2 2 tables, statistical independence is equivalent to a population

More information