methods for handling missing data
play

Methods for Handling Missing Data Joseph Hogan Brown University - PowerPoint PPT Presentation

Methods for Handling Missing Data Joseph Hogan Brown University MDEpiNet Conference Workshop October 22, 2018 Hogan (MDEpiNet) Missing Data October 22, 2018 1 / 160 Course Overview I 1 Introduction and Background Introduce case studies


  1. MDM for univariate sampling Missing values of Y are missing completely at random (MCAR) if R ⊥ ⊥ Y , or equivalently if f ( r | y ) = f ( r ) For univariate samples, this is also classified as missing at random (MAR). More on this distinction later. Hogan (MDEpiNet) Missing Data October 22, 2018 28 / 160

  2. MDM for univariate sampling Missing values of Y are missing not at random (MNAR) if there exists at least one value of y such that f ( r | y ) � = f ( r ) Or in words, if the probability of response is systematically higher/lower for particular values of y . Hogan (MDEpiNet) Missing Data October 22, 2018 29 / 160

  3. MDM for univariate sampling Under MAR, methods applied to the observed data only will generally yield valid inferences about the population. ◮ Estimates will be consistent ◮ Standard errors may be larger than if you had the full data Under MNAR, methods applied to observed data only generally will not yield valid inferences Hogan (MDEpiNet) Missing Data October 22, 2018 30 / 160

  4. MDM for multivariate sampling – regression Consider the setting where we are interested in the regression of Y on X . Let µ ( X ) = E ( Y | X ). Assume there are no other covariates available The model we have in mind is g { µ ( X ) } = β 0 + β 1 X Here, the function g tells you the type of regression model you are fitting (linear, logistic, etc.) The full data are ( Y 1 , X 1 , R 1 ) , ( Y 2 , X 2 , R 2 ) , . . . , ( Y n , X n , R n ) Hogan (MDEpiNet) Missing Data October 22, 2018 31 / 160

  5. MDM for multivariate sampling – regression In cases like this, we can define missing data mechanisms relative to the objective of inference. The Y ’s are missing at random if Y ⊥ ⊥ R | X In words, the MDM is a random deletion mechanism within distinct levels of X Another way to write this: f ( r | y , x ) = f ( r | x ) The deletion mechanism depends on X , but within levels of X it does not depend on Y Hogan (MDEpiNet) Missing Data October 22, 2018 32 / 160

  6. Examples of MAR in regression Let Y denote blood pressure, X denote gender (1 = F, 0 = M). The regression model of interest is E ( Y | X ) = β 0 + β 1 X , so that β 0 = mean BP among men, β 1 = mean difference. Let’s assume men have higher BP on average. Randomly delete BP for 20% of men and 40% of women. R does depend on Y , but only through X ◮ Men have higher BP ◮ Men less likely to be deleted ◮ ⇒ those with higher BP less likely to be deleted Within levels of X , deletion mechanism is completely random. Hogan (MDEpiNet) Missing Data October 22, 2018 33 / 160

  7. MAR in regression – some practical issues Revisit the MAR condition. If R ⊥ ⊥ Y | X , this also means f ( y | x , r ) = f ( y | x ) , or f ( y | x , r = 1) = f ( y | x , r = 0) The relationship between Y and X is the same whether R = 1 or R = 0. Consequence is that a regression fit to those with R = 1 gives valid estimates of regression parameters. (Standard errors will be higher relative to having all the data.) Hogan (MDEpiNet) Missing Data October 22, 2018 34 / 160

  8. MAR in regression – some practical issues Under MAR, the inferences are still valid even if ◮ The X distribution is different between those with missing and observed Y ’s Question you have to ask to (subjectively) assess MAR: Is the missing data mechanism a random deletion of Y ’s among people who have the same X values? Equivalent formulation of this question: Is the relationship between X and Y the same among those with missing and observed Y values? Hogan (MDEpiNet) Missing Data October 22, 2018 35 / 160

  9. MAR for regression when auxiliary variables are available In some cases we have information on more than just the X variables In a clinical trial we may be interested in E ( Y | X ) when X is a treatment group, but we have collected lots of baseline covariates V . In a longitudinal study, we may be interested in the mean outcome at the last measurement time, but we have accumulated information on the outcome at previous measurement times. When auxiliary information is available, we can use it in some cases to make MAR more plausible. Here MAR has a slightly different formulation. Hogan (MDEpiNet) Missing Data October 22, 2018 36 / 160

  10. MAR with auxiliary covariates The relationship of interest is g { µ ( X ) } = β 0 + β 1 X The full data are ( Y 1 , X 1 , V 1 , R 1 ) , ( Y 2 , X 2 , V 2 , R 2 ) , . . . , ( Y n , X n , V n , R n ) where Y is observed when R = 1 and is missing when R = 0. Hogan (MDEpiNet) Missing Data October 22, 2018 37 / 160

  11. MAR with auxiliary covariates Values of Y are missing at random (MAR) if Y ⊥ ⊥ R | ( X , V ) Two equivalent ways to write this are: f ( r | x , v , y ) = f ( r | x , v ) f ( y | x , v , r = 1) = f ( y | x , v , r = 0) The first says that within distinct levels defined by ( X , V ), missingness in Y is a random deletion mechanism The second says that the relationship between Y and ( X , V ) is the same whether Y is missing or not Hogan (MDEpiNet) Missing Data October 22, 2018 38 / 160

  12. MAR with auxiliaries – example Return to our BP example, but now assume V denotes income level. Recall, we are interested in the coefficient β 1 from E ( Y | X ) = β 0 + β 1 X and not the coefficient α 1 from E ( Y | X , V ) = α 0 + α 1 X + α 2 V Hogan (MDEpiNet) Missing Data October 22, 2018 39 / 160

  13. Missing data mechanisms for longitudinal data Need to define some notation for longitudinal data Y j = value of Y at time j R j = 1 if Y j observed, 0 otherwise = ( Y 1 , Y 2 , . . . , Y j ) Y j = outcome history up to time j = covariate history up to time j X j H j = ( X j , Y j − 1 ) Allows us to define MCAR, MAR, MNAR for longitudinal data Hogan (MDEpiNet) Missing Data October 22, 2018 40 / 160

  14. Missing data mechanisms for longitudinal data MAR Missing at random (MAR) If interest is in marginal means such as E ( Y j ), MAR means R j ⊥ ⊥ Y j | ( R j − 1 = 1 , H j ) Interpretation: ◮ Among those in follow up at time j, missingness is independent of the outcome Y j conditional on the previously observed Y ’s . ◮ Missingness does not depend on present or future Y ’s, given the past. Hogan (MDEpiNet) Missing Data October 22, 2018 41 / 160

  15. Missing data mechanisms for longitudinal data MAR Implications 1 Selection mechanism [ R j | R j − 1 = 1 , H J ] = [ R j | R j − 1 = 1 , H j ] ◮ Can model selection probability as a function of observed past 2 Imputation mechanism [ Y j | R j = 0 , R j − 1 = 1 , H j ] = [ Y j | R j = 1 , R j − 1 = 1 , H j ] ◮ Can impute missing Y j using a model of the observed Y j ◮ Critical: Must correctly specify observed-data model Hogan (MDEpiNet) Missing Data October 22, 2018 42 / 160

  16. LOCF We can characterize LOCF in this framework It is an imputation mechanism ◮ Missing Y j is equal to the most recently observed value of Y ◮ Missing value filled in with probability one (no variance) Formally Y j ∗ with probability one , [ Y j | R j = 0 , H j ] = where j ∗ = max k < j { R k = 1 } Not an MAR mechanism in general ◮ Conditional distribution of missing Y j not equal to that for observed Y j Hogan (MDEpiNet) Missing Data October 22, 2018 43 / 160

  17. Random effects and parametric models Assume a joint distribution for the repeated measures Model applies to the full data , hence cannot be checked Example: Multivariate normal ( Y 1 , . . . , Y J ) T ∼ N ( µ, Σ) , where µ J × 1 = E ( Y ) and Σ J × J = var( Y ) Special case: Random effects model ◮ Particular way of structuring mean and variance Hogan (MDEpiNet) Missing Data October 22, 2018 44 / 160

  18. Random effects and parametric models When do these models yield valid inference? Most parametric models valid if ◮ MAR holds ◮ All parts of the model are correctly specified These models have an implied distribution for the conditionals [ Y j | Y 1 , . . . , Y j − 1 ] Under MAR, the implied distribution applies to those with complete and incomplete data, but .... Parametric assumptions cannot be checked empirically Hogan (MDEpiNet) Missing Data October 22, 2018 45 / 160

  19. GEE I Assume a mean and variance structure for the repeated measures ◮ Not necessarily a full parametric model Assumed variance structure is the ‘working covariance’ With complete data ◮ Inferences are most efficient when covariance correctly specified ◮ Correct inference about time-specific means even if covariance mis-specified ◮ Reason: all information about time-specific means is already observed Hogan (MDEpiNet) Missing Data October 22, 2018 46 / 160

  20. GEE II With incomplete data ◮ Information about time-specific means relies on ‘imputation’ of missing observations ◮ These imputations come from the conditional distribution [ Y j | Y 1 , . . . , Y j − 1 ] ◮ The form of the conditional distribution depends on the working covariance Implication: Correct inference about time-specific means only when both mean and covariance are correctly specified ◮ Can get different treatment effects with different working covariances Hogan (MDEpiNet) Missing Data October 22, 2018 47 / 160

  21. Dependence of estimates on working covariance From Hogan et al., 2004 Statistics in Medicine Hogan (MDEpiNet) Missing Data October 22, 2018 48 / 160

  22. Structure of case studies 1 Introduce modeling approach 2 Relate modeling approach to missing data hierarchy 3 Illustrate on simple cases 4 Include a treatment comparison 5 Discussion of key points from case study Hogan (MDEpiNet) Missing Data October 22, 2018 49 / 160

  23. CASE STUDY I: MIXTURE MODEL ANALYSIS OF GROWTH HORMONE TRIAL Hogan (MDEpiNet) Missing Data October 22, 2018 50 / 160

  24. Outline of analysis Objective: Compare EG to EP at month 12 ◮ Variable: Y 3 Estimation of E ( Y 3 ) for EG arm only ◮ Ignoring baseline covariates ◮ Using information from baseline covariate Y 1 ◮ MAR and MNAR (sensitivity analysis) Treatment comparisons ◮ Expand to longitudinal case ◮ MAR – using regression imputation ◮ MNAR – sensitivity analysis Hogan (MDEpiNet) Missing Data October 22, 2018 51 / 160

  25. Estimate E ( Y ) from univariate sample Y3 R [1,] NA 0 [2,] 82.6 1 [3,] 118.7 1 [4,] 99.6 1 [5,] NA 0 [6,] 57.5 1 [7,] NA 0 [8,] NA 0 [9,] 87.5 1 [10,] 97.4 1 [11,] 21.4 1 [12,] 47.2 1 [13,] NA 0 [14,] 68.6 1 Hogan (MDEpiNet) Missing Data October 22, 2018 52 / 160

  26. Estimating E ( Y ) from univariate sample Model: E ( Y | R = 1) = µ 1 E ( Y | R = 0) = µ 0 (not identifiable) Target of estimation E ( Y ) = µ 1 P ( R = 1) + µ 0 P ( R = 0) Question: what to assume about µ 0 ? In a sense, we are going to impute a value for µ 0 , or impute values of the missing Y ’s that will lead to an estimate of µ 0 . Hogan (MDEpiNet) Missing Data October 22, 2018 53 / 160

  27. Parameterizing departures from MAR Target of estimation: E µ 0 ( Y ) = µ 1 P ( R = 1) + µ 0 P ( R = 0) = µ 1 + ( µ 0 − µ 1 ) P ( R = 0) Suggests sensitivity parameter ∆( µ 0 ) = µ 0 − µ 1 . Leads to E ∆ ( Y ) = µ 1 + ∆ P ( R = 0) Features of this format: Centered at MAR (∆ = 0) ∆ cannot be estimated from observed data Can vary ∆ for sensitivity analysis Allows Bayesian approach by placing prior on ∆ Hogan (MDEpiNet) Missing Data October 22, 2018 54 / 160

  28. Estimation under MNAR Recall model E ∆ ( Y ) = µ 1 + ∆ P ( R = 0) Estimate known quantities � n 1 = i R i � n 0 = i (1 − R i ) � P ( R = 0) = n 0 / ( n 1 + n 0 ) (1 / n 1 ) � µ 1 � = i Y i Have one unknown quantity ∆ = µ 0 − � µ 1 Hogan (MDEpiNet) Missing Data October 22, 2018 55 / 160

  29. Estimation under MAR Plug into model � � E ∆ ( Y ) = � µ 1 + ( µ 0 − � µ 1 ) P ( R = 0) � �� � � = � µ 1 + ∆ P ( R = 0) Interpretation Under MAR (∆ = 0), ◮ � E ∆ ( Y ) = � µ 1 ◮ Estimator is the observed-data mean Under MNAR (∆ � = 0), ◮ Shift observed-dta mean by ∆ P ( R = 0) ◮ Shift is proportional to fraction of missing observations Hogan (MDEpiNet) Missing Data October 22, 2018 56 / 160

  30. 105 100 95 y3.delta 90 85 80 75 70 -10 -5 0 5 10 delta Y [ R =1] = 88 . 3 � P ( R = 0) = 0 . 42 Hogan (MDEpiNet) Missing Data October 22, 2018 57 / 160

  31. Using information from baseline covariates Y1 Y3 R [1,] 35.0 NA 0 [2,] 74.5 82.6 1 [3,] 120.2 118.7 1 [4,] 84.8 99.6 1 [5,] 68.6 NA 0 [6,] 47.9 57.5 1 [7,] 39.0 NA 0 [8,] 52.0 NA 0 [9,] 92.9 87.5 1 [10,] 98.8 97.4 1 [11,] 48.6 21.4 1 [12,] 45.7 47.2 1 [13,] 63.4 NA 0 [14,] 64.2 68.6 1 Hogan (MDEpiNet) Missing Data October 22, 2018 58 / 160

  32. General model for Y Objective: Inference for E ( Y 3 ) Model – General form [ Y 3 | Y 1 , R = 1] ∼ F 1 ( y 3 | y 1 ) [ Y 3 | Y 1 , R = 0] ∼ F 0 ( y 3 | y 1 ) The general form encompasses all possible models that can be assumed for the observed and missing values of Y 3 The model F 0 cannot be estimated from data Hogan (MDEpiNet) Missing Data October 22, 2018 59 / 160

  33. The model under MAR and MNAR Under MAR, F 0 = F 1 . Suggests following strategy ◮ Fit model for F 1 using observed data; call it � F 1 ◮ Use this model to impute missing values of Y 3 Under MNAR, F 0 � = F 1 . Suggests following strategy ◮ Parameterize a model so that F 0 is related to F 1 through a sensitivity parameter ∆ ◮ Generically write this as F 0 = F ∆ 1 ◮ Use fitted version of F ∆ 1 to impute missing Y 3 Hogan (MDEpiNet) Missing Data October 22, 2018 60 / 160

  34. Regression parameterization of F 1 and F 0 Take the case of MAR first MAR implies [ Y 3 | Y 1 , R = 1] = [ Y 3 | Y 1 , R = 0] Assume regression model for [ Y | X , R = 1] E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 Assume a model for [ Y | X , R = 0] of similar form E ( Y 3 | Y 1 , R = 0) = α 0 + β 0 Y 1 Cannot estimate parameters from observed data Hogan (MDEpiNet) Missing Data October 22, 2018 61 / 160

  35. Regression parameterization of F 1 and F 0 Recall model E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 E ( Y 3 | Y 1 , R = 0) = α 0 + β 0 Y 1 Link the models. One way to do this: β 0 = β 1 + ∆ β α 0 = α 1 + ∆ α Under MAR: ∆ α = ∆ β = 0 Hogan (MDEpiNet) Missing Data October 22, 2018 62 / 160

  36. Caveats to using this (or any!) approach Recall model E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 E ( Y 3 | Y 1 , R = 0) = α 0 + β 0 Y 1 More general version of missing-data model E ( Y 3 | Y 1 , R = 0) = g ( Y 1 ; θ ) Do we know the form of g ? Do we know the value of θ ? Do we know that Y 1 is sufficient to predict Y 3 ? We are assuming we know all of these things. Hogan (MDEpiNet) Missing Data October 22, 2018 63 / 160

  37. Estimation of E ( Y 3 ) under MAR 1 Fit the model E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 α 1 , � ⇒ obtain � β 1 . 2 For those with R = 0, impute predicted value via � � Y 3 i = E ( Y 3 | Y 1 i , R i = 0) α 1 + � = � β 1 Y 1 i 3 Estimate overall mean as the mixture � � R i Y 3 i + (1 − R i ) � E ( Y 3 ) = (1 / n ) Y 3 i i Hogan (MDEpiNet) Missing Data October 22, 2018 64 / 160

  38. Regression imputation under MAR 140 + 120 100 + y3[r == 1] + + 80 + + + + + + 60 + + + + 40 20 60 80 100 120 y1[r == 1] Hogan (MDEpiNet) Missing Data October 22, 2018 65 / 160

  39. Sample means and imputed means under MAR n r Y 1 Y 3 R = 0 16 57 67 R = 1 22 78 88 MAR 38 79 Hogan (MDEpiNet) Missing Data October 22, 2018 66 / 160

  40. Some intuition behind this (simple) estimator Could base estimate purely on regression model E ( Y 3 ) = E Y 1 , R { E ( Y 3 | Y 1 , R ) } � � = E R E Y 1 | R { E ( Y 3 | Y 1 , R ) } = E R [ α 1 + β 1 E ( Y 1 | R )] = α 1 + β 1 E ( Y 1 ) Plug in the estimators for each term May be more efficient when regression model is correct Hogan (MDEpiNet) Missing Data October 22, 2018 67 / 160

  41. Some more details ... If we don’t want to use the regression model for Y 3 , can write E ( Y 3 ) = E ( Y 3 | R = 1) P ( R = 1) + E ( Y 3 | R = 0) P ( R = 0) , where the term in red is E ( Y 3 | R = 0) = E Y 1 | R =0 { E ( Y 3 | Y 1 , R = 0) } = E Y 1 | R =0 ( α 1 + β 1 Y 1 | R = 0) = α 1 + β 1 E ( Y 1 | R = 0) Hence � � [ R =1] [ R =0] � ˆ α 1 + � ˆ E ( Y 3 ) = Y P ( R = 1) + � β 1 Y P ( R = 0) 3 1 Hogan (MDEpiNet) Missing Data October 22, 2018 68 / 160

  42. Inference and treatment comparisons SE and CI: bootstrap ◮ Draw bootstrap sample ◮ Carry out imputation procedure ◮ Repeat for lots of bootstrap samples (say B ) ◮ Base SE and CI on the B bootstrapped estimators Why not multiple imputation? ◮ Estimators are linear ◮ Bootstrap takes care of missing data uncertainty here Treatment comparisons – coming later Hogan (MDEpiNet) Missing Data October 22, 2018 69 / 160

  43. Estimation of E ( Y 3 ) under MNAR Recall model E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 E ( Y 3 | Y 1 , R = 0) = α 0 + β 0 Y 1 More general version of model E ( Y 3 | Y 1 , R = 1) = g 1 ( Y 1 ; θ 1 ) E ( Y 3 | Y 1 , R = 0) = g 0 ( Y 1 ; θ 1 , ∆) = h { g 1 ( Y 1 ; θ 1 ) , ∆ } The h function relates missing-data and observed-data models. User needs to specify form of h The parameter ∆ should not be estimable from observed data Vary ∆ in a sensitivity analysis Hogan (MDEpiNet) Missing Data October 22, 2018 70 / 160

  44. Regression-based specification under MNAR Specify observed-data model E ( Y 3 | Y 1 , R = 1) = g 1 ( Y 1 , θ 1 ) = α 1 + β 1 Y 1 Specify missing-data model E ( Y 3 | Y 1 , R = 0) = h { g 1 ( Y 1 , θ 1 ) , ∆ } = ∆ + g 1 ( Y 1 , θ 1 ) = ∆ + ( α 1 + β 1 Y 1 ) Many other choices are possible Here, add a constant to the MAR imputation Have MAR when ∆ = 0, and MNAR otherwise Hogan (MDEpiNet) Missing Data October 22, 2018 71 / 160

  45. Estimation of E ( Y 3 ) under MNAR 1 Fit the model E ( Y 3 | Y 1 , R = 1) = α 1 + β 1 Y 1 α 1 , � ⇒ obtain � β 1 . 2 For those with R = 0, impute predicted value via � � Y 3 i = E ( Y 3 | Y 1 i , R i = 0) α 1 + � = ∆ + � β 1 Y 1 i 3 Estimate overall mean as the mixture � � R i Y 3 i + (1 − R i ) � E ( Y 3 ) = (1 / n ) Y 3 i i � � [ R =1] [ R =0] ˆ α 1 + � ˆ = Y P ( R = 1) + ∆ + � β 1 Y P ( R = 0) 3 1 Hogan (MDEpiNet) Missing Data October 22, 2018 72 / 160

  46. Sensitivity analysis based on varying ∆ What should be the ‘anchor point’? ◮ Usually appropriate to anchor analysis at MAR ◮ Examine effect of MNAR by varying ∆ away from 0 How to select a range for ∆? ◮ Will always be specific to application. ◮ Ensure that range is appropriate to context (see upcoming example) ◮ Can use data-driven range for ∆, e.g. based on SD Reporting final inferences ◮ ‘Stress test’ approach ◮ Inverted sensitivity analysis — find values of ∆ that would change substantive conclusions ◮ Average over plausible ∆ values Hogan (MDEpiNet) Missing Data October 22, 2018 73 / 160

  47. Calibrating ∆ How should the range and scale of ∆ be chosen? Direction ◮ ∆ > 0 ⇒ dropouts have higher mean ◮ ∆ < 0 ⇒ dropouts have lower mean Range and scale ◮ Residual variation in outcome quantified by SD of regression error [ Y 3 | Y 1 , R = 1] = α 1 + β 1 Y 1 + e σ 2 var( e ) = ◮ Suggests scaling ∆ in units of σ ◮ Will illustrate in longitudinal case Hogan (MDEpiNet) Missing Data October 22, 2018 74 / 160

  48. Moving to longitudinal setting Set-up for single treatment arm Illustrate ideas with analysis of GH data ◮ Compare treatments ◮ Illustrate sensitivity analysis under MNAR ◮ Discuss how to report results Hogan (MDEpiNet) Missing Data October 22, 2018 75 / 160

  49. Longitudinal case: notation Assume missing data pattern is monotone K = dropout time � = R j j E k ( Y j ) = E ( Y j | K = k ) When j > k , cannot estimate E k ( Y j ) from data Hogan (MDEpiNet) Missing Data October 22, 2018 76 / 160

  50. Longitudinal model with J = 3: Set up E k ( Y 1 ) = E ( Y 1 | K = k ) identified for k = 1 , 2 , 3 For other means, we have j = 2 j = 3 K = 1 E 1 ( Y 2 | Y 1 ) E 1 ( Y 3 | Y 1 , Y 2 ) K = 2 E 2 ( Y 2 | Y 1 ) E 2 ( Y 3 | Y 1 , Y 2 ) K = 3 E 3 ( Y 2 | Y 1 ) E 3 ( Y 3 | Y 1 , Y 2 ) Components in red cannot be estimated. Need assumptions Hogan (MDEpiNet) Missing Data October 22, 2018 77 / 160

  51. Longitudinal model with J = 3: MAR j = 2 j = 3 E 3 ( Y 3 | Y 1 , � K = 1 ω E 2 ( Y 2 | Y 1 ) + (1 − ω ) E 3 ( Y 2 | Y 1 ) Y 2 ) K = 2 E 2 ( Y 2 | Y 1 ) E 3 ( Y 3 | Y 1 , Y 2 ) K = 3 E 3 ( Y 2 | Y 1 ) E 3 ( Y 3 | Y 1 , Y 2 ) Here, ω is a weight such that 0 ≤ ω ≤ 1 Hogan (MDEpiNet) Missing Data October 22, 2018 78 / 160

  52. Longitudinal model with J = 3: MNAR j = 2 j = 3 E 3 ( Y 3 | Y 1 , � K = 1 ω E 2 ( Y 2 | Y 1 ) + (1 − ω ) E 3 ( Y 2 | Y 1 ) + ∆ 1 Y 2 ) + ∆ 2 K = 2 E 2 ( Y 2 | Y 1 ) E 3 ( Y 3 | Y 1 , Y 2 ) + ∆ 3 K = 3 E 3 ( Y 2 | Y 1 ) E 3 ( Y 3 | Y 1 , Y 2 ) Hogan (MDEpiNet) Missing Data October 22, 2018 79 / 160

  53. Procedure with several longitudinal measures Start by imputing those with missing data at j = 2 1 Fit model for E ( Y 2 | Y 1 , R 2 = 1) α (2) + β (2) E ( Y 2 | Y 1 , R 2 = 1) = 1 Y 1 α (2) , � β (2) ⇒ obtain � 1 This is a model that combines those with K = 2 and K = 3 2 Impute missing Y 2 as before α (2) + � � β (2) Y 2 i = ∆ + � 1 Y 1 i Hogan (MDEpiNet) Missing Data October 22, 2018 80 / 160

  54. Procedure with several longitudinal measures Now impute those with missing data at j = 3 1 Fit model for E ( Y 3 | Y 1 , Y 2 , R 3 = 1) α (3) + β (3) 1 Y 1 + β (3) E ( Y 3 | Y 1 , Y 2 , R 3 = 1) = 2 Y 2 α (3) , � β (3) 1 , � β (3) ⇒ obtain � 2 2 Impute missing Y 3 as follows: ◮ For those with Y 1 , Y 2 observed, α (3) + � β (3) β (3) � 1 Y 1 i + � Y 3 i = ∆ + � 2 Y 2 i ◮ For those with only Y 1 observed, α (3) + � β (3) β (3) � 1 Y 1 i + � � = ∆ + � Y 3 i Y 2 i 2 Hogan (MDEpiNet) Missing Data October 22, 2018 81 / 160

  55. Side note Recall imputation for those with only Y 1 observed: α (3) + � � β (3) 1 Y 1 i + � β (3) � Y 3 i = ∆ + � Y 2 i 2 This is really just using information from the observed Y 1 because α (2) + � � β (2) Y 2 i = ∆ + � 1 Y 1 i Hence the imputation is from the (linear) model of E ( Y 3 | Y 1 ) that is implied by the other imputation models. Hogan (MDEpiNet) Missing Data October 22, 2018 82 / 160

  56. Calibration of ∆ At each time point j , ∆ is actually a multiplier of the residual SD for the observed data regression [ Y j | Y 1 , . . . , Y j − 1 , R = 1] For example, the imputation model at j = 3 is actually α (3) + � � β (3) 1 Y 1 i + � β (3) Y 3 i = ∆ � σ 3 + � 2 Y 2 i where σ 2 3 = var( Y 3 | Y 1 , Y 2 , R 3 = 1) Generally will suppress this for clarity Hogan (MDEpiNet) Missing Data October 22, 2018 83 / 160

  57. Procedure with several longitudinal measures Final step: compute estimate of E ( Y 3 ) � � R 3 i Y 3 i + (1 − R 3 i ) � E ∆ ( Y 3 ) = (1 / n ) Y 3 i (∆) i [ K =3] [ K =2] Based on the imputations, this turns out to be a weighted average of Y , Y , 3 2 [ K =1] and Y . 1 Weights depend on dropout rates at each time coefficients in imputation models sensitivity parameter(s) Hogan (MDEpiNet) Missing Data October 22, 2018 84 / 160

  58. Analysis components Fitted regression models at each time point ◮ Can check validity of imputation Contour plots: vary ∆ separately by treatment ◮ Treatment effect estimates ◮ p -values Summary table ◮ Treatment effect, SE, p-value ◮ These are computed using bootstrap Hogan (MDEpiNet) Missing Data October 22, 2018 85 / 160

  59. Hogan (MDEpiNet) Missing Data October 22, 2018 86 / 160

  60. Hogan (MDEpiNet) Missing Data October 22, 2018 87 / 160

  61. Hogan (MDEpiNet) Missing Data October 22, 2018 88 / 160

  62. Hogan (MDEpiNet) Missing Data October 22, 2018 89 / 160

  63. Case Study 2: Chronic Schizophrenia Major breakthroughs have been made in the treatment of patients with psychotic symptoms. However, side effects associated with some medications have limited their usefulness. RIS-INT-3 (Marder and Meibach, 1994, Chouinard et al. , 1993) was a multi-center study designed to assess the effectiveness and adverse experiences of four fixed doses of risperidone compared to haliperidol and placebo in the treatment of chronic schizophrenia. Hogan (MDEpiNet) Missing Data October 22, 2018 90 / 160

  64. RIS-INT-3 Patients were required to have a PANSS (Positive and Negative Syndrome Scale) score between 60 and 120. Prior to randomization, one-week washout phase (all anti-psychotic medications discontinued). If acute psychotic symptoms occurred, patients randomized to a double-blind treatment phase, schedule to last 8 weeks. Patients randomized to one of 6 treatment groups: risperidone 2, 6, 10 or 16 mg, haliperidol 20 mg, or placebo. Dose titration occurred during the first week of the double-blind phase. Hogan (MDEpiNet) Missing Data October 22, 2018 91 / 160

  65. RSIP-INT-3 Patients scheduled for 5 post-baseline assessments at weeks 1,2,4,6, and 8 of the double-blind phase. Primary efficacy variable: PANSS score Patients who did not respond to treatment and discontinued therapy or those who completed the study were eligible to receive risperidone in an open-label extension study. 521 patients randomized to receive placebo ( n = 88), haliperidol 20 mg ( n = 87), risperidone 2mg ( n = 87), risperidone 6mg ( n = 86), risperidone 10 mg ( n = 86), or risperidone 16 mg ( n = 87). Hogan (MDEpiNet) Missing Data October 22, 2018 92 / 160

  66. Dropout and withdrawal Only 49% of patients completed the 8 week treatment period. The most common reason for discontinuation was “insufficient response.” Other main reasons included: adverse events, uncooperativeness, and withdrawal of consent. Hogan (MDEpiNet) Missing Data October 22, 2018 93 / 160

  67. Dropout and Withdrawal Placebo Haliperidol Risp 2mg Risp 6mg Risp 10mg Risp 16 mg ( n = 88) ( n = 87) ( n = 87) ( n = 86) ( n = 86) ( n = 87) Completed 27 31% 36 41% 36 41% 53 62% 48 56% 54 62% Withdrawn 61 69% 51 59% 51 59% 33 38% 38 44% 33 38% Lack of Efficacy 51 58% 36 41% 41 47% 12 14% 25 29% 18 21% Other 10 11% 15 17% 10 11% 21 24% 13 15% 15 17% Hogan (MDEpiNet) Missing Data October 22, 2018 94 / 160

  68. Central Question What is the difference in the mean PANSS scores at week 8 between risperidone at a specified dose level vs. placebo in the counterfactual world in which all patients were followed to that week? Hogan (MDEpiNet) Missing Data October 22, 2018 95 / 160

  69. Sample means and imputed means under MAR N INS HOST EPS Y 0 µ R � Placebo R = 0 61 3.9 10.5 3.3 94 ?? R = 1 27 3.7 8.1 3.2 89 78 Risperidone R = 0 51 3.8 10.9 3.5 98 ?? R = 1 36 3.8 8.1 2.8 87 71 Hogan (MDEpiNet) Missing Data October 22, 2018 96 / 160

  70. Sample means and imputed means under MAR N INS HOST EPS Y 0 µ R � Placebo R = 0 61 3.9 10.5 3.3 94 79 R = 1 27 3.7 8.1 3.2 89 78 Risperidone R = 0 51 3.8 10.9 3.5 98 74 R = 1 36 3.8 8.1 2.8 87 71 Hogan (MDEpiNet) Missing Data October 22, 2018 97 / 160

  71. Regression imputation under MAR Placebo Hogan (MDEpiNet) Missing Data October 22, 2018 98 / 160

  72. Regression imputation under MAR Risperidone Hogan (MDEpiNet) Missing Data October 22, 2018 99 / 160

  73. Sample means by dropout time: Aggregated data 120 100 PANSS 80 60 1 2 3 4 5 6 Visit Hogan (MDEpiNet) Missing Data October 22, 2018 100 / 160

Recommend


More recommend