Using Prior Wave Information and Paradata: Can They Help to Predict Response Outcomes and Call Sequence Length in a Longitudinal Study? Olga Maslovskaya, Gabriele Durrant and Peter W.F. Smith University of Southampton 14 March 2018, Washington DC
Introduction In face ‐ to ‐ face surveys interviewers make several visits (calls) to households to obtain response (this leads to a call sequence) Aim for survey practice : to identify (and ideally avoid) unsuccessful and long call sequences
Introduction 2 • Paper 1 – Wave 1 data only Durrant, Maslovskaya, Smith (2015): Journal of Survey Statistics and Methodology 3(3): 397 ‐ 424. • Paper 2 – Wave 1 and Wave 2 – longitudinal context Durrant, Maslovskaya, Smith (2017): Journal of Official Statistics 33(3): 801 ‐ 834 .
Main Research Questions • Can we predict final call sequence length and final outcome early on in the data collection process of the current wave by taking into account information from current and previous waves? • In other words: – Can we predict after, for example, the third call if a household is going to respond or not? – Can we predict how many calls it is going to take to reach the final outcome?
Further Research Questions Ability of ‘classical’ nonresponse models without call record data to predict nonresponse is often limited (R 2 values well below 10%) (Olson et al. 2012; Olson and Groves 2012; West and Groves 2013). • How predictive are the models which include call record data? • Does their ability to predict improve once more call record data are available (e.g. including earlier call information; including information from previous waves in a longitudinal study)? • How can these models best be used in survey practice?
Data UK Understanding Society Survey (Waves 1 and 2) Wave 1: call record data, interviewers’ observation data, survey data – only responding households Wave 2: call record data, interviewers’ observation data Analysis sample: 10,630 households (HHs with 4 or more calls) Maximum number of calls in Wave 1: 30 Methodological difficulty: Household IDs are different in different waves (no such thing as longitudinal household) so it is difficult to follow HHs, but possible to follow individuals in households.
Dependent Variables and Models
Modelling Strategy and Explanatory Variables
Assessment of Models Focus on ability of models to predict length and outcome • To compare different models and to assess quality of model prediction and • model fit Pseudo ‐ R 2 statistic (proportion of variation in the dependent variable that is explained by the model) Concepts from epidemiology to assess accuracy of models (Plewis et al 2012): o discrimination (sensitivity and specificity) o prediction (positive and negative predicted value) AUC (Area under the Curve) ( ROC (receiver operating characteristic) curve)
Assessment of Models
Area under the Curve (AUC) • A Receiver Operating Curve (ROC) summarises predictive power for all possible values c (cut ‐ off for π ), by plotting sensitivity as a function of (1 ‐ specificity) • The higher sensitivity the higher predictive power, given a particular specificity • The greater area under the curve (AUC) the greater the predictive power. AUC values range from 0.5 (no discrimination) to 1 (perfect discrimination).
Results
Results
Results
Results
Results
Results
Results
Results: Sensitivity
Results: Positive Predicted Values Model Short Short Long Long Successful Unsuccessful Successful Unsuccessful (n=5304) (n=1400) (n=2216) (n=1710) 1 49.9% 0.0% 0.0% 0.0% 2 50.5% 66.7% 37.5% 34.6% 3 50.9% 46.7% 37.2% 37.2% 4 51.8% 39.3% 32.6% 36.6% 5 53.3% 43.0% 38.3% 37.2% 6 53.7% 37.9% 35.0% 39.1% 7 54.8% 48.3% 36.8% 40.4% 8 62.1% 48.6% 42.2% 41.1% P�� � k|� � � k ), for k=1,2,3,4 Of the cases predicted to be long unsuccessful 41.1% are predicted correctly
Summary (1) • Basic geographic information not very predictive • Controlling for survey data from Wave 1 increases R ‐ squared but not other assessment indicators (classical nonresponse model not very predictive!) • Call record data from Wave 1 highly significant and improves prediction (but not by very much) Time of calls and time between calls are all significant variables but their impact on prediction limited • Interviewers’ observation variables : some significant in the models, slightly improve the predictive power
Summary (2) • Indicators of change (between Wave 1 and Wave 2): ₋ indicators of change in interviewers’ observations significant and improve prediction (but not by very much) ₋ Variable ‘household has changed household composition’ highly significant (but not very predictive) • Big increase in prediction once call outcomes of most recent calls in Wave 2 included! (the more calls the more useful as expected) • Modelling length and final outcome jointly improves prediction (variables in length model and in outcome model can be different)
Conclusions • Novelty is to model sequence length and to model length and outcome jointly and to condition on all current and previous information on a case (including survey data, call data, interviewer observations) • Most recent call outcome most predictive: could indicate that the response process depends on current circumstances more than previous history and characteristics • Can be implemented into survey practice quite easily , using standard methodology • Survey managers may wish to weigh up between the probability of a successful outcome versus sequence length
References Durrant G., Maslovskaya O. and P.W.F. Smith (2015) Modelling final outcome and length of call sequence to improve efficiency in interviewer call scheduling, Journal of Survey Statistics and Methodology 3(3): 397 ‐ 424 . Durrant,G., Maslovskaya, O. and P.W.F. Smith (2017) Using prior wave information and paradata: Can they help to predict response outcomes and call sequence length in a longitudinal study?. Journal of Official Statistics 33(3): 801 ‐ 834. Olson, K., and R.M. Groves ( 2012) An examination of within ‐ person variation in response propensity over the data collection field period. J ournal of Official Statistics 28: 29 ‐ 51. Olson, K., J.D. Smyth, and H.M. Wood (2012) Does giving people their preferred survey mode actually increase survey participation rates? An experimental examination. Public Opinion Quarterly 76: 611 – 635. Plewis I., Ketende S. and L. Calderwood (2012) Assessing the accuracy of response propensities in longitudinal studies, Survey Methodology 38(2): 167 ‐ 171. West, B. T. and R.M. Groves (2013) A propensity ‐ adjusted interviewer performance indicator. Public Opinion Quarterly 77 : 352 ‐ 374.
Recommend
More recommend