Response time in online stated choice experiments: the non-triviality of identifying fast and slow respondents

Danny Campbell, Morten Raun Mørkbak, Søren Bøye Olsen

Abstract

In this paper, we use paradata relating to the length of time respondents required in a self-administered online stated preference surveys. Although this issue has been previously explored, there is little guidance on how to identify and deal with ‘fast’ and ‘slow’ respondents. In this paper, we use scale-adjusted latent class models to address preference and variance heterogeneity and explore how class membership varies with response latency. To test our methodology, we use stated choice data collected via an online survey to establish German anglers’ preferences for fishing site attributes in Denmark. Results from our analysis corroborate that response latency has a bearing on the estimates of utility coefficients and the error variance. Although the results highlight the non-triviality of identifying fast and slow respondents, they signal the need to estimate a large number of candidate models to identify the most appropriate ‘fast’ and ‘slow’ thresholds. Not doing so is likely to lead to an inferior model and has repercussions for marginal willingness to pay estimates and choice predictions.
Original languageEnglish
JournalJournal of Environmental Economics and Policy
Volume6
Issue number1
Pages (from-to)17-35
Number of pages19
ISSN2160-6544
DOIs
Publication statusPublished - 2017

Fingerprint

Dive into the research topics of 'Response time in online stated choice experiments: the non-triviality of identifying fast and slow respondents'. Together they form a unique fingerprint.

Cite this