D-efficient or deficient? A robustness analysis of stated choice experimental designs

Theory and Decision 84 (2):215-238 (2018)
  Copy   BIBTEX

Abstract

This paper is motivated by the increasing popularity of efficient designs for stated choice experiments. The objective in efficient designs is to create a stated choice experiment that minimizes the standard errors of the estimated parameters. In order to do so, such designs require specifying prior values for the parameters to be estimated. While there is significant literature demonstrating the efficiency improvements of employing efficient designs, the bulk of the literature tests conditions where the priors used to generate the efficient design are assumed to be accurate. However, there is substantially less literature that compares how different design types perform under varying degree of error of the prior. The literature that does exist assumes small fractions are used, which is in contrast to computer-aided surveys that readily allow for large fractions. Further, the results in the literature are abstract in that there is no reference point to provide clear insight on the magnitude of any issue. Our objective is to analyze the robustness of different designs within a typical stated choice experiment context of a trade-off between price and quality. We use as an example transportation mode choice, where the key parameter to estimate is the value of time. Within this context, we test many designs to examine how robust efficient designs are against a misspecification of the prior parameters. The simple mode choice setting allows for insightful visualizations of the designs themselves and also an interpretable reference point for the range in which each design is robust. Not surprisingly, the D-efficient design is most efficient in the region where the true population VOT is near the prior used to generate the design: the prior is $20/h and the efficient range is $10–$30/h. However, the D-efficient design quickly becomes the most inefficient outside of this range, and the estimation significantly degrades above $50/h. The orthogonal and random designs are robust for a much larger range of VOT. The robustness of Bayesian efficient designs varies depending on the variance that the prior assumes. Implementing two-stage designs that first use a small sample to estimate priors are also not robust relative to uninformative designs. Arguably, the random design performs as well as any design, and it will perform even better if data cleaning is done to remove choice tasks where one alternative dominates the other.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 92,261

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Robustness and sensitivity of biological models.Jani Raerinne - 2013 - Philosophical Studies 166 (2):285-303.
Economic Modelling as Robustness Analysis.Jaakko Kuorikoski, Aki Lehtinen & Caterina Marchionni - 2010 - British Journal for the Philosophy of Science 61 (3):541-567.
Clear-cut designs versus the uniformity of experimental practice.Francesco Guala - 2001 - Behavioral and Brain Sciences 24 (3):412-413.
Robustness Analysis.Michael Weisberg - 2006 - Philosophy of Science 73 (5):730-742.
Farming Dwelling Thinking.David Strong - 2016 - Ethics and the Environment 21 (2):27-50.
Robustness and reality.Markus I. Eronen - 2015 - Synthese 192 (12):3961-3977.

Analytics

Added to PP
2017-12-14

Downloads
19 (#803,294)

6 months
6 (#530,265)

Historical graph of downloads
How can I increase my downloads?

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references