Examining the replicability of online experiments selected by a decision market

  • Felix Holzmeister
  • , Magnus Johannesson
  • , Colin F. Camerer
  • , Yiling Chen
  • , Teck Hua Ho
  • , Suzanne Hoogeveen
  • , Juergen Huber
  • , Noriko Imai
  • , Taisuke Imai
  • , Lawrence Jin
  • , Michael Kirchler
  • , Alexander Ly
  • , Benjamin Mandl
  • , Dylan Manfredi
  • , Gideon Nave
  • , Brian A. Nosek
  • , Thomas Pfeiffer
  • , Alexandra Sarafoglou
  • , Rene Schwaiger
  • , Eric Jan Wagenmakers
  • Viking Waldén, Anna Dreber*
*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Here we test the feasibility of using decision markets to select studies for replication and provide evidence about the replicability of online experiments. Social scientists (n = 162) traded on the outcome of close replications of 41 systematically selected MTurk social science experiments published in PNAS 2015–2018, knowing that the 12 studies with the lowest and the 12 with the highest final market prices would be selected for replication, along with 2 randomly selected studies. The replication rate, based on the statistical significance indicator, was 83% for the top-12 and 33% for the bottom-12 group. Overall, 54% of the studies were successfully replicated, with replication effect size estimates averaging 45% of the original effect size estimates. The replication rate varied between 54% and 62% for alternative replication indicators. The observed replicability of MTurk experiments is comparable to that of previous systematic replication projects involving laboratory experiments.

Original languageEnglish
Article numbere124
Pages (from-to)316–330
Number of pages15
JournalNature Human Behaviour
Volume9
Issue number2
Early online date19 Nov 2024
DOIs
Publication statusPublished - 2025

Bibliographical note

Publisher Copyright:
© The Author(s) 2024.

Funding

We thank A. Andevall for helping with the data collection and programming of experiments and R. Willer for helpful advice on defining the IP address check and exclusion criteria used to exclude individuals from participating to minimize low-quality participant data. For financial support, we thank the Austrian Science FWF (grant SFB F63 to J.H. and M.K.), Jan Wallander and Tom Hedelius Foundation (grants P21-0091 and P23-0098 to A.D.), Knut and Alice Wallenberg Foundation and Marianne and Marcus Wallenberg Foundation (Wallenberg Scholar grant to A.D.) and Riksbankens Jubileumsfond (grant P21-0168 to M.J.). The funders had no role in study design, data collection and analysis, decision to publish or preparation of the paper. One author (V.W.) is currently employed by Sveriges Riksbank but did this work before being employed by Sveriges Riksbank; the opinions expressed in this article are the sole responsibility of the authors and should not be interpreted as reflecting the views of Sveriges Riksbank.

FundersFunder number
Jan Wallanders och Tom Hedelius Stiftelse samt Tore Browaldhs Stiftelse (Jan Wallander and Tom Hedelius Foundation and Tore Browaldh Foundation)SFB F63
Austrian Science FWFP21-0091, P23-0098
Jan Wallander and Tom Hedelius Foundation
Knut and Alice Wallenberg foundation
Marianne and Marcus Wallenberg Foundation (Wallenberg Scholar grant)P21-0168
Riksbankens Jubileumsfond

    Fingerprint

    Dive into the research topics of 'Examining the replicability of online experiments selected by a decision market'. Together they form a unique fingerprint.

    Cite this