Predicting the replicability of social and behavioural science claims from the COVID-19 Preprint Replication Project with structured expert and novice groups

Peer Reviewed
14 February 2023

Alexandru Marcoci, Elan Simon Parsons, John R Kerr, Melissa Kline Struhl, Marta Kolczynska, Yang Liu, Zachary Loomas, Brianna Luis, Esteban Méndez, Olivia Miske, Carolin Nast, Brian A. Nosek, Thomas Pfeiffer, Angela Jones, W. Robert Reed, Jon Roozenbeek, Alexa R. Schlyfestone, Claudia R. Schneider, Andrew Soh, Anirudh Tagat, Melba Tutor, Andrew Tyner, Karolina Urbanska, Sander van der Linden, Ans Vercammen, Hansika Kapoor, Magnus Johannesson, David Peter Wilkinson, Jianhua Duan, Anna Abatayo, Ernest Baskin, henk berkman, Erin Michelle Buchanan, Sara Capitán, Tabaré Capitán, Ginny Chan, Kent Jason Go Cheng, Tom Coupe, Sarah Dryhurst, John Edlund, Nick Huntington-Klein, Timothy M. Errington, Anna Fedor, Fiona Fidler, James Field, Nicholas William Fox, Hannah Fraser, Alexandra L.J. Freeman, Anca Hanea, Felix Holzmeister, Sanghyun Hong, Raquel Huggins, Bonnie Wintle

Replication is an important “credibility control” mechanism for clarifying the reliability of published findings. However, replication is costly, and it is infeasible to replicate everything. Accurate, fast, lower cost alternatives such as eliciting predictions from experts or novices could accelerate credibility assessment and improve allocation of replication resources for important and uncertain findings. We elicited judgments from experts and novices on 100 claims from preprints about an emerging area of research (COVID-19 pandemic) using a new interactive structured elicitation protocol and we conducted 35 new replications. Participants’ average estimates were similar to the observed replication rate of 60%. After interacting with their peers, novices updated both their estimates and confidence in their judgements significantly more than experts and their accuracy improved more between elicitation rounds. Experts’ average accuracy was 0.54 (95% CI: [0.454, 0.628]) after interaction and they correctly classified 55% of claims; novices’ average accuracy was 0.55 (95% CI: [0.455, 0.628]), correctly classifying 61% of claims. The difference in accuracy between experts and novices was not significant and their judgments on the full set of claims were strongly correlated (r=.48). These results are consistent with prior investigations eliciting predictions about the replicability of published findings in established areas of research and suggest that expertise may not be required for credibility assessment of some research findings.

Files and links

Country
Publication reference
Marcoci, A., Wilkinson, D. P., Abatayo, A., Baskin, E., berkman, henk, Buchanan, E. M., Capitán, S., Capitán, T., Chan, G., Cheng, K. J. G., Coupe, T., Dryhurst, S., Duan, J., Edlund, J., Errington, T. M., Fedor, A., Fidler, F., Field, J., Fox, N. W., … Wintle, B. (2023). Predicting the replicability of social and behavioural science claims from the COVID-19 Preprint Replication Project with structured expert and novice groups. https://doi.org/10.31222/osf.io/xdsjf
Publication | 19 December 2023