Betting scientists correctly predict reproducibility of papers

Findings suggest markets could be used to help prioritise which experiments need repeating most urgently 

August 27, 2018
Two women wearing fancy hats
Source: Getty
Mirror, mirror: researchers are able to successfully predict reproducibility of findings

Academics are easily able to predict whether an experiment’s findings will be reproducible, according to a study that raises more questions about the reliability of research published in leading journals.

A major new investigation – undertaken by 12 research centres across the world in partnership with the Centre for Open Science (COS) and published in Nature Human Behaviour – sought to test the most significant findings in social science papers published in Science and Nature between 2010 and 2015.

Conducting replication experiments for 21 eligible studies, the collaborative team of researchers from five laboratories found that 13 showed evidence “consistent with” the findings of the original paper. As many as eight failed to do so, however, suggesting a reproducibility rate of 62 per cent.

To ensure “high statistical power”, average sample sizes for the replication studies were about five times larger than their originals. Strikingly, however, effect sizes were found to be about 50 per cent smaller on average in the replication tests than their original studies had promised.

Lily Hummer, an assistant research scientist at the COS and co-author of one of the replication studies, explained that the smaller effect sizes showed that “increasing power substantially is not sufficient to reproduce all published findings”.

Project leaders set up prediction markets – allowing other researchers to bet money on whether they thought each one of the papers would replicate or not – before conducting the experiments.

ADVERTISEMENT

Taking into account the bets of 206 researchers, the markets correctly predicted replication outcomes for 18 of the 21 studies. Furthermore, market beliefs about replication were highly correlated with the true replication effect sizes, authors noted.

Thomas Pfeiffer, professor in computation biology at the New Zealand Institute for Advanced Study, another of the project leaders, said that this suggested that “researchers have advance knowledge about the likelihood that some findings will replicate”.

Speaking to Times Higher Education, Brian Nosek, professor of psychology at the University of Virginia and executive director of the COS, said that this outcome did not necessarily indicate that academics were willingly or knowingly submitting papers with poor-quality datasets to journals.

“I wouldn’t infer that researchers are submitting work that they know to be irreproducible,” he said. “The market reflects the price that the whole community puts on the likelihood of replication, but there is a lot of variability between individuals.”

It was also likely that a “shift in culture” meant that scientists who might once have been defensive about criticisms against their findings are becoming increasingly aware of potential problems with reproducibility of data and, thus, more open to public discussion than they might have been even just a few years ago.

“It may be that researchers themselves will sharpen their intuitions about the plausibility of results,” Professor Nosek said. “When the culture was just rewarding finding the sexy result, regardless of its plausibility, there was little reason to consider whether it was reproducible. That is one of the biggest benefits of the changing norms.”

This is not the first time that betting has been used as an indicator for the likelihood of results. A previous study published by the COS asked researchers to submit their predictions for the reproducibility of 44 studies published in prominent psychology journals.

Each researcher was given $100 (£78) to “trade” with, and a total of 2,496 transactions were carried out – suggesting to researchers that prediction markets could be used as “decision markets” to help scientists prioritise which studies to attempt to replicate in the future.

ADVERTISEMENT

rachael.pells@timeshighereducation.com

Register to continue

Why register?

  • Registration is free and only takes a moment
  • Once registered, you can read 3 articles a month
  • Sign up for our newsletter
Please
or
to read this article.

Related articles

Sponsored

Helping universities adopt a digital mindset

Higher education leaders met at Microsoft’s Executive Briefing Center to discuss research, artificial intelligence and collaboration with industry
Promoted by Microsoft
Sponsored

One Health

Through the ‘One Health’ approach, CityU aims to develop and integrate interdisciplinary, problem-based research collaboration in all aspects of health-related issues.
Promoted by City University of Hong Kong
Sponsored

Student success in a technological world

Extending study spaces, enabling modular learning and using data insights to remove silos were three recommendations from a panel of university and industry thought leaders
Promoted by Salesforce
Sponsored
ADVERTISEMENT