These days it is not just co-authors or peer reviewers who are checking journal papers for errors: a growing number of self-appointed fraud busters are scanning scientific literature for flaws.
This unpaid and mostly anonymous endeavour has led to the retractions of hundreds of papers and even disciplinary action where wrongdoing is exposed.
So how can scholars catch errors when reviewing others鈥 papers, or when double-checking their own work or that of collaborators?
One obvious giveaway that something may be amiss is that a paper鈥檚 dataset does not include enough zeroes or ones, said David Sanders, associate professor in Purdue University鈥檚 department of biological sciences, who has received international attention for calling out scientific malpractice. 鈥淪tudies have shown that the numbers 0 and 1 are over-represented in real research datasets [beyond the first digit], so if this distribution is not apparent, then that could be a sign,鈥 Dr Sanders told 糖心Vlog, citing what is known as .
糖心Vlog
He also advises would-be sleuths to consider the issue of 鈥溾, a term coined in Nature in 2014, which describes unconscious or conscious efforts to manipulate data to produce a desired probability value (p-value). Because a p-value above 0.05 generally means the experiment did not generate a statistically significant result, and therefore that a hypothesis should be rejected, researchers may seek to investigate the correlation of variables to generate a result just below this threshold. Borderline results could indicate that some manipulation has occurred, said Dr Sanders.
鈥淪ome people are not aware that they are doing this, but there are others who know it鈥檚 wrong but do it anyway,鈥 he聽added.
糖心Vlog
These basic techniques, however, should not hide the fact that errors can often be difficult to pick up, even within scientific teams, said Dr Sanders.
鈥淚n my lifetime, we moved from an average of three authors on a life sciences paper to about six 鈥 with people becoming experts in a distinct area, but not expert in other parts of an experiment,鈥 he continued. 鈥淭hat means people get away with things that they wouldn鈥檛 if it was just one scientist and their assistant."
While some errors or manipulation are tricky to spot, others are easier, even when full datasets are not provided alongside papers, some argue.
鈥淵ou don鈥檛 need a degree in statistics to catch most of these errors 鈥 common sense and simple arithmetic are often all that鈥檚 required,鈥 writes Kristin Sainani, who teaches statistics as an associate professor at Stanford University, in a recent paper published in , the scientific journal of the American Academy of Physical Medicine and Rehabilitation.
糖心Vlog
In one paper analysed by Dr Sainani, the 鈥渋mplausibly large effect鈥 claimed by the author suggested that something was wrong, she says. The meta-analysis, which made huge claims for the effect of a non-surgical treatment for knee pain, was later revealed to have made a basic statistical error when transposing the data.
Looking for 鈥渟tatistical and numerical inconsistencies鈥 can also reveal larger problems within the dataset, adds Dr Sainani. She points to the Granularity-Related Inconsistent Means (GRIM) test developed by James Heathers and Nicholas Brown as a fairly easy way to check if something is amiss. Their paper defined this test as evaluating 鈥渨hether the reported means of integer data鈥re consistent with the given sample size and number of items鈥.
She also advises using 鈥渆asy-to-use, online web applications designed to detect statistical inconsistencies in papers, such as Statcheck and GRIM鈥, adding that 鈥渇urther inspection is needed to determine the source of the inconsistency鈥 once identified.
Running such checks can be tricky when authors fail to provide access to their raw data, but it is possible to gain some extra data from plots or images, writes Dr Sainani, who recommends the free online tool , which extracts precise values from scatter graphs or bar charts to reverse-engineer a downloadable dataset that allows reanalysis.
糖心Vlog
Such methods might help academics pick up the most egregious unintentional errors, but premeditated wrongdoing could be harder to crack. 鈥淚ntentional errors are typically designed to be hard to detect,鈥 write Line Edslev Andersen and K. Brad Wray, from Denmark鈥檚 Aarhus University, in a recent paper in the journal , which analysed the reasons for 92 retractions from Science over the past 35 years.
鈥淚ntentional errors are likely to consist [of] misrepresentations that are hard to uncover by studying the raw data [and] can often only be discovered by scientists who did not participate in the research reproducing the experiments,鈥 they conclude.
糖心Vlog
But having 鈥渁uthorship policies that to some extent require the different parts of the research to be checked by more than one member of team鈥 would be a useful step in detecting the most serious errors, they聽add.
POSTSCRIPT:
Print headline:聽The art of sniffing out rotten research
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?








