Paper reports data from PET/CT scan, years before it arrived

MedicineAuthors have retracted a study just three months after publishing it, upon realizing they made “several critical errors.”

For one, the authors didn’t actually collect the data they claim to in the title of the paper, which reported on methods to screen patients for recurrence of lung cancer. The authors included data from positron emission tomography/computed tomography (PET/CT), collected from 2003 to 2007 — but their institution didn’t have a PET/CT scanner until 2009. Instead, the authors had mistakenly reported the results of PET scans alone, which may not find tumors as effectively as PET/CT.

Here’s the retraction notice in Medicine, which explains the nature of the error in more detail. (Note: One of the authors supplied some missing text, in brackets.)

Continue reading Paper reports data from PET/CT scan, years before it arrived

PLOS ONE paper plagiarized from 17 articles — yes, 17

PLOSOneA PLOS ONE paper about chronic pain plagiarized from multiple sources — 17, in fact.

According to the retraction notice released by the journal last week, the paper contains “extensive verbatim use of text from other sources.”

How did this make it past the editors? The journal published the paper in 2012 — before it began screening papers for plagiarism, according to a spokesperson.

Here’s the retraction notice for “The Effect of Social Stress on Chronic Pain Perception in Female and Male Mice:”

Continue reading PLOS ONE paper plagiarized from 17 articles — yes, 17

How should journals update papers when new findings come out?

NEJM Logo

When authors get new data that revise a previous report, what should they do?

In the case of a 2015 lung cancer drug study in the New England Journal of Medicine (NEJM), the journal published a letter to the editor with the updated findings.

Shortly after the paper was published, a pharmaceutical company released new data showing the drug wasn’t quite as effective as it had seemed. Once the authors included the new data in their analysis, they adjusted their original response rate of 59%  — hailed as one of a few “encouraging results” in an NEJM editorial at the time of publication — to 45%, as they write in the letter. One of the authors told us they published the 2015 paper using less “mature” data because the drug’s benefits appeared so promising, raising questions about when to publish “exciting but still evolving data.”

It’s not a correction, as the original paper has not been changed; it doesn’t even contain a flag that it’s been updated. But among the online letters about the paper is one from the authors, “Update to Rociletinib Data with the RECIST Confirmed Response Rate,” which provides the new data and backstory:

Continue reading How should journals update papers when new findings come out?

Duplicated data gets corrected — not retracted — by psych journal

Psychological Science

A psychology journal is correcting a paper for reusing data. The editor told us the paper is a “piecemeal publication,” not a duplicate, and is distinct enough from the previous article that it is not “grounds for retraction.”

The authors tracked the health and mood of 65 patients over nine weeks. In one paper, they concluded that measures of physical well being and psychosocial well being positively predict one another; in the other (the now corrected paper), they concluded that health and mood (along with positive emotions) influence each other in a self-sustaining dynamic.

As a press release for the now-corrected paper put it: Continue reading Duplicated data gets corrected — not retracted — by psych journal

Heart researcher faked 70+ experiments, 100+ images

ori-logoA former researcher at the University of Michigan and the University of Chicago faked dozens of experiments and images over the course of six years, according to a new finding from the Office of Research Integrity (ORI).

Ricky Malhotra, who studied heart cells, admitted to committing misconduct at both institutions, the ORI said in its report of the case. The fakery involved three National Institutes of Health (NIH) grant applications, one NIH progress report, one paper, seven presentations, and one image file. Despite an investigation at the University of Michigan, where Malhotra was from 2005-2006, he continued this falsification at [University of Chicago], after the [University of Michigan] research misconduct investigation was completed,” according to the ORI. The agency found that he Continue reading Heart researcher faked 70+ experiments, 100+ images

Researcher who sued to stop retractions gets his sixth

Mario Saad
Mario Saad

A sixth retraction has appeared for a diabetes researcher who previously sued a publisher to try to stop his papers from being retracted.

Mario Saad‘s latest retraction, in PLOS Biology, stems from inadvertent duplications, according to the authors.  Though an investigation at Saad’s institution — the University of Campinas in Brazil — found no evidence of misconduct, a critic of the paper told The Scientist he does not believe that the issues with blots were inadvertent.

Previously, Saad sued the American Diabetes Association to remove four expressions of concern from his papers; they were later retracted, even though Unicamp recommended keeping three of them published.

Here’s the new retraction notice, for “Gut Microbiota Is a Key Modulator of Insulin Resistance in TLR 2 Knockout Mice:” Continue reading Researcher who sued to stop retractions gets his sixth

Context matters when replicating experiments, argues study

PNASBackground factors such as culture, location, population, or time of day affect the success rates of replication experiments, a new study suggests.

The study, published today in the Proceedings of the National Academy of Sciences, used data from the psychology replication project, which found only 39 out of 100 experiments live up to their original claims. The authors conclude that more “contextually sensitive” papers — those whose background factors are more likely to affect their replicability — are slightly less likely to be reproduced successfully.

They summarize their results in the paper:

Continue reading Context matters when replicating experiments, argues study

Editors say they missed “fairly obvious clues” of third party tampering, publish fake peer reviews

BJCP Cover

The editors of a journal that recently retracted a paper after the peer-review process was “compromised” have published the fake reviews, along with additional details about the case.

In the editorial titled “Organised crime against the academic peer review system,” Adam Cohen and other editors at the British Journal of Clinical Pharmacology say they missed “several fairly obvious clues that should have set alarm bells ringing.” For instance, the glowing reviews from supposed high-profile researchers at Ivy League institutions were returned within a few days, were riddled with grammar problems, and the authors had no previous publications. 

The case is one of many we’ve recently seen in which papers are pulled due to actions of a third party

The paper was submitted on August 5, 2015. From the beginning, the timing was suspect, Cohen — the director for the Centre for Human Drug Research in The Netherlands — and his colleagues note: Continue reading Editors say they missed “fairly obvious clues” of third party tampering, publish fake peer reviews

PLOS editors discussing authors’ decision to remove chronic fatigue syndrome data

After PLOS ONE allowed authors to remove a dataset from a paper on chronic fatigue syndrome, the editors are now “discussing the matter” with the researchers, given the journal’s requirements about data availability.

As Leonid Schneider reported earlier today, the 2015 paper was corrected May 18 to remove an entire dataset; the authors note that they were not allowed to publish anonymized patient data, but can release it to researchers upon request. The journal, however, requires that authors make their data fully available.

Here’s the correction notice: Continue reading PLOS editors discussing authors’ decision to remove chronic fatigue syndrome data

Software glitch — not intentional manipulation — sunk immunology paper, says author

kuo photo
A black box appears over the control lane on the left

New evidence suggests a retracted paper was felled not by intentional manipulation — as it first appeared — but by a software glitch.

In 2014, we reported that Biochemical Journal had retracted a paper on suspicion it contained “shoddy Photoshopping”  — someone appeared to have blacked out a control lane in one figure. Now there’s evidence that it wasn’t done on purpose: An investigation at Duke into eight papers, including the Biochemical Journal paper, did not find evidence of misconduct; lead author Paul Kuo, currently chair of surgery at Loyola Medicine, told us that a glitch in the software caused the black box. Nevertheless, the journal does not plan to un-retract the paper. Continue reading Software glitch — not intentional manipulation — sunk immunology paper, says author