A study characterizing subtypes of the bacteria that cause bacterial meningitis is being retracted after the authors didn’t have permission to publish the data, even though the data itself remain available in a public database.
The paper, in PLOS ONE, relied on a laboratory collection of patient samples. In October, the authors retracted it because they “did not have permission” from the laboratory “to publish the data in their current form.” The data — anonymized — are now available at PubMLST.
Do you think the write-up of scientific results has gotten more rosy over time? If so, you’re right — the use of positive language in science abstracts has increased by 880% since 1974, according to new findings reported in the British Medical Journal.
Although many scientists fear putting their data to the test of replication efforts, due to the embarrassment they’d feel if their findings couldn’t be repeated, a new study suggests those fears are unfounded.
The paper, published last week in PLOS ONE, found that scientists overestimate how much having non-replicable data will hurt their careers, and the community values those who are honest about what went wrong.
If you could help reduce the waste of tens of billions of dollars per year in research spending, you’d do it, right?
This is the second in a series of two guest posts about the havoc misidentified cell lines can wreak on research, from Leonard P. Freedman, president of the Global Biological Standards Institute. Freedman who published a paper last summer detailing the financial costs of non-reproducible research — namely, tens of billions of dollars per year. Some of that non-reproducible research is due to the use of contaminated or misidentified cell lines. He writes about one key step to tackling the problem: Ask every scientist to use a relatively inexpensive technique to validate the identity of their cell lines.
Meanwhile, we have to deal with the issue of all the previously published papers that relied on problematic cell lines, now contaminating the scientific literature. Scroll down to the bottom of the post to take a poll on what you think should be done about those papers.
As new frontiers of science emerge, from Pluto to proteins, the very cornerstone of the scientific process—reproducibility—has also reared its head as a huge problem. Estimates of irreproducibility rates of published peer-reviewed papers range from 51% to 89%. An analysis that two colleagues and I recently published in PLOS Biology suggests the U.S. spends $28 billion per year on non-reproducible preclinical research; global spending could be up to $60 billion per year. This lack of reproducibility typically results from cumulative errors or flaws in one or more of the following areas: biological reagents and reference materials, study design, laboratory protocols, and data analysis and reporting. Given the size, scale, and especially the complexity of reproducing preclinical research, there is no single magic bullet fix. This is a difficult issue for scientists to own up to, and for the public to grasp.
What if we told you that approximately 1 in 6 researchers working with human cells are using the wrong cell line? In other words, they believe they are studying the effects of a drug on breast cancer cells, for instance, but what they really have are cells from the bladder. That is the unfortunate reality in life science research today, affecting hundreds of labs. It’s a major source of problematic papers which cannot be replicated, wasting scientists’ time and funding.
We’re pleased to present a guest post from Amanda Capes-Davis, chair of the International Cell Line Authentication Committee (ICLAC), a voluntary scientific committee created to improve awareness of misidentified cell lines. She also collects news about cell line and culture contamination. This is the first in a series of two posts from guest authors about how problematic cell lines are contaminating the scientific literature, and how we can clean it up.
In 2010, I worked alongside Ian Freshney of Glasgow University and other colleagues to publish a list of cross-contaminated or otherwise misidentified cell lines in the International Journal of Cancer. This database of false cell lines is now curated by the International Cell Line Authentication Committee (ICLAC).
We’re presenting a Q&A session with Peter Wilmshurst, now a part-time consultant cardiologist who has spent decades embroiled in misconduct investigations as a whistleblower.
Retraction Watch: A UK judge recently upheld two findings of dishonesty by the Medical Practitioners Tribunal Service against Andrew Dowson, director of headache services at King’s College Hospital and your former co-investigator. Were you pleased with the verdict? (Last week, Dowson also began a four-month suspension from practicing medicine in the UK.)
Peter Wilmshurst: In part, because I was pleased that he was shown to be dishonest, because I knew that he was, which was why I reported him. But I wasn’t pleased in the sense that I don’t think the investigation dealt with all the issues involved in the Migraine Intervention with STARflex Technology (MIST) Trial.
A new project does the relatively straightforward task of comparing reported outcomes from clinical trials to what the researchers said they planned to measure before the trial began. And what they’ve found is a bit sad, albeit not entirely surprising.
As part of The Compare Project, author and medical doctor Ben Goldacre and his team have so far evaluated 36 clinical trials published by the top five medical journals (New England Journal of Medicine, the Journal of the American Medical Association, The Lancet, Annals of Internal Medicine, and British Medical Journal). Many of those trials included “switched outcomes,” meaning the authors didn’t report something they said they would, or included additional outcomes in the published paper, with no explanation for the change.
Here are the latest results from the project, according to its website:
The director of the Research Integrity & Compliance Review Office at Colorado State University (CSU) is the new head of the U.S. Office of Research Integrity, taking over a position that has been vacant for more than a year and a half.
Nearly three years ago, our co-founders Ivan Oransky and Adam Marcus penned a column in Lab Times suggesting ways for readers to report alleged scientific misconduct. They are now retracting that advice.