“Publications that are based on wrong data, methodological mistakes, or contain other types of severe errors can spoil the scientific record if they are not retracted. Retraction of publications is one of the effective ways to correct the scientific record. However, before a problematic publication can be retracted, the problem has to be found and brought to the attention of the people involved (the authors of the publication and editors of the journal). The earlier a problem with a published paper is detected, the earlier the publication can be retracted and the less wasted effort goes into new research that is based on disinformation within the scientific record. Therefore, it would be advantageous to have an early warning system that spots potential problems with published papers, or maybe even before based on a preprint version….”
“Outside of eLife and , to an extent , PLoS , no one of scale and weight in the commercial publishing sector has really climbed aboard the Open Science movement with a recognition of the sort of data and communication control that Open Science will require .
So what is that requirement ? In two words – Replicability and Retraction . …”
“We have examined retracted publications in different subject fields and attempted to analyse whether online free accessibility (Open Access) influences retraction by examining the scholarly literature published from 2000 through 2019, an incidence of the recent 20 years of publications. InCites, a research analytics tool developed by Clarivate Analytics®, in consultation with Web of Science, PubMed Central, and Retraction Watch databases were used to harvest data for the study. Retracted ‘Article’ and ‘Review’ publications were examined concerning their online accessibility mode (Toll Access and Open Access), based on non-parametric tests like Odds Ratio, Wilcoxon Signed Rank Test, Mann–Whitney U Test, Mann–Kendall and Sen’s methods. The Odds for OA articles to have retraction are about 1.62 as large (62% higher) compared with TA articles (95% CI 1.5, 1.7). 0.028% of OA publications are retracted compared with 0.017% TA publications. Retractions have occurred in all subject areas. In eight subject areas, the Odds for retraction of OA articles are larger compared with retraction of TA articles. In three subject areas, the Odds for retraction of OA articles are lesser compared with the retraction of TA articles. In the remaining 11 subject areas, no significant difference is observed. Post-retraction, though a decline is observed in the citation count of OA & TA publications (p?<?.01), yet the Odds for OA articles to get cited after retraction are about 1.21 as large (21% higher) compared with TA articles (95% CI 1.53, 1.72). TA publications are retracted earlier compared to OA publications (p?<?.01). We observed an increasing trend of retracted works published in both modes. However, the rate of retraction of OA publications is double than the rate of retraction of TA publications.
Abstract: This communication refers to the retractions of the two high profile COVID-19 papers of the top medical journals when the data analytics company declined to share the raw data of the papers. In this commentary, we emphasize that it is very pertinent for the journals to mandatorily ask the authors for sharing of the primary data. This will ensure data integrity and transparency of the research findings, and help in negating the publication frauds.
Here’s the full paper:
“Aries Systems Corporation, a leading technology workflow solutions provider to the scholarly publishing community, and scite, a platform for discovering and evaluating scientific articles via Smart Citations, are pleased to announce their partnership to facilitate the veracity of scientific references….”
“At the STM Association Annual Meeting in “virtual Frankfurt” last week, much of the focus was on how scholarly publishers are responding to the COVID crisis. Publishing executives reported how they have accelerated their editorial and peer review processes for COVID submissions, rightly taking pride in the contributions they have made to fighting the pandemic. They also emphasized again and again that they want to be more trusted. This is a formidable challenge in light of some recent failures. To achieve their objectives, publishers need to become more comfortable talking about their mistakes to prove convincingly that they are learning from them….
At the same time, I would encourage publishers to balance their celebrations with self-reflection. Scholarly publishers wish to see themselves as stewards of the scholarly record and of the transition to open science. To do so in a way that is compelling to all stakeholders, they must continuously increase the quality and rigor of their work, probe their processes for weaknesses, and make their work ever more resilient against potential points of failure. …
Today, the scholarly publishing sector looks to reestablish itself as a steward of the scholarly record and a trusted party to lead the transition to open science, and we need it in this type of role more than ever. Being entrusted with this role requires that publishers identify problems honestly and with humility, since trust is earned, or squandered, at a sector-wide level. The sector does not need triumphalism from leaders that enables their organizations to downplay festering problems. And, it does not need its boosters to selectively amplify concerns with preprints — when publishers should focus on their own shortcomings. The sector needs not only to ask for trust but also to make sure that it is continuously earning it every day.”
“According to an investigative report in The Guardian, Sapan Desai had been previously linked to highly ambitious (and dubious) claims. In 2008, he promoted a “next generation human augmentation device” called Neurodynamics Flow, which he said “can help you achieve what you never thought was possible,” claiming that “with its sophisticated programming, optimal neural induction points, and tried and true results, Neurodynamics Flow allows you to rise to the peak of human evolution.”
It is important to realize that concerns about the existence and validity of the Surgisphere databases surfaced only after the paper on hydroxychloroquine was published. The earlier NEJM paper on inhibitors of the renin-angiotensin system was never criticized, even though Surgisphere was the primary data and analytical source.
Why? The NEJM paper included data from 8,910 patients treated at 169 hospitals across three continents (Asia, Europe and North America), a database that may have seemed credible — even though Surgisphere had no track record of publications. In contrast, the Lancet paper cited data from 96,032 patients treated at 671 hospitals from six continents. It seems that the decision by the authors to include data from Australia and Africa represented a fatal strategic error, since these could be far more easily matched up with public records. When the data from these two regions failed to make sense, the paper unraveled. Conceivably, if the authors had not overreached and if they had merely confined their analysis to three continents, it is likely that the Lancet paper would have survived….
The possibility that fraudulent data would have been accepted — if it had not been for the excessive ambitions of the authors — is distressing beyond words. The implications for medical research are profound….
Many have criticized preprint servers because they allow the dissemination of data and information that has not been peer-reviewed. But can we continue to denigrate papers lacking peer review if the process failed us at this critical time? Some might still argue that peer review was highly effective in the two COVID-19 retractions; it simply occurred following (rather than prior to) publication. However, even the staunchest advocates of journals as gatekeepers must concede that the post-publication examination and analysis can occur whether the information is presented in a top-tier journal or on a preprint server….”
Two major study retractions in one month have left researchers wondering if the peer review process is broken.
Abstract: This study investigates whether negative citations in articles and comments posted on post-publication peer review platforms are both equally contributing to the correction of science. These 2 types of written evidence of disputes are compared by analyzing their occurrence in relation to articles that have already been retracted or corrected. We identi-fied retracted or corrected articles in a corpus of 72,069 articles coming from the Engineer-ing field, from 3 journals (Science, Tumor Biology, Cancer Research) and from 3 authors with many retractions to their credit (Sarkar, Schön, Voinnet). We used Scite to retrieve contradicting citations and PubPeer to retrieve the number of comments for each article, and then we considered them as traces left by scientists to contest published results. Our study shows that contradicting citations are very uncommon and that retracted or corrected articles are not more contradicted in scholarly articles than those that are neither retracted nor corrected but they do generate more comments on Pubpeer, presumably because of the possibility for contributors to remain anonymous. Moreover, post-publication peer review platforms, although external to the scientific publication process contribute more to the correction of science than negative citations. Consequently, post-publication peer review venues, and more specifically the comments found on it, although not contributing to the scientific literature, are a mechanism for correcting science. Lastly, we introduced the idea of strengthening the role of contradicting citations to rehabilitate the clear expression of judgment in scientific papers.
“Rather than being focused on the “Impact Factor,” perhaps authors should focus on the other “IF” or the “Integrity Factor.” I propose that we begin calculating the “Integrity Factor” for journals and perhaps this should be the number of retractions in, say, a 5? or 10?year period divided by the number of original research papers published. Authors could then pick a journal based upon its integrity rather than the impact….”
Abstract: Articles may be retracted when their findings are no longer considered trustworthy due to scientific misconduct or error, they plagiarize previously published work, or they are found to violate ethical guidelines. Using a novel measure that we call the “retraction index,” we found that the frequency of retraction varies among journals and shows a strong correlation with the journal impact factor. Although retractions are relatively rare, the retraction process is essential for correcting the literature and maintaining trust in the scientific process.