Retraction Watch

Tracking retractions as a window into the scientific process

The Retraction Watch Leaderboard

with 21 comments

Who has the most retractions? Here’s our unofficial list (see notes on methodology), which we’ll update as more information comes to light:

  1. Yoshitaka Fujii (total retractions: 183) Sources: Final report of investigating committee, our reporting
  2. Joachim Boldt (94) Sources: Editors in chief statement, additional coverage
  3. Diederik Stapel (58) Source: Our cataloging
  4. Adrian Maxim (48) Source: IEEE database
  5. Peter Chen (Chen-Yuan Chen) (43) Source: SAGE, our cataloging
  6. Hua Zhong (41) Source: Journal
  7. Shigeaki Kato (39) Source: Our cataloging
  8. James Hunton (37) Source: Our cataloging
  9. Hendrik Schön (36) Sources: PubMed and Thomson Scientific
  10. Hyung-In Moon (35) Source: Our cataloging
  11. Naoki Mori (32) Source: PubMed, our cataloging
  12. Tao Liu: (29) Source: Journal
  13. Cheng-Wu Chen (28) Source: our cataloging
  14. Gideon Goldstein (26)
  15. Scott Reuben (25)
  16. Gilson Khang (22) Sources: WebCitation.org, WebCitation.org, journal
  17. Friedhelm Herrmann (21)
  18. Noel Chia (21)
  19. Dipak Das (20) Click here for a full list of retracted papers
  20. Khalid Zaman (20)
  21. Jin Cheng (19)
  22. Bharat Aggarwal (18)
  23. Fazlul Sarkar (18)
  24. John Darsee (17)
  25. Wataru Matsuyama (17)
  26. Alirio Melendez (17)
  27. Robert Slutsky (17)
  28. Ulrich Lichtenthaler (16)
  29. Erin Potts-Kant (16)
  30. Pattium Chiranjeevi (15)

We note that all but one of the top 30 are men, which agrees with the general findings of a 2013 paper suggesting that men are more likely to commit fraud.

Notes:

Many accounts of the John Darsee story cite 80-plus retractions, which would place him third on the list, but Web of Science only lists 17, three of which are categorized as corrections. That’s not the only discrepancy. For example, Fujii has 138 retractions listed in Web of Science, compared to 183 as recommended by a university committee, while Reuben has 25, compared to the 22 named in this paper. We know that not everything ends up in Web of Science — Chen, for example, isn’t there at all — so we’ve used our judgment based on covering these cases to arrive at the highest numbers we could verify.

Shigeaki Kato is likely to end up with 43 retractions, based on the results of a university investigation.

All of this is a good reminder why the database we’re building with the generous support of the MacArthur Foundation and Arnold Foundation will be useful.

Like Retraction Watch? Consider supporting our growth. You can also follow us on Twitter, like us on Facebook, add us to your RSS reader, and sign up on our homepage for an email every time there’s a new post.

Written by Ivan Oransky

June 16th, 2015 at 11:09 am

Posted in

Comments
  • Dave Fernig June 16, 2015 at 3:48 pm

    Q number of those in the list have papers that look like they should be retracted but haven’t. I look forward to the “retraction curve”, is it hyperbolic or is it best described by some other function?

  • Distractingly Sexy June 16, 2015 at 6:01 pm

    Is it possible that crying in the lab and falling in love with the PI make you more honest?

    http://blogs.scientificamerican.com/voices/furor-over-tim-hunt-must-lead-to-systemic-change/

  • Elliott June 17, 2015 at 7:05 am

    Marion A. Brach, the other active part in them widely-known “Herrmann-Brach affair”, has 14 retractions: http://www.ncbi.nlm.nih.gov/pubmed/?term=brach+m+AND+retracted

  • Elliott June 17, 2015 at 10:39 am

    Jon Sudbø, formerly at the Norwegian Radium Hospital and the University of Oslo, has 12 retractions. The fabrications in his papers were revealed in 2006. See http://www.ncbi.nlm.nih.gov/pubmed/?term=sudbo+j+AND+retract* and https://en.wikipedia.org/wiki/Jon_Sudbø

  • Elliott June 17, 2015 at 11:06 am

    What about Khalid Zaman, who has 16 retractions – at least according to your own post (http://retractionwatch.com/2014/12/19/elsevier-retracting-16-papers-faked-peer-review/)?

  • Professor Karen Woolley PhD CMPP June 18, 2015 at 8:39 pm

    Curious…how many of these leading offenders worked for pharmaceutical/device companies? Given your reference to anesthesiology, it is interesting to note that the editor of “Anesthesia and Analgesia” published that “Anesthesia & Analgesia has experienced its share of fraud. Not a single case, including this one, has involved a study directly sponsored by a drug or device company. Sponsored studies are very closely audited, with each case report form checked against patient and laboratory data.” Keeping an eye on industry (‘the usual suspect’) is understandable, intense, somehow comforting, and helps sell books, but what if the real enemy isn’t within industry?

    • Steven Shafer November 16, 2015 at 7:25 pm

      Having written the quoted text, I have yet to see outright fraud in any study that was directly sponsored by a pharmaceutical company. I have significant issues with other kinds of misrepresentation. A classic example is Merck claiming that naproxen reduced risk of myocardial infarction, rather state the study finding that rofecoxib increased risk (see NEJM 353;26, 2005). The cases of fraud that I’ve personally handled, including Fujii, Boldt, and Reuben, were not directly sponsored by industry. One of the few Boldt studies NOT retracted was funded by pharma. This study had an actual IRB approval, underwent patient-level auditing by the company, and (no surprise) took years longer to complete than his fraudulent studies.

      One has to be alert for fraud and misrepresentation in any study. However, pharmaceutical companies have too much at risk to engage in outright fabrication.

      • Professor Karen Woolley PhD CMPP November 16, 2015 at 9:04 pm

        Hi Steven,
        Thank you for the additional information. To reduce the risk of retraction, industry sponsors are introducing publication audits (eg, to check authors meet authorship criteria, to check disclosures are made, to ensure authors had access to data etc…). For the first time, “auditing” appears in the Good Publication Practice guidelines (now in its third version – GPP3; published Annals of Internal Medicine 2015; disclosure – I’m a co-author). All sponsors, industry or otherwise, should consider a risk-based approach to publication audits. The possibility of an audit may help PREVENT (vs detect) publication misconduct. If most retractions occur with academic-sponsored research, then academia needs to find funds to conduct publication audits. Clinical trial audits are conducted by industry and others…why not publication audits?

  • Thomas Munro June 19, 2015 at 1:10 pm

    Some champion chemists will be feeling neglected:

    Hua Zhong: 41
    http://journals.iucr.org/e/issues/2010/01/00/me0404/index.html

    Tao Liu: 29
    http://journals.iucr.org/e/issues/2010/01/00/me0405/index.html

    Pattium Chiranjeevi: ~30 retracted out of ~70 fakes.
    google scholar {(retraction OR retracted) author:”chiranjeevi p”}
    http://www.rsc.org/chemistryworld/News/2008/March/25030801.asp

  • fernando pessoa June 20, 2015 at 11:57 am

    “We note that all but one of the top 25 are men”

    Marion A. Brach (14) Woman
    Silvia Bulfone-Paus (13) Woman

  • Thomas Munro June 21, 2015 at 8:25 am

    I’m very impressed by the rapid updates here, and I stand corrected on Chiranjeevi’s tally. As with Darsee and Fujii, many journals seem to have more important matters to attend to than retracting fakes. Perhaps you could have a leaderboard for journals: the number of fraudulent papers identified in formal investigations that remain unretracted.

    One more contender, who sadly won’t make the top ten:
    Gilson Khang: 21
    http://www.webcitation.org/6VD9lOA5o
    http://www.webcitation.org/6VD9x5Ewi

    These papers, all in “Tissue Engineering and Regenerative Medicine”, are mentioned in Grieneisen’s 2012 review. I’m glad I archived them, because they have since vanished.

    • Aleksey July 10, 2016 at 9:26 am

      Great idea about the retraction-resistant journals

  • Elliott June 22, 2015 at 12:34 pm

    What to do with co-authors? Probably, most of them are “innocent”, of course, and therefore shouldn’t be listed.

    However, about Hidenori Toyooka, co-author on several dozens of record holder Fuji’s retracted papers, RW wrote:

    The investigators do identify one co-author, Hidenori Toyooka, who appears to have known about the fabrication and yet still co-authored “dozens” of papers with Fujii. According to the report, Toyooka “recognized the suspicion” raised against his colleague in 2000, but “did not take any action.” (http://retractionwatch.com/2012/07/02/does-anesthesiology-have-a-problem-final-version-of-report-suggests-fujii-will-take-retraction-record-with-172/)

    Does he “deserve” to be listed?

    Another example: Elena Bulanova and Vadim Budagian, the two former postdocs of Silvia Bulfone-Paus, appeared as co-authors on twelve of the thirteen retracted Bulfone-Paus papers. In the formal investigation both were officially blamed responsible for the misconduct in all these twelve papers. See, for example: https://www.timeshighereducation.co.uk/news/new-retraction-of-paper-by-husband-and-wife-research-team/418431.article

    I think, those two definitely deserve to be listed.

    • Prof Henry August 11, 2015 at 2:37 am

      The “innocence” of co-authors is a difficult and problematic concept. Clearly, co-authorship confers many rewards – citations, acknowledgments, H-(and other) indexes, expanding lists of publications. It also confers responsibilities. The ICMJE “Recommendations for the Conduct, Reporting, Editing, and Publication of Scholarly Work in Medical Journals” (often referred to as the Vancouver Protocols) state this clearly:

      “Authorship confers credit and has important academic, social, and financial implications. Authorship also implies responsibility and accountability for published work.”

      Since the first edition in 1979, there have been statements voicing the expectation that those named as authors take responsibility for the content of papers bearing their names. Consistently, it has recommended that a “covering letter should contain a statement that the manuscript has been seen and approved by all authors”. By 1988, this had firmed up into the form of words that underlie Authorship Criteria in most national and international statements today:

      “All persons designated as authors should qualify for authorship. Each author should have participated sufficiently in the work to take public responsibility for the content.”

      For large-scale multi-centre projects, a caveat was added in about 1999:
      “Each author should have participated sufficiently in the work to take public responsibility for appropriate portions of the content. … Some journals now also request that one or more authors, referred to as “guarantors,” be identified as the persons who take responsibility for the integrity of the work as a whole, from inception to published article, and publish that information.”

      Nevertheless, it was reasserted that those named as authors must have met all the criteria for authorship and have given “final approval of the version to be published.” By 2013, the provision for assigning overall responsibility to a single ‘guarantor’ (even in large multi-centre, multi-group studies) had been abandoned and a stronger statement added as a 4th criterion for authorship:

      “4. Agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investi- gated and resolved.
      In addition to being accountable for the parts of the work he or she has done, an author should be able to identify which co-authors are responsible for specific other parts of the work. In addition, authors should have confidence in the integrity of the contributions of their co- authors.
      All those designated as authors should meet all four criteria for authorship, and all who meet the four criteria should be identified as authors. Those who do not meet all four criteria should be acknowledged—see Section II.A.3 below. These authorship criteria are intended to reserve the status of authorship for those who deserve credit and can take responsibility for the work.”

      It must be really tough if you have been “deceived” by one of your colleagues into believing that some data or other images, or key calculations were accurate when they were not but it seems to me that one of the guarantees on which we depend when we read co-authored papers is that integrity of those key elements in which scientific discoveries are based have been validated, subjected to at least some due diligence by those who have signed their names to the paper.

      I think that when co-authors understand the gravity of the consequences of not doing that, we might just have fewer cases of research misconduct appearing in the literature. I hope.

    • John February 23, 2016 at 11:51 am

      This brings up a very interesting topic for debate. Lots of discussion about relative credit for authorship and what it means to be first, second, or corresponding author. Not so much on how to allot credit for retractions and how much co-authors should have known or done.

  • MK June 24, 2015 at 11:25 am

    “Mastuyama” should be “Matsuyama”

  • Jason August 19, 2015 at 7:14 pm

    Following the Maryka Quik link it seems that you’ve counted both the retracted paper and the retraction letter. I actually count 8, not 16, retractions. You should give at least a quick check on all of these.

  • Thomas Munro August 26, 2015 at 12:22 pm

    I think Adrian Maxim’s correct tally is 48, as given in Grieneisen and Zhang 2012. If you broaden the search to “Maxim, A” they all show up on the IEEE search.

  • Andrew Moore January 25, 2016 at 3:37 am

    Given that systematic reviews have been at the forefront of picking up fraud resulting in subsequent retraction (See Tramer EJAnaesthesiol 2013 30: 195), it is worth seeing how many systematic reviews might blindly continue to add retracted data. As best I can see Fujii data still appears in systematic reviews – probably in this one, for example.

    Surg Laparosc Endosc Percutan Tech. 2013 Feb;23(1):79-87. doi: 10.1097/SLE.0b013e31827549e8.
    Comparison of the efficacy of ondansetron and granisetron to prevent postoperative nausea and vomiting after laparoscopic cholecystectomy: a systematic review and meta-analysis.
    Wu SJ1, Xiong XZ, Lin YX, Cheng NS.

    If they have included retracted data, shouldn’t the systematic review also be retracted?

    • Aleksey July 10, 2016 at 1:53 pm

      Great question, I think they should. I would retract at least half of all scientific papers published, as they are not reproducible and flawed.

  • KK August 18, 2016 at 7:01 pm

    would it be informative if you insert last update date at the bottom of this post?

  • Post a comment

    Threaded commenting powered by interconnect/it code.