Europe PMC

This website requires cookies, and the limited processing of your personal data in order to function. By using the site you are agreeing to this as outlined in our privacy notice and cookie policy.

Abstract 


We describe the ongoing citations to biomedical articles affected by scientific misconduct, and characterize the papers that cite these affected articles. The citations to 102 articles named in official findings of scientific misconduct during the period of 1993 and 2001 were identified through the Institute for Scientific Information Web of Science database. Using a stratified random sampling strategy, we performed a content analysis of 603 of the 5,393 citing papers to identify indications of awareness that the cited articles affected by scientific misconduct had validity issues, and to examine how the citing papers referred to the affected articles. Fewer than 5% of citing papers indicated any awareness that the cited article was retracted or named in a finding of misconduct. We also tested the hypothesis that affected articles would have fewer citations than a comparison sample; this was not supported. Most articles affected by misconduct were published in basic science journals, and we found little cause for concern that such articles may have affected clinical equipoise or clinical care.

Free full text 


Logo of nihpaAbout Author manuscriptsSubmit a manuscriptHHS Public Access; Author Manuscript; Accepted for publication in peer reviewed journal;
Sci Eng Ethics. Author manuscript; available in PMC 2014 Aug 22.
Published in final edited form as:
PMCID: PMC4141682
NIHMSID: NIHMS527184
PMID: 19597966

Analysis of Citations to Biomedical Articles Affected by Scientific Misconduct

Abstract

We describe the ongoing citations to biomedical articles affected by scientific misconduct, and characterize the papers that cite these affected articles. The citations to 102 articles named in official findings of scientific misconduct during the period of 1993 and 2001 were identified through the Institute for Scientific Information Web of Science database. Using a stratified random sampling strategy, we performed a content analysis of 603 of the 5,393 citing papers to identify indications of awareness that the cited articles affected by scientific misconduct had validity issues, and to examine how the citing papers referred to the affected articles. Fewer than 5% of citing papers indicated any awareness that the cited article was retracted or named in a finding of misconduct. We also tested the hypothesis that affected articles would have fewer citations than a comparison sample; this was not supported. Most articles affected by misconduct were published in basic science journals, and we found little cause for concern that such articles may have affected clinical equipoise or clinical care.

Keywords: Bibliometric analysis, Journalology, Journal citations, Quantitative content analysis, Retraction, Scientific misconduct

Introduction

A number of studies have identified and described the continuing use of retracted publications in the biomedical literature.Budd et al. (1998) found 235 articles published from 1966 through 1997 that were tagged in MEDLINE as retracted, noting that these articles were cited 2034 times after retraction, and that most of these citations were “implicitly positive” while 275/2034 were explicitly positive. Similarly, Nath et al. (2006) used MEDLINE to identify 395 articles published between 1982 and 2002 and subsequently retracted, and categorized the reasons for retraction as misconduct or unintentional error. They estimated that retractions are more than twice as likely to result from unintentional mistakes as from scientific misconduct.

Recently, Cokol et al. (2008) reported that retraction rates are increasing, although author unawareness is likely a factor in the continuing use of retracted literature (Budd et al. 1998, 1999; Snodgrass and Pfeifer 1992). In the past, retractions and corrections were not always accessible or obvious to database users (Snodgrass and Pfeifer 1992, Neale et al. 2007, Pfeifer and Snodgrass 1992). Journals and libraries have a variety of approaches to handling retractions and requests for retractions (Pfeifer and Snodgrass 1992; Atlas 2004; Friedman 1990; Parrish 1999). This landscape seems to be improving with more awareness of scientific misconduct and the need for policies for retracting affecting articles, (Sox and Rennie 2006) plus improvements in databases (Garfield et al. 2006; National Library of Medicine 2008; Norton and Saltman 2007).

Our study of the retracted and corrected publications took a different approach: we started with articles affected by scientific misconduct, and then identified if they were retracted or corrected, the extent that they were subsequently cited in other papers, and the way that citing authors used the information in these affected papers. The use of literature affected by misconduct is concerning because it has the potential to misdirect subsequent research and clinical care (Budd et al. 1999; Benos et al. 2005; Couzin and Unger 2006; Gardner et al. 2005; Katz 2006; Roberts et al. 2007).

Study Background

In a previous paper from the same study, we identified 102 published research articles that were named in official findings of scientific misconduct, and reported on whether they had been retracted or corrected per the administrative actions in the scientific misconduct report (Neale et al. 2007). To identify the 102 articles affected by scientific misconduct, the study methods involved a content analysis of all the “Findings of Scientific Misconduct” published in two public sources (the NIH Guide for Grants and Contracts, and the Annual Reports of the U.S. Office of Research Integrity) during the period of 1993–2001. We also determined the number of citations to these 102 affected articles by the retraction or correction status of each.

The purpose of this report is to characterize the papers that subsequently cited the 102 articles affected by scientific misconduct. Three study objectives are addressed: (1) Characterization of a sample of published papers that cited the articles affected by scientific misconduct using a quantitative content analysis methodology. Specifically we evaluated how the citing authors used the affected article, and determined if the citing paper contained an indication of awareness that it was citing an article affected by scientific misconduct. (2) Evaluation of the hypothesis that affected articles (i.e., named in findings of scientific misconduct) have fewer citations than those in a comparable group of articles unaffected by scientific misconduct. (3) Consideration of whether clinical practice might have been influenced by any of the 102 articles affected by scientific misconduct. To address this third objective, we used a case study approach to examine those papers that used an affected clinical article to directly support their own study purpose, and we also supplemented our observations with literature searching and expert opinion.

Methods

Articles Affected by Scientific Misconduct

We used the same sample of 102 articles affected by scientific misconduct that are described by (Neale et al. 2007). As previously detailed, we used the Cited References Search in the Institute for Scientific Information (ISI) Web of Science (Thomson Reuters 2009) to determine the number of citations by subsequent authors to the 102 articles affected by misconduct. The citation analysis was conducted during the week of May 17, 2005; as of that time, the Web of Science database listed 5,393 citations to the 102 articles (range: 0–592; median of 26 citations per article) (Neale et al. 2007).

Stratified Random Sample of Citing Papers

To address the first study objective (to characterize citing papers), we conducted a content analysis of papers that cited the articles affected by scientific misconduct. As it was not feasible to perform a content analysis on the population of the 5,393 citing papers, we developed a stratified random sampling strategy, stratified on the number of citations, that included at least one citation to each of the 86 affected articles that had a citation (16/102 affected articles did not have a citation at the time of data collection). If an affected article had 1–3 citations (n = 22), then we randomly selected 1 of these citing papers; if an affected article had ≥4 citations (n = 64), then we randomly selected 33% of these citing papers. This methodology also allowed citations to accrue for a minimum of three years after publication.

Using this stratified random sampling strategy, 603 articles were drawn from the population of 5,393 citing articles (excluding the approximately 65 articles used for the content analysis training). The following eligibility criteria were used to select the stratified random sample of citing articles for the content analysis: (1) only English-language articles; (2) only articles denoted in PubMed as a research study or a review article; (3) only articles for which we obtained a copy of the article (either by open access, institutional subscription, Loansome Doc or inter-library loan); (4) only citing papers published at least one year after the first official notification of the finding of scientific misconduct. When an ineligible citing paper was sampled (e.g. not in English; a letter to the editor; or unsuccessful efforts to obtain a copy of the article), then another was selected at random from the stratum; this occurred in <10% of sampled papers.

Content Analysis of Citing Papers

During the content analysis, each of the 603 sampled papers was read and the following information abstracted and recorded: (1) the type of study in the citing paper (human clinical, human basic science or animal); (2) whether the affected article was explicitly referenced such as a direct naming of the author or the study findings, or referred to implicitly such as being embedded in a string of citations; (3) evidence of awareness that the article was affected by scientific misconduct (e.g. citation of a retraction/correction or discussion of the scientific misconduct investigation); and (4) the nature of how the affected article was used in the citing paper (direct support/contrast, indirect support/contrast, benign use, and other; see Table 1 for operational definitions).

Table 1

Operational definitions for content analysis of how the affected articlea was used by the citing paper.b

Nature of use of affected article by citing paperExamples
Direct support: Citing author used invalid information from the affected article to directly support research (hypothesis, specific aim or study findings). This rating denotes a direct confirmation of the invalid findings in the affected article.The introduction section of a citing paper refers to an invalid finding from the affected article. The specific aim or hypothesis of the citing paper is directly derived from or linked to the invalid study finding from the affected article.
Indirect support: Citing author used invalid information from an affected paper in a discussion of the of the citing paper’s aims or findings, but did not directly confirm or support the new information.The introduction section of a citing paper refers to an invalid finding of an affected article. The new hypothesis or study aim may be related to the invalid study finding but the mechanism of action, organ system, or study environment is substantially different.
Direct Contrast: Citing author used invalid information from the affected article to directly contrast a new hypothesis, specific aim, or study finding. This rating denotes a direct challenge of the findings reported in the affected article.The discussion section of a citing paper refers to the invalid finding of an affected article. The specific aim or hypothesis of the citing paper directly relates and is linked to the invalid findings of the affected article, except that the results are contradictory.
Indirect contrast: Citing author uses invalid information from the affected article in a discussion of the citing paper’s aims or findings. However, the citing paper does not directly challenge the invalid information.The introduction of a citing paper refers to the invalid finding of an affected article. The specific aim or hypothesis of the study is different from the affected article but still linked by a broader goal or technology use and the citation reflects that an improved approach is being studied.
Awareness: Citing author indicated awareness that information from the affected article was potentially invalid and/or in need of a correction or retraction.The citing paper includes in its reference list both the affected article and its published corrigenda.
Benign: Citing author did not address the invalid information from the affected article; only referred to valid information in the article.This rating was used when valid information was cited from an affected article with an indexed correction.
Other: This rating was used with occurrences that did not fit well into the above categories. These unusual occurrences were discussed by the study team.The methods section of a citing paper indicated that a step in the procedure of a retracted article was replicated. We could not determine if this step is common (i.e. validated elsewhere) in this type of study or was unique only to the retracted article.
aAffected article: 86/102 articles affected by scientific misconduct had at least one citation
bCiting paper: One of the 603 papers that cited one of the affected articles (selected by a stratified random sampling strategy)

Content Analysis Training

Content analysis training was conducted in two phases. In the first phase, the principal investigator (AVN) and the study coordinator separately abstracted 20 citing papers using an open-ended data abstraction form. After a discussion of findings, a close-ended data abstraction form was developed and used in a second pilot study of 20 different citing papers. After another review and discussion, the content analysis data abstraction form was revised into its final form.

A majority of the 102 articles affected by misconduct were basic science studies (52.9% animal; 20.6% human), and 26.5% were human clinical studies. An advanced graduate student completing a Master of Science degree in our medical school’s biomedical sciences program was retained to read and abstract the 603 citing papers, most of which were published in basic science journals. The graduate student was trained by the study coordinator in the methods of the content analysis using 10 citing papers. We then conducted two analyses of the proportion of abstractions in which they had any disagreements. In the first analysis, they independently abstracted 15 citing papers and found disagreement on only one variable in each of three articles. Consensus for the coding of the three discrepancies was reached by discussion. In the second analysis, they each abstracted 21 new citing papers and found disagreement on one variable in each of 5 papers. Following more discussion about coding nuances, the graduate student read and independently abstracted the remaining citing papers sampled for the content analysis. Occasional questions were resolved with the study team.

Comparison Group

To test the hypothesis in the second study objective, we developed a matched comparison group to determine if the affected articles were cited at a comparable rate as similar publications that were not named in “Findings of Scientific Misconduct”. The methodology for identifying comparison articles was to select one article from the same year and journal issue as each affected article. Other eligibility criteria for selection as a comparison article were: (1) Different authors than affected article; (2) No evidence that the article was affected by misconduct, as indicated by corrections or retractions indexed in PubMed; (3) Same type of paper: original research report or review article; (4) Same type of study: basic science (animal or human in vitro) or human clinical (determined from the MEDLINE database, and the content analysis; (5) Indexed in the Institute for Scientific Information (ISI) Web of Science (Thomson Reuters 2009); and (Neale et al. 2007) Printed adjacent (preferably immediately after) the affected article. If the affected article was the last article in the issue, then the article just before it in that issue was selected. If the selected article did not meet all inclusion criteria, then the next adjacent article was sampled, until an eligible article found. The number of citations to this comparison group was determined from the Cited Reference Search in the ISI Web of Science database (Thomson Reuters 2009) during the period of 8/8/2005 – 8/23/2005. The distribution of the citation data was skewed and so we used the Wilcoxen signed rank test to compare the number of citations for the comparison group to that for the affected articles during the period of 05/17/2005 through 05/20/ 2005 (Neale et al. 2007).

Consideration of Effect on Clinical Practice

To consider how articles affected by misconduct possibly might have affected clinical care (the third study objective), we identified the four papers (Caen and Han 1993; Maier and Watkins 1999; Whitsett 1995; Zauli and Catani 1995) that used an affected article for direct support of a study purpose or finding to evaluate any evidence that they might have influenced clinical practice. We also used other qualitative approaches (literature searching and expert opinion) for insights into the clinical use of findings from such affected articles.

Results

Characterization of Citing Papers

The citation to an article affected by misconduct was embedded in a string of references in 61% of the 603 sampled citing papers; there was an explicit reference to an affected article in 39% of citing papers. Table 2 shows how the affected article was used by subsequent citing authors: 8.6% of citing papers explicitly used the affected article as either direct support or direct contrast of their study; and 54.1% implicitly used the affected article as indirect support or indirect contrast. (See Table 1 for definitions of how the citing papers used the affected articles.) One-third (32.8%) of citing papers did not address invalid information in the affected article (i.e. they used a portion of the affected article that was not affected by the misconduct, such as the literature review or a description of a methodological technique). Only 2.8% of the 603 citing articles referenced the corrigendum (i.e. retraction, erratum or a comment tag) to the article affected by scientific misconduct.

Table 2

Nature of use of 86/102 articles affected by scientific misconduct in 603 subsequent citing papers.a

Nature of UsebNumberPercent
Direct Support50c8.3
Indirect support30049.8
Direct contrast2d0.3
Indirect contrast264.3
Awareness172.8
Benign19832.8
Other101.7
Total603100.0
aA stratified random sample of 603 citing papers was drawn from a population of 5,393 citations that had accrued as of 08/8/2005 through 08/23/2005 [determined from the ISI Web of Science (Thomson Reuters 2009)]
bOperational definitions shown on Table 1
c4 of these 50 direct support articles were clinical studies
dNone of the direct contrast articles were clinical studies

Citations to Affected and Comparison Articles

Table 3 shows that the 102 articles affected by misconduct had a median of 26 citations, and the 102 comparison articles had a median of 27 citations (p = 0.08). Thus, the hypothesis that affected articles named in findings of scientific misconduct will have fewer citations than those in a comparison group (study objective 2) was not supported.

Table 3

Comparison of citations to articles affected by scientific misconducta and a matched comparison groupb

Number of articlesMedian citations
by subsequent authors
Total citations
Affected articles10225.55,393
Comparison articles10227.06,363

p = 0.08 (Wilcoxon signed rank test)

aCitations to articles affected by scientific misconduct were determined from the ISI Web of Science (Thomson Reuters 2009) during 05/17/2005 through 05/20/2005
bCitations to comparison group articles were determined from the ISI Web of Science (Thomson Reuters 2009) during 08/8/2005 through 08/23/2005

Influence of Affected Articles on Clinical Practice

In regards to study objective 3, Table 2 shows that 50 of the 603 sampled citing papers explicitly used an affected article for direct support of a study purpose; four of these 50 papers cited a clinical study. Neither of the two papers that explicitly used an affected paper for direct contrast were clinical studies. We reviewed the four citing papers for any evidence that they possibly might have affected clinical care. Each of these four papers (Caen and Han 1993; Whitsett 1995; Whitsett 1995; Zauli and Catani 1995) cited the same clinical article affected by scientific misconduct titled “Preliminary report: Effects of interleukin-1 on platelet counts”, and authored by Anand Tewari, William Buhles and H. Fletcher Starnes (Tewari et al. 1990). In part, the abstract of this article states the following:

“Recombinant human interleukin-1β was given in 5 daily intravenous infusions to ten patients with metastatic malignant disorders as part of an antineoplastic trial…. A 50% rise in platelets occurred in response to interleukin-1β…Interleukin-1β may therefore be beneficial in the treatment of conditions of thrombocytopenia associated with haematological disorders and chemotherapy for malignant disorders.” (Tewari et al. 1990).

As indicated in PubMed, Bhules and Starnes (1992) retracted their paper coauthored with Tewari on 08/22/1992. The retraction states that some arithmetical errors were made in summarizing the data and the overall conclusions as originally stated are valid, but they must retract the paper.

The four papers that cited theTewari et al. (1990) article were all review articles published after the 1992 retraction. The citing paper by Caen and Han (1993) was published in 1993, and it is possible that the retraction (Buhles and Starnes 1992) was posted around the time or even after Caen and Han (1993) submitted their manuscript for publication. The other three citing papers were published after longer time lags: both Whitsett’s (1995) paper and the Zauli and Catani (1995) paper were published in 1995. And finally, the Maier and Watkins (1999) review article was published in 1999. Three of these four review articles (Caen and Han 1993; Maier and Watkins 1999; Zauli and Catani 1995) used theTewari et al. (1990) article to support similar statements that interleukin-1 increases platelet production. Whitsett (1995) added that it “could increase platelet counts in cancer patients”, and Zauli and Catani (1995) stated that it “promotes platelet production in clinical trials.” Maier and Watkins (1999) citeTewari et al. (1990) with: “the administration of cytokines to humans produces reports of depressed mood”; this is curious because the Tewari article (Tewari et al. 1990) did not measure or discuss depression or any mood state.

To gain insight into the possible current use of interleukin-1 to increase platelet production, we searched the Clinical Practice Guidelines in Oncology™ published by the National Comprehensive Cancer Network, Inc. (2009), and did not identify any treatment guidelines suggesting interleukin-1 to increase platelets. We also consulted with four physicians and an advance practice nurse affiliated with our National Cancer Institute-designated comprehensive cancer center. Each said they were not aware that interleukin-1 was ever a standard of care to increase platelets. Based on this informal survey, we found no suggestion that theTewari et al. (1990) paper affected clinical oncology care.

Discussion

The judgment of scientific misconduct and the publication of subsequent corrigenda (notices of findings of misconduct, retractions or corrections) in PubMed did not seem to influence the number of citations to such affected articles, which were similar to those of a scientifically drawn comparison group. Indeed, fewer than 5% of citing authors evidenced any awareness that they were citing an article named in a judgment of scientific misconduct. Depending on the method for locating and accessing such affected papers, the authors of the citing paper may not have been exposed to the retraction (Neale et al. 2007). Recently, Eugene Garfield and colleagues (Garfield et al. 2006) noted that that the Science Citation Index in the ISI Web of Science offers a reliable approach to find retractions.

The majority of citing authors used an affected article in an implicit or indirect way, or referred to a portion of the affected article that was not discredited (e.g. the study methods). Four (all review articles) of the 603 sampled citing papers explicitly referred to one affected clinical study (Tewari et al. 1990) for direct support of their purpose, and it is unlikely that this single clinical study had a significant effect on clinical equipoise or clinical care.

Limitations

Several limitations to the study methodology are acknowledged. First, we were unable to determine the dates when retractions or errata were posted, and to identify which citations accrued before such postings. Second, the selection of the most appropriate comparison group was challenging. We used a matched comparison group strategy by selecting an adjacent article in the same issue. This reduced factors that could affect citations such as the journal impact factor, the journal’s open access status and the length of time since publication for citations to accrue. Third, it would be ideal to identify the citations to the population of articles affected by misconduct and to the comparison sample during the same brief time window; however the labor-intensive nature of the data collection resulted in slightly different time windows. The progression of work was such that we first identified the citations to the affected articles, then developed the comparison sample, and then identified the citations to the comparison sample.

Conclusion

Although most articles named in misconduct investigations have an identifiable corrigenda (Neale et al. 2007), few citing articles reference such retractions or corrections in their bibliography. Citing articles evidenced little awareness that they were using an article affected by scientific misconduct, which in part may be due to the past barriers to identifying retractions and corrections (Snodgrass and Pfeifer 1992; Neale et al. 2007). In spite of the ongoing use of discredited literature, science tends to be self-correcting through the processes of replication and post-publication peer review (Cokol et al. 2007; Couzin 2006; Poulton 2007), with occasional notable exceptions (Katz 2006). Although others have noted that citations diminish after retractions and public exposure of misconduct (Garfield and Welljams-Dorof 1990; Pfeifer and Snodgrass 1990), we found a similar level of citations to both the papers affected by scientific misconduct and those in a matched comparison group. Authors may not use all of the available information about the validity of some of the biomedical literature, yet there is little indication that the articles affected by scientific misconduct used in this research had an adverse effect on clinical practice. Nevertheless, the use of tainted literature likely still has the untoward consequences of wasting time and effort as well as undermining public trust in the integrity of the scientific enterprise (Sox and Rennie 2006; Benos et al. 2005; Horton 1999; Tobin 2000).

Acknowledgments

This research was supported by the Research on Research Integrity Program, on ORI/NIH collaboration, grant # R01 NS44487. We are grateful for the contributions of the study coordinator Justin Northrup, MPT, and Daniel Sapeika, MS, MD who performed the content analysis.

Contributor Information

Anne Victoria Neale, Department of Family Medicine and Public Health Sciences, Wayne State University School of Medicine, 101 E. Alexandrine, #249, Detroit, MI 48201, USA, ude.enyaw.dem@elaenv.

Rhonda K. Dailey, Department of Family Medicine and Public Health Sciences, Wayne State University School of Medicine, 101 E. Alexandrine, Detroit, MI 48201, USA.

Judith Abrams, Integrated Biostatistics Unit, Wayne State University School of Medicine, 4160 John R Street, Suite 716, Detroit, MI 48201, USA.

References

  • Atlas MC. Retraction policies of high-impact biomedical journals. Journal of the Medical Library Association. 2004;92:242–250. [Europe PMC free article] [Abstract] [Google Scholar]
  • Benos DJ, Fabres J, Farmer J, Gutierrez JP, Hennessy K, Kosek D, et al. Ethics and scientific publication. Advances in Physiology Education. 2005;29:59–74. [Abstract] [Google Scholar]
  • Budd JM, Sievert M, Schultz TR. Phenomena of retraction: Reasons for retraction and citations to the publications. Journal of the American Medical Association. 1998;280:296–297. [Abstract] [Google Scholar]
  • Budd JM, Sievert M, Schultz TR, Scoville C. Effects of article retraction on citation and practice in medicine. Bulletin of the Medical Library Association. 1999;87:437–443. [Europe PMC free article] [Abstract] [Google Scholar]
  • Buhles WC, Starnes HF. Retraction: Effects of interleukin-1 on platelet counts. The Lancet. 1992;340:496. [Abstract] [Google Scholar]
  • Caen JP, Han ZC. Control of megakaryocyte development: From basic data to clinical results. Comptes Rendus de l’Academie des Sciences. Série III, Sciences de la vie. 1993;316:925–930. [Abstract] [Google Scholar]
  • Cokol M, Iossifov I, Rodriguez-Esteban R. How many scientific papers should be retracted? European Molecular Biology Organization (EMBO) Reports. 2007;8:422–423. [Europe PMC free article] [Abstract] [Google Scholar]
  • Cokol M, Ozbay F, Rodriguez-Esteban R. Retraction rates are on the rise. EMBO Reports. 2008;9:2. [Europe PMC free article] [Abstract] [Google Scholar]
  • Couzin J. Stem cells…and how the problems eluded peer reviewers and editors. Science. 2006;311:23–24. [Abstract] [Google Scholar]
  • Couzin J, Unger K. Scientific misconduct. Cleaning up the paper trail. Science. 2006;312:38–43. [Abstract] [Google Scholar]
  • Friedman PJ. Correcting the literature following fraudulent publication. Journal of the American Medical Association. 1990;263:1416–1419. [Abstract] [Google Scholar]
  • Gardner W, Lidz CW, Hartwig KC. Authors’ reports about research integrity problems in clinical trials. Contemporary Clinical Trials. 2005;26:244–251. [Abstract] [Google Scholar]
  • Garfield E, McVeigh M, Muff M. Re: Research misconduct, retraction, and cleansing the medical literature: Lessons from the Poehlman case. Annals of Internal Medicine. 2006;145:472–473. [Abstract] [Google Scholar]
  • Garfield E, Welljams-Dorof A. The impact of fraudulent research on the scientific literature. The Stephen E. Breuning case. JAMA. 1990;263(10):1424–1426. [Abstract] [Google Scholar]
  • Horton R. Scientific misconduct: Exaggerated fear but still real and requiring a proportionate response. The Lancet. 1999;354(9172):7–8. [Abstract] [Google Scholar]
  • Katz TJ. Propagation of errors in review articles. Science. 2006;313:1236. [Abstract] [Google Scholar]
  • Maier SF, Watkins LR. Bidirectional communication between the brain and the immune system: Implications for behaviour. Animal Behaviour. 1999;57:741–751. [Google Scholar]
  • Nath SB, Marcus SC, Druss BG. Retractions in the research literature: Misconduct or mistakes? Medical Journal of Australia. 2006;185:152–154. [Abstract] [Google Scholar]
  • National Comprehensive Cancer Network, Inc. [Accessed 1 Jan 2009];NCCN Clinical Practice Guidelines in Oncology™ (v.1.2009, 09/10/08) 2009 [website]. http://www.nccn.org. [Google Scholar]
  • National Library of Medicine. [Accessed 12 Jan 2009];Fact sheet. Errata, retractions, partial retractions, corrected and republished articles, duplicate publications, comments (including author replies), updates, patient summaries, and republished (reprinted) articles policy for MEDLINE (updated 10/08/2008) 2008 [website]. http://www.nlm.nih.gov/pubs/factsheets/errata.html.
  • Neale AV, Northrup J, Dailey R, Marks E, Abrams J. Correction and use of literature affected by scientific misconduct. Science and Engineering Ethics. 2007;13:5–24. [Europe PMC free article] [Abstract] [Google Scholar]
  • Norton ML, Saltman DC. Corrections in an electronic environment. BMC Medicine. 2007;5:4. [Europe PMC free article] [Abstract] [Google Scholar]
  • Parrish DM. Scientific misconduct and correcting the scientific literature. Academic Medicine. 1999;72:221–230. [Abstract] [Google Scholar]
  • Pfeifer MP, Snodgrass GL. The continued use of retracted, invalid scientific literature. Journal of the American Medical Association. 1990;263:1420–1423. [Abstract] [Google Scholar]
  • Pfeifer MP, Snodgrass GL. Medical school libraries’ handling of articles that report invalid science. Academic Medicine. 1992;67:109–113. [Abstract] [Google Scholar]
  • Poulton A. Mistakes and misconduct in the research literature: Retractions just the tip of the iceberg. Medical Journal of Australia. 2007;186:323–324. [Abstract] [Google Scholar]
  • Roberts I, Smith R, Evans S. Doubts over head injury studies. British Medical Journal. 2007;334:392–394. [Europe PMC free article] [Abstract] [Google Scholar]
  • Snodgrass GL, Pfeifer MP. The characteristics of medical retraction notices. Bulletin of the Medical Library Association. 1992;80:328–334. [Europe PMC free article] [Abstract] [Google Scholar]
  • Sox HC, Rennie D. Research misconduct, retraction, and cleansing the medical literature: Lessons from the Poehlman case. Annals of Internal Medicine. 2006;144:609–613. [Abstract] [Google Scholar]
  • Tewari A, Buhles WC, Jr, Starnes HF., Jr Preliminary report: Effects of interleukin-1 on platelet counts. The Lancet. 1990;336:712–714. [Abstract] [Google Scholar]
  • Thomson Reuters. [Accessed 14 Jan 2009];Institute for Scientific Information Web of Science. 2009 [website]. Available at http://isiknowledge.com/WOS. [Google Scholar]
  • Tobin MJ. Reporting research, retraction of results, and responsibility. American Journal of Respiratory and Critical Care Medicine. 2000;162:773–774. [Abstract] [Google Scholar]
  • Whitsett CF. The role of hematopoietic growth factors in transfusion medicine. Transfusion Medicine II. Hematology/Oncology Clinics of North America. 1995;9:23–68. [Abstract] [Google Scholar]
  • Zauli G, Catani L. Human megakaryocyte biology and parthophysiology. Critical Reviews in Oncology/hematology. 1995;21:135–157. [Abstract] [Google Scholar]

Citations & impact 


Impact metrics

Jump to Citations

Citations of article over time

Alternative metrics

Altmetric item for https://www.altmetric.com/details/1432713
Altmetric
Discover the attention surrounding your research
https://www.altmetric.com/details/1432713

Smart citations by scite.ai
Smart citations by scite.ai include citation statements extracted from the full text of the citing article. The number of the statements may be higher than the number of citations provided by EuropePMC if one paper cites another multiple times or lower if scite has not yet processed some of the citing articles.
Explore citation contexts and check if this article has been supported or disputed.
https://scite.ai/reports/10.1007/s11948-009-9151-4

Supporting
Mentioning
Contrasting
2
61
0

Article citations


Go to all (19) article citations

Similar Articles 


To arrive at the top five similar articles we use a word-weighted algorithm to compare words from the Title and Abstract of each citation.


Funding 


Funders who supported this work.

NCI NIH HHS (1)

NINDS NIH HHS (3)