A Pilot Survey of Authors' Experiences With Poor Peer Review Practices
A Pilot Survey of Authors' Experiences With Poor Peer Review Practices
A Pilot Survey of Authors' Experiences With Poor Peer Review Practices
Title: A Pilot Survey of Authors’ Experiences with Poor Peer Review Practices
Corresponding author:
Kyle B. McCloskey, MS
Drexel University College of Medicine
2900 West Queen Lane
Philadelphia, Pennsylvania, 19129
Email: km3844@drexel.edu
Funding: This research received no specific grant from any funding agency in the public,
commercial or not-for-profit sectors.
Conflicts of Interest: In the last 3 years, JFM has received financial compensation for service on
several Data and Safety Monitoring Boards for the NIH and the American College of Radiology
Imaging Network, for service on a pharmacogenomics ethics advisory board for Merck, and as
an expert witness on behalf of plaintiffs on the issue of the definition of human subjects research.
JFM received partial salary support as moderator of the IRBForum.
(https://community.primr.org/home) by a grant from Public Responsibility in Medicine &
Research (PRIMR) from 2012 through 2020. KM has no conflicts of interest to disclose.
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 Objectives
2 To develop a typology of poor peer review practices (PPRP) and assess researchers’ experiences
3 with PPRP.
4 Design
6 Participants
7 We solicited 500 researchers funded by the NIH extramural grants in 2018 by direct email and
11 The total number of reported PPRP and a five-point scale to assess participants’ views about the
13 Results
14 The mean number of PPRP experienced per author was 12.5 of 28 (44.6%; range 0–27; 95% CI
15 = 11.2–13.8), with fourteen PPRP experienced by 50% or more of the sample. The number of
16 reported PPRP increased with age (P = 0.01) and total number of published peer-reviewed
18 compared to represented groups (P = 0.05). Most authors viewed the peer review process
19 favorably, with 67% (74/111) of authors responding “sometimes” or “often” to having received
20 insightful peer reviews that improved the quality of their final papers. However, a total of 57%
23 Conclusions
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 This study introduces a practical list of PPRP and a framework for a typology of PPRP, which
2 could serve as an educational tool for editors and reviewers and further our understanding of
3 poor peer review practices. Future researchers will expand authors’ experiences with
7 • The sample consisted primarily of experienced researchers from diverse fields, which
9 • This survey included a core set of 28 poor peer review practices and allowed respondents
10 to add other practices they had experienced, which helped to generate an extensive list of
12 • The generalizability of the prevalence of poor review types and the degree of negative
13 impact on authors should be interpreted with caution due to the low response rate and the
15
16 1. INTRODUCTION
17 Peer review is a fundamental tool of the editorial process that helps maintain scientific
18 integrity. Since its introduction in 1731 by the Royal Society of Edinburg, it has become the
19 “gold standard” for evaluating scholarly work by calling on independent reviewers to assess a
20 study's validity, quality, and originality.1 This process is intended to prevent the publication of
21 inaccurate or poor-quality research while providing expert feedback to authors to aid them in
1 Over the past few decades, however, a growing body of literature has begun to confirm
2 the various limitations of the peer review process, from biased or inexperienced reviewers to low
3 inter-reliability between reviews.4-9 The little guidance offered to peer reviewers on what
4 constitutes a “good” or “poor” peer review may also partly explain these shortcomings. While
5 more educational resources have become available for reviewers, it is still axiomatic that
6 reviewers occasionally provide ambiguous critiques or cross the line, making comments that
7 strike authors as bothersome or worse.9-13 For authors, these limitations can make peer review a
8 frustrating and mysterious process that deters them from disseminating their work. It is perhaps
9 axiomatic that authors set aside reviews for a few days after receipt to calm the agita and avoid
12 assess the types of poor peer review practices (PPRP) encountered by researchers, to develop a
13 typology of poor peer review practices, and to begin to assess the impact of poor peer review
14 practices on authors’ perceived abilities to disseminate their research. A pilot study was
15 performed to test the survey instrument before a larger-scale study. This knowledge could
16 potentially serve as an educational tool to elucidate issues hindering the peer review process
17 from its goal of promoting the quality and integrity of the sciences.
18
19 2. METHODS
21 Studies (CROSS).14 The institutional review board at the University of Pennsylvania deemed this
22 study exempt.
23
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
2 An invitation to complete the survey was emailed to a random sample of 500 researchers
3 funded by the National Institute of Health’s (NIH) extramural grants in 2018, with replacement
10 Washington, USA). The first section assessed the prevalence of 28 unique PPRP examples
11 through yes-no questions (e.g., “Have you experienced the following examples of poor peer
12 review practices?” – e.g., “Ad hominem attack,” “Unbalanced negative review,” “Unstructured
13 review”). The 28 unique PPRP examples were compiled based on the experience of one of the
14 authors (JFM), a literature review, and a scan of the Facebook user group
15 Reviewer2mustbestopped. Participants were also asked open-ended questions about whether they
16 experienced poor peer reviews not listed in the survey to capture any additional types of PPRP
18 The second section utilized Likert scale questions to assess the effect of PPRP on
19 participants' ability to conduct and disseminate their research. Participants were asked to rate six
20 statements on a five-point scale (e.g., “How would you rate the following statement: ‘I have
21 considered leaving academia after receiving unfair review.’? – “never,” “rarely,” “sometimes,”
22 “often,” “all the time”). The last section of the survey asked about demographic information
23 (age, gender, race, language, field of study, total number of peer-reviewed publications, category
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
2 underrepresented demographics, participants were asked the yes-no question: “Do you consider
4 The presentation of questions was randomized for each section except for demographic
5 questions. No incentives or prizes were offered for completing the survey, and consent was
6 required from each subject before entering the survey. Each respondent was restricted to only
7 one submission. The final instrument was pre-tested using two expert reviewers, which helped
8 identify missing topics in the survey and determine content and response process validity. Three
9 mailings were performed at one-week intervals throughout April 2022, and data was collected
10 for analysis at the end of April 2022. The survey instrument is provided in the supplementary
11 material.
12
14 Respondents reported PPRP experiences were summarized and compared across various
16 Whitney) test and a Cuzick extension of Wilcoxon rank-sum test for ordered groups (Stata 12.1,
17 StataCorp, © 2014). Two-sided P < .05 was considered significant, and consistent with the
18 exploratory nature of this analysis, P values were unadjusted for multiple comparisons. In
19 addition to statical analyses, the authors categorized the examples of PPRP reported by
20 participants according to similar themes to help process and understand the diversity of PPRP
21 examples.
22
23 3. RESULTS
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
2 A total of 112 researchers completed the surveys, roughly 10% of those solicited (Table).
3 The respondents were predominantly male (54%), held a Ph.D. (78%), > 50 years old (66%),
4 identified as white (87%), published more than 50 peer-reviewed papers in their career (64%),
5 were trained in the humanities or social sciences (54%) and conducted primarily empirical
6 research (61%).
8 Table. Univariate exploratory analysis of total poor peer review practices (PPRP)
2 3.2 Interactions between the number of poor peer reviews practices and other variables
3 The mean number of PPRPs experienced per author was 12.5 of 28 (44.6%; range 0–27;
4 95% CI = 11.2–13.8), with fourteen PPRPs experienced by 50% or more of the sample (Figure
5 1). As shown in the exploratory results presented in Table, reported PPRP increased with age (P
7 underrepresented groups (P = 0.05), and those who received fewer “helpful/positive” reviews (P
8 = 0.01). No statistical differences were found in the number of reported PPRPs between race and
9 gender.
10
11 Figure 1. Prevalence of poor peer review practices assessed through yes-no questions.
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
2 Most authors viewed the peer review process favorably, with 67% (74/111) of authors
3 responding “sometimes” or “often” to having received insightful peer reviews that improved the
4 quality of their final papers. However, a total of 57% (63/111) of respondents admitted to
5 previously abandoning a manuscript after receiving unfair peer reviews. Moreover, 72%
6 (81/112) of respondents admitted feeling discouraged after receiving unfair peer reviews, though
7 33% (37/112) said they are rarely discouraged. As shown in Table, the total number of PPRP
8 reported by respondents increased with the number of reported adverse consequences (p=0.001).
10
11 Figure 2. Authors’ experiences with poor peer review practices assessed through 5-point Likert-
12 style questions.
13
14 3.4 A typology of poor peer review practices
15 A typology of PPRP was developed to understand the extensive range of PPRP
16 experienced by respondents. Five notable themes in reported PPRP examples were summarized:
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 1) Ill-natured Reviews – Peer reviews that have an unprofessional disposition and appear
4 2) Erroneous Reviews – Peer reviews that contain factual errors, such as false assertions
8 that a different study should have been done, or requests to cite irrelevant (and
10 4) Inconsequential or Incoherent Reviews – Peer reviews that are overly particular and do
11 not impact the overall argument of the work or are incoherent due to being unstructured,
13 5) Editorial or Process Issues – Issues internal to the editorial process and negatively
14 impact authors’ ability to disseminate their work or result, such as an insufficient number
15 of peer reviewers, breaks of anonymity, or multiple rounds of peer review where new
17
18 4. DISCUSSION
19 To our knowledge, this is the first study to systematically analyze the types of poor peer
20 review practices and develop a typology of PPRP. These findings suggest that researchers may
22 researchers from diverse fields, which likely aided in capturing a wide variety of poor peer
23 review examples. Indeed, respondents who published more peer-reviewed manuscripts also
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 reported experiencing more PPRP. The list and typology of poor peer review practices
2 introduced in this study could serve as an educational tool for editors and reviewers and further
4 The most commonly reported practice was “Multiple rounds of peer review where new
5 criticisms are raised in subsequent rounds” and is classified as an Editorial or Process Issue.
6 While some editors may see this as a normal function of the peer review process,1,2 authors may
7 view it as a cumbersome issue that delays the dissemination of their work. New criticisms raised
8 on subsequent rounds of revisions could potentially reflect the low reliability between reviewers
9 and the inherent arbitrariness of peer review; however, there are many reasons why new
10 criticisms are raised throughout the review process, with some justified and others unnecessary.
11 Editors should thus consider such issues on a case-by-case basis. While Ill-natured, Erroneous,
12 and Incoherent Reviews are unequivocally poor peer review practices that do not warrant further
13 discussion, there is a need for further research on what constitutes a PPRP for Editorial or
15 This study also investigated the impact of the peer review process on authors' ability to
16 disseminate their research. Unsurprisingly, most respondents benefited from the peer review
17 process, with 58% of authors responding “often” to receiving insightful peer reviews that
18 improved the quality of their final papers. However, this finding was tempered by 57% of
19 authors who admitted to abandoning a manuscript after receiving unfair peer reviews. The
20 findings also suggest that authors who self-identify as underrepresented minorities in their
21 respective fields are more likely to report experiencing PPRP than represented groups. Previous
22 research found similar findings, with unprofessional peer reviews disproportionately harming
23 underrepresented groups in STEM.8 These findings confirm that the peer review process is
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 indeed a powerful tool of the editorial process that can improve a researcher’s work; however, it
3 The generalizability of the prevalence of poor review types and the degree of negative
4 impact on authors should be interpreted with caution due to the low response rate and the
5 potential for response bias, given the nature of survey studies. However, the low response rate is
6 less germane to the primary aim of generating a list and typology of poor peer review practices.
7 Another limitation is that this study focused on poor peer review practices and excluded any in-
8 depth exploration of constructive or helpful peer review practices. Further research is needed to
9 establish greater external validity of the prevalence of the types of poor peer review practices
10 generated in this study and their effect on researchers’ ability to disseminate their research.
11 Furthermore, while this pilot study confirmed the survey instrument's feasibility and efficacy, we
12 plan to expand the survey to explore constructive or helpful peer review practices and examine
14
15 5. CONCLUSION
16 This study introduces a practical list of poor peer review practices and a framework for a
17 typology of poor peer review practices, which could serve as an educational tool for editors and
18 reviewers and further our understanding of poor peer review practices. These findings also
19 suggest that researchers may encounter an extensive range of poor peer reviews of various types.
20 Further research is warranted on Editorial and Process Issues that are more ambiguous and,
21 consequently, difficult to assess whether their function promotes or hinders the peer review
22 process. Future researchers should also consider investigating authors’ experiences with
2 REFERENCES
4 doi: 10.1016/s0167-7799(02)01985-6
5 2. Chung KJ. Peer review and roles of the reviewer. Arch Craniofac Surg 2019;20(6):345-
7 3. Ali PA, Watson R. Peer review and the publication process. Nurs Open 2016;3(4):193-
9 4. Tennant JP, Ross-Hellauer T. The limitations to our understanding of peer review. Res
11 5. Smith R. Peer review: a flawed process at the heart of science and journals. J R Soc Med
20 9. Mavrogenis AF, Quaile A, Scarlat MM. The good, the bad and the rude peer-review. Int
22 10. Zimba O, Gasparyan AY. Peer review guidance: a primer for researchers. Reumatologia
1 11. Garfield JM, Kaye AD, Kolinsky DC, Urman RD. A systematic guide for peer reviewers
3 12. Larson BP, Chung KC. A systematic review of peer review for scientific manuscripts.
5 13. Song E, Ang L, Park JY, Jun EY, Kim KH, et al. (2021) A scoping review on biomedical
6 journal peer review guides for reviewers. PLOS ONE 16(5): e0251440.
7 https://doi.org/10.1371/journal.pone.0251440
8 14. Sharma A, Minh Duc NT, Luu Lam Thang T, et al. A Consensus-Based Checklist for
10 10.1007/s11606-021-06737-1
11 15. Freedom of Information Act Office: Contact Information NIH-Supported PIs. NIH.
12 https://www.nih.gov/institutes-nih/nih-office-director/office-communications-public-
13 liaison/freedom-information-act-office/contact-information-nih-supported-pis (accessed
15
16 AUTHOR CONTRIBUTIONS
17 Kyle McCloskey: Writing - original draft, Writing - review & editing, Conceptualization,
18 Visualization. Jon F. Merz: Writing - original draft, Writing - review & editing,
20
21 ACKNOWLEDGEMENTS
22 The authors thank scientist and bioethicist respondents for completing the survey, Jim Coyne,
23 and anonymous reviewers for comments. An abstract presenting these findings were posted at
bioRxiv preprint doi: https://doi.org/10.1101/2022.12.20.521261; this version posted December 28, 2022. The copyright holder for this preprint
(which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made
available under aCC-BY-ND 4.0 International license.
1 the Ninth International Congress on Peer Review and Scientific Publication, Chicago, IL, USA,
2 September 8-10, 2022. Responsibility for the work is solely that of the authors.