Development and Validation of a Scoring Rubric for Editorial Evaluation of Peer-review Quality: A Pilot Study.

Autor: Love JN; Georgetown University School of Medicine, Department of Emergency Medicine, Washington, District of Columbia., Messman AM; Wayne State University, Department of Emergency Medicine, Detroit, Michigan., Ilgen JS; University of Washington, Department of Emergency Medicine, Seattle, Washington., Merritt C; Alpert Medical School of Brown University, Department of Emergency Medicine, Providence, Rhode Island.; Alpert Medical School of Brown University, Department of Pediatrics & Emergency Medicine, Providence, Rhode Island., Coates WC; David Geffen School of Medicine at UCLA, Department of Emergency Medicine, Los Angeles, California., Ander DS; Emory University, Department of Emergency Medicine, Atlanta, Georgia., Way DP; Ohio State University College of Medicine, Department of Emergency Medicine, Columbus, Ohio.
Jazyk: angličtina
Zdroj: The western journal of emergency medicine [West J Emerg Med] 2024 Mar; Vol. 25 (2), pp. 254-263.
DOI: 10.5811/westjem.18432
Abstrakt: Introduction: Despite the importance of peer review to publications, there is no generally accepted approach for editorial evaluation of a peer review's value to a journal editor's decision-making. The graduate medical education editors of the Western Journal of Emergency Medicine Special Issue in Educational Research & Practice (Special Issue) developed and studied the holistic editor's scoring rubric (HESR) with the objective of assessing the quality of a review and an emphasis on the degree to which it informs a holistic appreciation for the submission under consideration.
Methods: Using peer-review guidelines from several journals, the Special Issue's editors formulated the rubric as descriptions of peer reviews of varying degree of quality from the ideal to the unacceptable. Once a review was assessed by each editor using the rubric, the score was submitted to a third party for blinding purposes. We compared the performance of the new rubric to a previously used semantic differential scale instrument. Kane's validity framework guided the evaluation of the new scoring rubric around three basic assumptions: improved distribution of scores; relative consistency rather than absolute inter-rater reliability across editors; and statistical evidence that editors valued peer reviews that contributed most to their decision-making.
Results: Ninety peer reviews were the subject of this study, all were assessed by two editors. Compared to the highly skewed distribution of the prior rating scale, the distribution of the new scoring rubric was bell shaped and demonstrated full use of the rubric scale. Absolute agreement between editors was low to moderate, while relative consistency between editor's rubric ratings was high. Finally, we showed that recommendations of higher rated peer reviews were more likely to concur with the editor's formal decision.
Conclusion: Early evidence regarding the HESR supports the use of this instrument in determining the quality of peer reviews as well as its relative importance in informing editorial decision-making.
Competing Interests: Conflicts of Interest: By the WestJEM article submission agreement, all authors are required to disclose all affiliations, funding sources and financial or management relationships that could be perceived as potential sources of bias. No author has professional or financial relationships with any companies that are relevant to this study. There are no conflicts of interest or sources of funding to declare.
Databáze: MEDLINE