IDEAS home Printed from https://ideas.repec.org/a/bla/jorssa/v184y2021i3p904-919.html
   My bibliography  Save this article

When zero may not be zero: A cautionary note on the use of inter‐rater reliability in evaluating grant peer review

Author

Listed:
  • Elena A. Erosheva
  • Patrícia Martinková
  • Carole J. Lee

Abstract

Considerable attention has focused on studying reviewer agreement via inter‐rater reliability (IRR) as a way to assess the quality of the peer review process. Inspired by a recent study that reported an IRR of zero in the mock peer review of top‐quality grant proposals, we use real data from a complete range of submissions to the National Institutes of Health and to the American Institute of Biological Sciences to bring awareness to two important issues with using IRR for assessing peer review quality. First, we demonstrate that estimating local IRR from subsets of restricted‐quality proposals will likely result in zero estimates under many scenarios. In both data sets, we find that zero local IRR estimates are more likely when subsets of top‐quality proposals rather than bottom‐quality proposals are considered. However, zero estimates from range‐restricted data should not be interpreted as indicating arbitrariness in peer review. On the contrary, despite different scoring scales used by the two agencies, when complete ranges of proposals are considered, IRR estimates are above 0.6 which indicates good reviewer agreement. Furthermore, we demonstrate that, with a small number of reviewers per proposal, zero estimates of IRR are possible even when the true value is not zero.

Suggested Citation

  • Elena A. Erosheva & Patrícia Martinková & Carole J. Lee, 2021. "When zero may not be zero: A cautionary note on the use of inter‐rater reliability in evaluating grant peer review," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 184(3), pages 904-919, July.
  • Handle: RePEc:bla:jorssa:v:184:y:2021:i:3:p:904-919
    DOI: 10.1111/rssa.12681
    as

    Download full text from publisher

    File URL: https://doi.org/10.1111/rssa.12681
    Download Restriction: no

    File URL: https://libkey.io/10.1111/rssa.12681?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Kevin Gross & Carl T Bergstrom, 2019. "Contest models highlight inherent inefficiencies of scientific funding competitions," PLOS Biology, Public Library of Science, vol. 17(1), pages 1-15, January.
    2. Rüdiger Mutz & Lutz Bornmann & Hans-Dieter Daniel, 2012. "Heterogeneity of Inter-Rater Reliabilities of Grant Peer Reviews and Its Determinants: A General Estimating Equations Approach," PLOS ONE, Public Library of Science, vol. 7(10), pages 1-10, October.
    3. Upali W. Jayasinghe & Herbert W. Marsh & Nigel Bond, 2003. "A multilevel cross‐classified modelling approach to peer review of grant proposals: the effects of assessor and researcher attributes on assessor ratings," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 166(3), pages 279-300, October.
    4. Michael R Martin & Andrea Kopstein & Joy M Janice, 2010. "An Analysis of Preliminary and Post-Discussion Priority Scores for Grant Applications Peer Reviewed by the Center for Scientific Review at the NIH," PLOS ONE, Public Library of Science, vol. 5(11), pages 1-6, November.
    5. repec:nas:journl:v:115:y:2018:p:2952-2957 is not listed on IDEAS
    6. Elise S. Brezis & Aliaksandr Birukou, 2020. "Arbitrariness in the peer review process," Scientometrics, Springer;Akadémiai Kiadó, vol. 123(1), pages 393-411, April.
    7. Adler, Seymour & Campion, Michael & Colquitt, Alan & Grubb, Amy & Murphy, Kevin & Ollander-Krane, Rob & Pulakos, Elaine D., 2016. "Getting Rid of Performance Ratings: Genius or Folly? A Debate," Industrial and Organizational Psychology, Cambridge University Press, vol. 9(2), pages 219-252, June.
    8. Adcock, Robert & Collier, David, 2001. "Measurement Validity: A Shared Standard for Qualitative and Quantitative Research," American Political Science Review, Cambridge University Press, vol. 95(3), pages 529-546, September.
    9. Mark D Lindner & Richard K Nakamura, 2015. "Examining the Predictive Validity of NIH Peer Review Scores," PLOS ONE, Public Library of Science, vol. 10(6), pages 1-12, June.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Gerald Schweiger & Adrian Barnett & Peter van den Besselaar & Lutz Bornmann & Andreas De Block & John P. A. Ioannidis & Ulf Sandstrom & Stijn Conix, 2024. "The Costs of Competition in Distributing Scarce Research Funds," Papers 2403.16934, arXiv.org.
    2. Klocek, Adam & Kollerová, Lenka & Havrdová, Egle & Kotrbová, Monika & Netík, Jan & Pour, Marek, 2024. "Effectiveness of the KiVa anti-bullying program in the Czech Republic: A cluster randomized control trial," Evaluation and Program Planning, Elsevier, vol. 106(C).
    3. Lubomír Štěpánek & Jana Dlouhá & Patrícia Martinková, 2023. "Item Difficulty Prediction Using Item Text Features: Comparison of Predictive Performance across Machine-Learning Algorithms," Mathematics, MDPI, vol. 11(19), pages 1-30, September.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Müge Simsek & Mathijs Vaan & Arnout Rijt, 2024. "Do grant proposal texts matter for funding decisions? A field experiment," Scientometrics, Springer;Akadémiai Kiadó, vol. 129(5), pages 2521-2532, May.
    2. David G Pina & Darko Hren & Ana Marušić, 2015. "Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research," PLOS ONE, Public Library of Science, vol. 10(6), pages 1-15, June.
    3. Stephen A Gallo & Afton S Carpenter & Scott R Glisson, 2013. "Teleconference versus Face-to-Face Scientific Peer Review of Grant Application: Effects on Review Outcomes," PLOS ONE, Public Library of Science, vol. 8(8), pages 1-9, August.
    4. Patrícia Martinková & Dan Goldhaber & Elena Erosheva, 2018. "Disparities in ratings of internal and external applicants: A case for model-based inter-rater reliability," PLOS ONE, Public Library of Science, vol. 13(10), pages 1-17, October.
    5. Axel Philipps, 2022. "Research funding randomly allocated? A survey of scientists’ views on peer review and lottery," Science and Public Policy, Oxford University Press, vol. 49(3), pages 365-377.
    6. Bayindir, Esra Eren & Gurdal, Mehmet Yigit & Saglam, Ismail, 2019. "A Game Theoretic Approach to Peer Review of Grant Proposals," Journal of Informetrics, Elsevier, vol. 13(4).
    7. Jens Jirschitzka & Aileen Oeberst & Richard Göllner & Ulrike Cress, 2017. "Inter-rater reliability and validity of peer reviews in an interdisciplinary field," Scientometrics, Springer;Akadémiai Kiadó, vol. 113(2), pages 1059-1092, November.
    8. Kyriaki Papadopoulou, 2020. "Comparative Review Of Performance Measurement Methods Effectiveness," Economics and Management, Faculty of Economics, SOUTH-WEST UNIVERSITY "NEOFIT RILSKI", BLAGOEVGRAD, vol. 17(1), pages 127-139.
    9. Schakel, Arjan Hille, 2009. "A Postfunctionalist Theory of Regional Government," MPRA Paper 21596, University Library of Munich, Germany.
    10. Elise S. Brezis & Aliaksandr Birukou, 2020. "Arbitrariness in the peer review process," Scientometrics, Springer;Akadémiai Kiadó, vol. 123(1), pages 393-411, April.
    11. Qianshuo Liu & David Pérez-Castrillo & Inés Macho-Stadler & Albert Banal-Estañol, 2021. "Similar-to-me Effects in the Grant Application Process: Applicants, Panelists, and the Likelihood of Obtaining Funds," Working Papers 1289, Barcelona School of Economics.
    12. Zim Nwokora & Riccardo Pelizzo, 2017. "Measuring Party System Change: A Systems Perspective," Research Africa Network Working Papers 17/048, Research Africa Network (RAN).
    13. Gustav Lidén, 2013. "What about theory? The consequences on a widened perspective of social theory," Quality & Quantity: International Journal of Methodology, Springer, vol. 47(1), pages 213-225, January.
    14. Patrícia Martinková & František Bartoš & Marek Brabec, 2023. "Assessing Inter-rater Reliability With Heterogeneous Variance Components Models: Flexible Approach Accounting for Contextual Variables," Journal of Educational and Behavioral Statistics, , vol. 48(3), pages 349-383, June.
    15. Bornmann, Lutz & Mutz, Rüdiger & Daniel, Hans-Dieter, 2008. "Latent Markov modeling applied to grant peer review," Journal of Informetrics, Elsevier, vol. 2(3), pages 217-228.
    16. Stephen A Gallo & Karen B Schmaling, 2022. "Peer review: Risk and risk tolerance," PLOS ONE, Public Library of Science, vol. 17(8), pages 1-20, August.
    17. Anju Devianee Keetharuth & Elizabeth Taylor Buck & Catherine Acquadro & Katrin Conway & Janice Connell & Michael Barkham & Jill Carlton & Thomas Ricketts & Rosemary Barber & John Brazier, 2018. "Integrating Qualitative and Quantitative Data in the Development of Outcome Measures: The Case of the Recovering Quality of Life (ReQoL) Measures in Mental Health Populations," IJERPH, MDPI, vol. 15(7), pages 1-14, June.
    18. Malte Luebker, 2019. "Can the Structure of Inequality Explain Fiscal Redistribution? Revisiting the Social Affinity Hypothesis," LIS Working papers 762, LIS Cross-National Data Center in Luxembourg.
    19. Christoph Carnehl & Marco Ottaviani & Justus Preusser, 2024. "Designing Scientific Grants," NBER Chapters, in: Entrepreneurship and Innovation Policy and the Economy, volume 4, pages 139-178, National Bureau of Economic Research, Inc.
    20. Gaëlle Vallée-Tourangeau & Ana Wheelock & Tushna Vandrevala & Priscilla Harries, 2022. "Peer reviewers’ dilemmas: a qualitative exploration of decisional conflict in the evaluation of grant applications in the medical humanities and social sciences," Palgrave Communications, Palgrave Macmillan, vol. 9(1), pages 1-11, December.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:bla:jorssa:v:184:y:2021:i:3:p:904-919. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Wiley Content Delivery (email available below). General contact details of provider: https://edirc.repec.org/data/rssssea.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.