IDEAS home Printed from https://ideas.repec.org/a/spr/scient/v126y2021i5d10.1007_s11192-021-03914-1.html
   My bibliography  Save this article

Generalizing research findings for enhanced reproducibility: an approach based on verbal alternative representations

Author

Listed:
  • Ron S. Kenett

    (The Hebrew University of Jerusalem
    The KPA Group
    Samuel Neaman Institute)

  • Abraham Rubinstein

    (The Hebrew University of Jerusalem)

Abstract

Research aims at generating research claims. The paper introduces a "border of meaning", abbreviated BOM, as a mode of representation of research findings that supplements statistical tests. The suggested approach was originally developed in a pedagogical context of promoting conceptual understanding in education. Here we aim at helping better understand research claims stated in a scientific paper. Considering new approaches to the presentation of findings, has an impact on the reproducibility of research. The BOM approach is demonstrated using examples from clinical research and translational medicine. Specifically, we map research findings into a list that delineates a demarcation line between alternative representations of the research claims, some with meaning equivalence and some with surface similarity. Such a mapping can be statistically evaluated by sin type error tests. Our main message is that findings should be presented and generalized with a BOM.

Suggested Citation

  • Ron S. Kenett & Abraham Rubinstein, 2021. "Generalizing research findings for enhanced reproducibility: an approach based on verbal alternative representations," Scientometrics, Springer;Akadémiai Kiadó, vol. 126(5), pages 4137-4151, May.
  • Handle: RePEc:spr:scient:v:126:y:2021:i:5:d:10.1007_s11192-021-03914-1
    DOI: 10.1007/s11192-021-03914-1
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s11192-021-03914-1
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s11192-021-03914-1?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to

    for a different version of it.

    References listed on IDEAS

    as
    1. Ron S. Kenett & Galit Shmueli, 2014. "On information quality," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 177(1), pages 3-38, January.
    2. Todd A. Kuffner & Stephen G. Walker, 2019. "Why are p-Values Controversial?," The American Statistician, Taylor & Francis Journals, vol. 73(1), pages 1-3, January.
    3. Monya Baker, 2016. "1,500 scientists lift the lid on reproducibility," Nature, Nature, vol. 533(7604), pages 452-454, May.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Pierpaolo D’Urso & Vincenzina Vitale, 2020. "Bayesian Networks Model Averaging for Bes Indicators," Social Indicators Research: An International and Interdisciplinary Journal for Quality-of-Life Measurement, Springer, vol. 151(3), pages 897-919, October.
    2. Kleber Neves & Pedro B Tan & Olavo B Amaral, 2022. "Are most published research findings false in a continuous universe?," PLOS ONE, Public Library of Science, vol. 17(12), pages 1-18, December.
    3. Dennis Bontempi & Leonard Nuernberg & Suraj Pai & Deepa Krishnaswamy & Vamsi Thiriveedhi & Ahmed Hosny & Raymond H. Mak & Keyvan Farahani & Ron Kikinis & Andrey Fedorov & Hugo J. W. L. Aerts, 2024. "End-to-end reproducible AI pipelines in radiology using the cloud," Nature Communications, Nature, vol. 15(1), pages 1-9, December.
    4. Chai, Daniel & Ali, Searat & Brosnan, Mark & Hasso, Tim, 2024. "Understanding researchers' perceptions and experiences in finance research replication studies: A pre-registered report," Pacific-Basin Finance Journal, Elsevier, vol. 86(C).
    5. Deer, Lachlan & Adler, Susanne J. & Datta, Hannes & Mizik, Natalie & Sarstedt, Marko, 2025. "Toward open science in marketing research," International Journal of Research in Marketing, Elsevier, vol. 42(1), pages 212-233.
    6. Pierpaolo D’Urso & Vincenzina Vitale, 2021. "Modeling Local BES Indicators by Copula-Based Bayesian Networks," Social Indicators Research: An International and Interdisciplinary Journal for Quality-of-Life Measurement, Springer, vol. 153(3), pages 823-847, February.
    7. Roberto Savona & Cristina Maria Alberini & Lucia Alessi & Iacopo Baussano & Petros Dellaportas & Ranieri Guerra & Sean Khozin & Andrea Modena & Sergio Pecorelli & Guido Rasi & Paolo Daniele Siviero & , 2023. "Towards a Framework for a New Research Ecosystem," Papers 2312.07065, arXiv.org, revised Dec 2023.
    8. Federica Cugnata & Silvia Salini, 2014. "Model-based approach for importance–performance analysis," Quality & Quantity: International Journal of Methodology, Springer, vol. 48(6), pages 3053-3064, November.
    9. repec:plo:pcbi00:1005412 is not listed on IDEAS
    10. Peter Harremoës, 2019. "Replication Papers," Publications, MDPI, vol. 7(3), pages 1-8, July.
    11. Hasso, Tim & Brosnan, Mark & Ali, Searat & Chai, Daniel, 2025. "Perceived problems, causes, and solutions of finance research reproducibility and replicability: A pre-registered report," Pacific-Basin Finance Journal, Elsevier, vol. 91(C).
    12. Fernando Hoces de la Guardia & Sean Grant & Edward Miguel, 2021. "A framework for open policy analysis," Science and Public Policy, Oxford University Press, vol. 48(2), pages 154-163.
    13. Antonella Lanati & Marinella Marzano & Caterina Manzari & Bruno Fosso & Graziano Pesole & Francesca De Leo, 2019. "Management at the service of research: ReOmicS, a quality management system for omics sciences," Palgrave Communications, Palgrave Macmillan, vol. 5(1), pages 1-13, December.
    14. Joel Ferguson & Rebecca Littman & Garret Christensen & Elizabeth Levy Paluck & Nicholas Swanson & Zenan Wang & Edward Miguel & David Birke & John-Henry Pezzuto, 2023. "Survey of open science practices and attitudes in the social sciences," Nature Communications, Nature, vol. 14(1), pages 1-13, December.
    15. Thomas F. Heston, 2024. "Redefining Significance: Robustness and Percent Fragility Indices in Biomedical Research," Stats, MDPI, vol. 7(2), pages 1-12, June.
    16. Erastus Karanja & Aditya Sharma & Ibrahim Salama, 2020. "What does MIS survey research reveal about diversity and representativeness in the MIS field? A content analysis approach," Scientometrics, Springer;Akadémiai Kiadó, vol. 122(3), pages 1583-1628, March.
    17. Coleman Shirley Y., 2016. "Data-Mining Opportunities for Small and Medium Enterprises with Official Statistics in the UK," Journal of Official Statistics, Sciendo, vol. 32(4), pages 849-865, December.
    18. Breznau, Nate & Rinke, Eike Mark & Wuttke, Alexander & Adem, Muna & Adriaans, Jule & Akdeniz, Esra & Alvarez-Benjumea, Amalia & Andersen, Henrik K. & Auer, Daniel & Azevedo, Flavio & Bahnsen, Oke & Ba, 2025. "The reliability of replications: a study in computational reproductions," EconStor Open Access Articles and Book Chapters, ZBW - Leibniz Information Centre for Economics, vol. 12(3), pages 1-23.
    19. Bor Luen Tang, 2023. "Some Insights into the Factors Influencing Continuous Citation of Retracted Scientific Papers," Publications, MDPI, vol. 11(4), pages 1-14, October.
    20. Domenico Piccolo & Rosaria Simone, 2019. "Rejoinder to the discussion of “The class of cub models: statistical foundations, inferential issues and empirical evidence”," Statistical Methods & Applications, Springer;Società Italiana di Statistica, vol. 28(3), pages 477-493, September.
    21. repec:plo:pbio00:2005561 is not listed on IDEAS
    22. Taavi Päll & Hannes Luidalepp & Tanel Tenson & Ülo Maiväli, 2023. "A field-wide assessment of differential expression profiling by high-throughput sequencing reveals widespread bias," PLOS Biology, Public Library of Science, vol. 21(3), pages 1-27, March.

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:scient:v:126:y:2021:i:5:d:10.1007_s11192-021-03914-1. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.