IDEAS home Printed from https://ideas.repec.org/a/plo/pmed00/1001489.html
   My bibliography  Save this article

Threats to Validity in the Design and Conduct of Preclinical Efficacy Studies: A Systematic Review of Guidelines for In Vivo Animal Experiments

Author

Listed:
  • Valerie C Henderson
  • Jonathan Kimmelman
  • Dean Fergusson
  • Jeremy M Grimshaw
  • Dan G Hackam

Abstract

Background: The vast majority of medical interventions introduced into clinical development prove unsafe or ineffective. One prominent explanation for the dismal success rate is flawed preclinical research. We conducted a systematic review of preclinical research guidelines and organized recommendations according to the type of validity threat (internal, construct, or external) or programmatic research activity they primarily address. Methods and Findings: We searched MEDLINE, Google Scholar, Google, and the EQUATOR Network website for all preclinical guideline documents published up to April 9, 2013 that addressed the design and conduct of in vivo animal experiments aimed at supporting clinical translation. To be eligible, documents had to provide guidance on the design or execution of preclinical animal experiments and represent the aggregated consensus of four or more investigators. Data from included guidelines were independently extracted by two individuals for discrete recommendations on the design and implementation of preclinical efficacy studies. These recommendations were then organized according to the type of validity threat they addressed. A total of 2,029 citations were identified through our search strategy. From these, we identified 26 guidelines that met our eligibility criteria—most of which were directed at neurological or cerebrovascular drug development. Together, these guidelines offered 55 different recommendations. Some of the most common recommendations included performance of a power calculation to determine sample size, randomized treatment allocation, and characterization of disease phenotype in the animal model prior to experimentation. Conclusions: By identifying the most recurrent recommendations among preclinical guidelines, we provide a starting point for developing preclinical guidelines in other disease domains. We also provide a basis for the study and evaluation of preclinical research practice. Background: The development process for new drugs is lengthy and complex. It begins in the laboratory, where scientists investigate the causes of diseases and identify potential new treatments. Next, promising interventions undergo preclinical research in cells and in animals (in vivo animal experiments) to test whether the intervention has the expected effect and to support the generalization (extension) of this treatment–effect relationship to patients. Drugs that pass these tests then enter clinical trials, where their safety and efficacy is tested in selected groups of patients under strictly controlled conditions. Finally, the government bodies responsible for drug approval review the results of the clinical trials, and successful drugs receive a marketing license, usually a decade or more after the initial laboratory work. Notably, only 11% of agents that enter clinical testing (investigational drugs) are ultimately licensed. Why Was This Study Done?: The frequent failure of investigational drugs during clinical translation is potentially harmful to trial participants. Moreover, the costs of these failures are passed onto healthcare systems in the form of higher drug prices. It would be good, therefore, to reduce the attrition rate of investigational drugs. One possible explanation for the dismal success rate of clinical translation is that preclinical research, the key resource for justifying clinical development, is flawed. To address this possibility, several groups of preclinical researchers have issued guidelines intended to improve the design and execution of in vivo animal studies. In this systematic review (a study that uses predefined criteria to identify all the research on a given topic), the authors identify the experimental practices that are commonly recommended in these guidelines and organize these recommendations according to the type of threat to validity (internal, construct, or external) that they address. Internal threats to validity are factors that confound reliable inferences about treatment–effect relationships in preclinical research. For example, experimenter expectation may bias outcome assessment. Construct threats to validity arise when researchers mischaracterize the relationship between an experimental system and the clinical disease it is intended to represent. For example, researchers may use an animal model for a complex multifaceted clinical disease that only includes one characteristic of the disease. External threats to validity are unseen factors that frustrate the transfer of treatment–effect relationships from animal models to patients. What Did the Researchers Do and Find?: The researchers identified 26 preclinical guidelines that met their predefined eligibility criteria. Twelve guidelines addressed preclinical research for neurological and cerebrovascular drug development; other disorders covered by guidelines included cardiac and circulatory disorders, sepsis, pain, and arthritis. Together, the guidelines offered 55 different recommendations for the design and execution of preclinical in vivo animal studies. Nineteen recommendations addressed threats to internal validity. The most commonly included recommendations of this type called for the use of power calculations to ensure that sample sizes are large enough to yield statistically meaningful results, random allocation of animals to treatment groups, and “blinding” of researchers who assess outcomes to treatment allocation. Among the 25 recommendations that addressed threats to construct validity, the most commonly included recommendations called for characterization of the properties of the animal model before experimentation and matching of the animal model to the human manifestation of the disease. Finally, six recommendations addressed threats to external validity. The most commonly included of these recommendations suggested that preclinical research should be replicated in different models of the same disease and in different species, and should also be replicated independently. What Do These Findings Mean?: This systematic review identifies a range of investigational recommendations that preclinical researchers believe address threats to the validity of preclinical efficacy studies. Many of these recommendations are not widely implemented in preclinical research at present. Whether the failure to implement them explains the frequent discordance between the results on drug safety and efficacy obtained in preclinical research and in clinical trials is currently unclear. These findings provide a starting point, however, for the improvement of existing preclinical research guidelines for specific diseases, and for the development of similar guidelines for other diseases. They also provide an evidence-based platform for the analysis of preclinical evidence and for the study and evaluation of preclinical research practice. These findings should, therefore, be considered by investigators, institutional review bodies, journals, and funding agents when designing, evaluating, and sponsoring translational research. Additional Information: Please access these websites via the online version of this summary at http://dx.doi.org/10.1371/journal.pmed.1001489.

Suggested Citation

  • Valerie C Henderson & Jonathan Kimmelman & Dean Fergusson & Jeremy M Grimshaw & Dan G Hackam, 2013. "Threats to Validity in the Design and Conduct of Preclinical Efficacy Studies: A Systematic Review of Guidelines for In Vivo Animal Experiments," PLOS Medicine, Public Library of Science, vol. 10(7), pages 1-14, July.
  • Handle: RePEc:plo:pmed00:1001489
    DOI: 10.1371/journal.pmed.1001489
    as

    Download full text from publisher

    File URL: https://journals.plos.org/plosmedicine/article?id=10.1371/journal.pmed.1001489
    Download Restriction: no

    File URL: https://journals.plos.org/plosmedicine/article/file?id=10.1371/journal.pmed.1001489&type=printable
    Download Restriction: no

    File URL: https://libkey.io/10.1371/journal.pmed.1001489?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. H Bart van der Worp & David W Howells & Emily S Sena & Michelle J Porritt & Sarah Rewell & Victoria O'Collins & Malcolm R Macleod, 2010. "Can Animal Models of Disease Reliably Inform Human Studies?," PLOS Medicine, Public Library of Science, vol. 7(3), pages 1-8, March.
    2. Story C. Landis & Susan G. Amara & Khusru Asadullah & Chris P. Austin & Robi Blumenstein & Eileen W. Bradley & Ronald G. Crystal & Robert B. Darnell & Robert J. Ferrante & Howard Fillit & Robert Finke, 2012. "A call for transparent reporting to optimize the predictive value of preclinical research," Nature, Nature, vol. 490(7419), pages 187-191, October.
    3. Emily S Sena & H Bart van der Worp & Philip M W Bath & David W Howells & Malcolm R Macleod, 2010. "Publication Bias in Reports of Animal Stroke Studies Leads to Major Overstatement of Efficacy," PLOS Biology, Public Library of Science, vol. 8(3), pages 1-8, March.
    4. Carol Kilkenny & William J Browne & Innes C Cuthill & Michael Emerson & Douglas G Altman, 2010. "Improving Bioscience Research Reporting: The ARRIVE Guidelines for Reporting Animal Research," PLOS Biology, Public Library of Science, vol. 8(6), pages 1-5, June.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Jonathan Kimmelman & Jeffrey S Mogil & Ulrich Dirnagl, 2014. "Distinguishing between Exploratory and Confirmatory Preclinical Research Will Improve Translation," PLOS Biology, Public Library of Science, vol. 12(5), pages 1-4, May.
    2. Judith van Luijk & Brenda Bakker & Maroeska M Rovers & Merel Ritskes-Hoitinga & Rob B M de Vries & Marlies Leenaars, 2014. "Systematic Reviews of Animal Studies; Missing Link in Translational Research?," PLOS ONE, Public Library of Science, vol. 9(3), pages 1-5, March.
    3. Pandora Pound & Christine J Nicol, 2018. "Retrospective harm benefit analysis of pre-clinical animal research for six treatment interventions," PLOS ONE, Public Library of Science, vol. 13(3), pages 1-26, March.
    4. Marc T Avey & David Moher & Katrina J Sullivan & Dean Fergusson & Gilly Griffin & Jeremy M Grimshaw & Brian Hutton & Manoj M Lalu & Malcolm Macleod & John Marshall & Shirley H J Mei & Michael Rudnicki, 2016. "The Devil Is in the Details: Incomplete Reporting in Preclinical Animal Research," PLOS ONE, Public Library of Science, vol. 11(11), pages 1-13, November.
    5. Lucile Vogt & Thomas S Reichlin & Christina Nathues & Hanno Würbel, 2016. "Authorization of Animal Experiments Is Based on Confidence Rather than Evidence of Scientific Rigor," PLOS Biology, Public Library of Science, vol. 14(12), pages 1-24, December.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Jonathan A Eisen & Emma Ganley & Catriona J MacCallum, 2014. "Open Science and Reporting Animal Studies: Who's Accountable?," PLOS Biology, Public Library of Science, vol. 12(1), pages 1-3, January.
    2. Jason A Miranda & Phil Stanley & Katrina Gore & Jamie Turner & Rebecca Dias & Huw Rees, 2014. "A Preclinical Physiological Assay to Test Modulation of Knee Joint Pain in the Spinal Cord: Effects of Oxycodone and Naproxen," PLOS ONE, Public Library of Science, vol. 9(8), pages 1-11, August.
    3. Constance Holman & Sophie K Piper & Ulrike Grittner & Andreas Antonios Diamantaras & Jonathan Kimmelman & Bob Siegerink & Ulrich Dirnagl, 2016. "Where Have All the Rodents Gone? The Effects of Attrition in Experimental Research on Cancer and Stroke," PLOS Biology, Public Library of Science, vol. 14(1), pages 1-12, January.
    4. Jonathan Kimmelman & Alex John London, 2011. "Predicting Harms and Benefits in Translational Trials: Ethics, Evidence, and Uncertainty," PLOS Medicine, Public Library of Science, vol. 8(3), pages 1-5, March.
    5. Malcolm R Macleod & Aaron Lawson McLean & Aikaterini Kyriakopoulou & Stylianos Serghiou & Arno de Wilde & Nicki Sherratt & Theo Hirst & Rachel Hemblade & Zsanett Bahor & Cristina Nunes-Fonseca & Aparn, 2015. "Risk of Bias in Reports of In Vivo Research: A Focus for Improvement," PLOS Biology, Public Library of Science, vol. 13(10), pages 1-12, October.
    6. Catriona J MacCallum, 2010. "Reporting Animal Studies: Good Science and a Duty of Care," PLOS Biology, Public Library of Science, vol. 8(6), pages 1-2, June.
    7. Konstantinos K Tsilidis & Orestis A Panagiotou & Emily S Sena & Eleni Aretouli & Evangelos Evangelou & David W Howells & Rustam Al-Shahi Salman & Malcolm R Macleod & John P A Ioannidis, 2013. "Evaluation of Excess Significance Bias in Animal Studies of Neurological Diseases," PLOS Biology, Public Library of Science, vol. 11(7), pages 1-10, July.
    8. Lucile Vogt & Thomas S Reichlin & Christina Nathues & Hanno Würbel, 2016. "Authorization of Animal Experiments Is Based on Confidence Rather than Evidence of Scientific Rigor," PLOS Biology, Public Library of Science, vol. 14(12), pages 1-24, December.
    9. Jonathan Kimmelman & Jeffrey S Mogil & Ulrich Dirnagl, 2014. "Distinguishing between Exploratory and Confirmatory Preclinical Research Will Improve Translation," PLOS Biology, Public Library of Science, vol. 12(5), pages 1-4, May.
    10. Kimberley E Wever & Carlijn R Hooijmans & Niels P Riksen & Thomas B Sterenborg & Emily S Sena & Merel Ritskes-Hoitinga & Michiel C Warlé, 2015. "Determinants of the Efficacy of Cardiac Ischemic Preconditioning: A Systematic Review and Meta-Analysis of Animal Studies," PLOS ONE, Public Library of Science, vol. 10(11), pages 1-17, November.
    11. Nathalie Percie du Sert & Viki Hurst & Amrita Ahluwalia & Sabina Alam & Marc T Avey & Monya Baker & William J Browne & Alejandra Clark & Innes C Cuthill & Ulrich Dirnagl & Michael Emerson & Paul Garne, 2020. "The ARRIVE guidelines 2.0: Updated guidelines for reporting animal research," PLOS Biology, Public Library of Science, vol. 18(7), pages 1-12, July.
    12. David Baker & Katie Lidster & Ana Sottomayor & Sandra Amor, 2014. "Two Years Later: Journals Are Not Yet Enforcing the ARRIVE Guidelines on Reporting Standards for Pre-Clinical Animal Studies," PLOS Biology, Public Library of Science, vol. 12(1), pages 1-6, January.
    13. Susanne Wieschowski & Diego S Silva & Daniel Strech, 2016. "Animal Study Registries: Results from a Stakeholder Analysis on Potential Strengths, Weaknesses, Facilitators, and Barriers," PLOS Biology, Public Library of Science, vol. 14(11), pages 1-12, November.
    14. Olivia Hogue & Tucker Harvey & Dena Crozier & Claire Sonneborn & Abagail Postle & Hunter Block-Beach & Eashwar Somasundaram & Francis J May & Monica Snyder Braun & Felicia L Pasadyn & Khandi King & Ca, 2022. "Statistical practice and transparent reporting in the neurosciences: Preclinical motor behavioral experiments," PLOS ONE, Public Library of Science, vol. 17(3), pages 1-17, March.
    15. Jennifer A Hirst & Jeremy Howick & Jeffrey K Aronson & Nia Roberts & Rafael Perera & Constantinos Koshiaris & Carl Heneghan, 2014. "The Need for Randomization in Animal Trials: An Overview of Systematic Reviews," PLOS ONE, Public Library of Science, vol. 9(6), pages 1-11, June.
    16. Carol Kilkenny & William J Browne & Innes C Cuthill & Michael Emerson & Douglas G Altman, 2010. "Improving Bioscience Research Reporting: The ARRIVE Guidelines for Reporting Animal Research," PLOS Biology, Public Library of Science, vol. 8(6), pages 1-5, June.
    17. Gillian L Currie & Helena N Angel-Scott & Lesley Colvin & Fala Cramond & Kaitlyn Hair & Laila Khandoker & Jing Liao & Malcolm Macleod & Sarah K McCann & Rosie Morland & Nicki Sherratt & Robert Stewart, 2019. "Animal models of chemotherapy-induced peripheral neuropathy: A machine-assisted systematic review and meta-analysis," PLOS Biology, Public Library of Science, vol. 17(5), pages 1-34, May.
    18. Ana Antonic & Emily S Sena & Jennifer S Lees & Taryn E Wills & Peta Skeers & Peter E Batchelor & Malcolm R Macleod & David W Howells, 2013. "Stem Cell Transplantation in Traumatic Spinal Cord Injury: A Systematic Review and Meta-Analysis of Animal Studies," PLOS Biology, Public Library of Science, vol. 11(12), pages 1-14, December.
    19. Julie E. Goodman & Catherine Petito Boyce & Sonja N. Sax & Leslie A. Beyer & Robyn L. Prueitt, 2015. "Rethinking Meta‐Analysis: Applications for Air Pollution Data and Beyond," Risk Analysis, John Wiley & Sons, vol. 35(6), pages 1017-1039, June.
    20. Pandora Pound & Christine J Nicol, 2018. "Retrospective harm benefit analysis of pre-clinical animal research for six treatment interventions," PLOS ONE, Public Library of Science, vol. 13(3), pages 1-26, March.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:plo:pmed00:1001489. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: plosmedicine (email available below). General contact details of provider: https://journals.plos.org/plosmedicine/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.