IDEAS home Printed from https://ideas.repec.org/a/sae/evarev/v41y2017i5p472-505.html
   My bibliography  Save this article

Implications of Small Samples for Generalization: Adjustments and Rules of Thumb

Author

Listed:
  • Elizabeth Tipton
  • Kelly Hallberg
  • Larry V. Hedges
  • Wendy Chan

Abstract

Background: Policy makers and researchers are frequently interested in understanding how effective a particular intervention may be for a specific population. One approach is to assess the degree of similarity between the sample in an experiment and the population. Another approach is to combine information from the experiment and the population to estimate the population average treatment effect (PATE). Method: Several methods for assessing the similarity between a sample and population currently exist as well as methods estimating the PATE. In this article, we investigate properties of six of these methods and statistics in the small sample sizes common in education research (i.e., 10–70 sites), evaluating the utility of rules of thumb developed from observational studies in the generalization case. Result: In small random samples, large differences between the sample and population can arise simply by chance and many of the statistics commonly used in generalization are a function of both sample size and the number of covariates being compared. The rules of thumb developed in observational studies (which are commonly applied in generalization) are much too conservative given the small sample sizes found in generalization. Conclusion: This article implies that sharp inferences to large populations from small experiments are difficult even with probability sampling. Features of random samples should be kept in mind when evaluating the extent to which results from experiments conducted on nonrandom samples might generalize.

Suggested Citation

  • Elizabeth Tipton & Kelly Hallberg & Larry V. Hedges & Wendy Chan, 2017. "Implications of Small Samples for Generalization: Adjustments and Rules of Thumb," Evaluation Review, , vol. 41(5), pages 472-505, October.
  • Handle: RePEc:sae:evarev:v:41:y:2017:i:5:p:472-505
    DOI: 10.1177/0193841X16655665
    as

    Download full text from publisher

    File URL: https://journals.sagepub.com/doi/10.1177/0193841X16655665
    Download Restriction: no

    File URL: https://libkey.io/10.1177/0193841X16655665?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Robert B. Olsen & Larry L. Orr & Stephen H. Bell & Elizabeth A. Stuart, 2013. "External Validity in Policy Evaluations That Choose Sites Purposively," Journal of Policy Analysis and Management, John Wiley & Sons, Ltd., vol. 32(1), pages 107-121, January.
    2. Colm O'Muircheartaigh & Larry V. Hedges, 2014. "Generalizing from unrepresentative experiments: a stratified propensity score approach," Journal of the Royal Statistical Society Series C, Royal Statistical Society, vol. 63(2), pages 195-210, February.
    3. repec:mpr:mprres:8128 is not listed on IDEAS
    4. Kosuke Imai & Gary King & Elizabeth A. Stuart, 2008. "Misunderstandings between experimentalists and observationalists about causal inference," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 171(2), pages 481-502, April.
    5. Elizabeth A. Stuart & Stephen R. Cole & Catherine P. Bradshaw & Philip J. Leaf, 2011. "The use of propensity scores to assess the generalizability of results from randomized trials," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 174(2), pages 369-386, April.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Wendy Chan, 2018. "Applications of Small Area Estimation to Generalization With Subclassification by Propensity Scores," Journal of Educational and Behavioral Statistics, , vol. 43(2), pages 182-224, April.
    2. Elizabeth A. Stuart & Anna Rhodes, 2017. "Generalizing Treatment Effect Estimates From Sample to Population: A Case Study in the Difficulties of Finding Sufficient Data," Evaluation Review, , vol. 41(4), pages 357-388, August.
    3. repec:mpr:mprres:8128 is not listed on IDEAS
    4. Elizabeth Tipton & Laura R. Peck, 2017. "A Design-Based Approach to Improve External Validity in Welfare Policy Evaluations," Evaluation Review, , vol. 41(4), pages 326-356, August.
    5. Elizabeth Tipton & Robert B. Olsen, "undated". "Enhancing the Generalizability of Impact Studies in Education," Mathematica Policy Research Reports 35d5625333dc480aba9765b3b, Mathematica Policy Research.
    6. Sharples, Linda D., 2018. "The role of statistics in the era of big data: Electronic health records for healthcare research," Statistics & Probability Letters, Elsevier, vol. 136(C), pages 105-110.
    7. Denis Fougère & Nicolas Jacquemet, 2020. "Policy Evaluation Using Causal Inference Methods," SciencePo Working papers Main hal-03455978, HAL.
    8. Elizabeth Tipton, 2021. "Beyond generalization of the ATE: Designing randomized trials to understand treatment effect heterogeneity," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 184(2), pages 504-521, April.
    9. Sarah A. Avellar & Jaime Thomas & Rebecca Kleinman & Emily Sama-Miller & Sara E. Woodruff & Rebecca Coughlin & T’Pring R. Westbrook, 2017. "External Validity: The Next Step for Systematic Reviews?," Evaluation Review, , vol. 41(4), pages 283-325, August.
    10. Elizabeth Tipton, 2013. "Stratified Sampling Using Cluster Analysis," Evaluation Review, , vol. 37(2), pages 109-139, April.
    11. Kellie Ottoboni & Jason Poulos, 2019. "Estimating population average treatment effects from experiments with noncompliance," Papers 1901.02991, arXiv.org, revised Aug 2020.
    12. Esterling, Kevin & Brady, David & Schwitzgebel, Eric, 2021. "The Necessity of Construct and External Validity for Generalized Causal Claims," OSF Preprints 2s8w5, Center for Open Science.
    13. Rajeev Dehejia & Cristian Pop-Eleches & Cyrus Samii, 2021. "From Local to Global: External Validity in a Fertility Natural Experiment," Journal of Business & Economic Statistics, Taylor & Francis Journals, vol. 39(1), pages 217-243, January.
    14. Esterling, Kevin M. & Brady, David & Schwitzgebel, Eric, 2023. "The Necessity of Construct and External Validity for Generalized Causal Claims," I4R Discussion Paper Series 18, The Institute for Replication (I4R).
    15. Das, Ashis & Friedman, Jed & Kandpal, Eeshani, 2014. "Does involvement of local NGOs enhance public service delivery ? cautionary evidence from a Malaria-prevention evaluation in India," Policy Research Working Paper Series 6931, The World Bank.
    16. David M. Phillippo & Anthony E. Ades & Sofia Dias & Stephen Palmer & Keith R. Abrams & Nicky J. Welton, 2018. "Methods for Population-Adjusted Indirect Comparisons in Health Technology Appraisal," Medical Decision Making, , vol. 38(2), pages 200-211, February.
    17. Andrews, Isaiah & Oster, Emily, 2019. "A simple approximation for evaluating external validity bias," Economics Letters, Elsevier, vol. 178(C), pages 58-62.
    18. Ottoboni Kellie N. & Poulos Jason V., 2020. "Estimating population average treatment effects from experiments with noncompliance," Journal of Causal Inference, De Gruyter, vol. 8(1), pages 108-130, January.
    19. Elizabeth Tipton, 2014. "How Generalizable Is Your Experiment? An Index for Comparing Experimental Samples and Populations," Journal of Educational and Behavioral Statistics, , vol. 39(6), pages 478-501, December.
    20. Peter Z. Schochet, "undated". "Statistical Theory for the RCT-YES Software: Design-Based Causal Inference for RCTs," Mathematica Policy Research Reports a0c005c003c242308a92c02dc, Mathematica Policy Research.
    21. Jacob Alex Klerman, 2017. "Editor in Chief’s Comment: External Validity in Systematic Reviews," Evaluation Review, , vol. 41(5), pages 391-402, October.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:sae:evarev:v:41:y:2017:i:5:p:472-505. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: SAGE Publications (email available below). General contact details of provider: .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.