IDEAS home Printed from https://ideas.repec.org/a/spr/reihed/v58y2017i4d10.1007_s11162-016-9429-8.html
   My bibliography  Save this article

Measuring teaching quality in higher education: assessing selection bias in course evaluations

Author

Listed:
  • Maarten Goos

    (Utrecht University School of Economics)

  • Anna Salomons

    (Utrecht University School of Economics)

Abstract

Student evaluations of teaching (SETs) are widely used to measure teaching quality in higher education and compare it across different courses, teachers, departments and institutions. Indeed, SETs are of increasing importance for teacher promotion decisions, student course selection, as well as for auditing practices demonstrating institutional performance. However, survey response is typically low, rendering these uses unwarranted if students who respond to the evaluation are not randomly selected along observed and unobserved dimensions. This paper is the first to fully quantify this problem by analyzing the direction and size of selection bias resulting from both observed and unobserved characteristics for over 3000 courses taught in a large European university. We find that course evaluations are upward biased, and that correcting for selection bias has non-negligible effects on the average evaluation score and on the evaluation-based ranking of courses. Moreover, this bias mostly derives from selection on unobserved characteristics, implying that correcting evaluation scores for observed factors such as student grades does not solve the problem. However, we find that adjusting for selection only has small impacts on the measured effects of observables on SETs, validating a large related literature which considers the observable determinants of evaluation scores without correcting for selection bias.

Suggested Citation

  • Maarten Goos & Anna Salomons, 2017. "Measuring teaching quality in higher education: assessing selection bias in course evaluations," Research in Higher Education, Springer;Association for Institutional Research, vol. 58(4), pages 341-364, June.
  • Handle: RePEc:spr:reihed:v:58:y:2017:i:4:d:10.1007_s11162-016-9429-8
    DOI: 10.1007/s11162-016-9429-8
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s11162-016-9429-8
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s11162-016-9429-8?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Steven G. Rivkin & Eric A. Hanushek & John F. Kain, 2005. "Teachers, Schools, and Academic Achievement," Econometrica, Econometric Society, vol. 73(2), pages 417-458, March.
    2. Braga, Michela & Paccagnella, Marco & Pellizzari, Michele, 2014. "Evaluating students’ evaluations of professors," Economics of Education Review, Elsevier, vol. 41(C), pages 71-88.
    3. Angrist, Joshua D. & Guryan, Jonathan, 2008. "Does teacher testing raise teacher quality? Evidence from state certification requirements," Economics of Education Review, Elsevier, vol. 27(5), pages 483-503, October.
    4. Giuseppe De Luca & Valeria Perotti, 2011. "Estimation of ordered response models with sample selection," Stata Journal, StataCorp LP, vol. 11(2), pages 213-239, June.
    5. Michael A. McPherson, 2006. "Determinants of How Students Evaluate Teachers," The Journal of Economic Education, Taylor & Francis Journals, vol. 37(1), pages 3-20, January.
    6. Eric A. Hanushek & John F. Kain & Steven G. Rivkin & Daniel M. O'Brien, 2005. "The Market for Teacher Quality," Discussion Papers 04-025, Stanford Institute for Economic Policy Research.
    7. Raj Chetty & John N. Friedman & Jonah E. Rockoff, 2014. "Measuring the Impacts of Teachers I: Evaluating Bias in Teacher Value-Added Estimates," American Economic Review, American Economic Association, vol. 104(9), pages 2593-2632, September.
    8. Krautmann, Anthony C. & Sander, William, 1999. "Grades and student evaluations of teachers," Economics of Education Review, Elsevier, vol. 18(1), pages 59-63, February.
    9. Kane, Thomas J. & Rockoff, Jonah E. & Staiger, Douglas O., 2008. "What does certification tell us about teacher effectiveness? Evidence from New York City," Economics of Education Review, Elsevier, vol. 27(6), pages 615-631, December.
    10. Erik Hanushek & F. Welch (ed.), 2006. "Handbook of the Economics of Education," Handbook of the Economics of Education, Elsevier, edition 1, volume 1, number 1, June.
    11. Gronau, Reuben, 1974. "Wage Comparisons-A Selectivity Bias," Journal of Political Economy, University of Chicago Press, vol. 82(6), pages 1119-1143, Nov.-Dec..
    12. Rosemary J. Avery & W. Keith Bryant & Alan Mathios & Hyojin Kang & Duncan Bell, 2006. "Electronic Course Evaluations: Does an Online Delivery System Influence Student Evaluations?," The Journal of Economic Education, Taylor & Francis Journals, vol. 37(1), pages 21-37, January.
    13. Imberman, Scott A. & Lovenheim, Michael F., 2016. "Does the market value value-added? Evidence from housing prices after a public release of school and teacher value-added," Journal of Urban Economics, Elsevier, vol. 91(C), pages 104-121.
    14. Alfonso Miranda & Sophia Rabe-Hesketh, 2006. "Maximum likelihood estimation of endogenous switching and sample selection models for binary, ordinal, and count variables," Stata Journal, StataCorp LP, vol. 6(3), pages 285-308, September.
    15. Charles F. Manski, 1989. "Anatomy of the Selection Problem," Journal of Human Resources, University of Wisconsin Press, vol. 24(3), pages 343-360.
    16. Raj Chetty & John N. Friedman & Jonah E. Rockoff, 2014. "Measuring the Impacts of Teachers II: Teacher Value-Added and Student Outcomes in Adulthood," American Economic Review, American Economic Association, vol. 104(9), pages 2633-2679, September.
    17. William E. Becker & William Bosshardt & Michael Watts, 2012. "How Departments of Economics Evaluate Teaching," The Journal of Economic Education, Taylor & Francis Journals, vol. 43(3), pages 325-333, July.
    18. Samer Kherfi, 2011. "Whose Opinion Is It Anyway? Determinants of Participation in Student Evaluation of Teaching," The Journal of Economic Education, Taylor & Francis Journals, vol. 42(1), pages 19-30, January.
    19. van der Klaauw, Bas & Koning, Ruud H, 2003. "Testing the Normality Assumption in the Sample Selection Model with an Application to Travel Demand," Journal of Business & Economic Statistics, American Statistical Association, vol. 21(1), pages 31-42, January.
    20. Patrick Puhani, 2000. "The Heckman Correction for Sample Selection and Its Critique," Journal of Economic Surveys, Wiley Blackwell, vol. 14(1), pages 53-68, February.
    21. Beleche, Trinidad & Fairris, David & Marks, Mindy, 2012. "Do course evaluations truly reflect student learning? Evidence from an objectively graded post-test," Economics of Education Review, Elsevier, vol. 31(5), pages 709-719.
    22. Brian A. Jacob & Lars Lefgren, 2008. "Can Principals Identify Effective Teachers? Evidence on Subjective Performance Evaluation in Education," Journal of Labor Economics, University of Chicago Press, vol. 26(1), pages 101-136.
    23. Cunha, Jesse M. & Miller, Trey, 2014. "Measuring value-added in higher education: Possibilities and limitations in the use of administrative data," Economics of Education Review, Elsevier, vol. 42(C), pages 64-77.
    24. Erik Hanushek & F. Welch (ed.), 2006. "Handbook of the Economics of Education," Handbook of the Economics of Education, Elsevier, edition 1, volume 2, number 2, June.
    25. Heckman, James J, 1978. "Dummy Endogenous Variables in a Simultaneous Equation System," Econometrica, Econometric Society, vol. 46(4), pages 931-959, July.
    26. Florian Hoffmann & Philip Oreopoulos, 2009. "Professor Qualities and Student Achievement," The Review of Economics and Statistics, MIT Press, vol. 91(1), pages 83-92, February.
    27. Scott E. Carrell & James E. West, 2010. "Does Professor Quality Matter? Evidence from Random Assignment of Students to Professors," Journal of Political Economy, University of Chicago Press, vol. 118(3), pages 409-432, June.
    28. Heckman, James, 2013. "Sample selection bias as a specification error," Applied Econometrics, Russian Presidential Academy of National Economy and Public Administration (RANEPA), vol. 31(3), pages 129-137.
    29. Meng, Chun-Lo & Schmidt, Peter, 1985. "On the Cost of Partial Observability in the Bivariate Probit Model," International Economic Review, Department of Economics, University of Pennsylvania and Osaka University Institute of Social and Economic Research Association, vol. 26(1), pages 71-85, February.
    30. Michael A. McPherson & R. Todd Jewell, 2007. "Leveling the Playing Field: Should Student Evaluation Scores be Adjusted?," Social Science Quarterly, Southwestern Social Science Association, vol. 88(3), pages 868-881, September.
    31. Arabmazar, Abbas & Schmidt, Peter, 1982. "An Investigation of the Robustness of the Tobit Estimator to Non-Normality," Econometrica, Econometric Society, vol. 50(4), pages 1055-1063, July.
    32. Francis Vella, 1998. "Estimating Models with Sample Selection Bias: A Survey," Journal of Human Resources, University of Wisconsin Press, vol. 33(1), pages 127-169.
    33. Anna Salomons & Maarten Goos, 2014. "Measuring Teaching Quality in Higher Education: Assessing the Problem of Selection Bias in Course Evaluations," Working Papers 14-16, Utrecht School of Economics.
    34. Pigini Claudia, 2015. "Bivariate Non-Normality in the Sample Selection Model," Journal of Econometric Methods, De Gruyter, vol. 4(1), pages 1-22, January.
    35. Jonah E. Rockoff, 2004. "The Impact of Individual Teachers on Student Achievement: Evidence from Panel Data," American Economic Review, American Economic Association, vol. 94(2), pages 247-252, May.
    36. Devesh Kapur & Megan Crowley, 2008. "Beyond the ABCs: Higher Education and Developing Countries," Working Papers 139, Center for Global Development.
    37. Van de Ven, Wynand P. M. M. & Van Praag, Bernard M. S., 1981. "The demand for deductibles in private health insurance : A probit model with sample selection," Journal of Econometrics, Elsevier, vol. 17(2), pages 229-252, November.
    38. Keane, Michael P, 1992. "A Note on Identification in the Multinomial Probit Model," Journal of Business & Economic Statistics, American Statistical Association, vol. 10(2), pages 193-200, April.
    39. Michael A McPherson & R Todd Jewell & Myungsup Kim, 2009. "What Determines Student Evaluation Scores? A Random Effects Analysis of Undergraduate Economics Classes," Eastern Economic Journal, Palgrave Macmillan;Eastern Economic Association, vol. 35(1), pages 37-51.
    40. Langbein, Laura, 2008. "Management by results: Student evaluation of faculty teaching and the mis-measurement of performance," Economics of Education Review, Elsevier, vol. 27(4), pages 417-428, August.
    41. Heckman, James J, 1974. "Shadow Prices, Market Wages, and Labor Supply," Econometrica, Econometric Society, vol. 42(4), pages 679-694, July.
    42. Paul Isely & Harinder Singh, 2005. "Do Higher Grades Lead to Favorable Student Evaluations?," The Journal of Economic Education, Taylor & Francis Journals, vol. 36(1), pages 29-42, January.
    43. Ewing, Andrew M., 2012. "Estimating the impact of relative expected grade on student evaluations of teachers," Economics of Education Review, Elsevier, vol. 31(1), pages 141-154.
    44. Bruce A. Weinberg & Masanori Hashimoto & Belton M. Fleisher, 2009. "Evaluating Teaching in Higher Education," The Journal of Economic Education, Taylor & Francis Journals, vol. 40(3), pages 227-261, July.
    45. Berg, Nathan, 2005. "Non-response bias," MPRA Paper 26373, University Library of Munich, Germany.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Neckermann, Susanne & Turmunkh, Uyanga & van Dolder, Dennie & Wang, Tong V., 2022. "Nudging student participation in online evaluations of teaching: Evidence from a field experiment," European Economic Review, Elsevier, vol. 141(C).
    2. Oana Eugenia & Gogu Emilia & Roman Monica & Marin Erika, 2022. "Students’ Perceptions on the Quality of the Economics Higher Education in Romania," Journal of Social and Economic Statistics, Sciendo, vol. 11(1-2), pages 14-35, December.
    3. Amalia Vanacore & Maria Sole Pellegrino, 2019. "How Reliable are Students’ Evaluations of Teaching (SETs)? A Study to Test Student’s Reproducibility and Repeatability," Social Indicators Research: An International and Interdisciplinary Journal for Quality-of-Life Measurement, Springer, vol. 146(1), pages 77-89, November.
    4. Marco Taliento, 2022. "The Triple Mission of the Modern University: Component Interplay and Performance Analysis from Italy," World, MDPI, vol. 3(3), pages 1-24, July.
    5. Angelo Antoci & Irene Brunetti & Pierluigi Sacco & Mauro Sodini, 2021. "Student evaluation of teaching, social influence dynamics, and teachers’ choices: An evolutionary model," Journal of Evolutionary Economics, Springer, vol. 31(1), pages 325-348, January.
    6. Cannon, Edmund & Cipriani, Giam Pietro, 2021. "Gender Differences in Student Evaluations of Teaching: Identification and Consequences," IZA Discussion Papers 14387, Institute of Labor Economics (IZA).
    7. José M. Ramírez-Hurtado & Alfredo G. Hernández-Díaz & Ana D. López-Sánchez & Víctor E. Pérez-León, 2021. "Measuring Online Teaching Service Quality in Higher Education in the COVID-19 Environment," IJERPH, MDPI, vol. 18(5), pages 1-14, March.
    8. Amal Said Al-Amri & Priya Mathew & Yong Zulina Zubairi & Rohana Jani, 2020. "Optimal Standards to Measure the Quality of Higher Education Institutions in Oman: Stakeholders’ Perception," SAGE Open, , vol. 10(3), pages 21582440209, July.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Anna Salomons & Maarten Goos, 2014. "Measuring Teaching Quality in Higher Education: Assessing the Problem of Selection Bias in Course Evaluations," Working Papers 14-16, Utrecht School of Economics.
    2. Braga, Michela & Paccagnella, Marco & Pellizzari, Michele, 2014. "Evaluating students’ evaluations of professors," Economics of Education Review, Elsevier, vol. 41(C), pages 71-88.
    3. Rieger, Matthias & Voorvelt, Katherine, 2016. "Gender, ethnicity and teaching evaluations: Evidence from mixed teaching teamsAuthor-Name: Wagner, Natascha," Economics of Education Review, Elsevier, vol. 54(C), pages 79-94.
    4. Wagner, N. & Rieger, M. & Voorvelt, K.J., 2016. "Gender, ethnicity and teaching evaluations : Evidence from mixed teaching teams," ISS Working Papers - General Series 617, International Institute of Social Studies of Erasmus University Rotterdam (ISS), The Hague.
    5. Naven, Matthew, 2019. "Human-Capital Formation During Childhood and Adolescence: Evidence from School Quality and Postsecondary Success in California," MPRA Paper 97716, University Library of Munich, Germany.
    6. Donghun Cho & Joonmo Cho, 2017. "Does More Accurate Knowledge of Course Grade Impact Teaching Evaluation?," Education Finance and Policy, MIT Press, vol. 12(2), pages 224-240, Spring.
    7. Tian, Zhilei & Wei, Yi & Li, Fang, 2019. "Who are better teachers? The effects of tenure-track and part-time faculty on student achievement," China Economic Review, Elsevier, vol. 53(C), pages 140-151.
    8. Gorry, Devon, 2017. "The impact of grade ceilings on student grades and course evaluations: Evidence from a policy change," Economics of Education Review, Elsevier, vol. 56(C), pages 133-140.
    9. Beleche, Trinidad & Fairris, David & Marks, Mindy, 2012. "Do course evaluations truly reflect student learning? Evidence from an objectively graded post-test," Economics of Education Review, Elsevier, vol. 31(5), pages 709-719.
    10. Ingo E. Isphording & Ulf Zölitz, 2020. "The value of a peer," ECON - Working Papers 342, Department of Economics - University of Zurich.
    11. Dhushyanth Raju, 2017. "Public School Teacher Management in Sri Lanka," South Asia Economic Journal, Institute of Policy Studies of Sri Lanka, vol. 18(1), pages 39-63, March.
    12. Rita Ginja & Julie Riise & Barton Willage & Alexander L.P. Willén, 2022. "Does Your Doctor Matter? Doctor Quality and Patient Outcomes," CESifo Working Paper Series 9788, CESifo.
    13. Rodríguez, Rosa & Rubio, Gonzalo, 2016. "Teaching quality and academic research," International Review of Economics Education, Elsevier, vol. 23(C), pages 10-27.
    14. Angelo Antoci & Irene Brunetti & Pierluigi Sacco & Mauro Sodini, 2021. "Student evaluation of teaching, social influence dynamics, and teachers’ choices: An evolutionary model," Journal of Evolutionary Economics, Springer, vol. 31(1), pages 325-348, January.
    15. Pau Balart & Antonio Cabrales, 2015. "La evaluación docente basada en el resultado como vía de mejora del sistema educativo," Studies on the Spanish Economy eee2015-13, FEDEA.
    16. Marine de Talancé, 2015. "Better Teachers, Better Results? Evidence from Rural Pakistan," Working Papers DT/2015/21, DIAL (Développement, Institutions et Mondialisation).
    17. Wedel, Katharina, 2021. "Instruction time and student achievement: The moderating role of teacher qualifications," Economics of Education Review, Elsevier, vol. 85(C).
    18. Omar Paccagnella, 2011. "Anchoring vignettes with sample selection due to non‐response," Journal of the Royal Statistical Society Series A, Royal Statistical Society, vol. 174(3), pages 665-687, July.
    19. Boring, Anne, 2017. "Gender biases in student evaluations of teaching," Journal of Public Economics, Elsevier, vol. 145(C), pages 27-41.
    20. Goel, Deepti & Barooah, Bidisha, 2018. "Drivers of Student Performance: Evidence from Higher Secondary Public Schools in Delhi," GLO Discussion Paper Series 231, Global Labor Organization (GLO).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:reihed:v:58:y:2017:i:4:d:10.1007_s11162-016-9429-8. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.