IDEAS home Printed from https://ideas.repec.org/p/arx/papers/2504.21400.html
   My bibliography  Save this paper

Who Gets the Callback? Generative AI and Gender Bias

Author

Listed:
  • Sugat Chaturvedi
  • Rochana Chaturvedi

Abstract

Generative artificial intelligence (AI), particularly large language models (LLMs), is being rapidly deployed in recruitment and for candidate shortlisting. We audit several mid-sized open-source LLMs for gender bias using a dataset of 332,044 real-world online job postings. For each posting, we prompt the model to recommend whether an equally qualified male or female candidate should receive an interview callback. We find that most models tend to favor men, especially for higher-wage roles. Mapping job descriptions to the Standard Occupational Classification system, we find lower callback rates for women in male-dominated occupations and higher rates in female-associated ones, indicating occupational segregation. A comprehensive analysis of linguistic features in job ads reveals strong alignment of model recommendations with traditional gender stereotypes. To examine the role of recruiter identity, we steer model behavior by infusing Big Five personality traits and simulating the perspectives of historical figures. We find that less agreeable personas reduce stereotyping, consistent with an agreeableness bias in LLMs. Our findings highlight how AI-driven hiring may perpetuate biases in the labor market and have implications for fairness and diversity within firms.

Suggested Citation

  • Sugat Chaturvedi & Rochana Chaturvedi, 2025. "Who Gets the Callback? Generative AI and Gender Bias," Papers 2504.21400, arXiv.org.
  • Handle: RePEc:arx:papers:2504.21400
    as

    Download full text from publisher

    File URL: http://arxiv.org/pdf/2504.21400
    File Function: Latest version
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. David Deming & Lisa B. Kahn, 2018. "Skill Requirements across Firms and Labor Markets: Evidence from Job Postings for Professionals," Journal of Labor Economics, University of Chicago Press, vol. 36(S1), pages 337-369.
    2. Thomas Le Barbanchon & Roland Rathelot & Alexandra Roulet, 2021. "Gender Differences in Job Search: Trading off Commute against Wage," The Quarterly Journal of Economics, President and Fellows of Harvard College, vol. 136(1), pages 381-426.
    3. Andreas Fuster & Paul Goldsmith‐Pinkham & Tarun Ramadorai & Ansgar Walther, 2022. "Predictably Unequal? The Effects of Machine Learning on Credit Markets," Journal of Finance, American Finance Association, vol. 77(1), pages 5-47, February.
    4. Benjamin S. Manning & Kehang Zhu & John J. Horton, 2024. "Automated Social Science: Language Models as Scientist and Subjects," Papers 2404.11794, arXiv.org, revised Apr 2024.
    5. Booth, Alison & Leigh, Andrew, 2010. "Do employers discriminate by gender? A field experiment in female-dominated occupations," Economics Letters, Elsevier, vol. 107(2), pages 236-238, May.
    6. Zhang, Shuo & Kuhn, Peter J., 2024. "Measuring Bias in Job Recommender Systems: Auditing the Algorithms," IZA Discussion Papers 17245, Institute of Labor Economics (IZA).
    7. Patrick Kline & Evan K Rose & Christopher R Walters, 2022. "Systemic Discrimination Among Large U.S. Employers [“Teachers and Student Achievement in the Chicago Public High Schools,”]," The Quarterly Journal of Economics, Oxford University Press, vol. 137(4), pages 1963-2036.
    8. Alexander Bick & Adam Blandin & David Deming, 2023. "The Rapid Adoption of Generative AI," On the Economy 98843, Federal Reserve Bank of St. Louis.
    9. Argyle, Lisa P. & Busby, Ethan C. & Fulda, Nancy & Gubler, Joshua R. & Rytting, Christopher & Wingate, David, 2023. "Out of One, Many: Using Language Models to Simulate Human Samples," Political Analysis, Cambridge University Press, vol. 31(3), pages 337-351, July.
    10. Anja Lambrecht & Catherine Tucker, 2019. "Algorithmic Bias? An Empirical Study of Apparent Gender-Based Discrimination in the Display of STEM Career Ads," Management Science, INFORMS, vol. 65(7), pages 2966-2981, July.
    11. Valentin Hofmann & Pratyusha Ria Kalluri & Dan Jurafsky & Sharese King, 2024. "AI generates covertly racist decisions about people based on their dialect," Nature, Nature, vol. 633(8028), pages 147-154, September.
    12. Haoran He & David Neumark & Qian Weng, 2021. "Do Workers Value Flexible Jobs? A Field Experiment," Journal of Labor Economics, University of Chicago Press, vol. 39(3), pages 709-738.
    13. John J. Horton, 2023. "Large Language Models as Simulated Economic Agents: What Can We Learn from Homo Silicus?," NBER Working Papers 31122, National Bureau of Economic Research, Inc.
    14. John J. Horton, 2023. "Large Language Models as Simulated Economic Agents: What Can We Learn from Homo Silicus?," Papers 2301.07543, arXiv.org.
    15. Shuo Zhang & Peter J. Kuhn, 2024. "Measuring Bias in Job Recommender Systems: Auditing the Algorithms," NBER Working Papers 32889, National Bureau of Economic Research, Inc.
    16. Benjamin S. Manning & Kehang Zhu & John J. Horton, 2024. "Automated Social Science: Language Models as Scientist and Subjects," NBER Working Papers 32381, National Bureau of Economic Research, Inc.
    17. Philip Oreopoulos, 2011. "Why Do Skilled Immigrants Struggle in the Labor Market? A Field Experiment with Thirteen Thousand Resumes," American Economic Journal: Economic Policy, American Economic Association, vol. 3(4), pages 148-171, November.
    18. Danielle Li & Lindsey R. Raymond & Peter Bergman, 2020. "Hiring as Exploration," NBER Working Papers 27736, National Bureau of Economic Research, Inc.
    19. Marianne Bertrand & Sendhil Mullainathan, 2004. "Are Emily and Greg More Employable Than Lakisha and Jamal? A Field Experiment on Labor Market Discrimination," American Economic Review, American Economic Association, vol. 94(4), pages 991-1013, September.
    20. Laura K. Gee, 2019. "The More You Know: Information Effects on Job Application Rates in a Large Field Experiment," Management Science, INFORMS, vol. 67(5), pages 2077-2094, May.
    21. Kuhn, Peter & Shen, Kailing & Zhang, Shuo, 2020. "Gender-targeted job ads in the recruitment process: Facts from a Chinese job board," Journal of Development Economics, Elsevier, vol. 147(C).
    22. Greenwald, Daniel L. & Howell, Sabrina T. & Li, Cangyuan & Yimfor, Emmanuel, 2024. "Regulatory arbitrage or random errors? Implications of race prediction algorithms in fair lending analysis," Journal of Financial Economics, Elsevier, vol. 157(C).
    23. Matteo Tranchero & Cecil-Francis Brenninkmeijer & Arul Murugan & Abhishek Nagaraj, 2024. "Theorizing with Large Language Models," NBER Working Papers 33033, National Bureau of Economic Research, Inc.
    24. repec:feb:natura:0058 is not listed on IDEAS
    25. Abraham, Lisa & Hallermeier, Johannes & Stein, Alison, 2024. "Words matter: Experimental evidence from job applications," Journal of Economic Behavior & Organization, Elsevier, vol. 225(C), pages 348-391.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Jiafu An & Difang Huang & Chen Lin & Mingzhu Tai, 2024. "Measuring Gender and Racial Biases in Large Language Models," Papers 2403.15281, arXiv.org.
    2. Alejandro Lopez-Lira & Yuehua Tang & Mingyin Zhu, 2025. "The Memorization Problem: Can We Trust LLMs' Economic Forecasts?," Papers 2504.14765, arXiv.org.
    3. Alejandro Lopez-Lira, 2025. "Can Large Language Models Trade? Testing Financial Theories with LLM Agents in Market Simulations," Papers 2504.10789, arXiv.org.
    4. Felipe A. Csaszar & Harsh Ketkar & Hyunjin Kim, 2024. "Artificial Intelligence and Strategic Decision-Making: Evidence from Entrepreneurs and Investors," Papers 2408.08811, arXiv.org.
    5. Anthony Edo & Nicolas Jacquemet & Constantine Yannelis, 2019. "Language skills and homophilous hiring discrimination: Evidence from gender and racially differentiated applications," Review of Economics of the Household, Springer, vol. 17(1), pages 349-376, March.
    6. Eva O. Arceo-Gomez & Raymundo M. Campos-Vazquez, 2014. "Race and Marriage in the Labor Market: A Discrimination Correspondence Study in a Developing Country," American Economic Review, American Economic Association, vol. 104(5), pages 376-380, May.
    7. Baert, Stijn, 2017. "Hiring Discrimination: An Overview of (Almost) All Correspondence Experiments Since 2005," GLO Discussion Paper Series 61, Global Labor Organization (GLO).
    8. Jung Ho Choi & Joseph Pacelli & Kristina M. Rennekamp & Sorabh Tomar, 2023. "Do Jobseekers Value Diversity Information? Evidence from a Field Experiment and Human Capital Disclosures," Journal of Accounting Research, Wiley Blackwell, vol. 61(3), pages 695-735, June.
    9. Piopiunik, Marc & Schwerdt, Guido & Simon, Lisa & Woessmann, Ludger, 2020. "Skills, signals, and employability: An experimental investigation," European Economic Review, Elsevier, vol. 123(C).
    10. Marianne Bertrand & Esther Duflo, 2016. "Field Experiments on Discrimination," NBER Working Papers 22014, National Bureau of Economic Research, Inc.
    11. Nir Chemaya & Daniel Martin, 2023. "Perceptions and Detection of AI Use in Manuscript Preparation for Academic Journals," Papers 2311.14720, arXiv.org, revised Jan 2024.
    12. Nir Chemaya & Daniel Martin, 2024. "Perceptions and detection of AI use in manuscript preparation for academic journals," PLOS ONE, Public Library of Science, vol. 19(7), pages 1-16, July.
    13. repec:osf:osfxxx:r3qng_v1 is not listed on IDEAS
    14. Melo, Vitor & Rocha, Hugo Vaca Pereira & Sigaud, Liam & Warren, Patrick L. & Gaddis, S. Michael, 2024. "Understanding Discrimination in College Admissions: A Field Experiment," SocArXiv 5ctms, Center for Open Science.
    15. Haoran He & Sherry Xin Li & Yuling Han, 2023. "Labor Market Discrimination against Family Responsibilities: A Correspondence Study with Policy Change in China," Journal of Labor Economics, University of Chicago Press, vol. 41(2), pages 361-387.
    16. Rajeev Darolia & Cory Koedel & Paco Martorell & Katie Wilson & Francisco Perez-Arce, 2016. "Race and gender effects on employer interest in job applicants: new evidence from a resume field experiment," Applied Economics Letters, Taylor & Francis Journals, vol. 23(12), pages 853-856, August.
    17. Valfort, Marie-Anne, 2020. "Anti-Muslim discrimination in France: Evidence from a field experiment," World Development, Elsevier, vol. 135(C).
    18. repec:osf:socarx:5ctms_v1 is not listed on IDEAS
    19. Benny, Liza & Bhalotra, Sonia & Fernández, Manuel, 2021. "Occupation flexibility and the graduate gender wage gap in the UK," ISER Working Paper Series 2021-05, Institute for Social and Economic Research.
    20. Gaddis, S. Michael, 2018. "An Introduction to Audit Studies in the Social Sciences," SocArXiv e5hfc, Center for Open Science.
    21. Qinyue Luo & Liaoliang Zhang, 2025. "Age Discrimination in Hiring: Evidence from Online Job Ads," RF Berlin - CReAM Discussion Paper Series 2511, Rockwool Foundation Berlin (RF Berlin) - Centre for Research and Analysis of Migration (CReAM).
    22. Aliya Amirova & Theodora Fteropoulli & Nafiso Ahmed & Martin R Cowie & Joel Z Leibo, 2024. "Framework-based qualitative analysis of free responses of Large Language Models: Algorithmic fidelity," PLOS ONE, Public Library of Science, vol. 19(3), pages 1-33, March.

    More about this item

    NEP fields

    This paper has been announced in the following NEP Reports:

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:arx:papers:2504.21400. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: arXiv administrators (email available below). General contact details of provider: http://arxiv.org/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.