IDEAS home Printed from https://ideas.repec.org/p/arx/papers/2404.08816.html

Measuring the Quality of Answers in Political Q&As with Large Language Models

Author

Listed:
  • R. Michael Alvarez
  • Jacob Morrier

Abstract

This article proposes a new approach for assessing the quality of answers in political question-and-answer sessions. We measure the quality of an answer based on how easily and accurately it can be recognized in a random set of candidate answers given the question's text. This measure reflects the answer's relevance and depth of engagement with the question. Like semantic search, we can implement this approach by training a language model on the corpus of observed questions and answers without additional human-labeled data. We showcase and validate our methodology within the context of the Question Period in the Canadian House of Commons. Our analysis reveals that while some answers have a weak semantic connection to questions, hinting at some evasion or obfuscation, they are generally at least moderately relevant, far exceeding what we would expect from random replies. We also find a meaningful correlation between answer quality and the party affiliation of the members of Parliament asking the questions.

Suggested Citation

  • R. Michael Alvarez & Jacob Morrier, 2024. "Measuring the Quality of Answers in Political Q&As with Large Language Models," Papers 2404.08816, arXiv.org, revised Feb 2025.
  • Handle: RePEc:arx:papers:2404.08816
    as

    Download full text from publisher

    File URL: http://arxiv.org/pdf/2404.08816
    File Function: Latest version
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Laurer, Moritz & van Atteveldt, Wouter & Casas, Andreu & Welbers, Kasper, 2024. "Less Annotating, More Classifying: Addressing the Data Scarcity Issue of Supervised Machine Learning with Deep Transfer Learning and BERT-NLI," Political Analysis, Cambridge University Press, vol. 32(1), pages 84-100, January.
    2. Bestvater, Samuel E. & Monroe, Burt L., 2023. "Sentiment is Not Stance: Target-Aware Opinion Classification for Political Text Analysis," Political Analysis, Cambridge University Press, vol. 31(2), pages 235-256, April.
    3. Wang, Yu, 2023. "Topic Classification for Political Texts with Pretrained Language Models," Political Analysis, Cambridge University Press, vol. 31(4), pages 662-668, October.
    4. Widmann, Tobias & Wich, Maximilian, 2023. "Creating and Comparing Dictionary, Word Embedding, and Transformer-Based Models to Measure Discrete Emotions in German Political Text," Political Analysis, Cambridge University Press, vol. 31(4), pages 626-641, October.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Leek, Lauren Caroline & Bischl, Simeon, 2024. "How Central Bank Independence Shapes Monetary Policy Communication: A Large Language Model Application," SocArXiv yrhka, Center for Open Science.
    2. Leek, Lauren & Bischl, Simeon, 2025. "How central bank independence shapes monetary policy communication: A Large Language Model application," European Journal of Political Economy, Elsevier, vol. 87(C).
    3. repec:osf:socarx:yrhka_v1 is not listed on IDEAS
    4. Yang, Yiwen & Lin, Yi-Wei & Cheng, Li-Chen, 2025. "Impact of real-time public sentiment on herding behavior in Taiwan's stock market: Insights across investor types and industries," International Review of Economics & Finance, Elsevier, vol. 102(C).
    5. repec:osf:osfxxx:r3qng_v1 is not listed on IDEAS
    6. Benjamin Born & Nora Lamersdorf & Jana-Lynn Schuster & Sascha Steffen, 2025. "From Tweets to Transactions: High-Frequency Inflation Expectations, Consumption, and Stock Returns," CESifo Working Paper Series 12361, CESifo.
    7. Gründler, Klaus & Potrafke, Niklas & Wochner, Timo, 2025. "Outside employment and parliamentary priorities," Journal of Economic Behavior & Organization, Elsevier, vol. 237(C).
    8. Karimi Motahhar, Vahid & Gruca, Thomas S. & Tavakoli, Mohammad Hosein, 2025. "Emotions and the status quo: The anti-incumbency bias in political prediction markets," International Journal of Forecasting, Elsevier, vol. 41(2), pages 571-579.
    9. Yukthakiran Matla & Rohith Rao Yannamaneni & George Pappas, 2024. "Globalizing Food Items Based on Ingredient Consumption," Sustainability, MDPI, vol. 16(17), pages 1-22, August.
    10. Jonathan Adkins & Ali Al Bataineh & Majd Khalaf, 2024. "Identifying Persons of Interest in Digital Forensics Using NLP-Based AI," Future Internet, MDPI, vol. 16(11), pages 1-19, November.
    11. Cova, Joshua & Schmitz, Luuk, 2024. "A primer for the use of classifier and generative large language models in social science research," OSF Preprints r3qng, Center for Open Science.
    12. Pablo Picazo-Sanchez, 2025. "Breaking the silence: the rise of hate speech in Spanish politics," Journal of Computational Social Science, Springer, vol. 8(3), pages 1-26, August.
    13. Mottee, Lara K. & Hong, Wei-Ting & Nelson, John D., 2025. "Exploring aspects of transport-induced gentrification across the project lifecycle: An NLP approach," Journal of Transport Geography, Elsevier, vol. 128(C).
    14. Luca Bellodi & Massimo Morelli & Jorg L. Spenkuch & Edoardo Teso & Matia Vannoni & Guo Xu, 2026. "Personnel is Policy: Delegation and Political Misalignment in the Rulemaking Process," NBER Working Papers 34932, National Bureau of Economic Research, Inc.
    15. Kelvin Leong & Anna Sung, 2024. "Gender stereotypes in artificial intelligence within the accounting profession using large language models," Humanities and Social Sciences Communications, Palgrave Macmillan, vol. 11(1), pages 1-11, December.
    16. Stephanie Ng & James Zhang & Samson Yu & Asim Bhatti & Kathryn Backholer & C. P. Lim, 2025. "Stance classification: a comparative study and use case on Australian parliamentary debates," Journal of Computational Social Science, Springer, vol. 8(2), pages 1-37, May.

    More about this item

    NEP fields

    This paper has been announced in the following NEP Reports:

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:arx:papers:2404.08816. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: arXiv administrators (email available below). General contact details of provider: http://arxiv.org/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.