IDEAS home Printed from https://ideas.repec.org/p/arx/papers/2512.08270.html

Reasoning Models Ace the CFA Exams

Author

Listed:
  • Jaisal Patel
  • Yunzhe Chen
  • Kaiwen He
  • Keyi Wang
  • David Li
  • Kairong Xiao
  • Xiao-Yang Liu

Abstract

Previous research has reported that large language models (LLMs) demonstrate poor performance on the Chartered Financial Analyst (CFA) exams. However, recent reasoning models have achieved strong results on graduate-level academic and professional examinations across various disciplines. In this paper, we evaluate state-of-the-art reasoning models on a set of mock CFA exams consisting of 980 questions across three Level I exams, two Level II exams, and three Level III exams. Using the same pass/fail criteria from prior studies, we find that most models clear all three levels. The models that pass, ordered by overall performance, are Gemini 3.0 Pro, Gemini 2.5 Pro, GPT-5, Grok 4, Claude Opus 4.1, and DeepSeek-V3.1. Specifically, Gemini 3.0 Pro achieves a record score of 97.6% on Level I. Performance is also strong on Level II, led by GPT-5 at 94.3%. On Level III, Gemini 2.5 Pro attains the highest score with 86.4% on multiple-choice questions while Gemini 3.0 Pro achieves 92.0% on constructed-response questions.

Suggested Citation

  • Jaisal Patel & Yunzhe Chen & Kaiwen He & Keyi Wang & David Li & Kairong Xiao & Xiao-Yang Liu, 2025. "Reasoning Models Ace the CFA Exams," Papers 2512.08270, arXiv.org.
  • Handle: RePEc:arx:papers:2512.08270
    as

    Download full text from publisher

    File URL: http://arxiv.org/pdf/2512.08270
    File Function: Latest version
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Xiao-Yang Liu & Guoxuan Wang & Hongyang Yang & Daochen Zha, 2023. "FinGPT: Democratizing Internet-scale Data for Financial Large Language Models," Papers 2307.10485, arXiv.org, revised Nov 2023.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Yichen Luo & Yebo Feng & Jiahua Xu & Paolo Tasca & Yang Liu, 2025. "LLM-Powered Multi-Agent System for Automated Crypto Portfolio Management," Papers 2501.00826, arXiv.org, revised Jan 2025.
    2. Thanos Konstantinidis & Giorgos Iacovides & Mingxue Xu & Tony G. Constantinides & Danilo Mandic, 2024. "FinLlama: Financial Sentiment Classification for Algorithmic Trading Applications," Papers 2403.12285, arXiv.org.
    3. Artur Kulpa & Grzegorz Wojarnik, 2025. "Prompt Engineering in Finance: An LLM-Based Multi-Agent Architecture for Decision Support," European Research Studies Journal, European Research Studies Journal, vol. 0(3), pages 1201-1217.
    4. Wentao Zhang & Mingxuan Zhao & Jincheng Gao & Jieshun You & Huaiyu Jia & Yilei Zhao & Bo An & Shuo Sun, 2026. "AlphaForgeBench: Benchmarking End-to-End Trading Strategy Design with Large Language Models," Papers 2602.18481, arXiv.org.
    5. Yuan Li & Bingqiao Luo & Qian Wang & Nuo Chen & Xu Liu & Bingsheng He, 2024. "A Reflective LLM-based Agent to Guide Zero-shot Cryptocurrency Trading," Papers 2407.09546, arXiv.org.
    6. Haochen Luo & Zhengzhao Lai & Junjie Xu & Yifan Li & Tang Pok Hin & Yuan Zhang & Chen Liu, 2026. "From Natural Language to Executable Option Strategies via Large Language Models," Papers 2603.16434, arXiv.org.
    7. Jun Han & Shuo Zhang & Wei Li & Zhi Yang & Yifan Dong & Tu Hu & Jialuo Yuan & Xiaomin Yu & Yumo Zhu & Fangqi Lou & Xin Guo & Zhaowei Liu & Tianyi Jiang & Ruichuan An & Jingping Liu & Biao Wu & Rongze , 2026. "QuantaAlpha: An Evolutionary Framework for LLM-Driven Alpha Mining," Papers 2602.07085, arXiv.org, revised Apr 2026.
    8. Guojun Xiong & Zhiyang Deng & Keyi Wang & Yupeng Cao & Haohang Li & Yangyang Yu & Xueqing Peng & Mingquan Lin & Kaleb E Smith & Xiao-Yang Liu & Jimin Huang & Sophia Ananiadou & Qianqian Xie, 2025. "FLAG-Trader: Fusion LLM-Agent with Gradient-based Reinforcement Learning for Financial Trading," Papers 2502.11433, arXiv.org, revised Feb 2025.
    9. Haohang Li & Yupeng Cao & Yangyang Yu & Shashidhar Reddy Javaji & Zhiyang Deng & Yueru He & Yuechen Jiang & Zining Zhu & Koduvayur Subbalakshmi & Guojun Xiong & Jimin Huang & Lingfei Qian & Xueqing Pe, 2024. "INVESTORBENCH: A Benchmark for Financial Decision-Making Tasks with LLM-based Agent," Papers 2412.18174, arXiv.org.
    10. Masanori Hirano & Kentaro Imajo, 2024. "The Construction of Instruction-tuned LLMs for Finance without Instruction Data Using Continual Pretraining and Model Merging," Papers 2409.19854, arXiv.org.
    11. Kefan Chen & Hussain Ahmad & Diksha Goel & Claudia Szabo, 2025. "3S-Trader: A Multi-LLM Framework for Adaptive Stock Scoring, Strategy, and Selection in Portfolio Optimization," Papers 2510.17393, arXiv.org.
    12. Junzhe Jiang & Chang Yang & Aixin Cui & Sihan Jin & Ruiyu Wang & Bo Li & Xiao Huang & Dongning Sun & Xinrun Wang, 2025. "FinMaster: A Holistic Benchmark for Mastering Full-Pipeline Financial Workflows with LLMs," Papers 2505.13533, arXiv.org.
    13. Masoud Soleimani, 2025. "LLM-Generated Counterfactual Stress Scenarios for Portfolio Risk Simulation via Hybrid Prompt-RAG Pipeline," Papers 2512.07867, arXiv.org.
    14. Zuoyou Jiang & Li Zhao & Rui Sun & Ruohan Sun & Zhongjian Li & Jing Li & Daxin Jiang & Zuo Bai & Cheng Hua, 2025. "Alpha-R1: Alpha Screening with LLM Reasoning via Reinforcement Learning," Papers 2512.23515, arXiv.org.
    15. Saber Talazadeh & Dragan Perakovic, 2024. "SARF: Enhancing Stock Market Prediction with Sentiment-Augmented Random Forest," Papers 2410.07143, arXiv.org.
    16. David Kuo Chuen Lee & Chong Guan & Yinghui Yu & Qinxu Ding, 2024. "A Comprehensive Review of Generative AI in Finance," FinTech, MDPI, vol. 3(3), pages 1-19, September.
    17. Fernando Spadea & Oshani Seneviratne, 2025. "Aligning Language Models with Investor and Market Behavior for Financial Recommendations," Papers 2510.15993, arXiv.org.
    18. Ameer Tamoor Khan & Shuai Li & Xinwei Cao, 2025. "Bridging finance and AI: a comprehensive survey of large language models in financial system," Digital Finance, Springer, vol. 7(4), pages 679-701, December.
    19. Gongao Zhang & Haijiang Zeng & Lu Jiang, 2026. "Uni-FinLLM: A Unified Multimodal Large Language Model with Modular Task Heads for Micro-Level Stock Prediction and Macro-Level Systemic Risk Assessment," Papers 2601.02677, arXiv.org.
    20. Wo Long & Wenxin Zeng & Xiaoyu Zhang & Ziyao Zhou, 2025. "Integrating Large Language Models and Reinforcement Learning for Sentiment-Driven Quantitative Trading," Papers 2510.10526, arXiv.org.

    More about this item

    NEP fields

    This paper has been announced in the following NEP Reports:

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:arx:papers:2512.08270. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: arXiv administrators (email available below). General contact details of provider: http://arxiv.org/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.