Author
Listed:
- Y. J. Sun
- Z. Y. Lu
- C. H. Liao
- T. S. Dai
- S. M. Yuan
Abstract
Traditional stock selection and optimization strategies are significantly restricted by the unpredictable nature of time-varying financial markets. However, flexible machine learning (ML) models can adapt to environmental changes and make timely investment decisions. In this study, we employ fundamental analyses using feedforward neural networks (FNN) to select stocks and propose a new deep reinforcement learning (DRL) method to further adjust allocations. To our knowledge, this is the first ML model to offer complete portfolio management functionalities with financial interpretability based on fundamental analysis. The reward function of our DRL model is based on the relative return to an equal-weighted (EW) portfolio, ensuring the stability of trading policy learning. We construct the policy network using temporal convolutional networks (TCN) that incorporate concatenated weekly (biweekly) aggregated information to learn stock price time series. Our model employs proximal policy optimization (PPO) to select allocation strategies, minimizing the impact of significant position changes and associated transaction costs. The resulting model outperforms traditional convolutional neural networks (CNN) and other commonly used policy gradient models. Evaluation using S&P 100 index data from 2009 to 2021 demonstrates that each component and combination significantly enhances portfolio performance.
Suggested Citation
Y. J. Sun & Z. Y. Lu & C. H. Liao & T. S. Dai & S. M. Yuan, 2026.
"Optimizing stock portfolios with deep reinforcement learning after FNN-based fundamental analysis,"
Quantitative Finance, Taylor & Francis Journals, vol. 26(1), pages 137-159, January.
Handle:
RePEc:taf:quantf:v:26:y:2026:i:1:p:137-159
DOI: 10.1080/14697688.2025.2604770
Download full text from publisher
As the access to this document is restricted, you may want to
for a different version of it.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:taf:quantf:v:26:y:2026:i:1:p:137-159. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Chris Longhurst (email available below). General contact details of provider: http://www.tandfonline.com/RQUF20 .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.