Author
Listed:
- Yoav Ger
- Eliya Nachmani
- Lior Wolf
- Nitzan Shahar
Abstract
Reinforcement learning (RL) models are used extensively to study human behavior. These rely on normative models of behavior and stress interpretability over predictive capabilities. More recently, neural network models have emerged as a descriptive modeling paradigm that is capable of high predictive power yet with limited interpretability. Here, we seek to augment the expressiveness of theoretical RL models with the high flexibility and predictive power of neural networks. We introduce a novel framework, which we term theoretical-RNN (t-RNN), whereby a recurrent neural network is trained to predict trial-by-trial behavior and to infer theoretical RL parameters using artificial data of RL agents performing a two-armed bandit task. In three studies, we then examined the use of our approach to dynamically predict unseen behavior along with time-varying theoretical RL parameters. We first validate our approach using synthetic data with known RL parameters. Next, as a proof-of-concept, we applied our framework to two independent datasets of humans performing the same task. In the first dataset, we describe differences in theoretical RL parameters dynamic among clinical psychiatric vs. healthy controls. In the second dataset, we show that the exploration strategies of humans varied dynamically in response to task phase and difficulty. For all analyses, we found better performance in the prediction of actions for t-RNN compared to the stationary maximum-likelihood RL method. We discuss the use of neural networks to facilitate the estimation of latent RL parameters underlying choice behavior.Author summary: Currently, neural network models fitted directly to behavioral human data are thought to dramatically outperform theoretical computational models in terms of predictive accuracy. However, these networks do not provide a clear theoretical interpretation of the mechanisms underlying the observed behavior. Generating plausible theoretical explanations for observed human data is a major goal in computational neuroscience. Here, we provide a proof-of-concept for a novel method where a recurrent neural network (RNN) is trained on artificial data generated from a known theoretical model to predict both trial-by-trial actions and theoretical parameters. We then freeze the RNN weights and use it to predict both actions and theoretical parameters of empirical data. We first validate our approach using synthetic data where the theoretical parameters are known. We then show, using two empirical datasets, that our approach allows dynamic estimation of latent parameters while providing better action predictions compared to theoretical models fitted with a maximum-likelihood approach. This proof-of-concept suggests that neural networks can be trained to predict meaningful time-varying theoretical parameters.
Suggested Citation
Yoav Ger & Eliya Nachmani & Lior Wolf & Nitzan Shahar, 2024.
"Harnessing the flexibility of neural networks to predict dynamic theoretical parameters underlying human choice behavior,"
PLOS Computational Biology, Public Library of Science, vol. 20(1), pages 1-22, January.
Handle:
RePEc:plo:pcbi00:1011678
DOI: 10.1371/journal.pcbi.1011678
Download full text from publisher
Most related items
These are the items that most often cite the same works as this one and are cited by the same works as this one.
- Amir Bagherpour, 2021.
"How computer simulations enhance geopolitical decisionāmaking: A commentary on Lustick and Tetlock 2021,"
Futures & Foresight Science, John Wiley & Sons, vol. 3(2), June.
- Amir Dezfouli & Bernard W Balleine, 2019.
"Learning the structure of the world: The adaptive nature of state-space and action representations in multi-stage decision-making,"
PLOS Computational Biology, Public Library of Science, vol. 15(9), pages 1-22, September.
- Shiva Farashahi & Alireza Soltani, 2021.
"Computational mechanisms of distributed value representations and mixed learning strategies,"
Nature Communications, Nature, vol. 12(1), pages 1-18, December.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:plo:pcbi00:1011678. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: ploscompbiol (email available below). General contact details of provider: https://journals.plos.org/ploscompbiol/ .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.