Author
Listed:
- Zhang, Qianwei
- Zhang, Xinran
Abstract
Reinforcement learning, as a powerful framework for analyzing strategic dynamics in evolutionary games, has gained significant traction in game theory research. In this study, we propose a dual-reputation incentive mechanism that integrates individual and group reputation metrics within the spatial Prisoner's Dilemma paradigm, aiming to elucidate how adaptive Q-learning drives the evolution of cooperation. Our approach combines traditional game payoffs with reputation-based rewards through a novel Q-learning reward function, strategically decomposing reputation into two components: individual rewards (quantifying an agent’s behavioral history) and group rewards (reflecting the collective reputation of their local neighborhood). Simulations demonstrate that when individual reputation rewards are prioritized, agents optimize long-term gains by dynamically adjusting strategies under strong motivational incentives, which ultimately enhances global cooperation levels. Microscopic analysis reveals that individual reputation incentives promote high-density cooperator clusters and facilitate cooperative behavior propagation. Furthermore, when a high weight is assigned to individual reputation rewards, evolutionary analysis demonstrates that cooperative Q-values consistently exceeds defective ones, indicating the emergence of cooperation as an evolutionarily stable strategy. This research provides theoretical insights for designing reputation-aware reinforcement learning systems to foster cooperation in real-world social dilemmas.
Suggested Citation
Zhang, Qianwei & Zhang, Xinran, 2025.
"Q-learning driven cooperative evolution with dual-reputation incentive mechanisms,"
Applied Mathematics and Computation, Elsevier, vol. 507(C).
Handle:
RePEc:eee:apmaco:v:507:y:2025:i:c:s0096300325003169
DOI: 10.1016/j.amc.2025.129590
Download full text from publisher
As the access to this document is restricted, you may want to
for a different version of it.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:apmaco:v:507:y:2025:i:c:s0096300325003169. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: https://www.journals.elsevier.com/applied-mathematics-and-computation .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.