IDEAS home Printed from https://ideas.repec.org/a/eee/apmaco/v463y2024ics0096300323005337.html
   My bibliography  Save this article

Interaction state Q-learning promotes cooperation in the spatial prisoner's dilemma game

Author

Listed:
  • Yang, Zhengzhi
  • Zheng, Lei
  • Perc, Matjaž
  • Li, Yumeng

Abstract

Many recent studies have used reinforcement learning methods to investigate the behavior of agents in evolutionary games. Q-learning, in particular, has become a mainstream method during this development. Here we introduce Q-learning agents into the evolutionary prisoner's dilemma game on a square lattice. Specifically, we associate the state space of Q-learning agents with the strategies of their neighbors, and we introduce a neighboring reward information sharing mechanism. We thus provide Q-learning agents with the payoff information of their neighbors, in addition to their strategies, which has not been done in previous studies. Through simulations, we show that considering neighborhood payoff information can significantly promote cooperation in the population. Moreover, we show that for an appropriate strength of neighborhood payoff information sharing, a chessboard pattern emerges on the lattice. We analyze in detail the reasons for the emergence of the chessboard pattern and the increase in cooperation frequency, and we also provide a theoretical analysis based on the pair approximation method. We hope that our research will inspire effective approaches for resolving social dilemmas by means of sharing more information among reinforcement learning agents during evolutionary games.

Suggested Citation

  • Yang, Zhengzhi & Zheng, Lei & Perc, Matjaž & Li, Yumeng, 2024. "Interaction state Q-learning promotes cooperation in the spatial prisoner's dilemma game," Applied Mathematics and Computation, Elsevier, vol. 463(C).
  • Handle: RePEc:eee:apmaco:v:463:y:2024:i:c:s0096300323005337
    DOI: 10.1016/j.amc.2023.128364
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0096300323005337
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.amc.2023.128364?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Cao, Xian-Bin & Du, Wen-Bo & Rong, Zhi-Hai, 2010. "The evolutionary public goods game on scale-free networks with heterogeneous investment," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 389(6), pages 1273-1280.
    2. Takahiro Ezaki & Yutaka Horita & Masanori Takezawa & Naoki Masuda, 2016. "Reinforcement Learning Explains Conditional Cooperation and Its Moody Cousin," PLOS Computational Biology, Public Library of Science, vol. 12(7), pages 1-13, July.
    3. Christoph Hauert & Michael Doebeli, 2004. "Spatial structure often inhibits the evolution of cooperation in the snowdrift game," Nature, Nature, vol. 428(6983), pages 643-646, April.
    4. Li, Yumeng & Zhang, Jun & Perc, Matjaž, 2018. "Effects of compassion on the evolution of cooperation in spatial social dilemmas," Applied Mathematics and Computation, Elsevier, vol. 320(C), pages 437-443.
    5. Li, Yumeng & Wang, Hanchen & Du, Wenbo & Perc, Matjaž & Cao, Xianbin & Zhang, Jun, 2019. "Resonance-like cooperation due to transaction costs in the prisoner’s dilemma game," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 521(C), pages 248-257.
    6. Oriol Vinyals & Igor Babuschkin & Wojciech M. Czarnecki & Michaël Mathieu & Andrew Dudzik & Junyoung Chung & David H. Choi & Richard Powell & Timo Ewalds & Petko Georgiev & Junhyuk Oh & Dan Horgan & M, 2019. "Grandmaster level in StarCraft II using multi-agent reinforcement learning," Nature, Nature, vol. 575(7782), pages 350-354, November.
    7. Theodor Cimpeanu & Francisco C. Santos & The Anh Han, 2023. "Does Spending More Always Ensure Higher Cooperation? An Analysis of Institutional Incentives on Heterogeneous Networks," Dynamic Games and Applications, Springer, vol. 13(4), pages 1236-1255, December.
    8. Ding, Hong & Zhang, Geng-shun & Wang, Shi-hao & Li, Juan & Wang, Zhen, 2019. "Q-learning boosts the evolution of cooperation in structured population by involving extortion," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 536(C).
    9. Cimpeanu, Theodor & Di Stefano, Alessandro & Perret, Cedric & Han, The Anh, 2023. "Social diversity reduces the complexity and cost of fostering fairness," Chaos, Solitons & Fractals, Elsevier, vol. 167(C).
    10. Du, Wen-Bo & Zheng, Hao-Ran & Hu, Mao-Bin, 2008. "Evolutionary prisoner’s dilemma game on weighted scale-free networks," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 387(14), pages 3796-3800.
    11. Wang, Shengxian & Chen, Xiaojie & Xiao, Zhilong & Szolnoki, Attila, 2022. "Decentralized incentives for general well-being in networked public goods game," Applied Mathematics and Computation, Elsevier, vol. 431(C).
    12. Wang, Hanchen & Sun, Yichun & Zheng, Lei & Du, Wenbo & Li, Yumeng, 2018. "The public goods game on scale-free networks with heterogeneous investment," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 509(C), pages 396-404.
    13. Lee, Hsuan-Wei & Cleveland, Colin & Szolnoki, Attila, 2022. "Mercenary punishment in structured populations," Applied Mathematics and Computation, Elsevier, vol. 417(C).
    14. Marco Alberto Javarone & Daniele Marinazzo, 2017. "Evolutionary dynamics of group formation," PLOS ONE, Public Library of Science, vol. 12(11), pages 1-10, November.
    15. David Silver & Julian Schrittwieser & Karen Simonyan & Ioannis Antonoglou & Aja Huang & Arthur Guez & Thomas Hubert & Lucas Baker & Matthew Lai & Adrian Bolton & Yutian Chen & Timothy Lillicrap & Fan , 2017. "Mastering the game of Go without human knowledge," Nature, Nature, vol. 550(7676), pages 354-359, October.
    16. Wang, Chaoqian & Szolnoki, Attila, 2023. "Inertia in spatial public goods games under weak selection," Applied Mathematics and Computation, Elsevier, vol. 449(C).
    17. Geng, Yini & Liu, Yifan & Lu, Yikang & Shen, Chen & Shi, Lei, 2022. "Reinforcement learning explains various conditional cooperation," Applied Mathematics and Computation, Elsevier, vol. 427(C).
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Sun, Jiaqin & Fan, Ruguo & Luo, Ming & Zhang, Yingqing & Dong, Lili, 2018. "The evolution of cooperation in spatial prisoner’s dilemma game with dynamic relationship-based preferential learning," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 512(C), pages 598-611.
    2. Yang, Han-Xin & Yang, Jing, 2019. "Reputation-based investment strategy promotes cooperation in public goods games," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 523(C), pages 886-893.
    3. Lee, Hsuan-Wei & Cleveland, Colin & Szolnoki, Attila, 2023. "Group-size dependent synergy in heterogeneous populations," Chaos, Solitons & Fractals, Elsevier, vol. 167(C).
    4. Molnar, Grant & Hammond, Caroline & Fu, Feng, 2023. "Reactive means in the iterated Prisoner’s dilemma," Applied Mathematics and Computation, Elsevier, vol. 458(C).
    5. Ping Zhu & Guiyi Wei, 2014. "Stochastic Heterogeneous Interaction Promotes Cooperation in Spatial Prisoner's Dilemma Game," PLOS ONE, Public Library of Science, vol. 9(4), pages 1-10, April.
    6. Qinghu Liao & Wenwen Dong & Boxin Zhao, 2023. "A New Strategy to Solve “the Tragedy of the Commons” in Sustainable Grassland Ecological Compensation: Experience from Inner Mongolia, China," Sustainability, MDPI, vol. 15(12), pages 1-24, June.
    7. Yu, Fengyuan & Wang, Jianwei & Chen, Wei & He, Jialu, 2023. "Increased cooperation potential and risk under suppressed strategy differentiation," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 621(C).
    8. Zha, Jiajing & Li, Cong & Fan, Suohai, 2022. "The effect of stability-based strategy updating on cooperation in evolutionary social dilemmas," Applied Mathematics and Computation, Elsevier, vol. 413(C).
    9. Lee, Hsuan-Wei & Cleveland, Colin & Szolnoki, Attila, 2021. "Small fraction of selective cooperators can elevate general wellbeing significantly," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 582(C).
    10. Quan, Ji & Zhou, Yawen & Wang, Xianjia & Yang, Jian-Bo, 2020. "Information fusion based on reputation and payoff promotes cooperation in spatial public goods game," Applied Mathematics and Computation, Elsevier, vol. 368(C).
    11. Chen, Qiao & Chen, Tong & Wang, Yongjie, 2019. "Cleverly handling the donation information can promote cooperation in public goods game," Applied Mathematics and Computation, Elsevier, vol. 346(C), pages 363-373.
    12. Zhu, Wenqiang & Pan, Qiuhui & Song, Sha & He, Mingfeng, 2023. "Effects of exposure-based reward and punishment on the evolution of cooperation in prisoner’s dilemma game," Chaos, Solitons & Fractals, Elsevier, vol. 172(C).
    13. Lee, Hsuan-Wei & Cleveland, Colin & Szolnoki, Attila, 2023. "Restoring spatial cooperation with myopic agents in a three-strategy social dilemma," Applied Mathematics and Computation, Elsevier, vol. 458(C).
    14. Li, Wenqing & Ni, Shaoquan, 2022. "Train timetabling with the general learning environment and multi-agent deep reinforcement learning," Transportation Research Part B: Methodological, Elsevier, vol. 157(C), pages 230-251.
    15. Geng, Yini & Liu, Yifan & Lu, Yikang & Shen, Chen & Shi, Lei, 2022. "Reinforcement learning explains various conditional cooperation," Applied Mathematics and Computation, Elsevier, vol. 427(C).
    16. Qingyan Li & Tao Lin & Qianyi Yu & Hui Du & Jun Li & Xiyue Fu, 2023. "Review of Deep Reinforcement Learning and Its Application in Modern Renewable Power System Control," Energies, MDPI, vol. 16(10), pages 1-23, May.
    17. Michael Curry & Alexander Trott & Soham Phade & Yu Bai & Stephan Zheng, 2022. "Analyzing Micro-Founded General Equilibrium Models with Many Agents using Deep Reinforcement Learning," Papers 2201.01163, arXiv.org, revised Feb 2022.
    18. Dong Liu & Feng Xiao & Jian Luo & Fan Yang, 2023. "Deep Reinforcement Learning-Based Holding Control for Bus Bunching under Stochastic Travel Time and Demand," Sustainability, MDPI, vol. 15(14), pages 1-18, July.
    19. Jia, Chun-Xiao & Liu, Run-Ran, 2022. "A moderate self-interest preference promotes cooperation in spatial public goods game," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 589(C).
    20. Song, Sha & Pan, Qiuhui & Zhu, Wenqiang & He, Mingfeng, 2023. "Evolution of cooperation in games with dual attribute strategy," Chaos, Solitons & Fractals, Elsevier, vol. 175(P1).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:apmaco:v:463:y:2024:i:c:s0096300323005337. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: https://www.journals.elsevier.com/applied-mathematics-and-computation .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.