IDEAS home Printed from https://ideas.repec.org/a/plo/pone00/0313951.html
   My bibliography  Save this article

How social reinforcement learning can lead to metastable polarisation and the voter model

Author

Listed:
  • Benedikt V Meylahn
  • Janusz M Meylahn

Abstract

Previous explanations for the persistence of polarization of opinions have typically included modelling assumptions that predispose the possibility of polarization (i.e., assumptions allowing a pair of agents to drift apart in their opinion such as repulsive interactions or bounded confidence). An exception is a recent simulation study showing that polarization is persistent when agents form their opinions using social reinforcement learning. Our goal is to highlight the usefulness of reinforcement learning in the context of modeling opinion dynamics, but that caution is required when selecting the tools used to study such a model. We show that the polarization observed in the model of the simulation study cannot persist indefinitely, and exhibits consensus asymptotically with probability one. By constructing a link between the reinforcement learning model and the voter model, we argue that the observed polarization is metastable. Finally, we show that a slight modification in the learning process of the agents changes the model from being non-ergodic to being ergodic. Our results show that reinforcement learning may be a powerful method for modelling polarization in opinion dynamics, but that the tools (objects to study such as the stationary distribution, or time to absorption for example) appropriate for analysing such models crucially depend on their properties (such as ergodicity, or transience). These properties are determined by the details of the learning process and may be difficult to identify based solely on simulations.

Suggested Citation

  • Benedikt V Meylahn & Janusz M Meylahn, 2024. "How social reinforcement learning can lead to metastable polarisation and the voter model," PLOS ONE, Public Library of Science, vol. 19(12), pages 1-23, December.
  • Handle: RePEc:plo:pone00:0313951
    DOI: 10.1371/journal.pone.0313951
    as

    Download full text from publisher

    File URL: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0313951
    Download Restriction: no

    File URL: https://journals.plos.org/plosone/article/file?id=10.1371/journal.pone.0313951&type=printable
    Download Restriction: no

    File URL: https://libkey.io/10.1371/journal.pone.0313951?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Tinggui Chen & Qianqian Li & Peihua Fu & Jianjun Yang & Chonghuan Xu & Guodong Cong & Gongfa Li, 2020. "Public Opinion Polarization by Individual Revenue from the Social Preference Theory," IJERPH, MDPI, vol. 17(3), pages 1-29, February.
    2. Borgers, Tilman & Sarin, Rajiv, 1997. "Learning Through Reinforcement and Replicator Dynamics," Journal of Economic Theory, Elsevier, vol. 77(1), pages 1-14, November.
    3. Sven Banischa & Ricardo Lima & Tanya Araújo, 2012. "Agent based models and opinion dynamics as markov chains," Working Papers Department of Economics 2012/10, ISEG - Lisbon School of Economics and Management, Department of Economics, Universidade de Lisboa.
    4. Adri'an Carro & Ra'ul Toral & Maxi San Miguel, 2016. "The noisy voter model on complex networks," Papers 1602.06935, arXiv.org, revised Apr 2016.
    5. Guillaume Deffuant & David Neau & Frederic Amblard & Gérard Weisbuch, 2000. "Mixing beliefs among interacting agents," Advances in Complex Systems (ACS), World Scientific Publishing Co. Pte. Ltd., vol. 3(01n04), pages 87-98.
    6. J. M. Meylahn & L. Janssen & Hassan Zargarzadeh, 2022. "Limiting Dynamics for Q-Learning with Memory One in Symmetric Two-Player, Two-Action Games," Complexity, Hindawi, vol. 2022, pages 1-20, November.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Agnieszka Kowalska-Styczeń & Krzysztof Malarz, 2020. "Noise induced unanimity and disorder in opinion formation," PLOS ONE, Public Library of Science, vol. 15(7), pages 1-22, July.
    2. Lücke, Marvin & Heitzig, Jobst & Koltai, Péter & Molkenthin, Nora & Winkelmann, Stefanie, 2023. "Large population limits of Markov processes on random networks," Stochastic Processes and their Applications, Elsevier, vol. 166(C).
    3. Wu, Dong & Zou, Fan, 2025. "Dominant design selected by users: Dynamic interaction and convergence of users," Technovation, Elsevier, vol. 140(C).
    4. Lee, Woosub & Yang, Seong-Gyu & Kim, Beom Jun, 2022. "The effect of media on opinion formation," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 595(C).
    5. Ponti, Giovanni, 2000. "Cycles of Learning in the Centipede Game," Games and Economic Behavior, Elsevier, vol. 30(1), pages 115-141, January.
    6. Philippe Jehiel & Aviman Satpathy, 2024. "Learning to be Indifferent in Complex Decisions: A Coarse Payoff-Assessment Model," Papers 2412.09321, arXiv.org, revised Dec 2024.
    7. Antonio Cabrales & Rosemarie Nagel & Roc Armenter, 2007. "Equilibrium selection through incomplete information in coordination games: an experimental study," Experimental Economics, Springer;Economic Science Association, vol. 10(3), pages 221-234, September.
    8. Sven Banisch & Eckehard Olbrich, 2021. "An Argument Communication Model of Polarization and Ideological Alignment," Journal of Artificial Societies and Social Simulation, Journal of Artificial Societies and Social Simulation, vol. 24(1), pages 1-1.
    9. Dehai Liu & Hongyi Li & Weiguo Wang & Chuang Zhou, 2015. "Scenario forecast model of long term trends in rural labor transfer based on evolutionary games," Journal of Evolutionary Economics, Springer, vol. 25(3), pages 649-670, July.
    10. Ianni, A., 2002. "Reinforcement learning and the power law of practice: some analytical results," Discussion Paper Series In Economics And Econometrics 203, Economics Division, School of Social Sciences, University of Southampton.
    11. Shang, Lihui & Zhao, Mingming & Ai, Jun & Su, Zhan, 2021. "Opinion evolution in the Sznajd model on interdependent chains," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 565(C).
    12. Hopkins, Ed, 1999. "Learning, Matching, and Aggregation," Games and Economic Behavior, Elsevier, vol. 26(1), pages 79-110, January.
    13. Tsakas, Elias & Voorneveld, Mark, 2009. "The target projection dynamic," Games and Economic Behavior, Elsevier, vol. 67(2), pages 708-719, November.
    14. Floriana Gargiulo & José J Ramasco, 2012. "Influence of Opinion Dynamics on the Evolution of Games," PLOS ONE, Public Library of Science, vol. 7(11), pages 1-7, November.
    15. Lu, Xi & Mo, Hongming & Deng, Yong, 2015. "An evidential opinion dynamics model based on heterogeneous social influential power," Chaos, Solitons & Fractals, Elsevier, vol. 73(C), pages 98-107.
    16. Antonio Morales, 2005. "On the Role of the Group Composition for Achieving Optimality," Annals of Operations Research, Springer, vol. 137(1), pages 387-397, July.
    17. DeJong, D.V. & Blume, A. & Neumann, G., 1998. "Learning in Sender-Receiver Games," Other publications TiSEM 4a8b4f46-f30b-4ad2-bb0c-1, Tilburg University, School of Economics and Management.
    18. Gaunersdorfer, A. & Hommes, C.H. & Wagener, F.O.O., 2000. "Bifurcation Routes to Volatility Clustering," CeNDEF Working Papers 00-04, Universiteit van Amsterdam, Center for Nonlinear Dynamics in Economics and Finance.
    19. Norman, Thomas W.L., 2009. "Rapid evolution under inertia," Games and Economic Behavior, Elsevier, vol. 66(2), pages 865-879, July.
    20. Fernando Lozano & Jaime Lozano & Mario García, 2007. "An artificial economy based on reinforcement learning and agent based modeling," Documentos de Trabajo 3907, Universidad del Rosario.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:plo:pone00:0313951. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: plosone (email available below). General contact details of provider: https://journals.plos.org/plosone/ .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.