IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v11y2023i10p2379-d1151476.html
   My bibliography  Save this article

A Multi-Agent Adaptive Co-Evolution Method in Dynamic Environments

Author

Listed:
  • Yan Li

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

  • Huazhi Zhang

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

  • Weiming Xu

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

  • Jianan Wang

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

  • Jialu Wang

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

  • Suyu Wang

    (School of Mechanical Electronic & Information Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China)

Abstract

It is challenging to ensure satisfying co-evolution efficiency for the multi-agents in dynamic environments since during Actor-Critic training there is a high probability of falling into local optimality, failing to adapt to the suddenly changed environment quickly. To solve this problem, this paper proposes a multi-agent adaptive co-evolution method in dynamic environments (ACE-D) based on the classical multi-agent reinforcement learning method MADDPG, which effectively realizes self-adaptive new environments and co-evolution in dynamic environments. First, an experience screening policy is introduced based on the MADDPG method to reduce the negative influence of original environment experience on exploring new environments. Then, an adaptive weighting policy is applied to the policy network, which accordingly generates benchmarks for varying environments and assigns higher weights to those policies that are more beneficial for new environments exploration, so that to save time while promoting adaptability of the agents. Finally, different types of dynamic environments with complexity at different levels are built to verify the co-evolutionary effects of the two policies separately and the ACE-D method comprehensively. The experimental results demonstrate that, compared with a range of other methods, the ACE-D method has obvious advantages helping multi-agent adapt to dynamic environments and preventing them from falling into local optima, with more than 25% improvement in stable reward and more than 23% improvement in training efficiency. The ACE-D method is valuable and commendable to promote the co-evolutionary effect of multi-agent in dynamic environments.

Suggested Citation

  • Yan Li & Huazhi Zhang & Weiming Xu & Jianan Wang & Jialu Wang & Suyu Wang, 2023. "A Multi-Agent Adaptive Co-Evolution Method in Dynamic Environments," Mathematics, MDPI, vol. 11(10), pages 1-18, May.
  • Handle: RePEc:gam:jmathe:v:11:y:2023:i:10:p:2379-:d:1151476
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/11/10/2379/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/11/10/2379/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Tianhao Wu & Mingzhi Jiang & Lin Zhang, 2020. "Cooperative Multiagent Deep Deterministic Policy Gradient (CoMADDPG) for Intelligent Connected Transportation with Unsignalized Intersection," Mathematical Problems in Engineering, Hindawi, vol. 2020, pages 1-12, July.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.

      Corrections

      All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:11:y:2023:i:10:p:2379-:d:1151476. See general information about how to correct material in RePEc.

      If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

      If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

      If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

      For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

      Please note that corrections may take a couple of weeks to filter through the various RePEc services.

      IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.