IDEAS home Printed from https://ideas.repec.org/a/gam/jeners/v18y2025i19p5172-d1760704.html
   My bibliography  Save this article

Research on Dynamic Energy Management Optimization of Park Integrated Energy System Based on Deep Reinforcement Learning

Author

Listed:
  • Xinjian Jiang

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

  • Lei Zhang

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

  • Fuwang Li

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

  • Zhiru Li

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

  • Zhijian Ling

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

  • Zhenghui Zhao

    (Shanxi Energy Internet Research Institute, Taiyuan 030000, China)

Abstract

Under the background of energy transition, the Integrated Energy System (IES) of the park has become a key carrier for enhancing the consumption capacity of renewable energy due to its multi-energy complementary characteristics. However, the high proportion of wind and solar resource access and the fluctuation of diverse loads have led to the system facing dual uncertainty challenges, and traditional optimization methods are difficult to adapt to the dynamic and complex dispatching requirements. To this end, this paper proposes a new dynamic energy management method based on Deep Reinforcement Learning (DRL) and constructs an IES hybrid integer nonlinear programming model including wind power, photovoltaic, combined heat and power generation, and storage of electric heat energy, with the goal of minimizing the operating cost of the system. By expressing the dispatching process as a Markov decision process, a state space covering wind and solar output, multiple loads and energy storage states is defined, a continuous action space for unit output and energy storage control is constructed, and a reward function integrating economic cost and the penalty for renewable energy consumption is designed. The Deep Deterministic Policy Gradient (DDPG) and Deep Q-Network (DQN) algorithms were adopted to achieve policy optimization. This study is based on simulation rather than experimental validation, which aligns with the exploratory scope of this research. The simulation results show that the DDPG algorithm achieves an average weekly operating cost of 532,424 yuan in the continuous action space scheduling, which is 8.6% lower than that of the DQN algorithm, and the standard deviation of the cost is reduced by 19.5%, indicating better robustness. Under the fluctuation of 10% to 30% on the source-load side, the DQN algorithm still maintains a cost fluctuation of less than 4.5%, highlighting the strong adaptability of DRL to uncertain environments. Therefore, this method has significant theoretical and practical value for promoting the intelligent transformation of the energy system.

Suggested Citation

  • Xinjian Jiang & Lei Zhang & Fuwang Li & Zhiru Li & Zhijian Ling & Zhenghui Zhao, 2025. "Research on Dynamic Energy Management Optimization of Park Integrated Energy System Based on Deep Reinforcement Learning," Energies, MDPI, vol. 18(19), pages 1-28, September.
  • Handle: RePEc:gam:jeners:v:18:y:2025:i:19:p:5172-:d:1760704
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1996-1073/18/19/5172/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1996-1073/18/19/5172/
    Download Restriction: no
    ---><---

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jeners:v:18:y:2025:i:19:p:5172-:d:1760704. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.