Author
Listed:
- Wei Niu
(State Grid Liaoning Electric Power Co., Ltd. Dalian Power Supply Company, Dalian 234643, China)
- Jifeng Li
(State Grid Liaoning Electric Power Co., Ltd. Dalian Power Supply Company, Dalian 234643, China)
- Zongle Ma
(State Grid Liaoning Electric Power Co., Ltd. Dalian Power Supply Company, Dalian 234643, China)
- Wenliang Yin
(School of Electrical and Electronic Engineering, Shandong University of Technology, Zhangdian District, Zibo 255000, China)
- Liang Feng
(School of Electrical and Electronic Engineering, Shandong University of Technology, Zhangdian District, Zibo 255000, China)
Abstract
This paper presents a deep reinforcement learning-based demand response (DR) optimization framework for active distribution networks under uncertainty and user heterogeneity. The proposed model utilizes a Double Deep Q-Network (Double DQN) to learn adaptive, multi-period DR strategies across residential, commercial, and electric vehicle (EV) participants in a 24 h rolling horizon. By incorporating a structured state representation—including forecasted load, photovoltaic (PV) output, dynamic pricing, historical DR actions, and voltage states—the agent autonomously learns control policies that minimize total operational costs while maintaining grid feasibility and voltage stability. The physical system is modeled via detailed constraints, including power flow balance, voltage magnitude bounds, PV curtailment caps, deferrable load recovery windows, and user-specific availability envelopes. A case study based on a modified IEEE 33-bus distribution network with embedded PV and DR nodes demonstrates the framework’s effectiveness. Simulation results show that the proposed method achieves significant cost savings (up to 35% over baseline), enhances PV absorption, reduces load variance by 42%, and maintains voltage profiles within safe operational thresholds. Training curves confirm smooth Q-value convergence and stable policy performance, while spatiotemporal visualizations reveal interpretable DR behavior aligned with both economic and physical system constraints. This work contributes a scalable, model-free approach for intelligent DR coordination in smart grids, integrating learning-based control with physical grid realism. The modular design allows for future extension to multi-agent systems, storage coordination, and market-integrated DR scheduling. The results position Double DQN as a promising architecture for operational decision-making in AI-enabled distribution networks.
Suggested Citation
Wei Niu & Jifeng Li & Zongle Ma & Wenliang Yin & Liang Feng, 2025.
"Multi-Time-Scale Demand Response Optimization in Active Distribution Networks Using Double Deep Q-Networks,"
Energies, MDPI, vol. 18(18), pages 1-26, September.
Handle:
RePEc:gam:jeners:v:18:y:2025:i:18:p:4795-:d:1745480
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jeners:v:18:y:2025:i:18:p:4795-:d:1745480. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.