IDEAS home Printed from https://ideas.repec.org/a/eee/renene/v200y2022icp433-448.html

Physical-model-free intelligent energy management for a grid-connected hybrid wind-microturbine-PV-EV energy system via deep reinforcement learning approach

Author

Listed:
  • Zhang, Bin
  • Hu, Weihao
  • Xu, Xiao
  • Li, Tao
  • Zhang, Zhenyuan
  • Chen, Zhe

Abstract

Renewable -based microgrid (MG) is recognized as an eco-friendly solution in the development of renewable energy (RE). Moreover, the MG energy management with high RE penetration faces complicate uncertainties due to the inaccuracy of predictions. Besides, the growing participation of electric vehicles (EVs) makes the traditional model-based methods even more infeasible. Considering uncertainties associated with RE, EVs, and electricity price, a model-free deep reinforcement learning (DRL), namely twin delayed deep deterministic policy gradient (TD3), is employed to develop an optimized control strategy to minimize operating costs and satisfy charging expectations. The energy management problem is first formulated as a Markov decision process. Then, TD3 solely relies on the limited observation to find the optimal continuous control strategy. The proposed method flexibly adjusts the operating and charging strategies of components according to RE output and electricity price. Its real-time optimized performance on three consecutive days along with electricity price is evaluated, indicating its practical potential for future application. Additionally, comparison results demonstrate that the proposed method reduces the total costs up to 15.27% and 4.24%, respectively, compared to traditional optimization and other DRL methods, which illustrates the superiority of the TD3 method on optimizing total costs of the considered MG.

Suggested Citation

  • Zhang, Bin & Hu, Weihao & Xu, Xiao & Li, Tao & Zhang, Zhenyuan & Chen, Zhe, 2022. "Physical-model-free intelligent energy management for a grid-connected hybrid wind-microturbine-PV-EV energy system via deep reinforcement learning approach," Renewable Energy, Elsevier, vol. 200(C), pages 433-448.
  • Handle: RePEc:eee:renene:v:200:y:2022:i:c:p:433-448
    DOI: 10.1016/j.renene.2022.09.125
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0960148122014847
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.renene.2022.09.125?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to

    for a different version of it.

    References listed on IDEAS

    as
    1. Harrison Fell & Alex Gilbert & Jesse D. Jenkins & Matto Mildenberger, 2022. "Nuclear power and renewable energy are both associated with national decarbonization," Nature Energy, Nature, vol. 7(1), pages 25-29, January.
    2. Ceusters, Glenn & Rodríguez, Román Cantú & García, Alberte Bouso & Franke, Rüdiger & Deconinck, Geert & Helsen, Lieve & Nowé, Ann & Messagie, Maarten & Camargo, Luis Ramirez, 2021. "Model-predictive control and reinforcement learning in multi-energy system case studies," Applied Energy, Elsevier, vol. 303(C).
    3. Emrani, Anisa & Berrada, Asmae & Bakhouya, Mohamed, 2022. "Optimal sizing and deployment of gravity energy storage system in hybrid PV-Wind power plant," Renewable Energy, Elsevier, vol. 183(C), pages 12-27.
    4. Chen, Changming & Wu, Xueyan & Li, Yan & Zhu, Xiaojun & Li, Zesen & Ma, Jien & Qiu, Weiqiang & Liu, Chang & Lin, Zhenzhi & Yang, Li & Wang, Qin & Ding, Yi, 2021. "Distributionally robust day-ahead scheduling of park-level integrated energy system considering generalized energy storages," Applied Energy, Elsevier, vol. 302(C).
    5. Zhang, Guozhou & Hu, Weihao & Cao, Di & Huang, Qi & Chen, Zhe & Blaabjerg, Frede, 2021. "A novel deep reinforcement learning enabled sparsity promoting adaptive control method to improve the stability of power systems with wind energy penetration," Renewable Energy, Elsevier, vol. 178(C), pages 363-376.
    6. Hua, Haochen & Qin, Yuchao & Hao, Chuantong & Cao, Junwei, 2019. "Optimal energy management strategies for energy Internet via deep reinforcement learning approach," Applied Energy, Elsevier, vol. 239(C), pages 598-609.
    7. Antonopoulos, Ioannis & Robu, Valentin & Couraud, Benoit & Kirli, Desen & Norbu, Sonam & Kiprakis, Aristides & Flynn, David & Elizondo-Gonzalez, Sergio & Wattam, Steve, 2020. "Artificial intelligence and machine learning approaches to energy demand-side response: A systematic review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 130(C).
    8. Hao, Ran & Lu, Tianguang & Ai, Qian & Wang, Zhe & Wang, Xiaolong, 2020. "Distributed online learning and dynamic robust standby dispatch for networked microgrids," Applied Energy, Elsevier, vol. 274(C).
    9. Hossain, Md Alamgir & Pota, Hemanshu Roy & Squartini, Stefano & Abdou, Ahmed Fathi, 2019. "Modified PSO algorithm for real-time energy management in grid-connected microgrids," Renewable Energy, Elsevier, vol. 136(C), pages 746-757.
    10. Pinciroli, Luca & Baraldi, Piero & Ballabio, Guido & Compare, Michele & Zio, Enrico, 2022. "Optimization of the Operation and Maintenance of renewable energy systems by Deep Reinforcement Learning," Renewable Energy, Elsevier, vol. 183(C), pages 752-763.
    11. Wu, Jingda & He, Hongwen & Peng, Jiankun & Li, Yuecheng & Li, Zhanjiang, 2018. "Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus," Applied Energy, Elsevier, vol. 222(C), pages 799-811.
    12. Chau, Ka Yin & Moslehpour, Massoud & Tu, Yu-Te & Tai, Nguyen Tan & Tien, Nguyen Hoang & Huy, Pham Quang, 2022. "Exploring the impact of green energy and consumption on the sustainability of natural resources: Empirical evidence from G7 countries," Renewable Energy, Elsevier, vol. 196(C), pages 1241-1249.
    13. Balderrama, Sergio & Lombardi, Francesco & Riva, Fabio & Canedo, Walter & Colombo, Emanuela & Quoilin, Sylvain, 2019. "A two-stage linear programming optimization framework for isolated hybrid microgrids in a rural context: The case study of the “El Espino” community," Energy, Elsevier, vol. 188(C).
    14. Velik, Rosemarie & Nicolay, Pascal, 2014. "Grid-price-dependent energy management in microgrids using a modified simulated annealing triple-optimizer," Applied Energy, Elsevier, vol. 130(C), pages 384-395.
    15. Li, Xuan & Zhang, Wei, 2022. "Physics-informed deep learning model in wind turbine response prediction," Renewable Energy, Elsevier, vol. 185(C), pages 932-944.
    16. Gomes, I.L.R. & Melicio, R. & Mendes, V.M.F., 2021. "A novel microgrid support management system based on stochastic mixed-integer linear programming," Energy, Elsevier, vol. 223(C).
    17. Yeh, Wei-Chang & He, Min-Fan & Huang, Chia-Ling & Tan, Shi-Yi & Zhang, Xianyong & Huang, Yaohong & Li, Li, 2020. "New genetic algorithm for economic dispatch of stand-alone three-modular microgrid in DongAo Island," Applied Energy, Elsevier, vol. 263(C).
    18. David Silver & Aja Huang & Chris J. Maddison & Arthur Guez & Laurent Sifre & George van den Driessche & Julian Schrittwieser & Ioannis Antonoglou & Veda Panneershelvam & Marc Lanctot & Sander Dieleman, 2016. "Mastering the game of Go with deep neural networks and tree search," Nature, Nature, vol. 529(7587), pages 484-489, January.
    19. Prevedello, Giulio & Werth, Annette, 2021. "The benefits of sharing in off-grid microgrids: A case study in the Philippines," Applied Energy, Elsevier, vol. 303(C).
    20. Zhang, Xizheng & Wang, Zeyu & Lu, Zhangyu, 2022. "Multi-objective load dispatch for microgrid with electric vehicles using modified gravitational search and particle swarm optimization algorithm," Applied Energy, Elsevier, vol. 306(PA).
    21. Volodymyr Mnih & Koray Kavukcuoglu & David Silver & Andrei A. Rusu & Joel Veness & Marc G. Bellemare & Alex Graves & Martin Riedmiller & Andreas K. Fidjeland & Georg Ostrovski & Stig Petersen & Charle, 2015. "Human-level control through deep reinforcement learning," Nature, Nature, vol. 518(7540), pages 529-533, February.
    22. Torkan, Ramin & Ilinca, Adrian & Ghorbanzadeh, Milad, 2022. "A genetic algorithm optimization approach for smart energy management of microgrids," Renewable Energy, Elsevier, vol. 197(C), pages 852-863.
    23. Cao, Di & Zhao, Junbo & Hu, Weihao & Ding, Fei & Yu, Nanpeng & Huang, Qi & Chen, Zhe, 2022. "Model-free voltage control of active distribution system with PVs using surrogate model-based deep reinforcement learning," Applied Energy, Elsevier, vol. 306(PA).
    24. Manzano, J.M. & Salvador, J.R. & Romaine, J.B. & Alvarado-Barrios, L., 2022. "Economic predictive control for isolated microgrids based on real world demand/renewable energy data and forecast errors," Renewable Energy, Elsevier, vol. 194(C), pages 647-658.
    25. Tomin, Nikita & Shakirov, Vladislav & Kozlov, Aleksander & Sidorov, Denis & Kurbatsky, Victor & Rehtanz, Christian & Lora, Electo E.S., 2022. "Design and optimal energy management of community microgrids with flexible renewable energy sources," Renewable Energy, Elsevier, vol. 183(C), pages 903-921.
    26. Xu, Xiao & Hu, Weihao & Liu, Wen & Du, Yuefang & Huang, Rui & Huang, Qi & Chen, Zhe, 2021. "Look-ahead risk-constrained scheduling for an energy hub integrated with renewable energy," Applied Energy, Elsevier, vol. 297(C).
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Xiong, Kang & Hu, Weihao & Cao, Di & Li, Sichen & Zhang, Guozhou & Liu, Wen & Huang, Qi & Chen, Zhe, 2023. "Coordinated energy management strategy for multi-energy hub with thermo-electrochemical effect based power-to-ammonia: A multi-agent deep reinforcement learning enabled approach," Renewable Energy, Elsevier, vol. 214(C), pages 216-232.
    2. Kim, Hyung Joon & Lee, Jae Yong & Tak, Hyunwoo & Kim, Dongwoo, 2025. "Deep reinforcement learning-based residential building energy management incorporating power-to-heat technology for building electrification," Energy, Elsevier, vol. 317(C).
    3. Agha Kassab, Fadi & Celik, Berk & Locment, Fabrice & Sechilariu, Manuela & Liaquat, Sheroze & Hansen, Timothy M., 2025. "Microgrid sizing with EV flexibility: Cascaded MILP and embedded APSO-MILP approaches," Applied Energy, Elsevier, vol. 396(C).
    4. Hajialigol, Parisa & Nweye, Kingsley & Aghaei, Mohammadreza & Najafi, Behzad & Moazami, Amin & Nagy, Zoltan, 2025. "Enhancing self-consumption ratio in a smart microgrid by applying a reinforcement learning-based energy management system," Energy, Elsevier, vol. 335(C).
    5. Zhou, Yuekuan & Liu, Xiaohua & Zhao, Qianchuan, 2024. "A stochastic vehicle schedule model for demand response and grid flexibility in a renewable-building-e-transportation-microgrid," Renewable Energy, Elsevier, vol. 221(C).
    6. Gerald Jones & Xueping Li & Yulin Sun, 2024. "Robust Energy Management Policies for Solar Microgrids via Reinforcement Learning," Energies, MDPI, vol. 17(12), pages 1-22, June.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Omar Al-Ani & Sanjoy Das, 2022. "Reinforcement Learning: Theory and Applications in HEMS," Energies, MDPI, vol. 15(17), pages 1-37, September.
    2. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    3. Du, Yan & Zandi, Helia & Kotevska, Olivera & Kurte, Kuldeep & Munk, Jeffery & Amasyali, Kadir & Mckee, Evan & Li, Fangxing, 2021. "Intelligent multi-zone residential HVAC control strategy based on deep reinforcement learning," Applied Energy, Elsevier, vol. 281(C).
    4. Qi, Chunyang & Zhu, Yiwen & Song, Chuanxue & Yan, Guangfu & Xiao, Feng & Da wang, & Zhang, Xu & Cao, Jingwei & Song, Shixin, 2022. "Hierarchical reinforcement learning based energy management strategy for hybrid electric vehicle," Energy, Elsevier, vol. 238(PA).
    5. Pinciroli, Luca & Baraldi, Piero & Compare, Michele & Zio, Enrico, 2023. "Optimal operation and maintenance of energy storage systems in grid-connected microgrids by deep reinforcement learning," Applied Energy, Elsevier, vol. 352(C).
    6. Shi, Tao & Xu, Chang & Dong, Wenhao & Zhou, Hangyu & Bokhari, Awais & Klemeš, Jiří Jaromír & Han, Ning, 2023. "Research on energy management of hydrogen electric coupling system based on deep reinforcement learning," Energy, Elsevier, vol. 282(C).
    7. Zhang, Qin & Liu, Yu & Xiang, Yisha & Xiahou, Tangfan, 2024. "Reinforcement learning in reliability and maintenance optimization: A tutorial," Reliability Engineering and System Safety, Elsevier, vol. 251(C).
    8. Md Shafiullah & Akib Mostabe Refat & Md Ershadul Haque & Dewan Mabrur Hasan Chowdhury & Md Sanower Hossain & Abdullah G. Alharbi & Md Shafiul Alam & Amjad Ali & Shorab Hossain, 2022. "Review of Recent Developments in Microgrid Energy Management Strategies," Sustainability, MDPI, vol. 14(22), pages 1-30, November.
    9. Bo Hu & Jiaxi Li & Shuang Li & Jie Yang, 2019. "A Hybrid End-to-End Control Strategy Combining Dueling Deep Q-network and PID for Transient Boost Control of a Diesel Engine with Variable Geometry Turbocharger and Cooled EGR," Energies, MDPI, vol. 12(19), pages 1-15, September.
    10. Qingyan Li & Tao Lin & Qianyi Yu & Hui Du & Jun Li & Xiyue Fu, 2023. "Review of Deep Reinforcement Learning and Its Application in Modern Renewable Power System Control," Energies, MDPI, vol. 16(10), pages 1-23, May.
    11. Hao, Peng & Wei, Zhensong & Bai, Zhengwei & Barth, Matthew J., 2020. "Developing an Adaptive Strategy for Connected Eco-Driving Under Uncertain Traffic and Signal Conditions," Institute of Transportation Studies, Working Paper Series qt2fv5063b, Institute of Transportation Studies, UC Davis.
    12. Tambet Matiisen & Aqeel Labash & Daniel Majoral & Jaan Aru & Raul Vicente, 2022. "Do Deep Reinforcement Learning Agents Model Intentions?," Stats, MDPI, vol. 6(1), pages 1-17, December.
    13. Benjamin Heinbach & Peter Burggräf & Johannes Wagner, 2024. "gym-flp: A Python Package for Training Reinforcement Learning Algorithms on Facility Layout Problems," SN Operations Research Forum, Springer, vol. 5(1), pages 1-26, March.
    14. Sun, Alexander Y., 2020. "Optimal carbon storage reservoir management through deep reinforcement learning," Applied Energy, Elsevier, vol. 278(C).
    15. Elsisi, Mahmoud & Amer, Mohammed & Dababat, Alya’ & Su, Chun-Lien, 2023. "A comprehensive review of machine learning and IoT solutions for demand side energy management, conservation, and resilient operation," Energy, Elsevier, vol. 281(C).
    16. Yassine Chemingui & Adel Gastli & Omar Ellabban, 2020. "Reinforcement Learning-Based School Energy Management System," Energies, MDPI, vol. 13(23), pages 1-21, December.
    17. Biemann, Marco & Scheller, Fabian & Liu, Xiufeng & Huang, Lizhen, 2021. "Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control," Applied Energy, Elsevier, vol. 298(C).
    18. Yuhong Wang & Lei Chen & Hong Zhou & Xu Zhou & Zongsheng Zheng & Qi Zeng & Li Jiang & Liang Lu, 2021. "Flexible Transmission Network Expansion Planning Based on DQN Algorithm," Energies, MDPI, vol. 14(7), pages 1-21, April.
    19. Jiacheng Zhang & Haolan Zhang, 2025. "Towards Human-like Artificial Intelligence: A Review of Anthropomorphic Computing in AI and Future Trends," Mathematics, MDPI, vol. 13(13), pages 1-49, June.
    20. Brini, Alessio & Tantari, Daniele, 2023. "Deep reinforcement trading with predictable returns," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 622(C).

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:renene:v:200:y:2022:i:c:p:433-448. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/renewable-energy .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.