IDEAS home Printed from https://ideas.repec.org/a/eee/renene/v200y2022icp433-448.html
   My bibliography  Save this article

Physical-model-free intelligent energy management for a grid-connected hybrid wind-microturbine-PV-EV energy system via deep reinforcement learning approach

Author

Listed:
  • Zhang, Bin
  • Hu, Weihao
  • Xu, Xiao
  • Li, Tao
  • Zhang, Zhenyuan
  • Chen, Zhe

Abstract

Renewable -based microgrid (MG) is recognized as an eco-friendly solution in the development of renewable energy (RE). Moreover, the MG energy management with high RE penetration faces complicate uncertainties due to the inaccuracy of predictions. Besides, the growing participation of electric vehicles (EVs) makes the traditional model-based methods even more infeasible. Considering uncertainties associated with RE, EVs, and electricity price, a model-free deep reinforcement learning (DRL), namely twin delayed deep deterministic policy gradient (TD3), is employed to develop an optimized control strategy to minimize operating costs and satisfy charging expectations. The energy management problem is first formulated as a Markov decision process. Then, TD3 solely relies on the limited observation to find the optimal continuous control strategy. The proposed method flexibly adjusts the operating and charging strategies of components according to RE output and electricity price. Its real-time optimized performance on three consecutive days along with electricity price is evaluated, indicating its practical potential for future application. Additionally, comparison results demonstrate that the proposed method reduces the total costs up to 15.27% and 4.24%, respectively, compared to traditional optimization and other DRL methods, which illustrates the superiority of the TD3 method on optimizing total costs of the considered MG.

Suggested Citation

  • Zhang, Bin & Hu, Weihao & Xu, Xiao & Li, Tao & Zhang, Zhenyuan & Chen, Zhe, 2022. "Physical-model-free intelligent energy management for a grid-connected hybrid wind-microturbine-PV-EV energy system via deep reinforcement learning approach," Renewable Energy, Elsevier, vol. 200(C), pages 433-448.
  • Handle: RePEc:eee:renene:v:200:y:2022:i:c:p:433-448
    DOI: 10.1016/j.renene.2022.09.125
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0960148122014847
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.renene.2022.09.125?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Emrani, Anisa & Berrada, Asmae & Bakhouya, Mohamed, 2022. "Optimal sizing and deployment of gravity energy storage system in hybrid PV-Wind power plant," Renewable Energy, Elsevier, vol. 183(C), pages 12-27.
    2. Hao, Ran & Lu, Tianguang & Ai, Qian & Wang, Zhe & Wang, Xiaolong, 2020. "Distributed online learning and dynamic robust standby dispatch for networked microgrids," Applied Energy, Elsevier, vol. 274(C).
    3. Hossain, Md Alamgir & Pota, Hemanshu Roy & Squartini, Stefano & Abdou, Ahmed Fathi, 2019. "Modified PSO algorithm for real-time energy management in grid-connected microgrids," Renewable Energy, Elsevier, vol. 136(C), pages 746-757.
    4. Pinciroli, Luca & Baraldi, Piero & Ballabio, Guido & Compare, Michele & Zio, Enrico, 2022. "Optimization of the Operation and Maintenance of renewable energy systems by Deep Reinforcement Learning," Renewable Energy, Elsevier, vol. 183(C), pages 752-763.
    5. Prevedello, Giulio & Werth, Annette, 2021. "The benefits of sharing in off-grid microgrids: A case study in the Philippines," Applied Energy, Elsevier, vol. 303(C).
    6. Torkan, Ramin & Ilinca, Adrian & Ghorbanzadeh, Milad, 2022. "A genetic algorithm optimization approach for smart energy management of microgrids," Renewable Energy, Elsevier, vol. 197(C), pages 852-863.
    7. Antonopoulos, Ioannis & Robu, Valentin & Couraud, Benoit & Kirli, Desen & Norbu, Sonam & Kiprakis, Aristides & Flynn, David & Elizondo-Gonzalez, Sergio & Wattam, Steve, 2020. "Artificial intelligence and machine learning approaches to energy demand-side response: A systematic review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 130(C).
    8. Balderrama, Sergio & Lombardi, Francesco & Riva, Fabio & Canedo, Walter & Colombo, Emanuela & Quoilin, Sylvain, 2019. "A two-stage linear programming optimization framework for isolated hybrid microgrids in a rural context: The case study of the “El Espino” community," Energy, Elsevier, vol. 188(C).
    9. Yeh, Wei-Chang & He, Min-Fan & Huang, Chia-Ling & Tan, Shi-Yi & Zhang, Xianyong & Huang, Yaohong & Li, Li, 2020. "New genetic algorithm for economic dispatch of stand-alone three-modular microgrid in DongAo Island," Applied Energy, Elsevier, vol. 263(C).
    10. Ceusters, Glenn & Rodríguez, Román Cantú & García, Alberte Bouso & Franke, Rüdiger & Deconinck, Geert & Helsen, Lieve & Nowé, Ann & Messagie, Maarten & Camargo, Luis Ramirez, 2021. "Model-predictive control and reinforcement learning in multi-energy system case studies," Applied Energy, Elsevier, vol. 303(C).
    11. Hua, Haochen & Qin, Yuchao & Hao, Chuantong & Cao, Junwei, 2019. "Optimal energy management strategies for energy Internet via deep reinforcement learning approach," Applied Energy, Elsevier, vol. 239(C), pages 598-609.
    12. Wu, Jingda & He, Hongwen & Peng, Jiankun & Li, Yuecheng & Li, Zhanjiang, 2018. "Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus," Applied Energy, Elsevier, vol. 222(C), pages 799-811.
    13. Li, Xuan & Zhang, Wei, 2022. "Physics-informed deep learning model in wind turbine response prediction," Renewable Energy, Elsevier, vol. 185(C), pages 932-944.
    14. Zhang, Xizheng & Wang, Zeyu & Lu, Zhangyu, 2022. "Multi-objective load dispatch for microgrid with electric vehicles using modified gravitational search and particle swarm optimization algorithm," Applied Energy, Elsevier, vol. 306(PA).
    15. Volodymyr Mnih & Koray Kavukcuoglu & David Silver & Andrei A. Rusu & Joel Veness & Marc G. Bellemare & Alex Graves & Martin Riedmiller & Andreas K. Fidjeland & Georg Ostrovski & Stig Petersen & Charle, 2015. "Human-level control through deep reinforcement learning," Nature, Nature, vol. 518(7540), pages 529-533, February.
    16. Cao, Di & Zhao, Junbo & Hu, Weihao & Ding, Fei & Yu, Nanpeng & Huang, Qi & Chen, Zhe, 2022. "Model-free voltage control of active distribution system with PVs using surrogate model-based deep reinforcement learning," Applied Energy, Elsevier, vol. 306(PA).
    17. Xu, Xiao & Hu, Weihao & Liu, Wen & Du, Yuefang & Huang, Rui & Huang, Qi & Chen, Zhe, 2021. "Look-ahead risk-constrained scheduling for an energy hub integrated with renewable energy," Applied Energy, Elsevier, vol. 297(C).
    18. Harrison Fell & Alex Gilbert & Jesse D. Jenkins & Matto Mildenberger, 2022. "Nuclear power and renewable energy are both associated with national decarbonization," Nature Energy, Nature, vol. 7(1), pages 25-29, January.
    19. Chen, Changming & Wu, Xueyan & Li, Yan & Zhu, Xiaojun & Li, Zesen & Ma, Jien & Qiu, Weiqiang & Liu, Chang & Lin, Zhenzhi & Yang, Li & Wang, Qin & Ding, Yi, 2021. "Distributionally robust day-ahead scheduling of park-level integrated energy system considering generalized energy storages," Applied Energy, Elsevier, vol. 302(C).
    20. Zhang, Guozhou & Hu, Weihao & Cao, Di & Huang, Qi & Chen, Zhe & Blaabjerg, Frede, 2021. "A novel deep reinforcement learning enabled sparsity promoting adaptive control method to improve the stability of power systems with wind energy penetration," Renewable Energy, Elsevier, vol. 178(C), pages 363-376.
    21. Chau, Ka Yin & Moslehpour, Massoud & Tu, Yu-Te & Tai, Nguyen Tan & Tien, Nguyen Hoang & Huy, Pham Quang, 2022. "Exploring the impact of green energy and consumption on the sustainability of natural resources: Empirical evidence from G7 countries," Renewable Energy, Elsevier, vol. 196(C), pages 1241-1249.
    22. Velik, Rosemarie & Nicolay, Pascal, 2014. "Grid-price-dependent energy management in microgrids using a modified simulated annealing triple-optimizer," Applied Energy, Elsevier, vol. 130(C), pages 384-395.
    23. Gomes, I.L.R. & Melicio, R. & Mendes, V.M.F., 2021. "A novel microgrid support management system based on stochastic mixed-integer linear programming," Energy, Elsevier, vol. 223(C).
    24. David Silver & Aja Huang & Chris J. Maddison & Arthur Guez & Laurent Sifre & George van den Driessche & Julian Schrittwieser & Ioannis Antonoglou & Veda Panneershelvam & Marc Lanctot & Sander Dieleman, 2016. "Mastering the game of Go with deep neural networks and tree search," Nature, Nature, vol. 529(7587), pages 484-489, January.
    25. Manzano, J.M. & Salvador, J.R. & Romaine, J.B. & Alvarado-Barrios, L., 2022. "Economic predictive control for isolated microgrids based on real world demand/renewable energy data and forecast errors," Renewable Energy, Elsevier, vol. 194(C), pages 647-658.
    26. Tomin, Nikita & Shakirov, Vladislav & Kozlov, Aleksander & Sidorov, Denis & Kurbatsky, Victor & Rehtanz, Christian & Lora, Electo E.S., 2022. "Design and optimal energy management of community microgrids with flexible renewable energy sources," Renewable Energy, Elsevier, vol. 183(C), pages 903-921.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Xiong, Kang & Hu, Weihao & Cao, Di & Li, Sichen & Zhang, Guozhou & Liu, Wen & Huang, Qi & Chen, Zhe, 2023. "Coordinated energy management strategy for multi-energy hub with thermo-electrochemical effect based power-to-ammonia: A multi-agent deep reinforcement learning enabled approach," Renewable Energy, Elsevier, vol. 214(C), pages 216-232.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Omar Al-Ani & Sanjoy Das, 2022. "Reinforcement Learning: Theory and Applications in HEMS," Energies, MDPI, vol. 15(17), pages 1-37, September.
    2. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    3. Qi, Chunyang & Zhu, Yiwen & Song, Chuanxue & Yan, Guangfu & Xiao, Feng & Da wang, & Zhang, Xu & Cao, Jingwei & Song, Shixin, 2022. "Hierarchical reinforcement learning based energy management strategy for hybrid electric vehicle," Energy, Elsevier, vol. 238(PA).
    4. Pinciroli, Luca & Baraldi, Piero & Compare, Michele & Zio, Enrico, 2023. "Optimal operation and maintenance of energy storage systems in grid-connected microgrids by deep reinforcement learning," Applied Energy, Elsevier, vol. 352(C).
    5. Du, Yan & Zandi, Helia & Kotevska, Olivera & Kurte, Kuldeep & Munk, Jeffery & Amasyali, Kadir & Mckee, Evan & Li, Fangxing, 2021. "Intelligent multi-zone residential HVAC control strategy based on deep reinforcement learning," Applied Energy, Elsevier, vol. 281(C).
    6. Md Shafiullah & Akib Mostabe Refat & Md Ershadul Haque & Dewan Mabrur Hasan Chowdhury & Md Sanower Hossain & Abdullah G. Alharbi & Md Shafiul Alam & Amjad Ali & Shorab Hossain, 2022. "Review of Recent Developments in Microgrid Energy Management Strategies," Sustainability, MDPI, vol. 14(22), pages 1-30, November.
    7. Shi, Tao & Xu, Chang & Dong, Wenhao & Zhou, Hangyu & Bokhari, Awais & Klemeš, Jiří Jaromír & Han, Ning, 2023. "Research on energy management of hydrogen electric coupling system based on deep reinforcement learning," Energy, Elsevier, vol. 282(C).
    8. Yuhong Wang & Lei Chen & Hong Zhou & Xu Zhou & Zongsheng Zheng & Qi Zeng & Li Jiang & Liang Lu, 2021. "Flexible Transmission Network Expansion Planning Based on DQN Algorithm," Energies, MDPI, vol. 14(7), pages 1-21, April.
    9. Gokhale, Gargya & Claessens, Bert & Develder, Chris, 2022. "Physics informed neural networks for control oriented thermal modeling of buildings," Applied Energy, Elsevier, vol. 314(C).
    10. Neha Soni & Enakshi Khular Sharma & Narotam Singh & Amita Kapoor, 2019. "Impact of Artificial Intelligence on Businesses: from Research, Innovation, Market Deployment to Future Shifts in Business Models," Papers 1905.02092, arXiv.org.
    11. Yin, Linfei & He, Xiaoyu, 2023. "Artificial emotional deep Q learning for real-time smart voltage control of cyber-physical social power systems," Energy, Elsevier, vol. 273(C).
    12. Taejong Joo & Hyunyoung Jun & Dongmin Shin, 2022. "Task Allocation in Human–Machine Manufacturing Systems Using Deep Reinforcement Learning," Sustainability, MDPI, vol. 14(4), pages 1-18, February.
    13. Zeyue Sun & Mohsen Eskandari & Chaoran Zheng & Ming Li, 2022. "Handling Computation Hardness and Time Complexity Issue of Battery Energy Storage Scheduling in Microgrids by Deep Reinforcement Learning," Energies, MDPI, vol. 16(1), pages 1-20, December.
    14. Oleh Lukianykhin & Tetiana Bogodorova, 2021. "Voltage Control-Based Ancillary Service Using Deep Reinforcement Learning," Energies, MDPI, vol. 14(8), pages 1-22, April.
    15. Daniel Egan & Qilun Zhu & Robert Prucka, 2023. "A Review of Reinforcement Learning-Based Powertrain Controllers: Effects of Agent Selection for Mixed-Continuity Control and Reward Formulation," Energies, MDPI, vol. 16(8), pages 1-31, April.
    16. Fathy, Ahmed, 2023. "Bald eagle search optimizer-based energy management strategy for microgrid with renewable sources and electric vehicles," Applied Energy, Elsevier, vol. 334(C).
    17. Chen, Jiaxin & Shu, Hong & Tang, Xiaolin & Liu, Teng & Wang, Weida, 2022. "Deep reinforcement learning-based multi-objective control of hybrid power system combined with road recognition under time-varying environment," Energy, Elsevier, vol. 239(PC).
    18. Amirhosein Mosavi & Yaser Faghan & Pedram Ghamisi & Puhong Duan & Sina Faizollahzadeh Ardabili & Ely Salwana & Shahab S. Band, 2020. "Comprehensive Review of Deep Reinforcement Learning Methods and Applications in Economics," Mathematics, MDPI, vol. 8(10), pages 1-42, September.
    19. Zhang, Yihao & Chai, Zhaojie & Lykotrafitis, George, 2021. "Deep reinforcement learning with a particle dynamics environment applied to emergency evacuation of a room with obstacles," Physica A: Statistical Mechanics and its Applications, Elsevier, vol. 571(C).
    20. Isaías Gomes & Rui Melicio & Victor M. F. Mendes, 2021. "Assessing the Value of Demand Response in Microgrids," Sustainability, MDPI, vol. 13(11), pages 1-16, May.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:renene:v:200:y:2022:i:c:p:433-448. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/renewable-energy .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.