IDEAS home Printed from https://ideas.repec.org/a/gam/jeners/v18y2025i17p4489-d1731167.html
   My bibliography  Save this article

Deep Reinforcement Learning Approaches the MILP Optimum of a Multi-Energy Optimization in Energy Communities

Author

Listed:
  • Vinzent Vetter

    (Illwerke vkw Endowed Professorship for Energy Efficiency, Energy Research Centre, Vorarlberg University of Applied Sciences, Hochschulstrasse 1, 6850 Dornbirn, Austria)

  • Philipp Wohlgenannt

    (Illwerke vkw Endowed Professorship for Energy Efficiency, Energy Research Centre, Vorarlberg University of Applied Sciences, Hochschulstrasse 1, 6850 Dornbirn, Austria
    Faculty of Engineering and Science, University of Agder, Jon Lilletuns vei 9, 4879 Grimstad, Norway)

  • Peter Kepplinger

    (Illwerke vkw Endowed Professorship for Energy Efficiency, Energy Research Centre, Vorarlberg University of Applied Sciences, Hochschulstrasse 1, 6850 Dornbirn, Austria)

  • Elias Eder

    (Illwerke vkw Endowed Professorship for Energy Efficiency, Energy Research Centre, Vorarlberg University of Applied Sciences, Hochschulstrasse 1, 6850 Dornbirn, Austria)

Abstract

As energy systems transition toward high shares of variable renewable generation, local energy communities (ECs) are increasingly relevant for enabling demand-side flexibility and self-sufficiency. This shift is particularly evident in the residential sector, where the deployment of photovoltaic (PV) systems is rapidly growing. While mixed-integer linear programming (MILP) remains the standard for operational optimization and demand response in such systems, its computational burden limits scalability and responsiveness under real-time or uncertain conditions. Reinforcement learning (RL), by contrast, offers a model-free, adaptive alternative. However, its application to real-world energy system operation remains limited. This study explores the application of a Deep Q-Network (DQN) to a real residential EC, which has received limited attention in prior work. The system comprises three single-family homes sharing a centralized heating system with a thermal energy storage (TES), a PV installation, and a grid connection. We compare the performance of MILP and RL controllers across economic and environmental metrics. Relative to a reference scenario without TES, MILP and RL reduce energy costs by 10.06% and 8.78%, respectively, and both approaches yield lower total energy consumption and CO 2 -equivalent emissions. Notably, the trained RL agent achieves a near-optimal outcome while requiring only 22% of the MILP’s computation time. These results demonstrate that DQNs can offer a computationally efficient and practically viable alternative to MILP for real-time control in residential energy systems.

Suggested Citation

  • Vinzent Vetter & Philipp Wohlgenannt & Peter Kepplinger & Elias Eder, 2025. "Deep Reinforcement Learning Approaches the MILP Optimum of a Multi-Energy Optimization in Energy Communities," Energies, MDPI, vol. 18(17), pages 1-20, August.
  • Handle: RePEc:gam:jeners:v:18:y:2025:i:17:p:4489-:d:1731167
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1996-1073/18/17/4489/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1996-1073/18/17/4489/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Robert Riechel, 2016. "Zwischen Gebäude und Gesamtstadt: das Quartier als Handlungsraum in der lokalen Wärmewende," Vierteljahrshefte zur Wirtschaftsforschung / Quarterly Journal of Economic Research, DIW Berlin, German Institute for Economic Research, vol. 85(4), pages 89-101.
    2. Ren, Hongbo & Gao, Weijun, 2010. "A MILP model for integrated plan and evaluation of distributed energy systems," Applied Energy, Elsevier, vol. 87(3), pages 1001-1014, March.
    3. Langer, Lissy & Volling, Thomas, 2020. "An optimal home energy management system for modulating heat pumps and photovoltaic systems," Applied Energy, Elsevier, vol. 278(C).
    4. Langer, Lissy & Volling, Thomas, 2022. "A reinforcement learning approach to home energy management for modulating heat pumps and photovoltaic systems," Applied Energy, Elsevier, vol. 327(C).
    5. Charbonnier, Flora & Peng, Bei & Vienne, Julie & Stai, Eleni & Morstyn, Thomas & McCulloch, Malcolm, 2025. "Centralised rehearsal of decentralised cooperation: Multi-agent reinforcement learning for the scalable coordination of residential energy flexibility," Applied Energy, Elsevier, vol. 377(PA).
    6. Aguilera, José Joaquín & Padullés, Roger & Meesenburg, Wiebke & Markussen, Wiebke Brix & Zühlsdorf, Benjamin & Elmegaard, Brian, 2024. "Operation optimization in large-scale heat pump systems: A scheduling framework integrating digital twin modelling, demand forecasting, and MILP," Applied Energy, Elsevier, vol. 376(PB).
    7. Cosic, Armin & Stadler, Michael & Mansoor, Muhammad & Zellinger, Michael, 2021. "Mixed-integer linear programming based optimization strategies for renewable energy communities," Energy, Elsevier, vol. 237(C).
    8. Wang, Zhe & Hong, Tianzhen, 2020. "Reinforcement learning for building controls: The opportunities and challenges," Applied Energy, Elsevier, vol. 269(C).
    9. Giulia Palma & Leonardo Guiducci & Marta Stentati & Antonio Rizzo & Simone Paoletti, 2024. "Reinforcement Learning for Energy Community Management: A European-Scale Study," Energies, MDPI, vol. 17(5), pages 1-19, March.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Dominik Latoń & Jakub Grela & Andrzej Ożadowicz, 2024. "Applications of Deep Reinforcement Learning for Home Energy Management Systems: A Review," Energies, MDPI, vol. 17(24), pages 1-30, December.
    2. Langer, Lissy & Volling, Thomas, 2022. "A reinforcement learning approach to home energy management for modulating heat pumps and photovoltaic systems," Applied Energy, Elsevier, vol. 327(C).
    3. Michael Bachseitz & Muhammad Sheryar & David Schmitt & Thorsten Summ & Christoph Trinkl & Wilfried Zörner, 2024. "PV-Optimized Heat Pump Control in Multi-Family Buildings Using a Reinforcement Learning Approach," Energies, MDPI, vol. 17(8), pages 1-16, April.
    4. Schmitz, Simon & Brucke, Karoline & Kasturi, Pranay & Ansari, Esmail & Klement, Peter, 2024. "Forecast-based and data-driven reinforcement learning for residential heat pump operation," Applied Energy, Elsevier, vol. 371(C).
    5. Panagiotis Michailidis & Iakovos Michailidis & Elias Kosmatopoulos, 2025. "Reinforcement Learning for Optimizing Renewable Energy Utilization in Buildings: A Review on Applications and Innovations," Energies, MDPI, vol. 18(7), pages 1-40, March.
    6. Kim, Hyung Joon & Lee, Jae Yong & Tak, Hyunwoo & Kim, Dongwoo, 2025. "Deep reinforcement learning-based residential building energy management incorporating power-to-heat technology for building electrification," Energy, Elsevier, vol. 317(C).
    7. Pergantis, Elias N. & Priyadarshan, & Theeb, Nadah Al & Dhillon, Parveen & Ore, Jonathan P. & Ziviani, Davide & Groll, Eckhard A. & Kircher, Kevin J., 2024. "Field demonstration of predictive heating control for an all-electric house in a cold climate," Applied Energy, Elsevier, vol. 360(C).
    8. Yin, Linfei & Xiong, Yi, 2024. "Incremental learning user profile and deep reinforcement learning for managing building energy in heating water," Energy, Elsevier, vol. 313(C).
    9. Yassine Chemingui & Adel Gastli & Omar Ellabban, 2020. "Reinforcement Learning-Based School Energy Management System," Energies, MDPI, vol. 13(23), pages 1-21, December.
    10. Yanfeng Liu & Yaxing Wang & Xi Luo, 2020. "Design and Operation Optimization of Distributed Solar Energy System Based on Dynamic Operation Strategy," Energies, MDPI, vol. 14(1), pages 1-26, December.
    11. Wang, Xuan & Shu, Gequn & Tian, Hua & Wang, Rui & Cai, Jinwen, 2020. "Operation performance comparison of CCHP systems with cascade waste heat recovery systems by simulation and operation optimisation," Energy, Elsevier, vol. 206(C).
    12. Gokhale, Gargya & Claessens, Bert & Develder, Chris, 2022. "Physics informed neural networks for control oriented thermal modeling of buildings," Applied Energy, Elsevier, vol. 314(C).
    13. Chen, Yen-Haw & Lu, Su-Ying & Chang, Yung-Ruei & Lee, Ta-Tung & Hu, Ming-Che, 2013. "Economic analysis and optimal energy management models for microgrid systems: A case study in Taiwan," Applied Energy, Elsevier, vol. 103(C), pages 145-154.
    14. Huang, Jinbo & Li, Zhigang & Wu, Q.H., 2017. "Coordinated dispatch of electric power and district heating networks: A decentralized solution using optimality condition decomposition," Applied Energy, Elsevier, vol. 206(C), pages 1508-1522.
    15. Mariuzzo, Ivan & Fina, Bernadette & Stroemer, Stefan & Corinaldesi, Carlo & Raugi, Marco, 2025. "Grid-friendly optimization of energy communities through enhanced multiple participation," Renewable and Sustainable Energy Reviews, Elsevier, vol. 208(C).
    16. Yuanyuan He & Luxin Wan & Manli Zhang & Huijuan Zhao, 2022. "Regional Renewable Energy Installation Optimization Strategies with Renewable Portfolio Standards in China," Sustainability, MDPI, vol. 14(17), pages 1-18, August.
    17. Chen, Yizhong & He, Li & Li, Jing, 2017. "Stochastic dominant-subordinate-interactive scheduling optimization for interconnected microgrids with considering wind-photovoltaic-based distributed generations under uncertainty," Energy, Elsevier, vol. 130(C), pages 581-598.
    18. Yutong Zhao & Shuang Zeng & Yifeng Ding & Lin Ma & Zhao Wang & Anqi Liang & Hongbo Ren, 2024. "Cost–Benefit Analysis of Distributed Energy Systems Considering the Monetization of Indirect Benefits," Sustainability, MDPI, vol. 16(2), pages 1-13, January.
    19. Luca Brunelli & Emiliano Borri & Anna Laura Pisello & Andrea Nicolini & Carles Mateu & Luisa F. Cabeza, 2024. "Thermal Energy Storage in Energy Communities: A Perspective Overview through a Bibliometric Analysis," Sustainability, MDPI, vol. 16(14), pages 1-27, July.
    20. Davide Coraci & Silvio Brandi & Marco Savino Piscitelli & Alfonso Capozzoli, 2021. "Online Implementation of a Soft Actor-Critic Agent to Enhance Indoor Temperature Control and Energy Efficiency in Buildings," Energies, MDPI, vol. 14(4), pages 1-26, February.

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jeners:v:18:y:2025:i:17:p:4489-:d:1731167. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.