IDEAS home Printed from https://ideas.repec.org/a/eee/appene/v389y2025ics0306261925004647.html
   My bibliography  Save this article

Reinforcement learning for data center energy efficiency optimization: A systematic literature review and research roadmap

Author

Listed:
  • Kahil, Hussain
  • Sharma, Shiva
  • Välisuo, Petri
  • Elmusrati, Mohammed

Abstract

With today’s challenges posed by climate change, global attention is increasingly focused on reducing energy consumption within sustainable communities. As significant energy consumers, data centers represent a crucial area for research in energy efficiency optimization. To address this issue, various algorithms have been employed to develop sophisticated solutions for data center systems. Recently, Reinforcement Learning (RL) and its advanced counterpart, Deep Reinforcement Learning (DRL), have demonstrated promising potential in improving data center energy efficiency. However, a comprehensive review of the deployment of these algorithms remains limited. In this systematic review, we explore the application of RL/DRL algorithms for optimizing data center energy efficiency, with a focus on optimizing the operation of cooling systems and Information and Communication Technology (ICT) processes, including task scheduling, resource allocation, virtual machine (VM) consolidation/placement, and network traffic control. Following the Preferred Reporting Items for Systematic review and Meta-Analysis (PRISMA) protocol, we provide a detailed overview of the methodologies and objectives of 65 identified studies, along with an in-depth analysis of their energy-related results. We also summarize key aspects of these studies, including benchmark comparisons, experimental setups, datasets, and implementation platforms. Additionally, we present a structured qualitative comparison of the Markov Decision Process (MDP) elements for joint optimization studies. Our findings highlight vital research gaps, including the lack of real-time validation for developed algorithms and the absence of multi-scale standardized metrics for reporting energy efficiency improvements. Furthermore, we propose joint optimization of multi-system objectives as a promising direction for future research.

Suggested Citation

  • Kahil, Hussain & Sharma, Shiva & Välisuo, Petri & Elmusrati, Mohammed, 2025. "Reinforcement learning for data center energy efficiency optimization: A systematic literature review and research roadmap," Applied Energy, Elsevier, vol. 389(C).
  • Handle: RePEc:eee:appene:v:389:y:2025:i:c:s0306261925004647
    DOI: 10.1016/j.apenergy.2025.125734
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0306261925004647
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.apenergy.2025.125734?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Zhang, Yingbo & Tang, Hong & Li, Hangxin & Wang, Shengwei, 2025. "Unlocking the flexibilities of data centers for smart grid services: Optimal dispatch and design of energy storage systems under progressive loading," Energy, Elsevier, vol. 316(C).
    2. Guo, Yuxiang & Qu, Shengli & Wang, Chuang & Xing, Ziwen & Duan, Kaiwen, 2024. "Optimal dynamic thermal management for data center via soft actor-critic algorithm with dynamic control interval and combined-value state space," Applied Energy, Elsevier, vol. 373(C).
    3. Kuldeep Kurte & Jeffrey Munk & Olivera Kotevska & Kadir Amasyali & Robert Smith & Evan McKee & Yan Du & Borui Cui & Teja Kuruganti & Helia Zandi, 2020. "Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses," Sustainability, MDPI, vol. 12(18), pages 1-38, September.
    4. Jin, Chaoqiang & Bai, Xuelian & Yang, Chao & Mao, Wangxin & Xu, Xin, 2020. "A review of power consumption models of servers in data centers," Applied Energy, Elsevier, vol. 265(C).
    5. Qiankun Chang & Yuanfeng Huang & Kaiyan Liu & Xin Xu & Yaohua Zhao & Song Pan, 2024. "Optimization Control Strategies and Evaluation Metrics of Cooling Systems in Data Centers: A Review," Sustainability, MDPI, vol. 16(16), pages 1-41, August.
    6. Biemann, Marco & Scheller, Fabian & Liu, Xiufeng & Huang, Lizhen, 2021. "Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control," Applied Energy, Elsevier, vol. 298(C).
    7. Liu, Wenyu & Yan, Yuejun & Sun, Yimeng & Mao, Hongju & Cheng, Ming & Wang, Peng & Ding, Zhaohao, 2023. "Online job scheduling scheme for low-carbon data center operation: An information and energy nexus perspective," Applied Energy, Elsevier, vol. 338(C).
    8. Na Huang & Xiang Li & Quanming Xu & Ronghao Chen & Huidong Chen & Aidong Chen, 2023. "Artificial Intelligence-Based Temperature Twinning and Pre-Control for Data Center Airflow Organization," Energies, MDPI, vol. 16(16), pages 1-15, August.
    9. Ayas Shaqour & Aya Hagishima, 2022. "Systematic Review on Deep Reinforcement Learning-Based Energy Management for Different Building Types," Energies, MDPI, vol. 15(22), pages 1-27, November.
    10. Alessandro Liberati & Douglas G Altman & Jennifer Tetzlaff & Cynthia Mulrow & Peter C Gøtzsche & John P A Ioannidis & Mike Clarke & P J Devereaux & Jos Kleijnen & David Moher, 2009. "The PRISMA Statement for Reporting Systematic Reviews and Meta-Analyses of Studies That Evaluate Health Care Interventions: Explanation and Elaboration," PLOS Medicine, Public Library of Science, vol. 6(7), pages 1-28, July.
    11. He, Hongwen & Meng, Xiangfei & Wang, Yong & Khajepour, Amir & An, Xiaowen & Wang, Renguang & Sun, Fengchun, 2024. "Deep reinforcement learning based energy management strategies for electrified vehicles: Recent advances and perspectives," Renewable and Sustainable Energy Reviews, Elsevier, vol. 192(C).
    12. Mahbod, Muhammad Haiqal Bin & Chng, Chin Boon & Lee, Poh Seng & Chui, Chee Kong, 2022. "Energy saving evaluation of an energy efficient data center using a model-free reinforcement learning approach," Applied Energy, Elsevier, vol. 322(C).
    13. Volodymyr Mnih & Koray Kavukcuoglu & David Silver & Andrei A. Rusu & Joel Veness & Marc G. Bellemare & Alex Graves & Martin Riedmiller & Andreas K. Fidjeland & Georg Ostrovski & Stig Petersen & Charle, 2015. "Human-level control through deep reinforcement learning," Nature, Nature, vol. 518(7540), pages 529-533, February.
    14. Zhang, Qingang & Zeng, Wei & Lin, Qinjie & Chng, Chin-Boon & Chui, Chee-Kong & Lee, Poh-Seng, 2023. "Deep reinforcement learning towards real-world dynamic thermal management of data centers," Applied Energy, Elsevier, vol. 333(C).
    15. Ce Chi & Kaixuan Ji & Penglei Song & Avinab Marahatta & Shikui Zhang & Fa Zhang & Dehui Qiu & Zhiyong Liu, 2021. "Cooperatively Improving Data Center Energy Efficiency Based on Multi-Agent Deep Reinforcement Learning," Energies, MDPI, vol. 14(8), pages 1-32, April.
    16. Anders S. G. Andrae & Tomas Edler, 2015. "On Global Electricity Usage of Communication Technology: Trends to 2030," Challenges, MDPI, vol. 6(1), pages 1-41, April.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Ayas Shaqour & Aya Hagishima, 2022. "Systematic Review on Deep Reinforcement Learning-Based Energy Management for Different Building Types," Energies, MDPI, vol. 15(22), pages 1-27, November.
    2. Han, Ouzhu & Ding, Tao & Yang, Miao & Jia, Wenhao & He, Xinran & Ma, Zhoujun, 2024. "A novel 4-level joint optimal dispatch for demand response of data centers with district autonomy realization," Applied Energy, Elsevier, vol. 358(C).
    3. Guo, Yuxiang & Qu, Shengli & Wang, Chuang & Xing, Ziwen & Duan, Kaiwen, 2024. "Optimal dynamic thermal management for data center via soft actor-critic algorithm with dynamic control interval and combined-value state space," Applied Energy, Elsevier, vol. 373(C).
    4. Xue, Lin & Wang, Jianxue & Li, Haotian & Yong, Weizhen & Zhang, Yao, 2025. "Online energy conservation scheduling for geo-distributed data centers with hybrid data-driven and knowledge-driven approach," Energy, Elsevier, vol. 322(C).
    5. Mahbod, Muhammad Haiqal Bin & Chng, Chin Boon & Lee, Poh Seng & Chui, Chee Kong, 2022. "Energy saving evaluation of an energy efficient data center using a model-free reinforcement learning approach," Applied Energy, Elsevier, vol. 322(C).
    6. Liao, Chenxin & Miyata, Shohei & Qu, Ming & Akashi, Yasunori, 2025. "Year-round operational optimization of HVAC systems using hierarchical deep reinforcement learning for enhancing indoor air quality and reducing energy consumption," Applied Energy, Elsevier, vol. 390(C).
    7. Chen, Siliang & Liang, Xinbin & Zhang, Zheming & Zheng, Fei & Jin, Xinqiao & Du, Zhimin, 2025. "A general energy-aware framework with multi-modal information and multi-task coordination for smart management towards net-zero emissions in energy system," Renewable and Sustainable Energy Reviews, Elsevier, vol. 212(C).
    8. Blad, C. & Bøgh, S. & Kallesøe, C. & Raftery, Paul, 2023. "A laboratory test of an Offline-trained Multi-Agent Reinforcement Learning Algorithm for Heating Systems," Applied Energy, Elsevier, vol. 337(C).
    9. Wang, Zixuan & Xiao, Fu & Ran, Yi & Li, Yanxue & Xu, Yang, 2024. "Scalable energy management approach of residential hybrid energy system using multi-agent deep reinforcement learning," Applied Energy, Elsevier, vol. 367(C).
    10. Omar Al-Ani & Sanjoy Das, 2022. "Reinforcement Learning: Theory and Applications in HEMS," Energies, MDPI, vol. 15(17), pages 1-37, September.
    11. Gao, Yuan & Hu, Zehuan & Yamate, Shun & Otomo, Junichiro & Chen, Wei-An & Liu, Mingzhe & Xu, Tingting & Ruan, Yingjun & Shang, Juan, 2025. "Unlocking predictive insights and interpretability in deep reinforcement learning for Building-Integrated Photovoltaic and Battery (BIPVB) systems," Applied Energy, Elsevier, vol. 384(C).
    12. Dimitrios Vamvakas & Panagiotis Michailidis & Christos Korkas & Elias Kosmatopoulos, 2023. "Review and Evaluation of Reinforcement Learning Frameworks on Smart Grid Applications," Energies, MDPI, vol. 16(14), pages 1-38, July.
    13. Xiao, Jiang-Wen & Yang, Yan-Bing & Cui, Shichang & Wang, Yan-Wu, 2023. "Cooperative online schedule of interconnected data center microgrids with shared energy storage," Energy, Elsevier, vol. 285(C).
    14. Gao, Yuan & Matsunami, Yuki & Miyata, Shohei & Akashi, Yasunori, 2022. "Multi-agent reinforcement learning dealing with hybrid action spaces: A case study for off-grid oriented renewable building energy system," Applied Energy, Elsevier, vol. 326(C).
    15. Matteo Manganelli & Alessandro Soldati & Luigi Martirano & Seeram Ramakrishna, 2021. "Strategies for Improving the Sustainability of Data Centers via Energy Mix, Energy Conservation, and Circular Energy," Sustainability, MDPI, vol. 13(11), pages 1-25, May.
    16. Seppo Sierla & Heikki Ihasalo & Valeriy Vyatkin, 2022. "A Review of Reinforcement Learning Applications to Control of Heating, Ventilation and Air Conditioning Systems," Energies, MDPI, vol. 15(10), pages 1-25, May.
    17. Lu, Ruyuan & Li, Xin & Chen, Ronghao & Lei, Aimin & Ma, Xiaoming, 2024. "An Alternative Reinforcement Learning (ARL) control strategy for data center air-cooled HVAC systems," Energy, Elsevier, vol. 308(C).
    18. Milena Dalton & Benjamin Sanderson & Leanne J Robinson & Caroline S E Homer & William Pomat & Margie Danchin & Stefanie Vaccher, 2023. "Impact of COVID-19 on routine childhood immunisations in low- and middle-income countries: A scoping review," PLOS Global Public Health, Public Library of Science, vol. 3(8), pages 1-17, August.
    19. Ludoviko Zirimenya & Fatima Mahmud-Ajeigbe & Ruth McQuillan & You Li, 2020. "A systematic review and meta-analysis to assess the association between urogenital schistosomiasis and HIV/AIDS infection," PLOS Neglected Tropical Diseases, Public Library of Science, vol. 14(6), pages 1-13, June.
    20. Desalegne Amare & Fentie Ambaw Getahun & Endalkachew Worku Mengesha & Getenet Dessie & Melashu Balew Shiferaw & Tegenaw Asemamaw Dires & Kefyalew Addis Alene, 2023. "Effectiveness of healthcare workers and volunteers training on improving tuberculosis case detection: A systematic review and meta-analysis," PLOS ONE, Public Library of Science, vol. 18(3), pages 1-13, March.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:appene:v:389:y:2025:i:c:s0306261925004647. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/wps/find/journaldescription.cws_home/405891/description#description .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.