IDEAS home Printed from https://ideas.repec.org/a/eee/energy/v263y2023ipbs0360544222025658.html
   My bibliography  Save this article

Cross temporal-spatial transferability investigation of deep reinforcement learning control strategy in the building HVAC system level

Author

Listed:
  • Fang, Xi
  • Gong, Guangcai
  • Li, Guannan
  • Chun, Liang
  • Peng, Pei
  • Li, Wenqiang
  • Shi, Xing

Abstract

Model free based DRL control strategies have achieved positive effects on the HVAC system optimal control. However, developing deep reinforcement learning (DRL) control strategies for different building HVAC systems is time-consuming and laborious. To address this issue, this study proposes a transfer learning and deep reinforcement learning (TL-DRL) integrated framework to achieve the DRL control strategy transfer in the building HVAC system level. Deep Q-learning (DQN) is first pre-trained in the source building until it converges to an optimal strategy. Then, the well pre-trained DQN parameters of the first few layers are transferred to the target DQN. Finally, the target DQN parameters of the last few layers are fine-tuned in the target building. An EnergyPlus-Python co-simulation testbed is developed to investigate the cross temporal-spatial transferability of DQN control strategy in the building HVAC system level. Results indicate that the proposed TL-DRL framework can effectively improve the training efficiency of control strategy by about 13.28% when transferring the first two layers compared to that of the DRL baseline models trained from scratch, while simultaneously maintaining energy consumption and indoor air temperature in an acceptable range. The proposed TL-DRL framework provides a preliminary direction for the scalability of intelligent HVAC control strategies.

Suggested Citation

  • Fang, Xi & Gong, Guangcai & Li, Guannan & Chun, Liang & Peng, Pei & Li, Wenqiang & Shi, Xing, 2023. "Cross temporal-spatial transferability investigation of deep reinforcement learning control strategy in the building HVAC system level," Energy, Elsevier, vol. 263(PB).
  • Handle: RePEc:eee:energy:v:263:y:2023:i:pb:s0360544222025658
    DOI: 10.1016/j.energy.2022.125679
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0360544222025658
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.energy.2022.125679?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Inayat, Abrar & Raza, Mohsin, 2019. "District cooling system via renewable energy sources: A review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 107(C), pages 360-373.
    2. Li, Wenzhuo & Wang, Shengwei & Koo, Choongwan, 2021. "A real-time optimal control strategy for multi-zone VAV air-conditioning systems adopting a multi-agent based distributed optimization method," Applied Energy, Elsevier, vol. 287(C).
    3. Wang, Zhe & Hong, Tianzhen, 2020. "Reinforcement learning for building controls: The opportunities and challenges," Applied Energy, Elsevier, vol. 269(C).
    4. Biemann, Marco & Scheller, Fabian & Liu, Xiufeng & Huang, Lizhen, 2021. "Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control," Applied Energy, Elsevier, vol. 298(C).
    5. Gao, Yixiang & Li, Shuhui & Fu, Xingang & Dong, Weizhen & Lu, Bing & Li, Zhongwen, 2020. "Energy management and demand response with intelligent learning for multi-thermal-zone buildings," Energy, Elsevier, vol. 210(C).
    6. Afram, Abdul & Janabi-Sharifi, Farrokh, 2015. "Gray-box modeling and validation of residential HVAC system for control system design," Applied Energy, Elsevier, vol. 137(C), pages 134-150.
    7. Yang, Ting & Zhao, Liyuan & Li, Wei & Wu, Jianzhong & Zomaya, Albert Y., 2021. "Towards healthy and cost-effective indoor environment management in smart homes: A deep reinforcement learning approach," Applied Energy, Elsevier, vol. 300(C).
    8. Wang, Tianjing & Tang, Yong, 2022. "Transfer-Reinforcement-Learning-Based rescheduling of differential power grids considering security constraints," Applied Energy, Elsevier, vol. 306(PB).
    9. Fang, Xi & Gong, Guangcai & Li, Guannan & Chun, Liang & Li, Wenqiang & Peng, Pei, 2021. "A hybrid deep transfer learning strategy for short term cross-building energy prediction," Energy, Elsevier, vol. 215(PB).
    10. Fan, Cheng & Sun, Yongjun & Xiao, Fu & Ma, Jie & Lee, Dasheng & Wang, Jiayuan & Tseng, Yen Chieh, 2020. "Statistical investigations of transfer learning-based methodology for short-term building energy predictions," Applied Energy, Elsevier, vol. 262(C).
    11. Xiao, Tong & Xu, Peng & He, Ruikai & Sha, Huajing, 2022. "Status quo and opportunities for building energy prediction in limited data Context—Overview from a competition," Applied Energy, Elsevier, vol. 305(C).
    12. Afroz, Zakia & Shafiullah, GM & Urmee, Tania & Higgins, Gary, 2018. "Modeling techniques used in building HVAC control systems: A review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 83(C), pages 64-84.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Fan, Cheng & Lei, Yutian & Sun, Yongjun & Mo, Like, 2023. "Novel transformer-based self-supervised learning methods for improved HVAC fault diagnosis performance with limited labeled data," Energy, Elsevier, vol. 278(PB).

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Omar Al-Ani & Sanjoy Das, 2022. "Reinforcement Learning: Theory and Applications in HEMS," Energies, MDPI, vol. 15(17), pages 1-37, September.
    2. Homod, Raad Z. & Togun, Hussein & Kadhim Hussein, Ahmed & Noraldeen Al-Mousawi, Fadhel & Yaseen, Zaher Mundher & Al-Kouz, Wael & Abd, Haider J. & Alawi, Omer A. & Goodarzi, Marjan & Hussein, Omar A., 2022. "Dynamics analysis of a novel hybrid deep clustering for unsupervised learning by reinforcement of multi-agent to energy saving in intelligent buildings," Applied Energy, Elsevier, vol. 313(C).
    3. Dongsu Kim & Yongjun Lee & Kyungil Chin & Pedro J. Mago & Heejin Cho & Jian Zhang, 2023. "Implementation of a Long Short-Term Memory Transfer Learning (LSTM-TL)-Based Data-Driven Model for Building Energy Demand Forecasting," Sustainability, MDPI, vol. 15(3), pages 1-23, January.
    4. Zhang, Qingang & Zeng, Wei & Lin, Qinjie & Chng, Chin-Boon & Chui, Chee-Kong & Lee, Poh-Seng, 2023. "Deep reinforcement learning towards real-world dynamic thermal management of data centers," Applied Energy, Elsevier, vol. 333(C).
    5. Zhang, Yunfei & Zhou, Zhihua & Du, Yahui & Shen, Jun & Li, Zhenxing & Yuan, Jianjuan, 2023. "A data transfer method based on one dimensional convolutional neural network for cross-building load prediction," Energy, Elsevier, vol. 277(C).
    6. Ayas Shaqour & Aya Hagishima, 2022. "Systematic Review on Deep Reinforcement Learning-Based Energy Management for Different Building Types," Energies, MDPI, vol. 15(22), pages 1-27, November.
    7. Li, Yanxue & Wang, Zixuan & Xu, Wenya & Gao, Weijun & Xu, Yang & Xiao, Fu, 2023. "Modeling and energy dynamic control for a ZEH via hybrid model-based deep reinforcement learning," Energy, Elsevier, vol. 277(C).
    8. Pinto, Giuseppe & Kathirgamanathan, Anjukan & Mangina, Eleni & Finn, Donal P. & Capozzoli, Alfonso, 2022. "Enhancing energy management in grid-interactive buildings: A comparison among cooperative and coordinated architectures," Applied Energy, Elsevier, vol. 310(C).
    9. Tang, Lingfeng & Xie, Haipeng & Wang, Xiaoyang & Bie, Zhaohong, 2023. "Privacy-preserving knowledge sharing for few-shot building energy prediction: A federated learning approach," Applied Energy, Elsevier, vol. 337(C).
    10. Clara Ceccolini & Roozbeh Sangi, 2022. "Benchmarking Approaches for Assessing the Performance of Building Control Strategies: A Review," Energies, MDPI, vol. 15(4), pages 1-30, February.
    11. Yue, Naihua & Caini, Mauro & Li, Lingling & Zhao, Yang & Li, Yu, 2023. "A comparison of six metamodeling techniques applied to multi building performance vectors prediction on gymnasiums under multiple climate conditions," Applied Energy, Elsevier, vol. 332(C).
    12. Di Natale, L. & Svetozarevic, B. & Heer, P. & Jones, C.N., 2022. "Physically Consistent Neural Networks for building thermal modeling: Theory and analysis," Applied Energy, Elsevier, vol. 325(C).
    13. Li, Guannan & Li, Fan & Ahmad, Tanveer & Liu, Jiangyan & Li, Tao & Fang, Xi & Wu, Yubei, 2022. "Performance evaluation of sequence-to-sequence-Attention model for short-term multi-step ahead building energy predictions," Energy, Elsevier, vol. 259(C).
    14. Shen, Rendong & Zhong, Shengyuan & Wen, Xin & An, Qingsong & Zheng, Ruifan & Li, Yang & Zhao, Jun, 2022. "Multi-agent deep reinforcement learning optimization framework for building energy system with renewable energy," Applied Energy, Elsevier, vol. 312(C).
    15. Zhuang, Dian & Gan, Vincent J.L. & Duygu Tekler, Zeynep & Chong, Adrian & Tian, Shuai & Shi, Xing, 2023. "Data-driven predictive control for smart HVAC system in IoT-integrated buildings with time-series forecasting and reinforcement learning," Applied Energy, Elsevier, vol. 338(C).
    16. Gao, Yuan & Matsunami, Yuki & Miyata, Shohei & Akashi, Yasunori, 2022. "Multi-agent reinforcement learning dealing with hybrid action spaces: A case study for off-grid oriented renewable building energy system," Applied Energy, Elsevier, vol. 326(C).
    17. Lei, Yue & Zhan, Sicheng & Ono, Eikichi & Peng, Yuzhen & Zhang, Zhiang & Hasama, Takamasa & Chong, Adrian, 2022. "A practical deep reinforcement learning framework for multivariate occupant-centric control in buildings," Applied Energy, Elsevier, vol. 324(C).
    18. Blad, C. & Bøgh, S. & Kallesøe, C. & Raftery, Paul, 2023. "A laboratory test of an Offline-trained Multi-Agent Reinforcement Learning Algorithm for Heating Systems," Applied Energy, Elsevier, vol. 337(C).
    19. Coraci, Davide & Brandi, Silvio & Hong, Tianzhen & Capozzoli, Alfonso, 2023. "Online transfer learning strategy for enhancing the scalability and deployment of deep reinforcement learning control in smart buildings," Applied Energy, Elsevier, vol. 333(C).
    20. Fan, Cheng & Lei, Yutian & Sun, Yongjun & Piscitelli, Marco Savino & Chiosa, Roberto & Capozzoli, Alfonso, 2022. "Data-centric or algorithm-centric: Exploiting the performance of transfer learning for improving building energy predictions in data-scarce context," Energy, Elsevier, vol. 240(C).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:energy:v:263:y:2023:i:pb:s0360544222025658. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/energy .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.