File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Multi-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control

TitleMulti-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control
Authors
KeywordsDeep reinforcement learning
Multi-source transfer
Multi-zone building HVAC control
Transfer learning
Issue Date1-Nov-2024
PublisherElsevier
Citation
Energy and Buildings, 2024, v. 322 How to Cite?
Abstract

Deep reinforcement learning (DRL) control methods have shown great potential for optimal HVAC control, but they require significant time and data to learn effective policies. By employing transfer learning (TL) with pre-trained models, the need to learn the data from scratch is avoided, saving time and resources. However, there are two main critical issues with this approach: the inappropriate selection of the source domain resulting in worse control performance and inefficient utilization of multi-source domain control experience. To address these challenges, a multi-source transfer learning and deep reinforcement learning (MTL-DRL) integrated framework is proposed for efficient HVAC system control. In order to select appropriate source domains, the contribution of various source domains to the target task is quantified first, followed by a comprehensive evaluation of transfer performance based on average energy consumption and average temperature deviation. The well-pretrained DRL parameters from the optimal multi-source transfer set are then sequentially transferred to the target DRL controller. Results from a series of transfer experiments between buildings with different thermal zones and weather conditions indicate that the MTL-DRL framework significantly reduces the training time of HVAC control, with improvements of up to 20% compared to DRL baseline models trained from scratch. Additionally, the MTL-DRL method leads to reductions in average energy consumption ranging from 1.43% to 3.12% and average temperature deviation up to 14.32%. The impact of the source domain transfer sequence on the performance of the DRL-based control method is also discussed. Overall, the proposed framework presents a promising solution for enhancing DRL-based HVAC control methods by reducing training time and energy consumption while maintaining occupants’ comfort.


Persistent Identifierhttp://hdl.handle.net/10722/360754
ISSN
2023 Impact Factor: 6.6
2023 SCImago Journal Rankings: 1.632

 

DC FieldValueLanguage
dc.contributor.authorHou, Fangli-
dc.contributor.authorCheng, Jack C.P.-
dc.contributor.authorKwok, Helen H.L.-
dc.contributor.authorMa, Jun-
dc.date.accessioned2025-09-13T00:36:12Z-
dc.date.available2025-09-13T00:36:12Z-
dc.date.issued2024-11-01-
dc.identifier.citationEnergy and Buildings, 2024, v. 322-
dc.identifier.issn0378-7788-
dc.identifier.urihttp://hdl.handle.net/10722/360754-
dc.description.abstract<p>Deep reinforcement learning (DRL) control methods have shown great potential for optimal HVAC control, but they require significant time and data to learn effective policies. By employing transfer learning (TL) with pre-trained models, the need to learn the data from scratch is avoided, saving time and resources. However, there are two main critical issues with this approach: the inappropriate selection of the source domain resulting in worse control performance and inefficient utilization of multi-source domain control experience. To address these challenges, a multi-source transfer learning and deep reinforcement learning (MTL-DRL) integrated framework is proposed for efficient HVAC system control. In order to select appropriate source domains, the contribution of various source domains to the target task is quantified first, followed by a comprehensive evaluation of transfer performance based on average energy consumption and average temperature deviation. The well-pretrained DRL parameters from the optimal multi-source transfer set are then sequentially transferred to the target DRL controller. Results from a series of transfer experiments between buildings with different thermal zones and weather conditions indicate that the MTL-DRL framework significantly reduces the training time of HVAC control, with improvements of up to 20% compared to DRL baseline models trained from scratch. Additionally, the MTL-DRL method leads to reductions in average energy consumption ranging from 1.43% to 3.12% and average temperature deviation up to 14.32%. The impact of the source domain transfer sequence on the performance of the DRL-based control method is also discussed. Overall, the proposed framework presents a promising solution for enhancing DRL-based HVAC control methods by reducing training time and energy consumption while maintaining occupants’ comfort.</p>-
dc.languageeng-
dc.publisherElsevier-
dc.relation.ispartofEnergy and Buildings-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectDeep reinforcement learning-
dc.subjectMulti-source transfer-
dc.subjectMulti-zone building HVAC control-
dc.subjectTransfer learning-
dc.titleMulti-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control-
dc.typeArticle-
dc.identifier.doi10.1016/j.enbuild.2024.114696-
dc.identifier.scopuseid_2-s2.0-85202209872-
dc.identifier.volume322-
dc.identifier.eissn1872-6178-
dc.identifier.issnl0378-7788-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats