File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1016/j.enbuild.2024.114696
- Scopus: eid_2-s2.0-85202209872
- Find via

Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: Multi-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control
| Title | Multi-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control |
|---|---|
| Authors | |
| Keywords | Deep reinforcement learning Multi-source transfer Multi-zone building HVAC control Transfer learning |
| Issue Date | 1-Nov-2024 |
| Publisher | Elsevier |
| Citation | Energy and Buildings, 2024, v. 322 How to Cite? |
| Abstract | Deep reinforcement learning (DRL) control methods have shown great potential for optimal HVAC control, but they require significant time and data to learn effective policies. By employing transfer learning (TL) with pre-trained models, the need to learn the data from scratch is avoided, saving time and resources. However, there are two main critical issues with this approach: the inappropriate selection of the source domain resulting in worse control performance and inefficient utilization of multi-source domain control experience. To address these challenges, a multi-source transfer learning and deep reinforcement learning (MTL-DRL) integrated framework is proposed for efficient HVAC system control. In order to select appropriate source domains, the contribution of various source domains to the target task is quantified first, followed by a comprehensive evaluation of transfer performance based on average energy consumption and average temperature deviation. The well-pretrained DRL parameters from the optimal multi-source transfer set are then sequentially transferred to the target DRL controller. Results from a series of transfer experiments between buildings with different thermal zones and weather conditions indicate that the MTL-DRL framework significantly reduces the training time of HVAC control, with improvements of up to 20% compared to DRL baseline models trained from scratch. Additionally, the MTL-DRL method leads to reductions in average energy consumption ranging from 1.43% to 3.12% and average temperature deviation up to 14.32%. The impact of the source domain transfer sequence on the performance of the DRL-based control method is also discussed. Overall, the proposed framework presents a promising solution for enhancing DRL-based HVAC control methods by reducing training time and energy consumption while maintaining occupants’ comfort. |
| Persistent Identifier | http://hdl.handle.net/10722/360754 |
| ISSN | 2023 Impact Factor: 6.6 2023 SCImago Journal Rankings: 1.632 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Hou, Fangli | - |
| dc.contributor.author | Cheng, Jack C.P. | - |
| dc.contributor.author | Kwok, Helen H.L. | - |
| dc.contributor.author | Ma, Jun | - |
| dc.date.accessioned | 2025-09-13T00:36:12Z | - |
| dc.date.available | 2025-09-13T00:36:12Z | - |
| dc.date.issued | 2024-11-01 | - |
| dc.identifier.citation | Energy and Buildings, 2024, v. 322 | - |
| dc.identifier.issn | 0378-7788 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/360754 | - |
| dc.description.abstract | <p>Deep reinforcement learning (DRL) control methods have shown great potential for optimal HVAC control, but they require significant time and data to learn effective policies. By employing transfer learning (TL) with pre-trained models, the need to learn the data from scratch is avoided, saving time and resources. However, there are two main critical issues with this approach: the inappropriate selection of the source domain resulting in worse control performance and inefficient utilization of multi-source domain control experience. To address these challenges, a multi-source transfer learning and deep reinforcement learning (MTL-DRL) integrated framework is proposed for efficient HVAC system control. In order to select appropriate source domains, the contribution of various source domains to the target task is quantified first, followed by a comprehensive evaluation of transfer performance based on average energy consumption and average temperature deviation. The well-pretrained DRL parameters from the optimal multi-source transfer set are then sequentially transferred to the target DRL controller. Results from a series of transfer experiments between buildings with different thermal zones and weather conditions indicate that the MTL-DRL framework significantly reduces the training time of HVAC control, with improvements of up to 20% compared to DRL baseline models trained from scratch. Additionally, the MTL-DRL method leads to reductions in average energy consumption ranging from 1.43% to 3.12% and average temperature deviation up to 14.32%. The impact of the source domain transfer sequence on the performance of the DRL-based control method is also discussed. Overall, the proposed framework presents a promising solution for enhancing DRL-based HVAC control methods by reducing training time and energy consumption while maintaining occupants’ comfort.</p> | - |
| dc.language | eng | - |
| dc.publisher | Elsevier | - |
| dc.relation.ispartof | Energy and Buildings | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject | Deep reinforcement learning | - |
| dc.subject | Multi-source transfer | - |
| dc.subject | Multi-zone building HVAC control | - |
| dc.subject | Transfer learning | - |
| dc.title | Multi-source transfer learning method for enhancing the deployment of deep reinforcement learning in multi-zone building HVAC control | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1016/j.enbuild.2024.114696 | - |
| dc.identifier.scopus | eid_2-s2.0-85202209872 | - |
| dc.identifier.volume | 322 | - |
| dc.identifier.eissn | 1872-6178 | - |
| dc.identifier.issnl | 0378-7788 | - |
