Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control
Published in NeurIPS CCAI & ACM BuildSys'24, 2024
Recommended citation: Liu, Hsin-Yu. Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control. NeurIPS CCAI & ACM BuildSys'24. Nov. 2024. https://dl.acm.org/doi/pdf/10.1145/3671127.3698163
This paper proposes an adaptive policy regularization approach for transferring policies from offline datasets to online deployment in HVAC control. The method uses weighted increased simple moving average Q-value estimators to stabilize policy updates and improve safety during online fine-tuning.
Results show significant improvements over baseline methods in HVAC control benchmarks.
title: "Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control"
collection: publications
permalink: /publication/adaptive-policy-regularization-2024
date: 2024-11-01
venue: "NeurIPS CCAI & ACM BuildSys'24"
paperurl: ''
paperurl: 'https://dl.acm.org/doi/pdf/10.1145/3671127.3698163'
citation: "Liu, Hsin-Yu. Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control. NeurIPS CCAI & ACM BuildSys'24. Nov. 2024."
This paper proposes an adaptive policy regularization approach for transferring policies from offline datasets to online deployment in HVAC control. The method uses weighted increased simple moving average Q-value estimators to stabilize policy updates and improve safety during online fine-tuning.
Results show significant improvements over baseline methods in HVAC control benchmarks.
