Publications

Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control

Published in NeurIPS CCAI & ACM BuildSys'24, 2024

This paper proposes an adaptive policy regularization approach for transferring policies from offline datasets to online deployment in HVAC control. The method uses weighted increased simple moving average Q-value estimators to stabilize policy updates and improve safety during online fine-tuning.

Recommended citation: Liu, Hsin-Yu. Adaptive Policy Regularization for Offline-to-Online Reinforcement Learning in HVAC Control. NeurIPS CCAI & ACM BuildSys'24. Nov. 2024. https://dl.acm.org/doi/pdf/10.1145/3671127.3698163

Policy Regularization in Model-Free Building Control via Comprehensive Approaches from Offline to Online Reinforcement Learning

Published in Ph.D. Dissertation, University of California San Diego, 2024

This dissertation develops a novel policy regularization framework for reinforcement learning in HVAC control systems, focusing on safe and efficient operation in real-world settings. Contributions include methods for offline-to-online policy regularization, open-source building batch RL datasets for benchmarking, and empirical results demonstrating energy and performance improvements in building control tasks.

Recommended citation: Liu, Hsin-Yu. Policy Regularization in Model-Free Building Control via Comprehensive Approaches from Offline to Online Reinforcement Learning. Ph.D. Dissertation, University of California San Diego, Jun. 2024. https://escholarship.org/content/qt0b23889v/qt0b23889v_noSplash_60b699fcef9c1cfc9988c4c5c4249a14.pdf