A Relearning Approach to Reinforcement Learning for control of Smart Buildings



Published Nov 3, 2020
Avisek Naug Marcos Q'uiñones -Grueiro Gautam Biswas


This paper demonstrates that continual relearning of control policies using incremental deep reinforcement learning (RL) can improve policy learning for non-stationary processes. This approach has been demonstrated in a data-driven “smart building environment” that we use as a test-bed for developing HVAC controllers for reducing energy consumption of large buildings on our university campus. The non-stationarity in building operations and weather patterns makes it imperative to develop control strategies that are adaptive to changing conditions. On-policy RL algorithms, such as Proximal Policy Optimization (PPO) represent an approach for addressing this non-stationarity, but they cannot be applied to safety-critical systems. As an alternative, we develop an incremental RL technique that simultaneously reduces building energy consumption without sacrificing overall comfort. We compare the performance of our incremental RL controller to that of a static RL controller that does not implement the relearning function. The performance of the static controller diminishes significantly over time, but the relearning controller adjusts to changing conditions while ensuring comfort and optimal energy performance.

How to Cite

Naug, A., Q’uiñones -Grueiro, M., & Biswas, G. (2020). A Relearning Approach to Reinforcement Learning for control of Smart Buildings. Annual Conference of the PHM Society, 12(1), 14. https://doi.org/10.36001/phmconf.2020.v12i1.1296
Abstract 77 | PDF Downloads 64



Smart Buildings, Reinforcement Learning, Energy Optimization, Control, LSTM

Technical Papers