Abstract
In this paper, we develop a grid-interactive multi-zone building controller based on a deep reinforcement learning (RL) approach. The controller is designed to facilitate building operation during normal conditions and demand response events, while ensuring occupants comfort and energy efficiency. We leverage a continuous action space RL formulation, and devise a two-stage global-local RL training framework. In the first stage, a global fast policy search is performed using a gradient-free RL algorithm. In the second stage, a local fine-tuning is conducted using a policy gradient method. In contrast to the state-of-the-art model predictive control (MPC) approach, the proposed RL controller does not require complex computation during real-time operation and can adapt to nonlinear building models. We illustrate the controller performance numerically using a five-zone commercial building.
Original language | American English |
---|---|
Pages | 4155-4162 |
Number of pages | 8 |
DOIs | |
State | Published - 25 May 2021 |
Event | 2021 American Control Conference, ACC 2021 - Virtual, New Orleans, United States Duration: 25 May 2021 → 28 May 2021 |
Conference
Conference | 2021 American Control Conference, ACC 2021 |
---|---|
Country/Territory | United States |
City | Virtual, New Orleans |
Period | 25/05/21 → 28/05/21 |
Bibliographical note
See NREL/CP-2C00-78000 for preprintNREL Publication Number
- NREL/CP-2C00-80740
Keywords
- building control
- demand response
- reinforcement learning