skip to main content
10.1145/3427773.3427871acmotherconferencesArticle/Chapter ViewAbstractPublication PagessensysConference Proceedingsconference-collections
research-article
Open access

Towards Off-policy Evaluation as a Prerequisite for Real-world Reinforcement Learning in Building Control

Published: 17 November 2020 Publication History

Abstract

We present an initial study of off-policy evaluation (OPE), a problem prerequisite to real-world reinforcement learning (RL), in the context of building control. OPE is the problem of estimating a policy's performance without running it on the actual system, using historical data from the existing controller. It enables the control engineers to ensure a new, pretrained policy satisfies the performance requirements and safety constraints of a real-world system, prior to interacting with it. While many methods have been developed for OPE, no study has evaluated which ones are suitable for building operational data, which are generated by deterministic policies and have limited coverage of the state-action space. After reviewing existing works and their assumptions, we adopted the approximate model (AM) method. Furthermore, we used bootstrapping to quantify uncertainty and correct for bias. In a simulation study, we evaluated the proposed approach on 10 policies pretrained with imitation learning. On average, the AM method estimated the energy and comfort costs with 1.84% and 14.1% error, respectively.

References

[1]
Bingqing Chen, Zicheng Cai, and Mario Bergés. 2019. Gnu-rl: A precocial reinforcement learning solution for building hvac control using a differentiable mpc policy. In Proceedings of the 6th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation. 316--325.
[2]
Bingqing Chen, Zicheng Cai, and Mario Berges. 2020. Gnu-RL: A Practical and Scalable Reinforcement Learning Solution for Building HVAC Control using a Differentiable MPC Policy. Frontiers in Built Environment 6 (2020), 174.
[3]
Drury B Crawley, Linda K Lawrie, Curtis O Pedersen, and Frederick C Winkelmann. 2000. Energy plus: energy simulation program. ASHRAE journal 42, 4 (2000), 49--56.
[4]
Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. 2019. On the sample complexity of the linear quadratic regulator. Foundations of Computational Mathematics (2019), 1--47.
[5]
Gabriel Dulac-Arnold, Daniel Mankowitz, and Todd Hester. 2019. Challenges of real-world reinforcement learning. arXiv preprint arXiv:1904.12901 (2019).
[6]
Javier Garcıa and Fernando Fernández. 2015. A comprehensive survey on safe reinforcement learning. Journal of Machine Learning Research 16, 1 (2015), 1437--1480.
[7]
Ruoxi Jia, Ming Jin, Kaiyu Sun, Tianzhen Hong, and Costas Spanos. 2019. Advanced building control via deep reinforcement learning. Energy Procedia 158 (2019), 6158--6163.
[8]
Nan Jiang and Lihong Li. 2016. Doubly robust off-policy value evaluation for reinforcement learning. In International Conference on Machine Learning. 652--661.
[9]
M.Jin and J. Lavaei. 2018. Control-Theoretic Analysis of Smoothness for Stability-Certified Reinforcement Learning. In 2018 IEEE Conference on Decision and Control (CDC). 6840--6847.
[10]
Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
[11]
Doina Precup. 2000. Eligibility traces for off-policy policy evaluation. Computer Science Department Faculty Publication Series (2000), 80.
[12]
Samuel Privara, Jiří Cigler, Zdeněk Váňa, Frauke Oldewurtel, Carina Sagerschnig, and Eva Žáčeková. 2013. Building modeling as a crucial part for building predictive control. Energy and Buildings 56 (2013), 8--22.
[13]
Harald Steck and Tommi S Jaakkola. 2004. Bias-corrected bootstrap and model uncertainty. In Advances in Neural Information Processing Systems. 521--528.
[14]
Philip Thomas and Emma Brunskill. 2016. Data-efficient off-policy policy evaluation for reinforcement learning. In International Conference on Machine Learning. 2139--2148.
[15]
Cameron Voloshin, Hoang M Le, Nan Jiang, and Yisong Yue. 2019. Empirical Study of Off-Policy Policy Evaluation for Reinforcement Learning. arXiv preprint arXiv:1911.06854 (2019).
[16]
Zhe Wang and Tianzhen Hong. 2020. Reinforcement learning for building controls: The opportunities and challenges. Applied Energy 269 (2020), 115036.
[17]
Larry Wasserman. 2013. All of statistics: a concise course in statistical inference. Springer Science & Business Media.
[18]
Yukun Yuan, Kin Sum Liu, Sirajum Munir, Jonathan Francis, Charles Shelton, and Shan Lin. 2020. Leveraging Fine-Grained Occupancy Estimation Patterns for Effective HVAC Control. In 2020 IEEE/ACM Fifth International Conference on Internet-of-Things Design and Implementation (IoTDI). IEEE, 92--103.
[19]
Zhiang Zhang and Khee Poh Lam. 2018. Practical Implementation and Evaluation of Deep Reinforcement Learning Control for a Radiant Heating System. In Proceedings of the 5th Conference on Systems for Built Environments (Shenzen, China) (BuildSys '18). ACM, New York, NY, USA, 148--157.

Cited By

View all
  • (2023)Real-World Implementation of Reinforcement Learning Based Energy Coordination for a Cluster of HouseholdsProceedings of the 10th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation10.1145/3600100.3625681(347-351)Online publication date: 15-Nov-2023
  • (2022)Reinforcement Learning in Construction Engineering and Management: A ReviewJournal of Construction Engineering and Management10.1061/(ASCE)CO.1943-7862.0002386148:11Online publication date: Nov-2022
  • (2021)Enforcing Policy Feasibility Constraints through Differentiable Projection for Energy OptimizationProceedings of the Twelfth ACM International Conference on Future Energy Systems10.1145/3447555.3464874(199-210)Online publication date: 22-Jun-2021

Index Terms

  1. Towards Off-policy Evaluation as a Prerequisite for Real-world Reinforcement Learning in Building Control

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Other conferences
      RLEM'20: Proceedings of the 1st International Workshop on Reinforcement Learning for Energy Management in Buildings & Cities
      November 2020
      63 pages
      ISBN:9781450381932
      DOI:10.1145/3427773
      This work is licensed under a Creative Commons Attribution International 4.0 License.

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 17 November 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Building Control
      2. Off-policy Evaluation
      3. Reinforcement Learning

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Conference

      BuildSys '20
      Sponsor:

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)56
      • Downloads (Last 6 weeks)6
      Reflects downloads up to 24 Sep 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2023)Real-World Implementation of Reinforcement Learning Based Energy Coordination for a Cluster of HouseholdsProceedings of the 10th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation10.1145/3600100.3625681(347-351)Online publication date: 15-Nov-2023
      • (2022)Reinforcement Learning in Construction Engineering and Management: A ReviewJournal of Construction Engineering and Management10.1061/(ASCE)CO.1943-7862.0002386148:11Online publication date: Nov-2022
      • (2021)Enforcing Policy Feasibility Constraints through Differentiable Projection for Energy OptimizationProceedings of the Twelfth ACM International Conference on Future Energy Systems10.1145/3447555.3464874(199-210)Online publication date: 22-Jun-2021

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media