Augmenting Reinforcement Learning with a Planning Model for Optimizing Energy Demand Response

13Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

While reinforcement learning (RL) on humans has shown incredible promise, it often suffers from a scarcity of data and few steps. In instances like these, a planning model of human behavior may greatly help. We present an experimental setup for the development and testing of an Soft Actor Critic (SAC) V2 RL architecture for several different neural architectures for planning models: an autoML optimized LSTM, an OLS, and a baseline model. We present the effects of including a planning model in agent learning within a simulation of the office, currently reporting a limited success with the LSTM.

Cite

CITATION STYLE

APA

Spangher, L., Gokul, A., Khattar, M., Palakapilly, J., Agwan, U., Tawade, A., & Spanos, C. (2020). Augmenting Reinforcement Learning with a Planning Model for Optimizing Energy Demand Response. In RLEM 2020 - Proceedings of the 1st International Workshop on Reinforcement Learning for Energy Management in Buildings and Cities (pp. 39–42). Association for Computing Machinery, Inc. https://doi.org/10.1145/3427773.3427863

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free