Abstract
While reinforcement learning (RL) on humans has shown incredible promise, it often suffers from a scarcity of data and few steps. In instances like these, a planning model of human behavior may greatly help. We present an experimental setup for the development and testing of an Soft Actor Critic (SAC) V2 RL architecture for several different neural architectures for planning models: an autoML optimized LSTM, an OLS, and a baseline model. We present the effects of including a planning model in agent learning within a simulation of the office, currently reporting a limited success with the LSTM.
Author supplied keywords
Cite
CITATION STYLE
Spangher, L., Gokul, A., Khattar, M., Palakapilly, J., Agwan, U., Tawade, A., & Spanos, C. (2020). Augmenting Reinforcement Learning with a Planning Model for Optimizing Energy Demand Response. In RLEM 2020 - Proceedings of the 1st International Workshop on Reinforcement Learning for Energy Management in Buildings and Cities (pp. 39–42). Association for Computing Machinery, Inc. https://doi.org/10.1145/3427773.3427863
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.