Reinforcement learning for control of flexibility providers in a residential microgrid

38Citations
Citations of this article
81Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can be used to learn system dynamics. This study explores two model-free reinforcement learning (RL) techniques - policy iteration (PI) and fitted Q-iteration (FQI) for scheduling the operation of flexibility providers - battery and heat pump in a residential microgrid. The proposed algorithms are data-driven and can be easily generalised to fit the control of any flexibility provider without requiring expert knowledge to build a detailed model of the flexibility provider and/or microgrid. The algorithms are tested in multi-agent collaborative and single-agent stochastic microgrid settings - with the uncertainty due to lack of knowledge on future electricity consumption patterns and photovoltaic production. Simulation results show that PI outperforms FQI with a 7.2% increase in photovoltaic self-consumption in the multi-agent setting and a 3.7% increase in the single-agent setting. Both RL algorithms perform better than a rule-based controller, and compete with a model-based optimal controller, and are thus, a valuable alternative to model- and rule-based controllers.

Cite

CITATION STYLE

APA

Mbuwir, B. V., Geysen, D., Spiessens, F., & Deconinck, G. (2020). Reinforcement learning for control of flexibility providers in a residential microgrid. IET Smart Grid, 3(1), 98–107. https://doi.org/10.1049/iet-stg.2019.0196

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free