Active Power Correction Strategies Based on Deep Reinforcement Learning - Part II: A Distributed Solution for Adaptability

12Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This article is the second part of Active Power Correction Strategies Based on Deep Reinforcement Learning. In Part II, we consider the renewable energy scenarios plugged into the large-scale power grid and provide an adaptive algorithmic implementation to maintain power grid stability. Based on the robustness method in Part I, a distributed deep reinforcement learning method is proposed to overcome the influence of the increasing renewable energy penetration. A multi-agent system is implemented in multiple control areas of the power system, which conducts a fully cooperative stochastic game. Based on the Monte Carlo tree search mentioned in Part I, we select practical actions in each sub-control area to search the Nash equilibrium of the game. Based on the QMIX method, a structure of offline centralized training and online distributed execution is proposed to employ better practical actions in the active power correction control. Our proposed method is evaluated in the modified global competition scenario cases of '2020 Learning to Run a Power Network - Neurips Track 2'.

Cite

CITATION STYLE

APA

Chen, S., Duan, J., Bai, Y., Zhang, J., Shi, D., Wang, Z. W., … Sun, Y. (2022). Active Power Correction Strategies Based on Deep Reinforcement Learning - Part II: A Distributed Solution for Adaptability. CSEE Journal of Power and Energy Systems, 8(4), 1134–1144. https://doi.org/10.17775/CSEEJPES.2020.07070

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free