Actor-critic reinforcement learning for energy optimization in hybrid production environments

5Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

This paper presents a centralized approach for energy optimization in large scale industrial production systems based on an actor-critic reinforcement learning (ACRL) framework. The objective of the on-line capable self-learning algorithm is the optimization of the energy consumption of a production process while meeting certain manufacturing constraints like a demanded throughput. Our centralized ACRL algorithm works with two artificial neural networks (ANN) for function approximation using Gaussian radial-basis functions (RBF), one for the critic and another for the actor, respectively. This kind of actorcritic design enables the handling of both, a discrete and continuous state and action space, which is essential for hybrid systems where discrete and continuous actuator behavior is combined. The ACRL algorithm is exemplary validated on a dynamic simulation model of a bulk good system for the task of supplying bulk good to a subsequent dosing section while consuming as low energy as possible. The simulation results clearly show the applicability and capability of our machine learning (ML) approach for energy optimization in hybrid production environments.

Cite

CITATION STYLE

APA

Schwung, D., Schwung, A., & Ding, S. X. (2019). Actor-critic reinforcement learning for energy optimization in hybrid production environments. International Journal of Computing, 18(4), 360–371. https://doi.org/10.47839/ijc.18.4.1607

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free