A multi-objective multi-agent deep reinforcement learning approach to residential appliance scheduling

18Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.

Abstract

Residential buildings are large consumers of energy. They contribute significantly to the demand placed on the grid, particularly during hours of peak demand. Demand-side management is crucial to reducing this demand placed on the grid and increasing renewable utilisation. This research study presents a multi-objective tunable deep reinforcement learning algorithm for demand-side management of household appliances. The proposed tunable Deep Q-Network (DQN) algorithm learns a single policy that accounts for different preferences for multiple objectives present when scheduling appliances. These include electricity cost, peak demand, and punctuality. The tunable Deep Q-Network algorithm is compared to two rule-based approaches for appliance scheduling. When comparing the 1-month simulation results for the tunable DQN with an electricity cost rule-based benchmark method, the tunable DQN agent provides a statistically significant improvement of 30%, 18.2%, and 37.3% for the cost, peak power, and punctuality objectives. Moreover, the tunable Deep Q-Network can produce a range of appliance scheduling policies for different objective preferences without requiring any computationally intensive retraining. This is the key advantage of the proposed tunable Deep Q-Network algorithm for appliance scheduling.

Cite

CITATION STYLE

APA

Lu, J., Mannion, P., & Mason, K. (2022). A multi-objective multi-agent deep reinforcement learning approach to residential appliance scheduling. IET Smart Grid, 5(4), 260–280. https://doi.org/10.1049/stg2.12068

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free