Fully dynamic reorder policies with deep reinforcement learning for multi-echelon inventory management

4Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The operation of inventory systems plays an important role in the success of manufacturing companies, making it a highly relevant domain for optimization. In particular, the domain lends itself to being approached via Deep Reinforcement Learning (DRL) models due to it requiring sequential reorder decisions based on uncertainty to minimize cost. In this paper, we evaluate state-of-the-art optimization approaches to determine whether Deep Reinforcement Learning can be applied to the multi-echelon inventory optimization (MEIO) framework in a practically feasible manner to generate fully dynamic reorder policies. We investigate how it performs in comparison to an optimized static reorder policy, how robust it is when it comes to structural changes in the environment, and whether the use of DRL is safe in terms of risk in real-world applications. Our results show promising performance for DRL with potential for improvement in terms of minimizing risky behavior.

Cite

CITATION STYLE

APA

Hammler, P., Riesterer, N., & Braun, T. (2023). Fully dynamic reorder policies with deep reinforcement learning for multi-echelon inventory management. Informatik-Spektrum, 46(5–6), 240–251. https://doi.org/10.1007/s00287-023-01556-6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free