Resource usage of production workloads running on shared compute clusters often fluctuate significantly across time. While simultaneous spike in the resource usage between two workloads running on the same machine can create performance degradation, unused resources in a machine results in wastage and undesirable operational characteristics for a compute cluster. Prior works did not consider such temporal resource fluctuations or their alignment for scheduling decisions. Due to the variety of time-varying workloads and their complex resource usage characteristics, it is challenging to design well-defined heuristics for scheduling them optimally across different machines in a cluster. In this paper, we propose a Deep Reinforcement Learning (DRL) based approach to exploit various temporal resource usage patterns of time-varying workloads as well as a technique for creating equivalence classes among a large number of production workloads to improve scalability of our method. Validations with real production traces from Google and Alibaba show that our technique can significantly improve metrics for operational excellence (e.g. utilization, fragmentation, resource exhaustion etc.) for a cluster compared to the baselines.
CITATION STYLE
Mondal, S. S., Sheoran, N., & Mitra, S. (2021). Scheduling of Time-Varying Workloads Using Reinforcement Learning. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 10B, pp. 9000–9008). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i10.17088
Mendeley helps you to discover research relevant for your work.