Neural network based recommendation models are widely used to power many internet-scale applications including product recommendation and feed ranking. As the models become more complex and more training data is required during training, improving the training scalability of these recommendation models becomes an urgent need. However, improving the scalability without sacrificing the model quality is challenging. In this paper, we conduct an in-depth analysis of the scalability bottleneck in existing training architecture on large scale CPU clusters. Based on these observations, we propose a new training architecture called Hierarchical Training, which exploits both data parallelism and model parallelism for the neural network part of the model within a group. We implement hierarchical training with a two-layer design: a tagging system that decides the operator placement and a net transformation system that materializes the training plans, and integrate hierarchical training into existing training stack. We propose several optimizations to improve the scalability of hierarchical training including model architecture optimization, communication compression, and various system-level improvements. Extensive experiments at massive scale demonstrate that hierarchical training can speed up distributed recommendation model training by 1.9x without model quality drop.
CITATION STYLE
Huang, Y., Wei, X., Wang, X., Yang, J., Su, B. Y., Bharuka, S., … Langman, J. (2021). Hierarchical Training: Scaling Deep Recommendation Models on Large CPU Clusters. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 3050–3058). Association for Computing Machinery. https://doi.org/10.1145/3447548.3467084
Mendeley helps you to discover research relevant for your work.