Memory capacity is a key bottleneck for training large scale neural networks. Intel® Optane™ DC PMM (persistent memory modules) which are available as NVDIMMs are a disruptive technology that promises significantly higher read bandwidth than traditional SSDs at a lower cost per bit than traditional DRAM. In this work we show how to take advantage of this new memory technology to minimize the amount of DRAM required without compromising performance significantly. Specifically, we take advantage of the static nature of the underlying computational graphs in deep neural network applications to develop a profile guided optimization based on Integer Linear Programming (ILP) called AutoTM to optimally assign and move live tensors to either DRAM or NVDIMMs. Our approach can replace 50% to 80% of a system's DRAM with PMM while only losing a geometric mean 27.7% performance. This is a significant improvement over first-touch NUMA, which loses 71.9% of performance. The proposed ILP based synchronous scheduling technique also provides 2x performance over using DRAM as a hardware-controlled cache for very large networks.
CITATION STYLE
Hildebrand, M., Khan, J., Trika, S., Lowe-Power, J., & Akella, V. (2020). AutOTM: Automatic tensor movement in heterogeneous memory systems using integer linear programming. In International Conference on Architectural Support for Programming Languages and Operating Systems - ASPLOS (pp. 875–890). Association for Computing Machinery. https://doi.org/10.1145/3373376.3378465
Mendeley helps you to discover research relevant for your work.