Improving Efficiency in Neural Network Accelerator using Operands Hamming Distance Optimization

1Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Neural network accelerator is a key enabler for the on-device AI inference, for which energy efficiency is an important metric. The datapath energy, including the computation energy and the data movement energy among the arithmetic units, claims a significant part of the total accelerator energy. By revisiting the basic physics of the arithmetic logic circuits, we show that the datapath energy is highly correlated with the bit flips when streaming the input operands into the arithmetic units, defined as the hamming distance (HD) of the input operand matrices. Based on the insight, we propose a post-training optimization algorithm and a HD-aware training algorithm to co-design and co-optimize the accelerator and the network synergistically. The experimental results based on post-layout simulation with MobileNetV2 demonstrate on average 2.85 datapath energy reduction and up to 8.51 datapath energy reduction for certain layers.

Cite

CITATION STYLE

APA

Li, M., Li, Y., & Chandra, V. (2021). Improving Efficiency in Neural Network Accelerator using Operands Hamming Distance Optimization. In Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC (pp. 599–604). Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1145/3394885.3446242

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free