Emerging data intensive AI/ML workloads encounter memory and power wall when run on general-purpose compute cores. This has led to the development of a myriad of techniques to deal with such workloads, among which DNN accelerator architectures have found a prominent place. In this work, we propose a hardware-software co-design approach to achieve system-level benefits. We propose a quantized data-aware POSIT number representation that leads to a highly optimized DNN accelerator. We demonstrate this work on SOTA SIMBA architecture, extendable to any other accelerator. Our proposal reduces the buffer/storage requirements within the architecture and reduces the data transfer cost between the main memory and the DNN accelerator. We have investigated the impact of using integer, IEEE floating point, and posit multipliers for LeNet, ResNet and VGG NNs trained and tested on MNIST, CIFAR10 and ImageNet datasets, respectively. Our system-level analysis shows that the proposed approximate-fixed-posit multiplier when implemented on SIMBA architecture, achieves on average ∼2.2× speed up, consumes ∼3.1× less energy and requires ∼3.2× less area, respectively, against the baseline SOTA architecture, without loss of accuracy (∼±1%).
CITATION STYLE
Glint, T., Prasad, K., Dagli, J., Gandhi, K., Gupta, A., Patel, V., … Mekie, J. (2023). Hardware-Software Codesign of DNN Accelerators Using Approximate Posit Multipliers. In Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC (pp. 469–474). Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1145/3566097.3567866
Mendeley helps you to discover research relevant for your work.