Hardware Implementation of a Softmax-Like Function for Deep Learning †

25Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.

Abstract

In this paper a simplified hardware implementation of a CNN softmax-like layer is proposed. Initially the softmax activation function is analyzed in terms of required numerical accuracy and certain optimizations are proposed. A proposed adaptable hardware architecture is evaluated in terms of the introduced error due to the proposed softmax-like function. The proposed architecture can be adopted to the accuracy required by the application by retaining or eliminating certain terms of the approximation thus allowing to explore accuracy for complexity trade-offs. Furthermore, the proposed circuits are synthesized in a 90 nm 1.0 V CMOS standard-cell library using Synopsys Design Compiler. Comparisons reveal that significant reduction is achieved in area × delay and power × delay products for certain cases, respectively, over prior art. Area and power savings are achieved with respect to performance and accuracy.

Cite

CITATION STYLE

APA

Kouretas, I., & Paliouras, V. (2020). Hardware Implementation of a Softmax-Like Function for Deep Learning †. Technologies, 8(3). https://doi.org/10.3390/technologies8030046

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free