Review and comparison of commonly used activation functions for deep neural networks

133Citations
Citations of this article
352Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The primary neural networks’ decision-making units are activation functions. Moreover, they evaluate the output of networks neural node; thus, they are essential for the performance of the whole network. Hence, it is critical to choose the most appropriate activation function in neural networks calculation. Acharya et al. (2018) suggest that numerous recipes have been formulated over the years, though some of them are considered deprecated these days since they are unable to operate properly under some conditions. These functions have a variety of characteristics, which are deemed essential to successfully learning. Their monotonicity, individual derivatives, and finite of their range are some of these characteristics. This research paper will evaluate the commonly used additive functions, such as swish, ReLU, Sigmoid, and so forth. This will be followed by their properties, own cons and pros, and particular formula application recommendations.

Cite

CITATION STYLE

APA

Szandała, T. (2021). Review and comparison of commonly used activation functions for deep neural networks. In Studies in Computational Intelligence (Vol. 903, pp. 203–224). Springer. https://doi.org/10.1007/978-981-15-5495-7_11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free