Benchmarking Comparison of Swish vs. Other Activation Functions on CIFAR-10 Imageset

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The choice of the most appropriate activation functions for artificial neural networks has a significant effect on the training time and task performance. Nowadays the most widely-used activation function is the Rectified Linear Unit (ReLU). Despite its “dying ReLU problem” and many attempts to replace it with something better, it is still considered as default choice to begin with creation of network. Two years ago a new, promising function has been described formulated by Google Brain Team. The proposed function - named Swish - was obtained using a combination of exhaustive and reinforcement learning-based search. According to the authors, simply replacing ReLUs with Swish units improves top-1 classification accuracy on ImageNet by 0.9% for Mobile NASNet-A and 0.6% for Inception-ResNet-v2. This paper describes an experiment on CIFAR-10 image set where Swish appears not to outperform ReLU.

Cite

CITATION STYLE

APA

Szandała, T. (2020). Benchmarking Comparison of Swish vs. Other Activation Functions on CIFAR-10 Imageset. In Advances in Intelligent Systems and Computing (Vol. 987, pp. 498–505). Springer Verlag. https://doi.org/10.1007/978-3-030-19501-4_49

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free