Asymmetric sigmoidal activation function for feed-forward artificial neural networks

0Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Artificial neural networks of the feed – forward kind, are an established technique under the supervised learning paradigm for the solution of learning tasks. The mathematical result that allows one to assert the usefulness of this technique is that these networks can approximate any continuous function to the desired degree. The requirement imposed on these networks is to have non-linear functions of a specific kind at the hidden nodes of the network. In general, sigmoidal non-linearities, called activation functions, are generally used. In this paper we propose an asymmetric activation function. The networks using the proposed activation function are compared against those using the generally used logistic and the hyperbolic tangent activation function for the solution of 12 function approximation problems. The results obtained allow us to infer that the proposed activation function, in general, reaches deeper minima of the error measures and has better generalization error values.

Cite

CITATION STYLE

APA

Sehrawat, R., Chandra, P., & Ghose, U. (2019). Asymmetric sigmoidal activation function for feed-forward artificial neural networks. International Journal of Innovative Technology and Exploring Engineering, 8(12), 852–858. https://doi.org/10.35940/ijitee.L3310.1081219

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free