Artificial neural networks of the feed – forward kind, are an established technique under the supervised learning paradigm for the solution of learning tasks. The mathematical result that allows one to assert the usefulness of this technique is that these networks can approximate any continuous function to the desired degree. The requirement imposed on these networks is to have non-linear functions of a specific kind at the hidden nodes of the network. In general, sigmoidal non-linearities, called activation functions, are generally used. In this paper we propose an asymmetric activation function. The networks using the proposed activation function are compared against those using the generally used logistic and the hyperbolic tangent activation function for the solution of 12 function approximation problems. The results obtained allow us to infer that the proposed activation function, in general, reaches deeper minima of the error measures and has better generalization error values.
CITATION STYLE
Sehrawat, R., Chandra, P., & Ghose, U. (2019). Asymmetric sigmoidal activation function for feed-forward artificial neural networks. International Journal of Innovative Technology and Exploring Engineering, 8(12), 852–858. https://doi.org/10.35940/ijitee.L3310.1081219
Mendeley helps you to discover research relevant for your work.