DocumentCode :
253389
Title :
A skewed derivative activation function for SFFANNs
Author :
Chandra, P. ; Sodhi, Sartaj Singh
Author_Institution :
Sch. of Inf. & Commun. Technol., Guru Gobind Singh Indraprastha Univ., New Delhi, India
fYear :
2014
fDate :
9-11 May 2014
Firstpage :
1
Lastpage :
6
Abstract :
In the current paper, a new activation function is proposed for usage in constructing sigmoidal feedforward artificial neural networks. The suitability of the proposed activation function is established. The proposed activation function has a skewed derivative whereas the usually utilized activation functions derivatives are symmetric about the y-axis (as for the log-sigmoid or the hyperbolic tangent function). The efficiency and efficacy of the usage of the proposed activation function is demonstrated on six function approximation tasks. The obtained results indicate that if a network using the proposed activation function in the hidden layer, is trained then it converges to deeper minima of the error functional, generalizes better and converges faster as compared to networks using the standard log-sigmoidal activation function at the hidden layer.
Keywords :
feedforward neural nets; function approximation; transfer functions; SFFANNs; function approximation tasks; hyperbolic tangent function; log-sigmoidal activation function; sigmoidal feedforward artificial neural networks; skewed derivative activation function; Function approximation; Neural networks; Silicon; Standards; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Recent Advances and Innovations in Engineering (ICRAIE), 2014
Conference_Location :
Jaipur
Print_ISBN :
978-1-4799-4041-7
Type :
conf
DOI :
10.1109/ICRAIE.2014.6909324
Filename :
6909324
Link To Document :
بازگشت