DocumentCode :
1748947
Title :
A novel adaptive activation function
Author :
Xu, Shuxiang ; Zhang, Ming
Author_Institution :
Sch. of Comput., Tasmania Univ., Launceston, Tas., Australia
Volume :
4
fYear :
2001
fDate :
2001
Firstpage :
2779
Abstract :
This paper deals with an experimental justification of a novel adaptive activation function for feedforward neural networks (FNNs). Simulation results reveal that FNNs with the proposed adaptive activation function present several advantages over traditional neuron-fixed feedforward networks such as much reduced network size, faster learning, and lessened approximation errors. Following the definition of the neuron-adaptive activation function, we conduct experiments with function approximation and financial data simulation, and depict the experimental outcomes that exhibit the advantages of FNN with our neuron-adaptive activation function over traditional FNN with fixed activation function
Keywords :
feedforward neural nets; financial data processing; function approximation; learning (artificial intelligence); transfer functions; adaptive activation function; feedforward neural networks; financial data processing; function approximation; learning; Artificial intelligence; Australia; Bismuth; Computational modeling; Computer networks; Electronic mail; Feedforward neural networks; Feedforward systems; Function approximation; Neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.938813
Filename :
938813
Link To Document :
بازگشت