DocumentCode
324546
Title
Generalization and comparison of Alopex learning algorithm and random optimization method for neural networks
Author
Peng, Pei-Yuan ; Sirag, David
Author_Institution
United Technol. Res. Center, East Hartford, CT, USA
Volume
2
fYear
1998
fDate
4-9 May 1998
Firstpage
1147
Abstract
For the minimum description of length, a penalty term is added to the cost function to reduce the network´s complexity. Alopex learning and random optimization method based neural networks are investigated. The Alopex algorithm is a stochastic learning algorithm for training neural networks of any topology, including feedback loops. The neurons are not restricted to any transfer function and the learning can use any error norm measure. The random optimization method by Matyas (1965) and its modified algorithm are studied and compared with the Alopex algorithm to some adaptive control problems. Simulation results show the pros and cons between two
Keywords
adaptive control; backpropagation; generalisation (artificial intelligence); neural nets; optimisation; parallel algorithms; Alopex learning algorithm; adaptive control; backpropagation; neural networks; random optimization; stochastic learning algorithm; underwater vehicles; Adaptive control; Cathode ray tubes; Cost function; Feedback loop; Network topology; Neural networks; Neurons; Optimization methods; Silver; Stochastic processes;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks Proceedings, 1998. IEEE World Congress on Computational Intelligence. The 1998 IEEE International Joint Conference on
Conference_Location
Anchorage, AK
ISSN
1098-7576
Print_ISBN
0-7803-4859-1
Type
conf
DOI
10.1109/IJCNN.1998.685934
Filename
685934
Link To Document