DocumentCode :
561188
Title :
Hybrid Evolution of Convolutional Networks
Author :
Cheung, Brian ; Sable, Carl
Author_Institution :
Dept. of Electr. Eng., Cooper Union, New York, NY, USA
Volume :
1
fYear :
2011
fDate :
18-21 Dec. 2011
Firstpage :
293
Lastpage :
297
Abstract :
With the increasing trend of neural network models towards larger structures with more layers, we expect a corresponding exponential increase in the number of possible architectures. In this paper, we apply a hybrid evolutionary search procedure to define the initialization and architectural parameters of convolutional networks, one of the first successful deep network models. We make use of stochastic diagonal Levenberg-Marquardt to accelerate the convergence of training, lowering the time cost of fitness evaluation. Using parameters found from the evolutionary search together with absolute value and local contrast normalization preprocessing between layers, we achieve the best known performance on several of the MNIST Variations, rectangles-image and convex image datasets.
Keywords :
evolutionary computation; learning (artificial intelligence); neural nets; search problems; MNIST Variations; architectural parameters; convex image datasets; convolutional networks; hybrid evolution; hybrid evolutionary search procedure; initialization parameters; neural network models; rectangles image datasets; stochastic diagonal Levenberg-Marquardt; Computer architecture; Jacobian matrices; Machine learning; Neural networks; Noise reduction; Object recognition; Training; convolutional networks; evolution; image classification; neural networks; second order methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning and Applications and Workshops (ICMLA), 2011 10th International Conference on
Conference_Location :
Honolulu, HI
Print_ISBN :
978-1-4577-2134-2
Type :
conf
DOI :
10.1109/ICMLA.2011.73
Filename :
6146987
Link To Document :
بازگشت