Title :
Hybrid Evolution of Convolutional Networks
Author :
Cheung, Brian ; Sable, Carl
Author_Institution :
Dept. of Electr. Eng., Cooper Union, New York, NY, USA
Abstract :
With the increasing trend of neural network models towards larger structures with more layers, we expect a corresponding exponential increase in the number of possible architectures. In this paper, we apply a hybrid evolutionary search procedure to define the initialization and architectural parameters of convolutional networks, one of the first successful deep network models. We make use of stochastic diagonal Levenberg-Marquardt to accelerate the convergence of training, lowering the time cost of fitness evaluation. Using parameters found from the evolutionary search together with absolute value and local contrast normalization preprocessing between layers, we achieve the best known performance on several of the MNIST Variations, rectangles-image and convex image datasets.
Keywords :
evolutionary computation; learning (artificial intelligence); neural nets; search problems; MNIST Variations; architectural parameters; convex image datasets; convolutional networks; hybrid evolution; hybrid evolutionary search procedure; initialization parameters; neural network models; rectangles image datasets; stochastic diagonal Levenberg-Marquardt; Computer architecture; Jacobian matrices; Machine learning; Neural networks; Noise reduction; Object recognition; Training; convolutional networks; evolution; image classification; neural networks; second order methods;
Conference_Titel :
Machine Learning and Applications and Workshops (ICMLA), 2011 10th International Conference on
Conference_Location :
Honolulu, HI
Print_ISBN :
978-1-4577-2134-2
DOI :
10.1109/ICMLA.2011.73