DocumentCode :
2928961
Title :
Using Different Cost Functions to Train Stacked Auto-Encoders
Author :
Amaral, T. ; Silva, Lynette M. ; Alexandre, Luis A. ; Kandaswamy, Chetak ; Santos, Jorge M. ; de Sa, Joaquim Marques
Author_Institution :
Inst. de Eng. Biomed. (INEB), Univ. do Porto, Porto, Portugal
fYear :
2013
fDate :
24-30 Nov. 2013
Firstpage :
114
Lastpage :
120
Abstract :
Deep neural networks comprise several hidden layers of units, which can be pre-trained one at a time via an unsupervised greedy approach. A whole network can then be trained (fine-tuned) in a supervised fashion. One possible pre-training strategy is to regard each hidden layer in the network as the input layer of an auto-encoder. Since auto-encoders aim to reconstruct their own input, their training must be based on some cost function capable of measuring reconstruction performance. Similarly, the supervised fine-tuning of a deep network needs to be based on some cost function that reflects prediction performance. In this work we compare different combinations of cost functions in terms of their impact on layer-wise reconstruction performance and on supervised classification performance of deep networks. We employed two classic functions, namely the cross-entropy (CE) cost and the sum of squared errors (SSE), as well as the exponential (EXP) cost, inspired by the error entropy concept. Our results were based on a number of artificial and real-world data sets.
Keywords :
entropy; greedy algorithms; learning (artificial intelligence); neural nets; pattern classification; CE cost function; EXP cost function; SSE cost function; artificial data sets; cross-entropy cost function; deep neural network pretraining strategy; error entropy concept; exponential cost function; hidden layer; input layer reconstruction; layer-wise reconstruction performance measurement; real-world data sets; stacked autoencoder training; sum of squared error cost function; supervised classification performance; unsupervised greedy approach; Cost function; Equations; Mathematical model; Neurons; Training; Tuning; Vectors; cost functions; deep neural networks; stacked auto-encoders;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Artificial Intelligence (MICAI), 2013 12th Mexican International Conference on
Conference_Location :
Mexico City
Print_ISBN :
978-1-4799-2604-6
Type :
conf
DOI :
10.1109/MICAI.2013.20
Filename :
6714656
Link To Document :
بازگشت