DocumentCode :
2260299
Title :
The inefficiency of batch training for large training sets
Author :
Wilson, D. Randall ; Martinez, Tony R.
Author_Institution :
Fonix Corp., USA
Volume :
2
fYear :
2000
fDate :
2000
Firstpage :
113
Abstract :
Multilayer perceptrons are often trained using error backpropagation (BP). BP training can be done in either a batch or continuous manner. Claims have frequently been made that batch training is faster and/or more “correct” than continuous training because it uses a better approximation of the true gradient for its weight updates. These claims are often supported by empirical evidence on very small data sets. These claims are untrue, however, for large training sets. This paper explains why batch training is much slower than continuous training for large training sets. Various levels of semi-batch training used on a 20,000-instance speech recognition task show a roughly linear increase in training time required with an increase in batch size
Keywords :
backpropagation; multilayer perceptrons; speech recognition; batch learning; error backpropagation; multilayer perceptrons; speech recognition; weight updates; Backpropagation; Computer networks; Equations; Frequency; Multilayer perceptrons; Speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
ISSN :
1098-7576
Print_ISBN :
0-7695-0619-4
Type :
conf
DOI :
10.1109/IJCNN.2000.857883
Filename :
857883
Link To Document :
بازگشت