DocumentCode :
2286409
Title :
Empirical modeling of very large data sets using neural networks
Author :
Owens, Aaron J.
Author_Institution :
DuPont Central Res. & Dev., Wilmington, DE, USA
Volume :
6
fYear :
2000
fDate :
2000
Firstpage :
302
Abstract :
Building empirical predictive models from very large data sets is challenging. One has to deal both with the `curse of dimensionality´ (hundreds or thousands of variables) and with `too many records´ (many thousands of instances). While neural networks [Rumelhart, et al., 1986] are widely recognized as universal function approximators [Cybenko, 1989], their training time rapidly increases with the number of variables and instances. I discuss practical methods for overcoming this problem so that neural network models can be developed for very large databases. The methods include: Dimensionality reduction with neural net modeling, PLS modeling, and bottleneck neural networks; Sub-sampling and re-sampling with many smaller data sets to reduce training time; Committee of networks to make the final prediction more robust and to estimate its uncertainty
Keywords :
database theory; learning (artificial intelligence); neural nets; very large databases; PLS modeling; bottleneck neural networks; committee of networks; dimensionality reduction; neural network models; neural networks; predictive models; universal function approximators; very large data sets; very large databases; Arithmetic; Artificial neural networks; Databases; Feedforward neural networks; Input variables; Neural networks; Predictive models; Research and development; Robustness; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
ISSN :
1098-7576
Print_ISBN :
0-7695-0619-4
Type :
conf
DOI :
10.1109/IJCNN.2000.859413
Filename :
859413
Link To Document :
بازگشت