DocumentCode :
285141
Title :
Universal property of learning curves under entropy loss
Author :
Amari, Shun-Ichi
Author_Institution :
Fac. of Eng., Tokyo Univ., Japan
Volume :
2
fYear :
1992
fDate :
7-11 Jun 1992
Firstpage :
368
Abstract :
A learning curve shows how fast a learning machine improves it behaviour as the number of training examples increases. A study of the universal asymptotic behaviour of learning curves for general dichotomy machines is presented. It is proved rigorously that the average predictive entropy <e*(t)> converges to zero as <e*(t)>~d/t as the number of t of training examples increases, where d is the number of modifiable parameters of a machine, irrespectively of the architecture of the machine
Keywords :
learning (artificial intelligence); neural nets; average predictive entropy; entropy loss; general dichotomy machines; learning curves; modifiable parameters; training examples; universal asymptotic behaviour; universal property; Annealing; Bayesian methods; Entropy; Error correction; Machine learning; Neural networks; Probability distribution; Stochastic processes;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
Type :
conf
DOI :
10.1109/IJCNN.1992.226960
Filename :
226960
Link To Document :
بازگشت