DocumentCode :
1749237
Title :
Input decay: simple and effective soft variable selection
Author :
Chapados, Nicolas ; Bengio, Yoshua
Author_Institution :
Dept. of Comput. Sci. & Oper. Res., Montreal Univ., Que., Canada
Volume :
2
fYear :
2001
fDate :
2001
Firstpage :
1233
Abstract :
To deal with the overfitting problems that occur when there are not enough examples compared to the number of input variables in supervised learning, traditional approaches are weight decay and greedy variable selection. An alternative that has recently started to attract attention is to keep all the variables but to put more emphasis on the “most useful” ones. We introduce a new regularization method called input decay that exerts more relative penalty on the parameters associated with the inputs that contribute less to the learned function. This method, like weight decay and variable selection, still requires to perform a kind of model selection. Successful comparative experiments with this new method were performed both on a simulated regression task and a real-world financial prediction task
Keywords :
financial data processing; learning (artificial intelligence); learning systems; financial prediction; input decay; model selection; overfitting; relative penalty; soft variable selection; supervised learning; Computational modeling; Computer networks; Computer science; Input variables; Linear regression; Machine learning algorithms; Neural networks; Operations research; Predictive models; Supervised learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.939537
Filename :
939537
Link To Document :
بازگشت