Title :
On comparison of adaptive regularization methods
Author :
Sigurdsson, Sigurdur ; Larsen, Jan ; Hansen, Lars Kai
Author_Institution :
Dept. of Math. Modeling, Tech. Univ. Denmark, Lyngby, Denmark
Abstract :
Modeling with flexible models, such as neural networks, requires careful control of the model complexity and generalization ability of the resulting model which finds expression in the ubiquitous bias-variance dilemma. Regularization is a tool for optimizing the model structure reducing variance at the expense of introducing extra bias. The overall objective of adaptive regularization is to tune the amount of regularization ensuring minimal generalization error. Regularization is a supplement to direct model selection techniques like step-wise selection and one would prefer a hybrid scheme; however, a very flexible regularization may substitute the need for selection procedures. This paper investigates recently suggested adaptive regularization schemes. Some methods focus directly on minimizing an estimate of the generalization error (either algebraic or empirical), whereas others start from different criteria, e.g., the Bayesian evidence. The evidence expresses basically the probability of the model, which is conceptually different from generalization error; however, asymptotically for large training data sets they will converge. First the basic model definition, training and generalization is presented. Next, different adaptive regularization schemes are reviewed and extended. Finally, the experimental section presents a comparative study concerning linear models for regression/time series problems
Keywords :
generalisation (artificial intelligence); neural nets; probability; statistical analysis; time series; Bayesian evidence; adaptive regularization methods; bias-variance dilemma; experiment; generalization; large training data sets; model complexity; model selection; neural networks; probability; regression; step-wise selection; time series; Bayesian methods; Cost function; Electronic mail; Loss measurement; Mathematical model; Neural networks; Predictive models; Signal processing; Training data; Vectors;
Conference_Titel :
Neural Networks for Signal Processing X, 2000. Proceedings of the 2000 IEEE Signal Processing Society Workshop
Conference_Location :
Sydney, NSW
Print_ISBN :
0-7803-6278-0
DOI :
10.1109/NNSP.2000.889413