DocumentCode
917037
Title
The AIC Criterion and Symmetrizing the Kullback–Leibler Divergence
Author
Seghouane, Abd-Krim ; Amari, Shun-Ichi
Author_Institution
Canberra Res. Lab., Nat. ICT Australia, Canberra, ACT
Volume
18
Issue
1
fYear
2007
Firstpage
97
Lastpage
106
Abstract
The Akaike information criterion (AIC) is a widely used tool for model selection. AIC is derived as an asymptotically unbiased estimator of a function used for ranking candidate models which is a variant of the Kullback-Leibler divergence between the true model and the approximating candidate model. Despite the Kullback-Leibler´s computational and theoretical advantages, what can become inconvenient in model selection applications is their lack of symmetry. Simple examples can show that reversing the role of the arguments in the Kullback-Leibler divergence can yield substantially different results. In this paper, three new functions for ranking candidate models are proposed. These functions are constructed by symmetrizing the Kullback-Leibler divergence between the true model and the approximating candidate model. The operations used for symmetrizing are the average, geometric, and harmonic means. It is found that the original AIC criterion is an asymptotically unbiased estimator of these three different functions. Using one of these proposed ranking functions, an example of new bias correction to AIC is derived for univariate linear regression models. A simulation study based on polynomial regression is provided to compare the different proposed ranking functions with AIC and the new derived correction with AICc
Keywords
information theory; linear systems; polynomials; regression analysis; Akaike information criterion; Kullback-Leibler divergence; asymptotically unbiased estimation; polynomial regression; univariate linear regression models; Australia Council; Bayesian methods; Laboratories; Least squares approximation; Linear regression; Maximum likelihood estimation; Parameter estimation; Polynomials; Probability; Solid modeling; Akaike information criterion (AIC); Kullback–Leibler divergence; geometric and harmonic means; model selection; Algorithms; Artificial Intelligence; Databases, Factual; Information Storage and Retrieval; Models, Theoretical; Pattern Recognition, Automated;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/TNN.2006.882813
Filename
4049836
Link To Document