Title :
Using AUC and accuracy in evaluating learning algorithms
Author :
Huang, Jin ; Ling, Charles X.
Author_Institution :
Dept. of Comput. Sci., Univ. of Western Ontario, London, Ont., Canada
fDate :
3/1/2005 12:00:00 AM
Abstract :
The area under the ROC (receiver operating characteristics) curve, or simply AUC, has been traditionally used in medical diagnosis since the 1970s. It has recently been proposed as an alternative single-number measure for evaluating the predictive ability of learning algorithms. However, no formal arguments were given as to why AUC should be preferred over accuracy. We establish formal criteria for comparing two different measures for learning algorithms and we show theoretically and empirically that AUC is a better measure (defined precisely) than accuracy. We then reevaluate well-established claims in machine learning based on accuracy using AUC and obtain interesting and surprising new results. For example, it has been well-established and accepted that Naive Bayes and decision trees are very similar in predictive accuracy. We show, however, that Naive Bayes is significantly better than decision trees in AUC. The conclusions drawn in this paper may make a significant impact on machine learning and data mining applications.
Keywords :
Bayes methods; data mining; decision trees; learning (artificial intelligence); pattern classification; statistical analysis; AUC; Naive Bayes; data mining; decision trees; learning algorithms; machine learning; medical diagnosis; receiver operating characteristics curve; Accuracy; Classification algorithms; Data mining; Decision trees; Error analysis; Helium; Machine learning; Machine learning algorithms; Medical diagnosis; Testing;
Journal_Title :
Knowledge and Data Engineering, IEEE Transactions on
DOI :
10.1109/TKDE.2005.50