Title :
Information Discriminant Analysis: Feature Extraction with an Information-Theoretic Objective
Author_Institution :
Univ. of California, Irvine
Abstract :
Using elementary information-theoretic tools, we develop a novel technique for linear transformation from the space of observations into a low-dimensional (feature) subspace for the purpose of classification. The technique is based on a numerical optimization of an information-theoretic objective function, which can be computed analytically. The advantages of the proposed method over several other techniques are discussed and the conditions under which the method reduces to linear discriminant analysis are given. We show that the novel objective function enjoys many of the properties of the mutual information and the Bayes error and we give sufficient conditions for the method to be Bayes-optimal. Since the objective function is maximized numerically, we show how the calculations can be accelerated to yield feasible solutions. The performance of the method compares favorably to other linear discriminant-based feature extraction methods on a number of simulated and real-world data sets.
Keywords :
Bayes methods; feature extraction; information theory; optimisation; pattern classification; Bayes error; Bayes-optimal; feature extraction; information discriminant analysis; information-theoretic objective function; linear discriminant analysis; linear transformation; low-dimensional subspace; numerical optimization; Acceleration; Data mining; Feature extraction; Information analysis; Information theory; Linear discriminant analysis; Mutual information; Principal component analysis; Sufficient conditions; Vectors; Bayes error.; Feature extraction; classification; entropy; information theory; linear discriminant analysis; mutual information;
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
DOI :
10.1109/TPAMI.2007.1156