Title :
Hybrid Dimensionality Reduction Method Based on Support Vector Machine and Independent Component Analysis
Author :
Sangwoo Moon ; Hairong Qi
Author_Institution :
Dept. of Electr. Eng. & Comput. Sci., Univ. of Tennessee, Knoxville, TN, USA
fDate :
5/1/2012 12:00:00 AM
Abstract :
This paper presents a new hybrid dimensionality reduction method to seek projection through optimization of both structural risk (supervised criterion) and data independence (unsupervised criterion). Classification accuracy is used as a metric to evaluate the performance of the method. By minimizing the structural risk, projection originated from the decision boundaries directly improves the classification performance from a supervised perspective. From an unsupervised perspective, projection can also be obtained based on maximum independence among features (or attributes) in data to indirectly achieve better classification accuracy over more intrinsic representation of the data. Orthogonality interrelates the two sets of projections such that minimum redundancy exists between the projections, leading to more effective dimensionality reduction. Experimental results show that the proposed hybrid dimensionality reduction method that satisfies both criteria simultaneously provides higher classification performance, especially for noisy data sets, in relatively lower dimensional space than various existing methods.
Keywords :
data reduction; independent component analysis; minimisation; pattern classification; risk management; support vector machines; classification accuracy; classification performance improvement; data independence; decision boundaries; hybrid dimensionality reduction method; independent component analysis; minimum redundancy; noisy data sets; structural risk minimization; supervised criterion; support vector machine; unsupervised criterion; Correlation; Kernel; Principal component analysis; Risk management; Robustness; Support vector machines; Vectors; Hybrid dimensionality reduction; independence maximization; projection; structural risk minimization;
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
DOI :
10.1109/TNNLS.2012.2189581