Title :
Reconstructable generalized maximum scatter difference discriminant analysis
Author :
Kai Huang ; Liqing Zhang
Author_Institution :
Dept. of Comput. Sci. & Eng., Shanghai Jiao Tong Univ., Shanghai, China
Abstract :
Dimensionality reduction is a key preprocessing step for many applications. Until our knowledge, unsupervised approaches such as PCA and ICA do not take label information of the original data into account, so a supervised approach such as Linear discriminant analysis (LDA) performs better on many classification tasks. Unfortunately, the classical LDA approach has shortcomings, such as the well-known small size problem, the heteroscedastic problem and the (C-1) low rank problem. The (C-1) low rank problem greatly limits the dimension of the extracted features. In addition, the calculation of the between-class and within-class scatter matrices in the classical LDA approach actually only takes account of the Mahalanobis distance like covariance distance of data centers and each data class, so if the dataset has very few classes or the data distribution of each class is not Gaussian-like but has some spatial structure in the feature space instead, classical LDA does not work well. In this paper we propose a dimensionality reduction approach which avoids the limitations of classical LDA and improves handling of the between-class scatter matrix. Our approach approach takes the distribution of data in each class into consideration to calculate the projection matrix. It does not assume that the data distribution of each class approximates Gaussian; each can have its own spatial structure. Experiments show that our method can obtain better projection directions than the classical LDA approach and greatly improve the classification accuracy. In addition, our approach is able to reconstruct the original signal well, while the classical LDA approach ignores the reconstruction property.
Keywords :
electrocardiography; feature extraction; matrix algebra; medical signal processing; signal classification; statistical analysis; C-1 low rank problem; ICA; LDA approach; Mahalanobis distance; PCA; between-class scatter matrix; data centers; data class; dimensionality reduction approach; feature extraction; heteroscedastic problem; independent component analysis; linear discriminant analysis; maximum scatter difference discriminant analysis; principal component analysis; small size problem; within-class scatter matrix; Covariance matrices; Eigenvalues and eigenfunctions; Equations; Feature extraction; Mathematical model; Principal component analysis; Vectors;
Conference_Titel :
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-6627-1
DOI :
10.1109/IJCNN.2014.6889486