DocumentCode :
3429422
Title :
Complementary feature splits for co-training
Author :
Salaheldin, Ahmed ; El-Gayar, Neamat
Author_Institution :
Center for Inf. Sci., Nile Univ., Giza, Egypt
fYear :
2012
fDate :
2-5 July 2012
Firstpage :
1303
Lastpage :
1308
Abstract :
In many data mining and machine learning applications, data may be easy to collect. However, labeling the data is often expensive, time consuming or difficult. Such applications give rise to semi-supervised learning techniques that combine the use of labelled and unlabelled data. Co-training is a popular semi-supervised learning algorithm that depends on splitting the features of a data set into two redundant and independent views. In many cases however such sets of features are not naturally present in the data or are unknown. In this paper we test feature splitting methods based on maximizing the confidence and the diversity of the views using genetic algorithms, and compare their performance against random splits. We also propose a new criterion that maximizes the complementary nature of the views. Experimental results on six different data sets show that our optimized splits enhance the performance of co-training over random splits and that the complementary split outperforms the confidence, diversity and random splits.
Keywords :
data mining; genetic algorithms; learning (artificial intelligence); complementary feature splits; cotraining; data mining; feature splitting methods; genetic algorithms; independent views; machine learning; optimized splits; random splits; semisupervised learning; unlabelled data; Accuracy; Breast cancer; Entropy; Genetic algorithms; Radial basis function networks; Supervised learning; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Information Science, Signal Processing and their Applications (ISSPA), 2012 11th International Conference on
Conference_Location :
Montreal, QC
Print_ISBN :
978-1-4673-0381-1
Electronic_ISBN :
978-1-4673-0380-4
Type :
conf
DOI :
10.1109/ISSPA.2012.6310494
Filename :
6310494
Link To Document :
بازگشت