DocumentCode :
3346393
Title :
Combining Distributed Classifies by Stacking
Author :
Wei Yanyan ; Li Taoshen ; Ge Zhihui
Author_Institution :
Sch. of Math. & Comput. Sci., Guangxi Univ. for Nat., Nanning, China
fYear :
2009
fDate :
14-17 Oct. 2009
Firstpage :
418
Lastpage :
421
Abstract :
Many current mining tasks analyze data in environments with distributed computing nodes. Classification in such scenario needs to perform local mining task in each data site and then integrate local classifiers to a global model of the data. However, integration strategy can influence the performance and complexity of the final model. In this paper, based on the formalization of combining multiple classifiers by stacking in Distributed Data Mining, a new strategy to from meta-level training set is proposed, which can describe the vote made by each base-level classifiers. The experiment results show that our method achieve better performance for those datasets with highly skewed class distribution.
Keywords :
data mining; distributed processing; pattern classification; base-level classifiers; distributed classifiers; distributed computing; distributed data mining; metalevel training set; Accuracy; Costs; Data mining; Distributed computing; Distributed decision making; Genetics; Mathematics; Probability distribution; Stacking; Voting; DDM; classification; combining classifiers; stacking;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Genetic and Evolutionary Computing, 2009. WGEC '09. 3rd International Conference on
Conference_Location :
Guilin
Print_ISBN :
978-0-7695-3899-0
Type :
conf
DOI :
10.1109/WGEC.2009.146
Filename :
5402861
Link To Document :
بازگشت