DocumentCode :
2985388
Title :
Multi-task Semi-supervised Semantic Feature Learning for Classification
Author :
Changying Du ; Fuzhen Zhuang ; Qing He ; Zhongzhi Shi
Author_Institution :
Key Lab. of Intell. Inf. Process., Inst. of Comput. Technol., Beijing, China
fYear :
2012
fDate :
10-13 Dec. 2012
Firstpage :
191
Lastpage :
200
Abstract :
Multi-task learning has proven to be useful to boost the learning of multiple related but different tasks. Meanwhile, latent semantic models such as LSA and LDA are popular and effective methods to extract discriminative semantic features of high dimensional dyadic data. In this paper, we present a method to combine these two techniques together by introducing a new matrix tri-factorization based formulation for semi-supervised latent semantic learning, which can incorporate labeled information into traditional unsupervised learning of latent semantics. Our inspiration for multi-task semantic feature learning comes from two facts, i.e., 1) multiple tasks generally share a set of common latent semantics, and 2) a semantic usually has a stable indication of categories no matter which task it is from. Thus to make multiple tasks learn from each other we wish to share the associations between categories and those common semantics among tasks. Along this line, we propose a novel joint Nonnegative matrix tri-factorization framework with the aforesaid associations shared among tasks in the form of a semantic-category relation matrix. Our new formulation for multi-task learning can simultaneously learn (1) discriminative semantic features of each task, (2) predictive structure and categories of unlabeled data in each task, (3) common semantics shared among tasks and specific semantics exclusive to each task. We give alternating iterative algorithm to optimize our objective and theoretically show its convergence. Finally extensive experiments on text data along with the comparison with various baselines and three state-of-the-art multi-task learning algorithms demonstrate the effectiveness of our method.
Keywords :
learning (artificial intelligence); matrix decomposition; multiprogramming; pattern classification; classification; discriminative semantic features; high dimensional dyadic data; joint nonnegative matrix trifactorization framework; latent semantic models; multi-task semi-supervised semantic feature learning; semantic-category relation matrix; Data mining; Data models; Feature extraction; Iterative methods; Joints; Optimization; Semantics; joint nonnegative matrix tri-factorization; multi-task learning; semantic feature learning; semi-supervised learning; text classification;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Data Mining (ICDM), 2012 IEEE 12th International Conference on
Conference_Location :
Brussels
ISSN :
1550-4786
Print_ISBN :
978-1-4673-4649-8
Type :
conf
DOI :
10.1109/ICDM.2012.15
Filename :
6413903
Link To Document :
بازگشت