Title :
A Regularized Maximum Figure-of-Merit (rMFoM) Approach to Supervised and Semi-Supervised Learning
Author :
Ma, Chengyuan ; Lee, Chin-Hui
Author_Institution :
Nuance Commun., Burlington, MA, USA
fDate :
7/1/2011 12:00:00 AM
Abstract :
We propose a regularized extension to supervised maximum figure-of-merit learning to improve its generalization capability and successfully extend it to semi-supervised learning. The proposed method can be used to approximate any objective function consisting of the commonly used performance metrics. We first derive detailed learning algorithms for supervised learning problems and then extend it to more general semi-supervised scenarios, where only a small part of the training data is labeled. The effectiveness of the proposed approach is justified by several text categorization experiments on different datasets. The novelty of this paper lies in several aspects: 1) Tikhonov regularization is used to alleviate potential overfitting of the maximum figure-of-merit criteria; 2) the regularized maximum figure-of-merit algorithm is successfully extended to semi-supervised learning tasks; 3) the proposed approach has good scalability to large-scale applications.
Keywords :
learning (artificial intelligence); text analysis; Tikhonov regularization; figure-of-merit criteria; generalization capability; learning algorithm; regularized maximum figure-of-merit algorithm; semi-supervised learning task; supervised learning problem; supervised maximum figure-of-merit learning; text categorization; Approximation methods; Error analysis; Hidden Markov models; Measurement; Support vector machines; Text categorization; Training; Deterministic annealing; discriminative training; maximum figure-of-merit; regularization; semi-supervised learning; supervised learning; text categorization; trust region method;
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
DOI :
10.1109/TASL.2010.2090146