Title :
Generalized Multiple Kernel Learning With Data-Dependent Priors
Author :
Qi Mao ; Tsang, Ivor W. ; Shenghua Gao ; Li Wang
Author_Institution :
Dept. of Electr. & Comput. Eng., Duke Univ., Durham, NC, USA
Abstract :
Multiple kernel learning (MKL) and classifier ensemble are two mainstream methods for solving learning problems in which some sets of features/views are more informative than others, or the features/views within a given set are inconsistent. In this paper, we first present a novel probabilistic interpretation of MKL such that maximum entropy discrimination with a noninformative prior over multiple views is equivalent to the formulation of MKL. Instead of using the noninformative prior, we introduce a novel data-dependent prior based on an ensemble of kernel predictors, which enhances the prediction performance of MKL by leveraging the merits of the classifier ensemble. With the proposed probabilistic framework of MKL, we propose a hierarchical Bayesian model to learn the proposed data-dependent prior and classification model simultaneously. The resultant problem is convex and other information (e.g., instances with either missing views or missing labels) can be seamlessly incorporated into the data-dependent priors. Furthermore, a variety of existing MKL models can be recovered under the proposed MKL framework and can be readily extended to incorporate these priors. Extensive experiments demonstrate the benefits of our proposed framework in supervised and semisupervised settings, as well as in tasks with partial correspondence among multiple views.
Keywords :
belief networks; generalisation (artificial intelligence); learning (artificial intelligence); maximum entropy methods; pattern classification; MKL model; classification model; classifier ensemble; data dependent prior; generalized multiple kernel learning; hierarchical Bayesian model; maximum entropy discrimination; noninformative prior; probabilistic framework; semisupervised learning; supervised learning; Bayes methods; Indexes; Kernel; Learning systems; Probabilistic logic; Training data; Vectors; Data fusion; dirty data; missing views; multiple kernel learning; partial correspondence; semisupervised learning;
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
DOI :
10.1109/TNNLS.2014.2334137