Abstract :
Multitask learning has been thoroughly proven to improve the generalization performance given a set of related tasks. Most multitask learning algorithm assume that all tasks are related. However, if all the tasks are not related, negative transfer of information occurs amongst the tasks, and the performance of traditional multitask learning algorithm worsens. Thus, we design an algorithm that simultaneously groups the related tasks and trains only the related task together. There are different approaches to train the related tasks in multi-task learning based on which information is shared across the tasks. These approaches either assume that the parameters of each of the tasks are situated close together, or assume that there is a common underlying latent space in the features of the tasks that is related. Most multi-task learning algorithm use either regularization method or matrix-variate priors. In our algorithm, the related tasks are tied together by a set of common features selected by each tasks. Thus, to train the related tasks together, we use spike and slab prior to select a common set of features for the related tasks, and a mixture of gaussians prior to select the set of related tasks. For validation, the developed algorithm is tested on toxicity prediction and hand written digit recognition data sets. The results show a significant improvement over multitask learning with feature selection for larger number of tasks. Further, the developed algorithm is also compared against another state of the art algorithm that similarly groups the related tasks together and proven to be better and more accurate.
Keywords :
Gaussian processes; feature extraction; handwritten character recognition; learning (artificial intelligence); mixture models; pattern classification; toxicology; Gaussian mixture; feature selection; generalization performance; hand written digit recognition data sets; information sharing; matrix-variate prior; multitask learning algorithm; regularization method; slab; spike; toxicity prediction; Bayes methods; Chemicals; Equations; Mathematical model; Slabs; Training; Vectors; Expectation Propagation; Groups of Tasks; Mixture of Gaussian Prior; Multi-task Learning; Spike and Slab Prior;