Title :
Fast AdaBoost training using weighted novelty selection
Author :
Seyedhosseini, Mojtaba ; Paiva, António R C ; Tasdizen, Tolga
Author_Institution :
Sci. Comput. & Imaging Inst., Univ. of Utah, Salt Lake City, UT, USA
fDate :
July 31 2011-Aug. 5 2011
Abstract :
In this paper, a new AdaBoost learning framework, called WNS-AdaBoost, is proposed for training discriminative models. The proposed approach significantly speeds up the learning process of adaptive boosting (AdaBoost) by reducing the number of data points. For this purpose, we introduce the weighted novelty selection (WNS) sampling strategy and combine it with AdaBoost to obtain an efficient and fast learning algorithm. WNS selects a representative subset of data thereby reducing the number of data points onto which AdaBoost is applied. In addition, WNS associates a weight with each selected data point such that the weighted subset approximates the distribution of all the training data. This ensures that AdaBoost can trained efficiently and with minimal loss of accuracy. The performance of WNS-AdaBoost is first demonstrated in a classification task. Then, WNS is employed in a probabilistic boosting-tree (PBT) structure for image segmentation. Results in these two applications show that the training time using WNS-AdaBoost is greatly reduced at the cost of only a few percent in accuracy.
Keywords :
image segmentation; learning (artificial intelligence); sampling methods; trees (mathematics); AdaBoost learning framework; PBT structure; WNS sampling strategy; WNS-AdaBoost; fast AdaBoost training; image segmentation; probabilistic boosting-tree structure; weighted novelty selection; Accuracy; Boosting; Image segmentation; Kernel; Probabilistic logic; Testing; Training;
Conference_Titel :
Neural Networks (IJCNN), The 2011 International Joint Conference on
Conference_Location :
San Jose, CA
Print_ISBN :
978-1-4244-9635-8
DOI :
10.1109/IJCNN.2011.6033366