DocumentCode :
3778722
Title :
Is noise always harmful? Visual learning from weakly-related data
Author :
Sheng-hua Zhong;Yan Liu;Kien A. Hua;Songtao Wu
Author_Institution :
College of Computer Science and Software Engineering, Shenzhen University, China
fYear :
2015
Firstpage :
181
Lastpage :
184
Abstract :
Noise exists universally in multimedia data, especially in Internet era. For example, tags from web users are often incomplete, arbitrary, and low relevant with the visual information. Intuitively, noise in the dataset is harmful to learning tasks, which implies that huge volumes of image tags from social media can´t be utilized directly. To collect the reliable training dataset, labor-intensive manual labeling and various learning based outlier detection techniques are widely used. This paper intends to discuss whether such kind of preprocessing is always needed. We focus on a very normal case in image classification that the available dataset includes a large amount of images weakly related to any target classes. We use deep models as the platform and design a series of experiments to compare the semi-supervised learning performance with/without weakly related unlabeled data. Fortunately, we validate that weakly related data is not always harmful, which is an encouraging finding for research on web image learning.
Keywords :
"Training","Standards","Machine learning","Erbium","Data models","Multimedia communication","Training data"
Publisher :
ieee
Conference_Titel :
Orange Technologies (ICOT), 2015 International Conference on
Type :
conf
DOI :
10.1109/ICOT.2015.7498518
Filename :
7498518
Link To Document :
بازگشت