شماره ركورد كنفرانس :
3297
عنوان مقاله :
Leveraging Deep Learning Representation for Search-based Image Annotation
عنوان به زبان ديگر :
Leveraging Deep Learning Representation for Search-based Image Annotation
پديدآورندگان :
Mohammadi Kashani Mahya Department of Computer Engineering Shahid Rajaee Teacher Training University Tehran , Amiri S. Hamid Department of Computer Engineering Shahid Rajaee Teacher Training University Tehran
كليدواژه :
Nearest Neighbors , CNN , Deep Learning , Relevant Tags , Image Annotation
عنوان كنفرانس :
نوزدهمين سمپوزيوم بين المللي هوش مصنوعي و پردازش سيگنال
چكيده لاتين :
Image annotation aims to assign some tags to an image
such that these tags provide a textual description for the content
of image. Search-based methods extract relevant tags for an image
based on the tags of nearest neighbor images in the training set. In
these methods, similarity of two images is determined based on the
distance between feature vectors of the images. Thus, it is essential
to extract informative feature vectors from images. In this paper,
we propose a framework that utilize deep learning to obtain visual
representation of images. We apply different architectures of
convolutional neural networks (CNN) to the input image and
obtain a single feature vector that is a rich representation for
visual content of the image. In this way, we eliminate the usage of
multiple feature vectors used in the state-of-the-art annotation
methods. We also integrate our feature extractors with a nearest
neighbors approach to obtain relevant tags of an image. Our
experiments on the standard datasets of image annotation
(including Corel5k, ESP Game, IAPR) demonstrate that our
approach reaches higher precision, recall and F1 than the state-ofthe-
art methods such as 2PKNN, TagProp, NMF-KNN and etc.