DocumentCode :
114326
Title :
BiModal latent dirichlet allocation for text and image
Author :
Xiaofeng Liao ; Qingshan Jiang ; Wei Zhang ; Kai Zhang
Author_Institution :
Shenzhen Inst. of Adv. Technol., Shenzhen, China
fYear :
2014
fDate :
26-28 April 2014
Firstpage :
736
Lastpage :
739
Abstract :
A BiModal Latent Dirichlet Allocation Model(BM-LDA) is proposed to learn a unified representation of data that comes from both the textual and visual modalities together. The model is able to form a unified representation that mixs both the textual and visual modalities. Based on the assumption, that the images and its surrounding text share a same topic, the model learns a posterior probability density in the space of latent variable of topics that bridging over the observed multi modality inputs. It maps the high dimensional space consist of the observed variables from both modalities to a low dimensional space of topcis. Experimental result on ImageCLEF data set, which consists of bi-modality data of images and surrounding text, shows our new BM-LDA model can get a fine representation for the multi-modality data, which is useful for tasks such as retrieval and classification.
Keywords :
Internet; image classification; image representation; probability; text analysis; BM-LDA; BiModal latent Dirichlet allocation model; ImageCLEF data set; bimodality image data; bimodality text data; classification; observed multimodality inputs; posterior probability density; retrieval; textual modalities; unified data representation; visual modalities; Data models; Image classification; Kernel; Mathematical model; Resource management; Standards; Visualization; BiModal Latent Dirichlet Allocation; Image; Multiple Modality; Text;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Information Science and Technology (ICIST), 2014 4th IEEE International Conference on
Conference_Location :
Shenzhen
Type :
conf
DOI :
10.1109/ICIST.2014.6920582
Filename :
6920582
Link To Document :
بازگشت