Title :
Automatic video annotation through search and mining
Author :
Moxley, Emily ; Mei, Tao ; Hua, Xian-Sheng ; Ma, Wei-Ying ; Manjunath, B.S.
Author_Institution :
Vision Res. Lab., Univ. of California, Santa Barbara, CA
fDate :
June 23 2008-April 26 2008
Abstract :
Conventional approaches to video annotation predominantly focus on supervised identification of a limited set of concepts, while unsupervised annotation with infinite vocabulary remains unexplored. This work aims to exploit the overlap in content of news video to automatically annotate by mining similar videos that reinforce, filter, and improve the original annotations. The algorithm employs a two-step process of search followed by mining. Given a query video consisting of visual content and speech-recognized transcripts, similar videos are first ranked in a multimodal search. Then, the transcripts associated with these similar videos are mined to extract keywords for the query. We conducted extensive experiments over the TRECVID 2005 corpus and showed the superiority of the proposed approach to using only the mining process on the original video for annotation. This work represents the first attempt at unsupervised automatic video annotation leveraging overlapping video content.
Keywords :
data mining; video retrieval; video signal processing; multimodal search; query keywords; speech-recognized transcripts; supervised identification; unsupervised annotation; unsupervised automatic video annotation; video mining; video query; video search; visual content transcripts; Asia; Automatic speech recognition; Data mining; Image databases; Machine learning; Noise figure; Support vector machines; Tagging; Video sharing; Vocabulary; data mining; video annotation; video search;
Conference_Titel :
Multimedia and Expo, 2008 IEEE International Conference on
Conference_Location :
Hannover
Print_ISBN :
978-1-4244-2570-9
Electronic_ISBN :
978-1-4244-2571-6
DOI :
10.1109/ICME.2008.4607527