DocumentCode :
249259
Title :
An augmented representation of activity in video using semantic-context information
Author :
Khoualed, Samir ; Chateau, Thierry ; Castellan, Umberto ; Samir, Chafik
Author_Institution :
ISIT, Univ. of Clermont, Clermont, France
fYear :
2014
fDate :
27-30 Oct. 2014
Firstpage :
4171
Lastpage :
4175
Abstract :
Learning and recognizing activity in videos is an especially important task in computer vision. However, it is hard to perform. In this paper, we propose a new method by combining local and global context information to extract a bag-of-words-like representation of a single space-time point. Each spacetime point is described by a bag of visual words that encodes its relationships with the remaining space-time points in the video, defining the space-time context. Experiments on the KTH benchmark of action recognition, show that our approach performs accurately compared to the state-of-the-art.
Keywords :
augmented reality; computer vision; video signal processing; augmented representation; bag-of-words-like representation; computer vision; global context information; local context information; semantic-context information; space-time context; Accuracy; Context; Shape; Support vector machines; Trajectory; Visualization; Vocabulary; Action recognition; SVM classification; semantic shape context; space-time interest point;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Image Processing (ICIP), 2014 IEEE International Conference on
Conference_Location :
Paris
Type :
conf
DOI :
10.1109/ICIP.2014.7025847
Filename :
7025847
Link To Document :
بازگشت