DocumentCode :
254698
Title :
A Piggyback Representation for Action Recognition
Author :
Wolf, Lars ; Hanani, Yair ; Hassner, Tal
Author_Institution :
Balvatnik Sch. of Comput. Sci., Tel Aviv Univ., Tel Aviv, Israel
fYear :
2014
fDate :
23-28 June 2014
Firstpage :
520
Lastpage :
525
Abstract :
In video understanding, the spatial patterns formed by local space-time interest points hold discriminative information. We encode these spatial regularities using a word2vec neural network, a recently proposed tool in the field of text processing. Then, building upon recent accumulator based image representation solutions, input videos are represented in a hybrid manner: the appearance of local space time interest points is used to collect and associate the learned descriptors, which capture the spatial patterns. Promising results are shown on recent action recognition benchmarks, using well established methods as the underlying appearance descriptors.
Keywords :
image representation; neural nets; object recognition; video signal processing; accumulator based image representation solutions; action recognition benchmarks; appearance descriptors; discriminative information; input videos; learned descriptors; local space time interest points; local space-time interest points; piggyback representation; spatial patterns; text processing; video understanding; word2vec neural network; Benchmark testing; Dictionaries; Encoding; Training; Trajectory; Vectors; Visualization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition Workshops (CVPRW), 2014 IEEE Conference on
Conference_Location :
Columbus, OH
Type :
conf
DOI :
10.1109/CVPRW.2014.81
Filename :
6910030
Link To Document :
بازگشت