DocumentCode :
3669621
Title :
Egocentric activity recognition using Histograms of Oriented Pairwise Relations
Author :
Ardhendu Behera;Matthew Chapman;Anthony G. Cohn;David C. Hogg
Author_Institution :
School of Computing, University of Leeds, LS2 9JT, U.K.
Volume :
2
fYear :
2014
Firstpage :
22
Lastpage :
30
Abstract :
This paper presents an approach for recognising activities using video from an egocentric (first-person view) setup. Our approach infers activity from the interactions of objects and hands. In contrast to previous approaches to activity recognition, we do not require to use an intermediate such as object detection, pose estimation, etc. Recently, it has been shown that modelling the spatial distribution of visual words corresponding to local features further improves the performance of activity recognition using the bag-of-visual words representation. Influenced and inspired by this philosophy, our method is based on global spatio-temporal relationships between visual words. We consider the interaction between visual words by encoding their spatial distances, orientations and alignments. These interactions are encoded using a histogram that we name the Histogram of Oriented Pairwise Relations (HOPR). The proposed approach is robust to occlusion and background variation and is evaluated on two challenging egocentric activity datasets consisting of manipulative task. We introduce a novel representation of activities based on interactions of local features and experimentally demonstrate its superior performance in comparison to standard activity representations such as bag-of-visual words.
Keywords :
"Feature extraction","Visualization","Histograms","Detectors","Wrist","Graphical models","Distribution functions"
Publisher :
ieee
Conference_Titel :
Computer Vision Theory and Applications (VISAPP), 2014 International Conference on
Type :
conf
Filename :
7294910
Link To Document :
بازگشت