DocumentCode :
1869956
Title :
Learning action dictionaries from video
Author :
Turaga, Pavan ; Chellappa, Rama
Author_Institution :
Center for Autom. Res., Univ. of Maryland, College Park, MD
fYear :
2008
fDate :
12-15 Oct. 2008
Firstpage :
1704
Lastpage :
1707
Abstract :
Summarizing the contents of a video containing human activities is an important problem in computer vision and has important applications in automated surveillance systems. Summarizing a video requires one to identify and learn a ´vocabulary´ of action-phrases corresponding to specific events and actions occurring in the video. We propose a generative model for dynamic scenes containing human activities as a composition of independent action-phrases - each of which is derived from an underlying vocabulary. Given a long video sequence, we propose a completely unsupervised approach to learn the vocabulary. Once the vocabulary is learnt, a video segment can be decomposed into a collection of phrases for summarization. We then describe methods to learn the correlations between activities and sequentiality of events. We also propose a novel method for building invariances to spatial transforms in the summarization scheme.
Keywords :
computer vision; image segmentation; learning (artificial intelligence); video surveillance; automated surveillance systems; computer vision; independent action-phrases; learning action dictionaries; spatial transforms; video segment decomposition; video sequence; Application software; Automation; Computer vision; Dictionaries; Educational institutions; Humans; Layout; Surveillance; Video sequences; Vocabulary; Activity Analysis; Video Summarization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on
Conference_Location :
San Diego, CA
ISSN :
1522-4880
Print_ISBN :
978-1-4244-1765-0
Electronic_ISBN :
1522-4880
Type :
conf
DOI :
10.1109/ICIP.2008.4712102
Filename :
4712102
Link To Document :
بازگشت