Title :
From video to text: Semantic driving scene understanding using a coarse-to-fine method
Author :
Fu, Huiyuan ; Ma, Huadong
Author_Institution :
Beijing Key Lab. of Intell. Telecommun. Software & Multimedia, Beijing Univ. of Posts & Telecommun., Beijing, China
Abstract :
Semantic understanding from video is one of the most challenging tasks in video analysis. However, it has not been taken enough attention. In this paper, we focus on understanding the semantics of video in the driving scene. We present a coarse-to-fine method to parse the driving scene, and obtain the high-level semantic information of the scene. In the coarse phase, we divide the captured frame into four separate parts based on edge density entropy and scene context. In the fine phase, we join multi-class object segmentation and detection algorithms together in a unified Conditional Random Filed (CRF) model for each part understanding. Moreover, the object probabilistic location prior knowledge based on training and previous edge density entropy result is also integrated into our approach for better object localization. Experimental results show that our proposed method is effective comparing to current state-of-the-art approaches.
Keywords :
edge detection; entropy; image segmentation; object detection; video signal processing; CRF model; coarse-to-fine method; conditional random filed model; edge density entropy; high-level semantic information; multiclass object detection algorithm; multiclass object segmentation algorithm; semantic driving scene; video analysis; Computer vision; Conferences; Entropy; Image segmentation; Probabilistic logic; Semantics; Training; Conditional Random Filed; Semantic understanding; detection; multi-class segmentation;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2012.6288151