Title :
Video segmentation by tracing discontinuities in a trajectory embedding
Author :
Fragkiadaki, Katerina ; Zhang, Geng ; Shi, Jianbo
Author_Institution :
Univ. of Pennsylvania, Philadelphia, PA, USA
Abstract :
Our goal is to segment a video sequence into moving objects and the world scene. In recent work, spectral embedding of point trajectories based on 2D motion cues accumulated from their lifespans, has shown to outperform factorization and per frame segmentation methods for video segmentation. The scale and kinematic nature of the moving objects and the background scene determine how close or far apart trajectories are placed in the spectral embedding. Such density variations may confuse clustering algorithms, causing over-fragmentation of object interiors. Therefore, instead of clustering in the spectral embedding, we propose detecting discontinuities of embedding density between spatially neighboring trajectories. Detected discontinuities are strong indicators of object boundaries and thus valuable for video segmentation. We propose a novel embedding discretization process that recovers from over-fragmentations by merging clusters according to discontinuity evidence along inter-cluster boundaries. For segmenting articulated objects, we combine motion grouping cues with a center-surround saliency operation, resulting in “context-aware”, spatially coherent, saliency maps. Figure-ground segmentation obtained from saliency thresholding, provides object connectedness constraints that alter motion based trajectory affinities, by keeping articulated parts together and separating disconnected in time objects. Finally, we introduce Gabriel graphs as effective per frame superpixel maps for converting trajectory clustering to dense image segmentation. Gabriel edges bridge large contour gaps via geometric reasoning without over-segmenting coherent image regions. We present experimental results of our method that outperform the state-of-the-art in challenging motion segmentation datasets.
Keywords :
embedded systems; image motion analysis; image segmentation; object detection; spectral analysis; video signal processing; 2D motion cues; Gabriel edges; Gabriel graphs; articulated objects segmentation; background scene; center-surround saliency operation; clustering algorithms; coherent image regions; context-aware maps; contour gaps; dense image segmentation; density variations; detected discontinuity; discontinuity detection; discontinuity evidence; embedding density; embedding discretization process; factorization; figure-ground segmentation; frame segmentation methods; geometric reasoning; inter-cluster boundary; kinematic nature; motion based trajectory affinity; motion grouping cues; motion segmentation datasets; moving objects; object boundary; object connectedness constraints; object interiors; over-fragmentations; point trajectory; saliency maps; saliency thresholding; spatially coherent maps; spatially neighboring trajectory; spectral embedding; superpixel maps; time objects; tracing discontinuity; trajectory clustering; trajectory embedding; video segmentation; Argon; Detectors; Educational institutions; Image segmentation; Motion segmentation; Tin; Trajectory;
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on
Conference_Location :
Providence, RI
Print_ISBN :
978-1-4673-1226-4
Electronic_ISBN :
1063-6919
DOI :
10.1109/CVPR.2012.6247883