DocumentCode :
1942984
Title :
Using Co-Occurrence and Segmentation to Learn Feature-Based Object Models from Video
Author :
Stepleton, Thomas ; Lee, Tai Sing
Author_Institution :
Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA
Volume :
1
fYear :
2005
fDate :
5-7 Jan. 2005
Firstpage :
129
Lastpage :
134
Abstract :
A number of recent systems for unsupervised feature- based learning of object models take advantage of cooccurrence: broadly, they search for clusters of discriminative features that tend to coincide across multiple still images or video frames. An intuition behind these efforts is that regularly co-occurring image features are likely to refer to physical traits of the same object, while features that do not often co-occur are more likely to belong to different objects. In this paper we discuss a refinement to these techniques in which multiple segmentations establish meaningful contexts for co-occurrence, or limit the spatial regions in which two features are deemed to co-occur. This approach can reduce the variety of image data necessary for model learning and simplify the incorporation of less discriminative features into the model.
Keywords :
feature extraction; image segmentation; learning (artificial intelligence); video signal processing; co-occurring image features; feature-based object models; multiple segmentations; unsupervised feature-based learning; video frames; Bandwidth; Cognition; Computer vision; Filtering; Frequency measurement; Image segmentation; Noise reduction; Robots; Training data; Video sequences;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Application of Computer Vision, 2005. WACV/MOTIONS '05 Volume 1. Seventh IEEE Workshops on
Conference_Location :
Breckenridge, CO
Print_ISBN :
0-7695-2271-8
Type :
conf
DOI :
10.1109/ACVMOT.2005.119
Filename :
4129471
Link To Document :
بازگشت