DocumentCode
3409166
Title
Growing semantically meaningful models for visual SLAM
Author
Flint, Alex ; Mei, Christopher ; Reid, Ian ; Murray, David
Author_Institution
Dept. Eng. Sci., Univ. of Oxford, Oxford, UK
fYear
2010
fDate
13-18 June 2010
Firstpage
467
Lastpage
474
Abstract
Though modern Visual Simultaneous Localisation and Mapping (vSLAM) systems are capable of localising robustly and efficiently even in the case of a monocular camera, the maps produced are typically sparse point-clouds that are difficult to interpret and of little use for higher-level reasoning tasks such as scene understanding or human- machine interaction. In this paper we begin to address this deficiency, presenting progress on expanding the competency of visual SLAM systems to build richer maps. Specifically, we concentrate on modelling indoor scenes using semantically meaningful surfaces and accompanying labels, such as “floor”, “wall”, and “ceiling” - an important step towards a representation that can support higher-level reasoning and planning. We leverage the Manhattan world assumption and show how to extract vanishing directions jointly across a video stream. We then propose a guided line detector that utilises known vanishing points to extract extremely subtle axis- aligned edges. We utilise recent advances in single view structure recovery to building geometric scene models and demonstrate our system operating on-line.
Keywords
SLAM (robots); computer vision; video streaming; geometric scene; monocular camera; reasoning tasks; sparse point-clouds; video stream; visual SLAM; visual simultaneous localisation and mapping; Buildings; Cameras; Clouds; Floors; Image edge detection; Image reconstruction; Layout; Photometry; Simultaneous localization and mapping; Surface reconstruction;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on
Conference_Location
San Francisco, CA
ISSN
1063-6919
Print_ISBN
978-1-4244-6984-0
Type
conf
DOI
10.1109/CVPR.2010.5540176
Filename
5540176
Link To Document