DocumentCode
3210281
Title
Alignment of continuous video onto 3D point clouds
Author
Zhao, W. ; Nister, D. ; Hsu, S.
Author_Institution
Sarnoff Corp., Princeton, NJ, USA
Volume
2
fYear
2004
fDate
June 27 2004-July 2 2004
Abstract
We propose a general framework for aligning continuous (oblique) video onto 3D sensor data. We align a point cloud computed from the video onto the point cloud directly obtained from a 3D sensor. This is in contrast to existing techniques where the 2D images are aligned to a 3D model derived from the 3D sensor data. Using point clouds enables the alignment for scenes full of objects that are difficult to model, for example, trees. To compute 3D point clouds from video, motion stereo is used along with a state-of-the-art algorithm for camera pose estimation. Our experiments with real data demonstrate the advantages of the proposed registration algorithm for texturing models in large-scale semi-urban environments. The capability to align video before a 3D model is built from the 3D sensor data opens up new possibilities for 3D modeling. We introduce a novel modeling-through-registration approach that fuses 3D information from both the 3D sensor and the video. Initial experiments with real data illustrate the potential of the proposed approach.
Keywords
image sequences; multidimensional signal processing; 3D point clouds; 3D sensor data; camera pose estimation; continuous video alignment; modeling-through-registration approach; registration algorithm; Cameras; Computer vision; Geometry; Image sensors; Iterative closest point algorithm; Laser radar; Layout; Solid modeling; Stereo vision; Three-dimensional displays;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Vision and Pattern Recognition, 2004. CVPR 2004. Proceedings of the 2004 IEEE Computer Society Conference on
Conference_Location
Washington, DC, USA
ISSN
1063-6919
Print_ISBN
0-7695-2158-4
Type
conf
DOI
10.1109/CVPR.2004.1315269
Filename
1315269
Link To Document