DocumentCode :
3748681
Title :
Semantically-Aware Aerial Reconstruction from Multi-modal Data
Author :
Randi Cabezas;Julian Straub;John W. Fisher
fYear :
2015
Firstpage :
2156
Lastpage :
2164
Abstract :
We consider a methodology for integrating multiple sensors along with semantic information to enhance scene representations. We propose a probabilistic generative model for inferring semantically-informed aerial reconstructions from multi-modal data within a consistent mathematical framework. The approach, called Semantically-Aware Aerial Reconstruction (SAAR), not only exploits inferred scene geometry, appearance, and semantic observations to obtain a meaningful categorization of the data, but also extends previously proposed methods by imposing structure on the prior over geometry, appearance, and semantic labels. This leads to more accurate reconstructions and the ability to fill in missing contextual labels via joint sensor and semantic information. We introduce a new multi-modal synthetic dataset in order to provide quantitative performance analysis. Additionally, we apply the model to real-world data and exploit OpenStreetMap as a source of semantic observations. We show quantitative improvements in reconstruction accuracy of large-scale urban scenes from the combination of LiDAR, aerial photography, and semantic data. Furthermore, we demonstrate the model´s ability to fill in for missing sensed data, leading to more interpretable reconstructions.
Keywords :
"Semantics","Three-dimensional displays","Geometry","Image reconstruction","Laser radar","Solid modeling","Probabilistic logic"
Publisher :
ieee
Conference_Titel :
Computer Vision (ICCV), 2015 IEEE International Conference on
Electronic_ISBN :
2380-7504
Type :
conf
DOI :
10.1109/ICCV.2015.249
Filename :
7410606
Link To Document :
بازگشت