DocumentCode
3690906
Title
Shared feature representations of LiDAR and optical images: Trading sparsity for semantic discrimination
Author
Manuel Campos-Taberner;Adriana Romero;Carlo Gatta;Gustau Camps-Valls
Author_Institution
Universitat de Valè
fYear
2015
fDate
7/1/2015 12:00:00 AM
Firstpage
4169
Lastpage
4172
Abstract
This paper studies the level of complementary information conveyed by extremely high resolution LiDAR and optical images. We pursue this goal following an indirect approach via unsupervised spatial-spectral feature extraction. We used a recently presented unsupervised convolutional neural network trained to enforce both population and lifetime spar-sity in the feature representation. We derived independent and joint feature representations, and analyzed the sparsity scores and the discriminative power. Interestingly, the obtained results revealed that the RGB+LiDAR representation is no longer sparse, and the derived basis functions merge color and elevation yielding a set of more expressive colored edge filters. The joint feature representation is also more discriminative when used for clustering and topological data visualization.
Keywords
"Laser radar","Feature extraction","Sociology","Statistics","Joints","Image color analysis","Semantics"
Publisher
ieee
Conference_Titel
Geoscience and Remote Sensing Symposium (IGARSS), 2015 IEEE International
ISSN
2153-6996
Electronic_ISBN
2153-7003
Type
conf
DOI
10.1109/IGARSS.2015.7326744
Filename
7326744
Link To Document