4.7 Article

Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks

Journal

Publisher

ELSEVIER
DOI: 10.1016/j.isprsjprs.2017.11.011

Keywords

Deep learning; Remote sensing; Semantic mapping; Data fusion

Funding

  1. Total-ONERA research project NAOMI

Ask authors/readers for more resources

In this work, we investigate various methods to deal with semantic labeling of very high resolution multi modal remote sensing data. Especially, we study how deep fully convolutional networks can be adapted to deal with multi-modal and multi-scale remote sensing data for semantic labeling. Our contributions are threefold: (a) we present an efficient multi-scale approach to leverage both a large spatial context and the high resolution data, (b) we investigate early and late fusion of Lidar and multispectral data, (c) we validate our methods on two public datasets with state-of-the-art results. Our results indicate that late fusion make it possible to recover errors steaming from ambiguous data, while early fusion allows for better joint-feature learning but at the cost of higher sensitivity to missing data. (C) 2017 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). Published by Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available