Effective Fusion of Multi-Modal Data with Group Convolutions for Semantic Segmentation of Aerial Imagery

Kaiqiang Chen, Kun Fu, Xin Gao, Menglong Yan, Wenkai Zhang, Yue Zhang, Xian Sun

Research output: Chapter in Book/Report/Conference proceedingConference contribution

15 Scopus citations

Abstract

In this paper, we achieve a semantic segmentation of aerial imagery based on the fusion of multi-modal data in an effective way. The multi-modal data contains a true orthophoto and the corresponding normalized Digital Surface Model (nDSM), which are stacked together before they are fed into a Convolutional Neural Network (CNN). Though the two modalities are fused at the early stage, their features are learned independently with group convolutions firstly and then the learned features of different modalities are fused at multiple scales with standard convolutions. Therefore, the multi-scale fusion of multi-modal features is completed in a single-branch convolutional network. In this way, the computational cost is reduced while the experimental results reveal that we can still get promising results.
Original languageEnglish (US)
Title of host publicationInternational Geoscience and Remote Sensing Symposium (IGARSS)
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages3911-3914
Number of pages4
ISBN (Print)9781538691540
DOIs
StatePublished - Jul 1 2019
Externally publishedYes

Fingerprint

Dive into the research topics of 'Effective Fusion of Multi-Modal Data with Group Convolutions for Semantic Segmentation of Aerial Imagery'. Together they form a unique fingerprint.

Cite this