IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing (Jan 2024)

Joint Network Combining Dual-Attention Fusion Modality and Two Specific Modalities for Land Cover Classification Using Optical and SAR Images

  • Xiao Liu,
  • Huijun Zou,
  • Shuxiang Wang,
  • Yuzhun Lin,
  • Xibing Zuo

DOI
https://doi.org/10.1109/JSTARS.2023.3347571
Journal volume & issue
Vol. 17
pp. 3236 – 3250

Abstract

Read online

Optical and synthetic aperture radar (SAR) images provide various complementary information on land properties, which can substantially improve the accuracy of land cover classification because of the fusion of their multisource information. However, excellent extraction of the discriminative information of optical and SAR images and effective integration of the interpretation information of multisource data remain challenging issues. In this study, we have proposed a novel joint network that combines a dual-attention fusion modality and two specific modalities to achieve superior land cover classification maps, which has two modes of encoding-level fusion (JoiTriNet-e) and decoding-level fusion (JoiTriNet-d). We first proposed a strategy for the fusion modality and specific modalities joint learning, the goal of which was to simultaneously find three mapping functions that project optical and SAR images separately and together into semantic maps. Two architectures were constructed using the proposed strategy, which improved the performance of multisource and single-source land cover classification. To aggregate the heterogeneous features of optical and SAR images more reasonably, we designed a multimodal dual-attention fusion module. Experiments were conducted on two multisource land cover datasets, among which comparison experiments highlighted the superiority and robustness of our model, and ablation experiments verified the effectiveness of the proposed architecture and module.

Keywords