International Journal of Applied Earth Observations and Geoinformation (Dec 2022)

DASFNet: Dense-Attention–Similarity-Fusion Network for scene classification of dual-modal remote-sensing images

  • Jianhui Jin,
  • Wujie Zhou,
  • Lv Ye,
  • Jingsheng Lei,
  • Lu Yu,
  • Xiaohong Qian,
  • Ting Luo

Journal volume & issue
Vol. 115
p. 103087

Abstract

Read online

Although significant progress has been made in scene classification of high-resolution remote-sensing images (HRRSIs), dual-modal HRRSI scene classification is still an active and challenging issue. In this study, we introduce an end-to-end dense-attention–similarity-fusion network (DASFNet) for dual-modal HRRSIs. Specifically, we propose a dense-attention map module based on graph convolution, which adaptively captures long-range semantic cues and further directs shallow-attention cues to the deep level to guide the generation of high-level feature attention cues. At the encoder stage, DASFNet uses feature similarity to explore the correlation between dual-modal features; a similarity-fusion module extracts complementary information by fusing features from different modalities. A multiscale context-feature-aggregation module is used to strengthen the feature embedding of any two spatial scales; this solves the of scale change problem. A large number of experiments on two HRRSI benchmark datasets for scene classification indicate that the proposed DASFNet outperforms the outstanding scene classification approaches.

Keywords