Electronic Research Archive (May 2023)

AENet: attention efficient network for cross-view image geo-localization

  • Jingqian Xu ,
  • Ma Zhu,
  • Baojun Qi,
  • Jiangshan Li,
  • Chunfang Yang

DOI
https://doi.org/10.3934/era.2023210
Journal volume & issue
Vol. 31, no. 7
pp. 4119 – 4138

Abstract

Read online

To address the problem that task-irrelevant objects such as cars, pedestrians and sky, will interfere with the extracted feature descriptors in cross-view image geo-localization, this paper proposes a novel method for cross-view image geo-localization, named as AENet. The method includes two main parts: an attention efficient network fusing channel and spatial attention mechanisms and a triplet loss function based on a multiple hard samples weighting strategy. In the first part, the EfficientNetV2 network is used to extract features from the images and preliminarily filter irrelevant features from the channel dimension, then the Triplet Attention layer is applied to further filter irrelevant features from the spatial dimension. In the second part, a multiple hard samples weighting strategy is proposed to enhance the learning of hard samples. Experimental results show that our proposed method significantly outperforms the state-of-the-art method on two existing benchmark datasets.

Keywords