Sensors (Oct 2021)

Joint Soft–Hard Attention for Self-Supervised Monocular Depth Estimation

  • Chao Fan,
  • Zhenyu Yin,
  • Fulong Xu,
  • Anying Chai,
  • Feiqing Zhang

DOI
https://doi.org/10.3390/s21216956
Journal volume & issue
Vol. 21, no. 21
p. 6956

Abstract

Read online

In recent years, self-supervised monocular depth estimation has gained popularity among researchers because it uses only a single camera at a much lower cost than the direct use of laser sensors to acquire depth. Although monocular self-supervised methods can obtain dense depths, the estimation accuracy needs to be further improved for better applications in scenarios such as autonomous driving and robot perception. In this paper, we innovatively combine soft attention and hard attention with two new ideas to improve self-supervised monocular depth estimation: (1) a soft attention module and (2) a hard attention strategy. We integrate the soft attention module in the model architecture to enhance feature extraction in both spatial and channel dimensions, adding only a small number of parameters. Unlike traditional fusion approaches, we use the hard attention strategy to enhance the fusion of generated multi-scale depth predictions. Further experiments demonstrate that our method can achieve the best self-supervised performance both on the standard KITTI benchmark and the Make3D dataset.

Keywords