IET Image Processing (Mar 2024)

Dual‐granularity feature fusion in visible‐infrared person re‐identification

  • Shuang Cai,
  • Shanmin Yang,
  • Jing Hu,
  • Xi Wu

DOI
https://doi.org/10.1049/ipr2.12999
Journal volume & issue
Vol. 18, no. 4
pp. 972 – 980

Abstract

Read online

Abstract Visible‐infrared person re‐identification (VI‐ReID) aims to recognize images of the same person captured in different modalities. Existing methods mainly focus on learning single‐granularity representations, which have limited discriminability and weak robustness. This paper proposes a novel dual‐granularity feature fusion network for VI‐ReID. Specifically, a dual‐branch module that extracts global and local features and then fuses them to enhance the representative ability is adopted. Furthermore, an identity‐aware modal discrepancy loss that promotes modality alignment by reducing the gap between features from visible and infrared modalities is proposed. Finally, considering the influence of non‐discriminative information in the modal‐shared features of RGB‐IR, a greyscale conversion is introduced to extract modality‐irrelevant discriminative features better. Extensive experiments on the SYSU‐MM01 and RegDB datasets demonstrate the effectiveness of the framework and superiority over state‐of‐the‐art methods.

Keywords