We found a match
Your institution may have rights to this item. Sign in to continue.
- Title
Dual‐granularity feature fusion in visible‐infrared person re‐identification.
- Authors
Cai, Shuang; Yang, Shanmin; Hu, Jing; Wu, Xi
- Abstract
Visible‐infrared person re‐identification (VI‐ReID) aims to recognize images of the same person captured in different modalities. Existing methods mainly focus on learning single‐granularity representations, which have limited discriminability and weak robustness. This paper proposes a novel dual‐granularity feature fusion network for VI‐ReID. Specifically, a dual‐branch module that extracts global and local features and then fuses them to enhance the representative ability is adopted. Furthermore, an identity‐aware modal discrepancy loss that promotes modality alignment by reducing the gap between features from visible and infrared modalities is proposed. Finally, considering the influence of non‐discriminative information in the modal‐shared features of RGB‐IR, a greyscale conversion is introduced to extract modality‐irrelevant discriminative features better. Extensive experiments on the SYSU‐MM01 and RegDB datasets demonstrate the effectiveness of the framework and superiority over state‐of‐the‐art methods.
- Subjects
COMPUTER vision; IMAGE retrieval
- Publication
IET Image Processing (Wiley-Blackwell), 2024, Vol 18, Issue 4, p972
- ISSN
1751-9659
- Publication type
Article
- DOI
10.1049/ipr2.12999