Beyond a strong baseline: cross-modality contrastive learning for visible-infrared person re-identification

被引:0
|
作者
Pengfei Fang
Yukang Zhang
Zhenzhong Lan
机构
[1] Southeast University,School of Computer Science and Engineering
[2] MOE Key Laboratory of Computer Network and Information Integration (Southeast University),School of Informatics
[3] Xiamen University,School of Engineering
[4] Westlake University,undefined
来源
关键词
Cross-modality; Person re-identification; Strong baseline; Cross-modality contrastive learning;
D O I
暂无
中图分类号
学科分类号
摘要
Cross-modality pedestrian image matching, which entails the matching of visible and infrared images, is a vital area in person re-identification (reID) due to its potential to facilitate person retrieval across a spectrum of lighting conditions. Despite its importance, this task presents considerable challenges stemming from two significant areas: cross-modality discrepancies due to the different imaging principles of spectrum cameras and within-class variations caused by the diverse viewpoints of large-scale distributed surveillance cameras. Unfortunately, the existing literature provides limited insights into effectively mitigating these issues, signifying a crucial research gap. In response to this, the present paper makes two primary contributions. First, we conduct a comprehensive study of training methodologies and subsequently present a strong baseline network designed specifically to address the complexities of the visible-infrared person reID task. This strong baseline network is paramount to the advancement of the field and to ensure the fair evaluation of algorithmic effectiveness. Second, we propose the Cross-Modality Contrastive Learning (CMCL) scheme, a novel approach to address the cross-modality discrepancies and enhance the quality of image embeddings across both modalities. CMCL incorporates intra-modality and inter-modality contrastive loss components, designed to improve the matching quality across the modalities. Thorough experiments show the superior performance of the baseline network, and the proposed CMCL can further bring performance over the baselines, outperforming the state-of-the-art methods considerably.
引用
收藏
相关论文
共 50 条
  • [1] Beyond a strong baseline: cross-modality contrastive learning for visible-infrared person re-identification
    Fang, Pengfei
    Zhang, Yukang
    Lan, Zhenzhong
    [J]. MACHINE VISION AND APPLICATIONS, 2023, 34 (06)
  • [2] Cross-modality consistency learning for visible-infrared person re-identification
    Shao, Jie
    Tang, Lei
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [3] Cross-Modality Transformer for Visible-Infrared Person Re-Identification
    Jiang, Kongzhu
    Zhang, Tianzhu
    Liu, Xiang
    Qian, Bingqiao
    Zhang, Yongdong
    Wu, Feng
    [J]. COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 480 - 496
  • [4] Cross-Modality Transformer With Modality Mining for Visible-Infrared Person Re-Identification
    Liang, Tengfei
    Jin, Yi
    Liu, Wu
    Li, Yidong
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8432 - 8444
  • [5] A cross-modality person re-identification method for visible-infrared images
    Sun, Yibo
    Wang, Rong
    Zhang, Qi
    Lin, Ronghui
    [J]. Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (06): : 2018 - 2025
  • [6] Discover Cross-Modality Nuances for Visible-Infrared Person Re-Identification
    Wu, Qiong
    Dai, Pingyang
    Chen, Jie
    Lin, Chia-Wen
    Wu, Yongjian
    Huang, Feiyue
    Zhong, Bineng
    Ji, Rongrong
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4328 - 4337
  • [7] Deep learning for visible-infrared cross-modality person re-identification: A comprehensive review
    Huang, Nianchang
    Liu, Jianan
    Miao, Yunqi
    Zhang, Qiang
    Han, Jungong
    [J]. INFORMATION FUSION, 2023, 91 : 396 - 411
  • [8] Visible-Infrared Person Re-Identification via Cross-Modality Interaction Transformer
    Feng, Yujian
    Yu, Jian
    Chen, Feng
    Ji, Yimu
    Wu, Fei
    Liu, Shangdon
    Jing, Xiao-Yuan
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7647 - 7659
  • [9] Counterfactual attention alignment for visible-infrared cross-modality person re-identification
    Sun, Zongzhe
    Zhao, Feng
    [J]. PATTERN RECOGNITION LETTERS, 2023, 168 : 79 - 85
  • [10] Cross-modality nearest neighbor loss for visible-infrared person re-identification
    Zhao, Sanyuan
    Qi, A.
    Gao, Yu
    [J]. Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (02): : 433 - 441