Beyond a strong baseline: cross-modality contrastive learning for visible-infrared person re-identification

被引:0
|
作者
Pengfei Fang
Yukang Zhang
Zhenzhong Lan
机构
[1] Southeast University,School of Computer Science and Engineering
[2] MOE Key Laboratory of Computer Network and Information Integration (Southeast University),School of Informatics
[3] Xiamen University,School of Engineering
[4] Westlake University,undefined
来源
关键词
Cross-modality; Person re-identification; Strong baseline; Cross-modality contrastive learning;
D O I
暂无
中图分类号
学科分类号
摘要
Cross-modality pedestrian image matching, which entails the matching of visible and infrared images, is a vital area in person re-identification (reID) due to its potential to facilitate person retrieval across a spectrum of lighting conditions. Despite its importance, this task presents considerable challenges stemming from two significant areas: cross-modality discrepancies due to the different imaging principles of spectrum cameras and within-class variations caused by the diverse viewpoints of large-scale distributed surveillance cameras. Unfortunately, the existing literature provides limited insights into effectively mitigating these issues, signifying a crucial research gap. In response to this, the present paper makes two primary contributions. First, we conduct a comprehensive study of training methodologies and subsequently present a strong baseline network designed specifically to address the complexities of the visible-infrared person reID task. This strong baseline network is paramount to the advancement of the field and to ensure the fair evaluation of algorithmic effectiveness. Second, we propose the Cross-Modality Contrastive Learning (CMCL) scheme, a novel approach to address the cross-modality discrepancies and enhance the quality of image embeddings across both modalities. CMCL incorporates intra-modality and inter-modality contrastive loss components, designed to improve the matching quality across the modalities. Thorough experiments show the superior performance of the baseline network, and the proposed CMCL can further bring performance over the baselines, outperforming the state-of-the-art methods considerably.
引用
收藏
相关论文
共 50 条
  • [31] Modality Synergy Complement Learning with Cascaded Aggregation for Visible-Infrared Person Re-Identification
    Zhang, Yiyuan
    Zhao, Sanyuan
    Kang, Yuhao
    Shen, Jianbing
    [J]. COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 462 - 479
  • [32] Learning enhancing modality-invariant features for visible-infrared person re-identification
    Zhang, La
    Zhao, Xu
    Du, Haohua
    Sun, Jian
    Wang, Jinqiao
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024,
  • [33] RGB-Infrared Cross-Modality Person Re-Identification
    Wu, Ancong
    Zheng, Wei-Shi
    Yu, Hong-Xing
    Gong, Shaogang
    Lai, Jianhuang
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 5390 - 5399
  • [34] Augmented Dual-Contrastive Aggregation Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Bin
    Ye, Mang
    Chen, Jun
    Wu, Zesen
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 2843 - 2851
  • [35] Occluded Visible-Infrared Person Re-Identification
    Feng, Yujian
    Ji, Yimu
    Wu, Fei
    Gao, Guangwei
    Gao, Yang
    Liu, Tianliang
    Liu, Shangdong
    Jing, Xiao-Yuan
    Luo, Jiebo
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1401 - 1413
  • [36] Dual Mutual Learning for Cross-Modality Person Re-Identification
    Zhang, Demao
    Zhang, Zhizhong
    Ju, Ying
    Wang, Cong
    Xie, Yuan
    Qu, Yanyun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5361 - 5373
  • [37] Cross-Modality Person Re-identification with Memory-Based Contrastive Embedding
    Cheng, De
    Wang, Xiaolong
    Wang, Nannan
    Wang, Zhen
    Wang, Xiaoyu
    Gao, Xinbo
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 425 - 432
  • [38] Attend to the Difference: Cross-Modality Person Re-Identification via Contrastive Correlation
    Zhang, Shizhou
    Yang, Yifei
    Wang, Peng
    Liang, Guoqiang
    Zhang, Xiuwei
    Zhang, Yanning
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 8861 - 8872
  • [39] Learning Progressive Modality-Shared Transformers for Effective Visible-Infrared Person Re-identification
    Lu, Hu
    Zou, Xuezhang
    Zhang, Pingping
    [J]. THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1835 - 1843
  • [40] Adversarial Decoupling and Modality-Invariant Representation Learning for Visible-Infrared Person Re-Identification
    Hu, Weipeng
    Liu, Bohong
    Zeng, Haitang
    Hou, Yanke
    Hu, Haifeng
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5095 - 5109