Text-augmented Multi-Modality contrastive learning for unsupervised visible-infrared person re-identification

被引:0
|
作者
Sun, Rui [1 ,2 ,3 ]
Huang, Guoxi [1 ,2 ]
Wang, Xuebin [1 ,2 ]
Du, Yun [1 ,2 ]
Zhang, Xudong [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, 485 Danxia Rd, Hefei 230009, Peoples R China
[2] Hefei Univ Technol, Anhui Key Lab Ind Safety & Emergency Technol, Hefei 230009, Peoples R China
[3] Minist Educ Peoples Republ China, Key Lab Knowledge Engn Big Data, Hefei 230009, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Unsupervised person re-identification; Text-augmented features; Multi-Modality; Contrastive learning;
D O I
10.1016/j.imavis.2024.105310
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visible-infrared person re-identification holds significant implications for intelligent security. Unsupervised methods can reduce the gap of different modalities without labels. Most previous unsupervised methods only train their models with image information, so that the model cannot obtain powerful deep semantic information. In this paper, we leverage CLIP to extract deep text information. We propose a Text-Image Alignment (TIA) module to align the image and text information and effectively bridge the gap between visible and infrared modality. We produce a Local-Global Image Match (LGIM) module to find homogeneous information. Specifically, we employ the Hungarian algorithm and Simulated Annealing (SA) algorithm to attain original information from image features while mitigating the interference of heterogeneous information. Additionally, we design a Changeable Cross-modality Alignment Loss (CCAL) to enable the model to learn modality-specific features during different training stages. Our method performs well and attains powerful robustness by targeted learning. Extensive experiments demonstrate the effectiveness of our approach, our method achieves a rank-1 accuracy that exceeds state-of-the-art approaches by approximately 10% on the RegDB.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Visible-Infrared Person Re-Identification With Modality-Specific Memory Network
    Li, Yulin
    Zhang, Tianzhu
    Liu, Xiang
    Tian, Qi
    Zhang, Yongdong
    Wu, Feng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 7165 - 7178
  • [42] Deep learning for visible-infrared cross-modality person re-identification: A comprehensive review
    Huang, Nianchang
    Liu, Jianan
    Miao, Yunqi
    Zhang, Qiang
    Han, Jungong
    INFORMATION FUSION, 2023, 91 : 396 - 411
  • [43] Unsupervised Visible-Infrared Person Re-Identification via Progressive Graph Matching and Alternate Learning
    Wu, Zesen
    Ye, Mang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9548 - 9558
  • [44] Modality Mitigation And Diverse Part Awareness for Visible-Infrared Person Re-identification
    Zhang, Meiling
    Li, Xin
    Wang, Qiang
    Guo, Hubo
    Huang, Zhihong
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT II, 2025, 15202 : 145 - 159
  • [45] Visible-Infrared Person Re-Identification via Homogeneous Augmented Tri-Modal Learning
    Ye, Mang
    Shen, Jianbing
    Shao, Ling
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 728 - 739
  • [46] Progressive Cross-Modal Association Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Yiming
    Hu, Weipeng
    Hu, Haifeng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 1290 - 1304
  • [47] Discover Cross-Modality Nuances for Visible-Infrared Person Re-Identification
    Wu, Qiong
    Dai, Pingyang
    Chen, Jie
    Lin, Chia-Wen
    Wu, Yongjian
    Huang, Feiyue
    Zhong, Bineng
    Ji, Rongrong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4328 - 4337
  • [48] A cross-modality person re-identification method for visible-infrared images
    Sun Y.
    Wang R.
    Zhang Q.
    Lin R.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (06): : 2018 - 2025
  • [49] Modality-perceptive harmonization network for visible-infrared person re-identification
    Zuo, Xutao
    Peng, Jinjia
    Cheng, Tianhang
    Wang, Huibing
    INFORMATION FUSION, 2025, 118
  • [50] Exploring modality enhancement and compensation spaces for visible-infrared person re-identification
    Cheng, Xu
    Deng, Shuya
    Yu, Hao
    IMAGE AND VISION COMPUTING, 2024, 146