Robust Duality Learning for Unsupervised Visible-Infrared Person Re-Identification

被引:0
|
作者
Li, Yongxiang [1 ]
Sun, Yuan [1 ]
Qin, Yang [1 ]
Peng, Dezhong [1 ,2 ]
Peng, Xi [1 ]
Hu, Peng [1 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Peoples R China
[2] Sichuan Natl Innovat New Vis UHD Video Technol Co, Chengdu 610095, Peoples R China
基金
中国国家自然科学基金;
关键词
Noise measurement; Noise; Overfitting; Adaptation models; Training; Predictive models; Semantics; Interference; Robustness; Optimization; Unsupervised VI-ReID; pseudo-label noise; noise correspondence; cluster consistency;
D O I
10.1109/TIFS.2025.3536613
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Unsupervised visible-infrared person re-identification (UVI-ReID) aims at retrieving pedestrian images of the same individual across distinct modalities, presenting challenges due to the inherent heterogeneity gap and the absence of cost-prohibitive annotations. Although existing methods employ self-training with clustering-generated pseudo-labels to bridge this gap, they always implicitly assume that these pseudo-labels are predicted correctly. In practice, however, this presumption is impossible to satisfy due to the difficulty of training a perfect model let alone without any ground truths, resulting in pseudo-labeling errors. Based on the observation, this study introduces a new learning paradigm for UVI-ReID considering Pseudo-Label Noise (PLN), which encompasses three challenges: noise overfitting, error accumulation, and noisy cluster correspondence. To conquer these challenges, we propose a novel robust duality learning framework (RoDE) for UVI-ReID to mitigate the adverse impact of noisy pseudo-labels. Specifically, for noise overfitting, we propose a novel Robust Adaptive Learning mechanism (RAL) to dynamically prioritize clean samples while deprioritizing noisy ones, thus avoiding overemphasizing noise. To circumvent error accumulation of self-training, where the model tends to confirm its mistakes, RoDE alternately trains dual distinct models using pseudo-labels predicted by their counterparts, thereby maintaining diversity and avoiding collapse into noise. However, this will lead to cross-cluster misalignment between the two distinct models, not to mention the misalignment between different modalities, resulting in dual noisy cluster correspondence and thus difficult to optimize. To address this issue, a Cluster Consistency Matching mechanism (CCM) is presented to ensure reliable alignment across distinct modalities as well as across different models by leveraging cross-cluster similarities. Extensive experiments on three benchmark datasets demonstrate the effectiveness of the proposed RoDE.
引用
收藏
页码:1937 / 1948
页数:12
相关论文
共 50 条
  • [1] Modality-agnostic learning for robust visible-infrared person re-identification
    Gong, Shengrong
    Li, Shuomin
    Xie, Gengsheng
    Yao, Yufeng
    Zhong, Shan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (03)
  • [2] Towards Grand Unified Representation Learning for Unsupervised Visible-Infrared Person Re-Identification
    Bin Yang
    Chen, Jun
    Ye, Mang
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11035 - 11045
  • [3] Two-stage contrastive learning for unsupervised visible-infrared person re-identification
    Zou, Yuan
    Zhu, Pengxu
    Yang, Jianwei
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (06)
  • [4] Shallow-Deep Collaborative Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Bin
    Chen, Jun
    Ye, Mang
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 16870 - 16879
  • [5] Dual Consistency-Constrained Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Bin
    Chen, Jun
    Chen, Cuiqun
    Ye, Mang
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 1767 - 1779
  • [6] Image-text feature learning for unsupervised visible-infrared person re-identification
    Guo, Jifeng
    Pang, Zhiqi
    IMAGE AND VISION COMPUTING, 2025, 158
  • [7] Unveiling the Power of CLIP in Unsupervised Visible-Infrared Person Re-Identification
    Chen, Zhong
    Zhang, Zhizhong
    Tan, Xin
    Qu, Yanyun
    Xie, Yuan
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3667 - 3675
  • [8] Occluded Visible-Infrared Person Re-Identification
    Feng, Yujian
    Ji, Yimu
    Wu, Fei
    Gao, Guangwei
    Gao, Yang
    Liu, Tianliang
    Liu, Shangdong
    Jing, Xiao-Yuan
    Luo, Jiebo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1401 - 1413
  • [9] Unsupervised Visible-Infrared Person Re-Identification via Progressive Graph Matching and Alternate Learning
    Wu, Zesen
    Ye, Mang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9548 - 9558
  • [10] Progressive Cross-Modal Association Learning for Unsupervised Visible-Infrared Person Re-Identification
    Yang, Yiming
    Hu, Weipeng
    Hu, Haifeng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 1290 - 1304