Adversarial Decoupling and Modality-Invariant Representation Learning for Visible-Infrared Person Re-Identification

被引:31
|
作者
Hu, Weipeng [1 ]
Liu, Bohong [1 ]
Zeng, Haitang [1 ]
Hou, Yanke [1 ]
Hu, Haifeng [1 ]
机构
[1] Sun Yat Sen Univ, Sch Elect & Informat Technol, Guangzhou 510006, Peoples R China
关键词
Representation learning; Feature extraction; Task analysis; Decorrelation; Cameras; Semantics; Lighting; Visible-infrared person re-identification; modality-invariant representations; orthogonal decorrelation; MATRICES;
D O I
10.1109/TCSVT.2022.3147813
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Visible-infrared person re-identification (RGB-IR ReID) has now attracted increasing attention due to its surveillance applications under low-light environments. However, the large intra-class variations between different domains are still a challenging issue in the field of computer vision. To address the above issue, we propose a novel adversarial Decoupling and Modality-invariant Representation learning (DMiR) method to explore potential spectrum-invariant yet identity-discriminative representations for cross-modality pedestrians. Our model consists of three key components, including Domain-related Representation Disentanglement (DrRD), Modality-invariant Discriminative Representation (MiDR) and Representation Orthogonal Decorrelation (ROD). First, two subnets named Identity-Net and Domain-Net are designed to extract identity-related features and domain-related features, respectively. Given this two-stream structure, the DrRD is introduced to achieve adversarial decoupling against domain-specific features via a min-max disentanglement process. Specifically, the classification objective function on Domain-Net is minimized to extract spectrum-specific information while maximizing it to reduce domain-specific information. Second, in Identity-Net, we introduce MiDR to enhance intra-class compactness and reduce domain variations by exploring positive and negative pair variations, semantic-wise differences, and pair-wise semantic variations. Finally, the correlation between the two decomposed features, i.e., identity-related features and domain-related features, may lead to the introduction of modal information in identity representations, and vice versa. Therefore, we present the ROD constraint to make the two decomposed features unrelated to each other, which can more effectively separate the two-component features and enhance feature representations. Practically, we construct ROD at the feature-level and parameter-level, and finally select feature-level ROD as the decorrelation strategy because of its superior decorrelation performance. The whole scheme leads to disentangling spectrum-dependent information, as well as purifying identity information. Extensive experiments are carried out on two mainstream RGB-IR ReID datasets, and the results demonstrate the effectiveness of our method.
引用
收藏
页码:5095 / 5109
页数:15
相关论文
共 50 条
  • [1] Learning enhancing modality-invariant features for visible-infrared person re-identification
    Zhang, La
    Zhao, Xu
    Du, Haohua
    Sun, Jian
    Wang, Jinqiao
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024,
  • [2] Keypoint-Guided Modality-Invariant Discriminative Learning for Visible-Infrared Person Re-identification
    Liang, Tengfei
    Jin, Yi
    Liu, Wu
    Feng, Songhe
    Wang, Tao
    Li, Yidong
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3965 - 3973
  • [3] Hybrid Modality Metric Learning for Visible-Infrared Person Re-Identification
    Zhang, La
    Guo, Haiyun
    Zhu, Kuan
    Qiao, Honglin
    Huang, Gaopan
    Zhang, Sen
    Zhang, Huichen
    Sun, Jian
    Wang, Jinqiao
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (01)
  • [4] Pose-Guided Modality-Invariant Feature Alignment for Visible-Infrared Object Re-Identification
    Liu, Min
    Sun, Yeqing
    Wang, Xueping
    Bian, Yuan
    Zhang, Zhu
    Wang, Yaonan
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 10
  • [5] Modality Unifying Network for Visible-Infrared Person Re-Identification
    Yu, Hao
    Cheng, Xu
    Peng, Wei
    Liu, Weihao
    Zhao, Guoying
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11151 - 11161
  • [6] Towards a Unified Middle Modality Learning for Visible-Infrared Person Re-Identification
    Zhang, Yukang
    Yan, Yan
    Lu, Yang
    Wang, Hanzi
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 788 - 796
  • [7] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [8] Cross-modality consistency learning for visible-infrared person re-identification
    Shao, Jie
    Tang, Lei
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [9] Beyond modality alignment: Learning part-level representation for visible-infrared person re-identification
    Zhang, Peng
    Wu, Qiang
    Yao, Xunxiang
    Xu, Jingsong
    [J]. IMAGE AND VISION COMPUTING, 2021, 108
  • [10] Style-Agnostic Representation Learning for Visible-Infrared Person Re-Identification
    Wu, Jianbing
    Liu, Hong
    Shi, Wei
    Liu, Mengyuan
    Li, Wenhao
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2263 - 2275