Learning enhancing modality-invariant features for visible-infrared person re-identification

被引:1
|
作者
Zhang, La [1 ]
Zhao, Xu [2 ]
Du, Haohua [3 ]
Sun, Jian [1 ]
Wang, Jinqiao [2 ]
机构
[1] Beijing Inst Technol, Beijing 100081, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
[3] Beihang Univ, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Visible-infrared person re-identification; Cross-modality; Feature learning; Feature distribution; RETRIEVAL; MODEL;
D O I
10.1007/s13042-024-02168-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To solve the task of visible-infrared person re-identification, most existing methods embed all images into a unified feature space through shared parameters, and then use a metric learning loss function to learn modality-invariant features. However, they may encounter the following two problems: For one thing, they mostly focus on modality-invariant features. In reality, some unique features within each modality can enhance feature discriminability but are often overlooked; For another, current metric learning loss functions mainly focus on feature discriminability and only align modality distributions implicitly, which leads to that the feature distributions from different modalities are still inconsistent in this unified feature space. Taking the foregoing into consideration, in this paper, we propose a novel end-to-end framework composed of two modules: the intra-modality enhancing module and the modality-invariant module. The former fully leverages modality-specific characteristics by establishing independent branches for each modality. It improves feature discriminability by further enhancing the intra-class compactness and inter-class discrepancy within each modality. The latter is designed with a cross-modality feature distribution consistency loss based on the Gaussian distribution assumption. It significantly alleviates the modality discrepancies by effectively and directly aligning the feature distribution in the unified feature space. As a result, the proposed framework can learn modality-invariant features with enhancing discriminability in each modality. Extensive experimental results on SYSU-MM01 and RegDB demonstrate the effectiveness of our method.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Adversarial Decoupling and Modality-Invariant Representation Learning for Visible-Infrared Person Re-Identification
    Hu, Weipeng
    Liu, Bohong
    Zeng, Haitang
    Hou, Yanke
    Hu, Haifeng
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5095 - 5109
  • [2] Keypoint-Guided Modality-Invariant Discriminative Learning for Visible-Infrared Person Re-identification
    Liang, Tengfei
    Jin, Yi
    Liu, Wu
    Feng, Songhe
    Wang, Tao
    Li, Yidong
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3965 - 3973
  • [3] Hybrid Modality Metric Learning for Visible-Infrared Person Re-Identification
    Zhang, La
    Guo, Haiyun
    Zhu, Kuan
    Qiao, Honglin
    Huang, Gaopan
    Zhang, Sen
    Zhang, Huichen
    Sun, Jian
    Wang, Jinqiao
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (01)
  • [4] Pose-Guided Modality-Invariant Feature Alignment for Visible-Infrared Object Re-Identification
    Liu, Min
    Sun, Yeqing
    Wang, Xueping
    Bian, Yuan
    Zhang, Zhu
    Wang, Yaonan
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 10
  • [5] Modality Unifying Network for Visible-Infrared Person Re-Identification
    Yu, Hao
    Cheng, Xu
    Peng, Wei
    Liu, Weihao
    Zhao, Guoying
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11151 - 11161
  • [6] Towards a Unified Middle Modality Learning for Visible-Infrared Person Re-Identification
    Zhang, Yukang
    Yan, Yan
    Lu, Yang
    Wang, Hanzi
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 788 - 796
  • [7] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [8] Cross-modality consistency learning for visible-infrared person re-identification
    Shao, Jie
    Tang, Lei
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [9] Modality Synergy Complement Learning with Cascaded Aggregation for Visible-Infrared Person Re-Identification
    Zhang, Yiyuan
    Zhao, Sanyuan
    Kang, Yuhao
    Shen, Jianbing
    [J]. COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 462 - 479
  • [10] An efficient framework for visible-infrared cross modality person re-identification
    Basaran, Emrah
    Gokmen, Muhittin
    Kamasak, Mustafa E.
    [J]. SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 87