MSIF: multi-spectrum image fusion method for cross-modality person re-identification

被引:0
|
作者
Qingshan Chen
Zhenzhen Quan
Yifan Zheng
Yujun Li
Zhi Liu
Mikhail G. Mozerov
机构
[1] Shandong University,School of Information Science and Engineering
[2] Universitat Autònoma de Barcelona,Computer Vision Center
关键词
Person ReID; Sketch; Infrared; Cross-modality; Multi-spectrum image fusion;
D O I
暂无
中图分类号
学科分类号
摘要
Sketch-RGB cross-modality person re-identification (ReID) is a challenging task that aims to match a sketch portrait drawn by a professional artist with a full-body photo taken by surveillance equipment to deal with situations where the monitoring equipment is damaged at the accident scene. However, sketch portraits only provide highly abstract frontal body contour information and lack other important features such as color, pose, behavior, etc. The difference in saliency between the two modalities brings new challenges to cross-modality person ReID. To overcome this problem, this paper proposes a novel dual-stream model for cross-modality person ReID, which is able to mine modality-invariant features to reduce the discrepancy between sketch and camera images end-to-end. More specifically, we propose a multi-spectrum image fusion (MSIF) method, which aims to exploit the image appearance changes brought by multiple spectrums and guide the network to mine modality-invariant commonalities during training. It only processes the spectrum of the input images without adding additional calculations and model complexity, which can be easily integrated into other models. Moreover, we introduce a joint structure via a generalized mean pooling (GMP) layer and a self-attention (SA) mechanism to balance background and texture information and obtain the regional features with a large amount of information in the image. To further shrink the intra-class distance, a weighted regularized triplet (WRT) loss is developed without introducing additional hyperparameters. The model was first evaluated on the PKU Sketch ReID dataset, and extensive experimental results show that the Rank-1/mAP accuracy of our method is 87.00%/91.12%, reaching the current state-of-the-art performance. To further validate the effectiveness of our approach in handling cross-modality person ReID, we conducted experiments on two commonly used IR-RGB datasets (SYSU-MM01 and RegDB). The obtained results show that our method achieves competitive performance. These results confirm the ability of our method to effectively process images from different modalities.
引用
收藏
页码:647 / 665
页数:18
相关论文
共 50 条
  • [1] MSIF: multi-spectrum image fusion method for cross-modality person re-identification
    Chen, Qingshan
    Quan, Zhenzhen
    Zheng, Yifan
    Li, Yujun
    Liu, Zhi
    Mozerov, Mikhail G.
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (02) : 647 - 665
  • [2] Self-attention Cross-modality Fusion Network for Cross-modality Person Re-identification
    Du, Peng
    Song, Yong-Hong
    Zhang, Xin-Yao
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (06): : 1457 - 1468
  • [3] Global and Part Feature Fusion for Cross-Modality Person Re-Identification
    Wang, Xianju
    Cordova, Ronald S.
    [J]. IEEE ACCESS, 2022, 10 : 122038 - 122046
  • [4] Modality interactive attention for cross-modality person re-identification
    Zou, Zilin
    Chen, Ying
    [J]. IMAGE AND VISION COMPUTING, 2024, 148
  • [5] Cross-Modality Person Re-identification Combined with Data Augmentation and Feature Fusion
    Song, Yu
    Wang, Banghai
    Cao, Ganggang
    [J]. Computer Engineering and Applications, 2024, 60 (04) : 133 - 141
  • [6] Cross-modality Person Re-identification Based on Joint Constraints of Image and Feature
    Zhang, Yu-Kang
    Tan, Lei
    Chen, Jing-Ying
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2021, 47 (08): : 1943 - 1950
  • [7] Cross-modality person re-identification via multi-task learning
    Huang, Nianchang
    Liu, Kunlong
    Liu, Yang
    Zhang, Qiang
    Han, Jungong
    [J]. PATTERN RECOGNITION, 2022, 128
  • [8] Cross-modality person re-identification via multi-task learning
    Huang, Nianchang
    Liu, Kunlong
    Liu, Yang
    Zhang, Qiang
    Han, Jungong
    [J]. Pattern Recognition, 2022, 128
  • [9] Deep Cross-Modality Alignment for Multi-Shot Person Re-IDentification
    Song, Zhichao
    Ni, Bingbing
    Yan, Yichao
    Ren, Zhe
    Xu, Yi
    Yang, Xiaokang
    [J]. PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 645 - 653
  • [10] The Multi-Layer Constrained Loss for Cross-Modality Person Re-Identification
    Sun, Zhanrui
    Zhu, Yongxin
    Song, Shijin
    Hou, Junjie
    Du, Sen
    Song, Yuefeng
    [J]. 2020 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND SIGNAL PROCESSING (AISP), 2020,