Multi-granular inter-frame relation exploration and global residual embedding for video-based person re-identification

被引:0
|
作者
机构
[1] [1,Zhu, Zhiqin
[2] Chen, Sixin
[3] Qi, Guanqiu
[4] Li, Huafeng
[5] 1,Gao, Xinbo
关键词
Adversarial machine learning;
D O I
10.1016/j.image.2024.117240
中图分类号
学科分类号
摘要
In recent years, the field of video-based person re-identification (re-ID) has conducted in-depth research on how to effectively utilize spatiotemporal clues, which has attracted attention for its potential in providing comprehensive view representations of pedestrians. However, although the discriminability and correlation of spatiotemporal features are often studied, the exploration of the complex relationships between these features has been relatively neglected. Especially when dealing with multi-granularity features, how to depict the different spatial representations of the same person under different perspectives becomes a challenge. To address this challenge, this paper proposes a multi-granularity inter-frame relationship exploration and global residual embedding network specifically designed to solve the above problems. This method successfully extracts more comprehensive and discriminative feature representations by deeply exploring the interactions and global differences between multi-granularity features. Specifically, by simulating the dynamic relationship of different granularity features in long video sequences and using a structured perceptual adjacency matrix to synthesize spatiotemporal information, cross-granularity information is effectively integrated into individual features. In addition, by introducing a residual learning mechanism, this method can also guide the diversified development of global features and reduce the negative impacts caused by factors such as occlusion. Experimental results verify the effectiveness of this method on three mainstream benchmark datasets, significantly surpassing state-of-the-art solutions. This shows that this paper successfully solves the challenging problem of how to accurately identify and utilize the complex relationships between multi-granularity spatiotemporal features in video-based person re-ID. © 2024 Elsevier B.V.
引用
下载
收藏
相关论文
共 50 条
  • [41] ALL FOR ONE: FRAME-WISE RANK LOSS FOR IMPROVING VIDEO-BASED PERSON RE-IDENTIFICATION
    Navaneet, K. L.
    Todi, Vasudha
    Babu, R. Venkatesh
    Chakraborty, Anirban
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2472 - 2476
  • [42] Watching You: Global-guided Reciprocal Learning for Video-based Person Re-identification
    Liu, Xuehu
    Zhang, Pingping
    Yu, Chenyang
    Lu, Huchuan
    Yang, Xiaoyun
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 13329 - 13338
  • [43] Video-based Person Re-identification Using Refined Attention Networks
    Rahman, Tanzila
    Rochan, Mrigank
    Wang, Yang
    2019 16TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS), 2019,
  • [44] Temporal Extension Topology Learning for Video-Based Person Re-identification
    Ning, Jiaqi
    Li, Fei
    Liu, Rujie
    Takeuchi, Shun
    Suzuki, Genta
    COMPUTER VISION - ACCV 2022 WORKSHOPS, 2023, 13848 : 213 - 225
  • [45] TEMPORALLY ALIGNED POOLING REPRESENTATION FOR VIDEO-BASED PERSON RE-IDENTIFICATION
    Gao, Changxin
    Wang, Jin
    Liu, Leyuan
    Yu, Jin-Gang
    Sang, Nong
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 4284 - 4288
  • [46] Learning Compact Appearance Representation for Video-Based Person Re-Identification
    Zhang, Wei
    Hu, Shengnan
    Liu, Kan
    Zha, Zhengjun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (08) : 2442 - 2452
  • [47] Learning Bidirectional Temporal Cues for Video-Based Person Re-Identification
    Zhang, Wei
    Yu, Xiaodong
    He, Xuanyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (10) : 2768 - 2776
  • [48] Video-Based Person Re-Identification Using Unsupervised Tracklet Matching
    Riachy, Chirine
    Khelifi, Fouad
    Bouridane, Ahmed
    IEEE ACCESS, 2019, 7 : 20596 - 20606
  • [49] Sequences consistency feature learning for video-based person re-identification
    Zhao, Kai
    Cheng, Deqiang
    Kou, Qiqi
    Li, Jiahan
    Liu, Ruihang
    ELECTRONICS LETTERS, 2022, 58 (04) : 142 - 144
  • [50] SANet: Statistic Attention Network for Video-Based Person Re-Identification
    Bai, Shutao
    Ma, Bingpeng
    Chang, Hong
    Huang, Rui
    Shan, Shiguang
    Chen, Xilin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (06) : 3866 - 3879