Cross-modal pedestrian re-recognition based on attention mechanism

被引:1
|
作者
Zhao, Yuyao [1 ]
Zhou, Hang [1 ]
Cheng, Hai [1 ]
Huang, Chunguang [1 ]
机构
[1] Heilongjiang Univ, Coll Elect & Engn, Harbin, Heilongjiang, Peoples R China
来源
VISUAL COMPUTER | 2024年 / 40卷 / 04期
基金
中国国家自然科学基金;
关键词
Person re-identification; Cross-modality; Attention mechanisms; Data integration;
D O I
10.1007/s00371-023-02926-7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Person re-identification, as an essential research direction in intelligent security, has gained the focus of researchers and scholars. In practical scenarios, visible light cameras depend highly on lighting conditions and have limited detection capability in poor light. Therefore, many scholars have gradually shifted their research goals to cross-modality person re-identification. However, there are few relevant studies, and challenges remain in resolving the differences in the images of different modalities. In order to solve these problems, this paper will use the research method based on the attention mechanism to narrow the difference between the two modes and guide the network in a more appropriate direction to improve the recognition performance of the network. Aiming at the problem of using the attention mechanism method can improve training efficiency. However, it is easy to cause the model training instability. This paper proposes a cross-modal pedestrian re-recognition method based on the attention mechanism. A new attention mechanism module is designed to allow the network to use less time to focus on more critical features of a person. In addition, a cross-modality hard center triplet loss is designed to supervise the model training better. The paper has conducted extensive experiments on the above two methods on two publicly available datasets, which obtained better performance than similar current methods and verified the effectiveness and feasibility of the proposed methods in this paper.
引用
收藏
页码:2405 / 2418
页数:14
相关论文
共 50 条
  • [21] Cross-modal orienting of visual attention
    Hillyard, Steven A.
    Stoermer, Viola S.
    Feng, Wenfeng
    Martinez, Antigona
    McDonald, John J.
    NEUROPSYCHOLOGIA, 2016, 83 : 170 - 178
  • [22] Visual question answering with attention transfer and a cross-modal gating mechanism
    Li, Wei
    Sun, Jianhui
    Liu, Ge
    Zhao, Linglan
    Fang, Xiangzhong
    PATTERN RECOGNITION LETTERS, 2020, 133 (133) : 334 - 340
  • [23] Cross-Modal Attention Mechanism for Weakly Supervised Video Anomaly Detection
    Sun, Wenwen
    Cao, Lin
    Guo, Yanan
    Du, Kangning
    BIOMETRIC RECOGNITION, CCBR 2023, 2023, 14463 : 437 - 446
  • [24] Multimodal Humor Detection Based on Cross-Modal Attention and Modal Maximum Correlation
    Quan, Zhibang
    Sun, Tao
    Su, Mengli
    Wei, Jishu
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 1064 - 1065
  • [25] Pedestrian Re-Recognition Algorithm Based on Optimization Deep Learning-Sequence Memory Model
    An, Feng-Ping
    COMPLEXITY, 2019, 2019
  • [26] Auditory Attention Detection via Cross-Modal Attention
    Cai, Siqi
    Li, Peiwen
    Su, Enze
    Xie, Longhan
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [27] Cross-modal body representation based on visual attention by saliency
    Hikita, Mai
    Fuke, Sawa
    Ogino, Masaki
    Asada, Minoru
    2008 IEEE/RSJ INTERNATIONAL CONFERENCE ON ROBOTS AND INTELLIGENT SYSTEMS, VOLS 1-3, CONFERENCE PROCEEDINGS, 2008, : 2041 - +
  • [28] Emotion recognition using cross-modal attention from EEG and facial expression
    Cui, Rongxuan
    Chen, Wanzhong
    Li, Mingyang
    Knowledge-Based Systems, 2024, 304
  • [29] Cross-Modal Learning with 3D Deformable Attention for Action Recognition
    Kim, Sangwon
    Ahn, Dasom
    Ko, Byoung Chul
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10231 - 10241
  • [30] PedFormer: Pedestrian Behavior Prediction via Cross-Modal Attention Modulation and Gated Multitask Learning
    Rasouli, Amir
    Kotseruba, Iuliia
    arXiv, 2022,