Cross-scale global attention feature pyramid network for person search

被引:3
|
作者
Li, Yang [1 ,2 ]
Xu, Huahu [2 ,3 ]
Bian, Minjie [3 ]
Xiao, Junsheng [2 ]
机构
[1] Shanghai Jianqiao Univ, Sch Informat Technol, Shanghai 201306, Peoples R China
[2] Shanghai Univ, Sch Comp Engn & Sci, Shanghai 200444, Peoples R China
[3] Shanghai Univ, Informat Off, Shanghai 200444, Peoples R China
关键词
Person search; Global attention; Feature pyramid network; Multi-scale; Fine-grained;
D O I
10.1016/j.imavis.2021.104332
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Person search aims to locate the target person in real unconstrained scene images. It faces many challenges such as multi-scale and fine-grained. To address the challenges, a novel cross-scale global attention feature pyramid network (CSGAFPN) is proposed. Firstly, we design a novel multi-head global attention module (MHGAM), which adopts cosine similarity and sparse query location methods to effectively capture cross-scale long-distance dependence. Then, we design the CSGAFPN, which extends top-down feature pyramid network with bottom-up connections and embeds MHGAMs to the connections. CSGAFPN can capture cross-scale long-distance global correlation from multi-scale feature maps, selectively strengthen important features and restrain less important features. CSGAFPN is applied for both person detection and person re-identification (reID) sub-tasks of person search, it can well handle the multi-scale and fine-grained challenges, and significantly improve person search performance. Furthermore, the output multi-scale feature maps of CSGAFPN are processed by an adaptive feature aggregation with attention (AFAA) layer to further improve the performance. Numerous exper-iments with two public person search datasets, CUHK-SYSU and PRW, show our CSGAFPN based approach ac-quires better performance than other state-of-the-art (SOTA) person search approaches. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] DCIFPN: Deformable cross-scale interaction feature pyramid network for object detection
    Xiao, Junrui
    Jiang, He
    Li, Zhikai
    Gu, Qingyi
    IET IMAGE PROCESSING, 2023, 17 (09) : 2596 - 2610
  • [2] RCNet: Reverse Feature Pyramid and Cross-scale Shift Network for Object Detection
    Zong, Zhuofan
    Cao, Qianggang
    Leng, Biao
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 5637 - 5645
  • [3] Feature-transferable pyramid network for cross-scale object detection in sar images
    Zhou Z.
    Cui Z.
    Cao Z.
    Yang J.
    Journal of Radars, 2021, 10 (04) : 544 - 558
  • [4] SleepFC: Feature Pyramid and Cross-Scale Context Learning for Sleep Staging
    Li, Wei
    Liu, Teng
    Xu, Baoguo
    Song, Aiguo
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2024, 32 : 2198 - 2208
  • [5] A cross-scale mixed attention network for smoke segmentation
    Yuan, Feiniu
    Shi, Yu
    Zhang, Lin
    Fang, Yuming
    DIGITAL SIGNAL PROCESSING, 2023, 134
  • [6] An efficient feature pyramid attention network for person re-identification
    Luo, Qian
    Shao, Jie
    Dang, Wanli
    Wang, Chao
    Cao, Libo
    Zhang, Tao
    IMAGE AND VISION COMPUTING, 2024, 145
  • [7] Asymmetric Cross-Scale Alignment for Text-Based Person Search
    Ji, Zhong
    Hu, Junhua
    Liu, Deyin
    Wu, Lin Yuanbo
    Zhao, Ye
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7699 - 7709
  • [8] GCPAN: an adaptive global cross-scale prior attention network for image super-resolution
    Shi, Mingzhu
    Kong, Siqi
    Zao, Bin
    Tan, Muxian
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (24): : 17671 - 17688
  • [9] A Spatial Cross-Scale Attention Network and Global Average Accuracy Loss for SAR Ship Detection
    Zhang, Lili
    Liu, Yuxuan
    Qu, Lele
    Cai, Jiannan
    Fang, Junpeng
    REMOTE SENSING, 2023, 15 (02)
  • [10] GCPAN: an adaptive global cross-scale prior attention network for image super-resolution
    Mingzhu Shi
    Siqi Kong
    Bin Zao
    Muxian Tan
    Neural Computing and Applications, 2023, 35 : 17671 - 17688