HTNet: A Hybrid Model Boosted by Triple Self-attention for Crowd Counting

被引:0
|
作者
Li, Yang [1 ]
Yin, Baoqun [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
Crowd Counting; Deep Learning; Self-Attention; Hybrid Model;
D O I
10.1007/978-981-99-8555-5_23
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The swift development of convolutional neural network (CNN) has enabled significant headway in crowd counting research. However, the fixed-size convolutional kernels of traditional methods make it difficult to handle problems such as drastic scale change and complex background interference. In this regard, we propose a hybrid crowd counting model to tackle existing challenges. Firstly, we leverage a global self-attention module (GAM) after CNN backbone to capture wider contextual information. Secondly, due to the gradual recovery of the feature map size in the decoding stage, the local self-attention module (LAM) is employed to reduce computational complexity. With this design, the model can fuse features from global and local perspectives to better cope with scale change. Additionally, to establish the interdependence between spatial and channel dimensions, we further design a novel channel self-attention module (CAM) and combine it with LAM. Finally, we construct a simple yet useful double head module that outputs a foreground segmentation map in addition to the intermediate density map, which are then multiplied together in a pixel-wise style to suppress background interference. The experimental results on several benchmark datasets demonstrate that our method achieves remarkable improvement.
引用
收藏
页码:290 / 301
页数:12
相关论文
共 50 条
  • [31] DCGSA: A global self-attention network with dilated convolution for crowd density map generating
    Zhu, Liping
    Li, Chengyang
    Wang, Bing
    Yuan, Kun
    Yang, Zhongguo
    NEUROCOMPUTING, 2020, 378 : 455 - 466
  • [32] Protein-protein interaction site prediction by model ensembling with hybrid feature and self-attention
    Cong, Hanhan
    Liu, Hong
    Cao, Yi
    Liang, Cheng
    Chen, Yuehui
    BMC BIOINFORMATICS, 2023, 24 (01)
  • [33] Traffic prediction using MSSBiLS with self-attention model
    Suvitha, D.
    Vijayalakshmi, M.
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (15):
  • [34] A Self-attention Network Based Node Embedding Model
    Nguyen, Dai Quoc
    Nguyen, Tu Dinh
    Phung, Dinh
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT III, 2021, 12459 : 364 - 377
  • [35] Keyword Transformer: A Self-Attention Model for Keyword Spotting
    Berg, Axel
    O'Connor, Mark
    Cruz, Miguel Tairum
    INTERSPEECH 2021, 2021, : 4249 - 4253
  • [36] Hybrid attention network based on progressive embedding scale-context for crowd counting
    Wang, Fusen
    Sang, Jun
    Wu, Zhongyuan
    Liu, Qi
    Sang, Nong
    INFORMATION SCIENCES, 2022, 591 : 306 - 318
  • [37] SAVE: Self-Attention on Visual Embedding for Zero-Shot Generic Object Counting
    Zgaren, Ahmed
    Bouachir, Wassim
    Bouguila, Nizar
    JOURNAL OF IMAGING, 2025, 11 (02)
  • [38] HybridHash: Hybrid Convolutional and Self-Attention Deep Hashing for Image Retrieval
    He, Chao
    Wei, Hongxi
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 824 - 832
  • [39] The implicit mathematical reasoning model combining self-attention and convolution
    Yao, Zhuangkai
    Zeng, Bi
    Hu, Huiting
    Wei, Pengfei
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (01) : 975 - 988
  • [40] Mineral Prospectivity Mapping Using Deep Self-Attention Model
    Bojun Yin
    Renguang Zuo
    Siquan Sun
    Natural Resources Research, 2023, 32 : 37 - 56