Attention-Guided Collaborative Counting

被引:10
|
作者
Mo, Hong [1 ]
Ren, Wenqi [2 ]
Zhang, Xiong [3 ]
Yan, Feihu [4 ]
Zhou, Zhong [1 ]
Cao, Xiaochun [2 ]
Wu, Wei [1 ]
机构
[1] Beihang Univ, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
[3] Neolix Autonomous Vehicle, Beijing 100016, Peoples R China
[4] Beijing Univ Civil Engn & Architecture, Sch Elect & Informat Engn, Beijing 100044, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Collaboration; Transformers; Task analysis; Head; Computational modeling; Computer vision; Crowd counting; attention-guided collaborative counting model; bi-directional transformer;
D O I
10.1109/TIP.2022.3207584
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing crowd counting designs usually exploit multi-branch structures to address the scale diversity problem. However, branches in these structures work in a competitive rather than collaborative way. In this paper, we focus on promoting collaboration between branches. Specifically, we propose an attention-guided collaborative counting module (AGCCM) comprising an attention-guided module (AGM) and a collaborative counting module (CCM). The CCM promotes collaboration among branches by recombining each branch's output into an independent count and joint counts with other branches. The AGM capturing the global attention map through a transformer structure with a pair of foreground-background related loss functions can distinguish the advantages of different branches. The loss functions do not require additional labels and crowd division. In addition, we design two kinds of bidirectional transformers (Bi-Transformers) to decouple the global attention to row attention and column attention. The proposed Bi-Transformers are able to reduce the computational complexity and handle images in any resolution without cropping the image into small patches. Extensive experiments on several public datasets demonstrate that the proposed algorithm performs favorably against the state-of-the-art crowd counting methods.
引用
收藏
页码:6306 / 6319
页数:14
相关论文
共 50 条
  • [11] A model of attention-guided visual perception and recognition
    Rybak, IA
    Gusakova, VI
    Golovan, AV
    Podladchikova, LN
    Shevtsova, NA
    VISION RESEARCH, 1998, 38 (15-16) : 2387 - 2400
  • [12] 3D Crowd Counting via Geometric Attention-Guided Multi-view Fusion
    Zhang, Qi
    Chan, Antoni B.
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (12) : 3123 - 3139
  • [13] Multiscale Attention-Guided Panoptic Segmentation Network
    Fu, Du
    Qu, Shaojun
    Fu, Ya
    Computer Engineering and Applications, 2023, 59 (22) : 223 - 232
  • [14] Attention-Guided Multispectral and Panchromatic Image Classification
    Shi, Cheng
    Dang, Yenan
    Fang, Li
    Lv, Zhiyong
    Shen, Huifang
    REMOTE SENSING, 2021, 13 (23)
  • [15] Attention-Guided GANs for Human Pose Transfer
    Zhang, Jinsong
    Zhao, Yuyang
    Li, Kun
    Liu, Yebin
    Yang, Jingyu
    Dai, Qionghai
    OPTOELECTRONIC IMAGING AND MULTIMEDIA TECHNOLOGY VI, 2019, 11187
  • [16] User Attention-guided Multimodal Dialog Systems
    Cui, Chen
    Wang, Wenjie
    Song, Xuemeng
    Huang, Minlie
    Xu, Xin-Shun
    Nie, Liqiang
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 445 - 454
  • [17] Attention-Guided Network for Semantic Video Segmentation
    Li, Jiangyun
    Zhao, Yikai
    Fu, Jun
    Wu, Jiajia
    Liu, Jing
    IEEE ACCESS, 2019, 7 : 140680 - 140689
  • [18] Attention-guided aggregation stereo matching network
    Zhang, Yaru
    Li, Yaqian
    Wu, Chao
    Liu, Bin
    IMAGE AND VISION COMPUTING, 2021, 106
  • [19] A model of attention-guided visual sparse coding
    Li, QY
    Shi, J
    Shi, ZZ
    ICCI 2005: FOURTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS - PROCEEDINGS, 2005, : 120 - 125
  • [20] Attention-Guided Wireless Channel Modeling and Generating
    He, Yawen
    Xu, Nan
    Cheng, Li
    Yuan, Haiwen
    APPLIED SCIENCES-BASEL, 2025, 15 (06):