Attention-Guided Collaborative Counting

被引:10
|
作者
Mo, Hong [1 ]
Ren, Wenqi [2 ]
Zhang, Xiong [3 ]
Yan, Feihu [4 ]
Zhou, Zhong [1 ]
Cao, Xiaochun [2 ]
Wu, Wei [1 ]
机构
[1] Beihang Univ, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
[3] Neolix Autonomous Vehicle, Beijing 100016, Peoples R China
[4] Beijing Univ Civil Engn & Architecture, Sch Elect & Informat Engn, Beijing 100044, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Collaboration; Transformers; Task analysis; Head; Computational modeling; Computer vision; Crowd counting; attention-guided collaborative counting model; bi-directional transformer;
D O I
10.1109/TIP.2022.3207584
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing crowd counting designs usually exploit multi-branch structures to address the scale diversity problem. However, branches in these structures work in a competitive rather than collaborative way. In this paper, we focus on promoting collaboration between branches. Specifically, we propose an attention-guided collaborative counting module (AGCCM) comprising an attention-guided module (AGM) and a collaborative counting module (CCM). The CCM promotes collaboration among branches by recombining each branch's output into an independent count and joint counts with other branches. The AGM capturing the global attention map through a transformer structure with a pair of foreground-background related loss functions can distinguish the advantages of different branches. The loss functions do not require additional labels and crowd division. In addition, we design two kinds of bidirectional transformers (Bi-Transformers) to decouple the global attention to row attention and column attention. The proposed Bi-Transformers are able to reduce the computational complexity and handle images in any resolution without cropping the image into small patches. Extensive experiments on several public datasets demonstrate that the proposed algorithm performs favorably against the state-of-the-art crowd counting methods.
引用
收藏
页码:6306 / 6319
页数:14
相关论文
共 50 条
  • [21] Real-time Speaker counting in a cocktail party scenario using Attention-guided Convolutional Neural Network
    Yousefi, Midia
    Hansen, John H. L.
    INTERSPEECH 2021, 2021, : 1484 - 1488
  • [22] SiamAGN: Siamese attention-guided network for visual tracking
    Wei, Bingbing
    Chen, Hongyu
    Ding, Qinghai
    Luo, Haibo
    NEUROCOMPUTING, 2022, 512 : 69 - 82
  • [23] Efficient Attention-guided Transformer for brain stroke segmentation
    Soliman, Ahmed
    Zafari-Ghadim, Yalda
    Rashed, Essam A.
    Mabrok, Mohamed
    9TH INTERNATIONAL CONFERENCE ON MULTIMEDIA AND IMAGE PROCESSING, ICMIP 2024, 2024, : 127 - 131
  • [24] Attention-Guided Low-Rank Tensor Completion
    Truong Thanh Nhat Mai
    Lam, Edmund Y.
    Lee, Chul
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 9818 - 9833
  • [25] A Explainable Attention-Guided Iris Presentation Attack Detector
    Chen, Cunjian
    Ross, Arun
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS (WACVW 2021), 2021, : 97 - 106
  • [26] Attention-guided Face Alignment Based on Inverted Residuals
    Tong, Xinjie
    Wang, Teng
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 759 - 763
  • [27] Attention-guided Progressive Mapping for Profile Face Recognition
    Huang, Junyang
    Ding, Changxing
    2021 INTERNATIONAL JOINT CONFERENCE ON BIOMETRICS (IJCB 2021), 2021,
  • [28] Object Detection by Attention-Guided Feature Fusion Network
    Shi, Yuxuan
    Fan, Yue
    Xu, Siqi
    Gao, Yue
    Gao, Ran
    SYMMETRY-BASEL, 2022, 14 (05):
  • [29] Attention-guided Low light enhancement CNN Network
    Liang, Xiwen
    Yan, Xiaoning
    Xu, Nenghua
    Chen, Xiaoyan
    Feng, Hao
    JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2023, 9 (04):
  • [30] Heterogeneous Face Recognition with Attention-guided Feature Disentangling
    Yang, Shanmin
    Yang, Xiao
    Lin, Yi
    Cheng, Peng
    Zhang, Yi
    Zhang, Jianwei
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4137 - 4145