Efficient Attention: Attention with Linear Complexities

被引:285
|
作者
Shen Zhuoran [1 ]
Zhang Mingyuan [2 ]
Zhao Haiyu [2 ]
Yi Shuai [2 ]
Li Hongsheng [3 ]
机构
[1] 4244 Univ Way NE 85406, Seattle, WA 98105 USA
[2] SenseTime Int, 182 Cecil St,36-02 Frasers Tower, Singapore 069547, Singapore
[3] Chinese Univ Hong Kong, Sha Tin, Hong Kong, Peoples R China
关键词
D O I
10.1109/WACV48630.2021.00357
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dot-product attention has wide applications in computer vision and natural language processing. However, its memory and computational costs grow quadratically with the input size. Such growth prohibits its application on high-resolution inputs. To remedy this drawback, this paper proposes a novel efficient attention mechanism equivalent to dot-product attention but with substantially less memory and computational costs. Its resource efficiency allows more widespread and flexible integration of attention modules into a network, which leads to better accuracies. Empirical evaluations demonstrated the effectiveness of its advantages. Efficient attention modules brought significant performance boosts to object detectors and instance segmenters on MS-COCO 2017. Further, the resource efficiency democratizes attention to complex models, where high costs prohibit the use of dot-product attention. As an exemplar, a model with efficient attention achieved state-of-the-art accuracies for stereo depth estimation on the Scene Flow dataset. Code is available at https://github.com/cmsflash/efficient-attention.
引用
收藏
页码:3530 / 3538
页数:9
相关论文
共 50 条
  • [31] Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution
    Chu, Yuezhong
    Qiao, Yunan
    Liu, Heng
    Han, Jungong
    COGNITIVE COMPUTATION, 2022, 14 (03) : 1140 - 1151
  • [32] Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution
    Yuezhong Chu
    Yunan Qiao
    Heng Liu
    Jungong Han
    Cognitive Computation, 2022, 14 : 1140 - 1151
  • [33] IDS-attention: an efficient algorithm for intrusion detection systems using attention mechanism
    Laghrissi, FatimaEzzahra
    Douzi, Samira
    Douzi, Khadija
    Hssina, Badr
    JOURNAL OF BIG DATA, 2021, 8 (01)
  • [34] COVID-19 and the Future with Digital Mental Health: Need for Attention to Complexities
    Malla, Ashok
    Joober, Ridha
    CANADIAN JOURNAL OF PSYCHIATRY-REVUE CANADIENNE DE PSYCHIATRIE, 2021, 66 (01): : 14 - 16
  • [35] PPLA-Transformer: An Efficient Transformer for Defect Detection with Linear Attention Based on Pyramid Pooling
    Song, Xiaona
    Tian, Yubo
    Liu, Haichao
    Wang, Lijun
    Niu, Jinxing
    SENSORS, 2025, 25 (03)
  • [36] Minimum attention control for linear systemsA linear programming approach
    M. C. F. Donkers
    P. Tabuada
    W. P. M. H. Heemels
    Discrete Event Dynamic Systems, 2014, 24 : 199 - 218
  • [37] LA-ResUNet: An Efficient Linear Attention Mechanism in ResUNet for the Semantic Segmentation of Pulmonary Nodules
    Sarah Prithvika, P. C.
    Jani Anbarasi, L.
    IEEE ACCESS, 2024, 12 : 182894 - 182907
  • [38] Beyond Self-Attention: External Attention Using Two Linear Layers for Visual Tasks
    Guo, Meng-Hao
    Liu, Zheng-Ning
    Mu, Tai-Jiang
    Hu, Shi-Min
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5436 - 5447
  • [39] Selective attention and the formation of linear decision boundaries
    McKinley, SC
    Nosofsky, RM
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 1996, 22 (02) : 294 - 317
  • [40] Linear normalization attention neural Hawkes process
    Song, Zhi-yan
    Liu, Jian-wei
    Yang, Jie
    Zhang, Lu-ning
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (01): : 1025 - 1039