Efficient Attention: Attention with Linear Complexities

被引:285
|
作者
Shen Zhuoran [1 ]
Zhang Mingyuan [2 ]
Zhao Haiyu [2 ]
Yi Shuai [2 ]
Li Hongsheng [3 ]
机构
[1] 4244 Univ Way NE 85406, Seattle, WA 98105 USA
[2] SenseTime Int, 182 Cecil St,36-02 Frasers Tower, Singapore 069547, Singapore
[3] Chinese Univ Hong Kong, Sha Tin, Hong Kong, Peoples R China
关键词
D O I
10.1109/WACV48630.2021.00357
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dot-product attention has wide applications in computer vision and natural language processing. However, its memory and computational costs grow quadratically with the input size. Such growth prohibits its application on high-resolution inputs. To remedy this drawback, this paper proposes a novel efficient attention mechanism equivalent to dot-product attention but with substantially less memory and computational costs. Its resource efficiency allows more widespread and flexible integration of attention modules into a network, which leads to better accuracies. Empirical evaluations demonstrated the effectiveness of its advantages. Efficient attention modules brought significant performance boosts to object detectors and instance segmenters on MS-COCO 2017. Further, the resource efficiency democratizes attention to complex models, where high costs prohibit the use of dot-product attention. As an exemplar, a model with efficient attention achieved state-of-the-art accuracies for stereo depth estimation on the Scene Flow dataset. Code is available at https://github.com/cmsflash/efficient-attention.
引用
收藏
页码:3530 / 3538
页数:9
相关论文
共 50 条
  • [1] PARAMETER-EFFICIENT VISION TRANSFORMER WITH LINEAR ATTENTION
    Zhao, Youpeng
    Tang, Huadong
    Jiang, Yingying
    Yong, A.
    Wu, Qiang
    Wang, Jun
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1275 - 1279
  • [2] Efficient Linear Attention for Fast and Accurate Keypoint Matching
    Suwanwimolkul, Suwichaya
    Komorita, Satoshi
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 330 - 341
  • [3] ScatterFormer: Efficient Voxel Transformer with Scattered Linear Attention
    He, Chenhang
    Li, Ruihuang
    Zhang, Guowen
    Zhang, Lei
    COMPUTER VISION - ECCV 2024, PT XXIX, 2025, 15087 : 74 - 92
  • [4] Agent Attention: On the Integration of Softmax and Linear Attention
    Han, Dongchen
    Ye, Tianzhu
    Han, Yizeng
    Xia, Zhuofan
    Pan, Siyuan
    Wan, Pengfei
    Song, Shiji
    Huang, Gao
    COMPUTER VISION - ECCV 2024, PT L, 2025, 15108 : 124 - 140
  • [5] Attention deficit/hyperactivity disorder: complexities and controversies
    Schonwald, Alison
    Lechner, Erin
    CURRENT OPINION IN PEDIATRICS, 2006, 18 (02) : 189 - 195
  • [8] Efficient conformer-based speech recognition with linear attention
    Li, Shengqiang
    Xu, Menglong
    Zhang, Xiao-Lei
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 448 - 453
  • [9] AGE DIFFERENCES IN INFANTS ATTENTION TO PATTERNS OF DIFFERENT COMPLEXITIES
    BRENNAN, WM
    AMES, EW
    MOORE, RW
    SCIENCE, 1966, 151 (3708) : 354 - &
  • [10] Linear-Time Self Attention with Codeword Histogram for Efficient Recommendation
    Wu, Yongji
    Lian, Defu
    Gong, Neil Zhenqiang
    Yin, Lu
    Yin, Mingyang
    Zhou, Jingren
    Yang, Hongxia
    PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 1262 - 1273