PeLK: Parameter-efficient Large Kernel ConvNets with Peripheral Convolution

被引:10
|
作者
Chen, Honghao [1 ,2 ,5 ]
Chu, Xiangxiang [3 ]
Ren, Yongjian [1 ,2 ]
Zhao, Xin [1 ,2 ]
Huang, Kaiqi [1 ,2 ,4 ]
机构
[1] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[3] Meituan, Beijing, Peoples R China
[4] CAS Ctr Excellence Brain Sci & Intelligence Techn, Beijing, Peoples R China
[5] Meituan Inc, Beijing, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52733.2024.00531
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, some large kernel convnets strike back with appealing performance and efficiency. However, given the square complexity of convolution, scaling up kernels can bring about an enormous amount of parameters and the proliferated parameters can induce severe optimization problem. Due to these issues, current CNNs compromise to scale up to 51 x 51 in the form of stripe convolution ( i.e., 51 x 5 + 5 x 51) and start to saturate as the kernel size continues growing. In this paper, we delve into addressing these vital issues and explore whether we can continue scaling up kernels for more performance gains. Inspired by human vision, we propose a human-like peripheral convolution that efficiently reduces over 90% parameter count of dense grid convolution through parameter sharing, and manage to scale up kernel size to extremely large. Our peripheral convolution behaves highly similar to human, reducing the complexity of convolution from O(K-2) to O(log K) without backfiring performance. Built on this, we propose Parameter-efficient Large Kernel Network (PeLK). Our PeLK outperforms modern vision Transformers and ConvNet architectures like Swin, ConvNeXt, RepLKNet and SLaK on various vision tasks including ImageNet classification, semantic segmentation on ADE20K and object detection on MS COCO. For the first time, we successfully scale up the kernel size of CNNs to an unprecedented 101 x 101 and demonstrate consistent improvements.
引用
收藏
页码:5557 / 5567
页数:11
相关论文
共 50 条
  • [21] Parameter-Efficient Model Adaptation for Vision Transformers
    He, Xuehai
    Li, Chuanyuan
    Zhang, Pengchuan
    Yang, Jianwei
    Wang, Xin Eric
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 817 - 825
  • [22] PET: Parameter-efficient Knowledge Distillation on Transformer
    Jeon, Hyojin
    Park, Seungcheol
    Kim, Jin-Gee
    Kang, U.
    PLOS ONE, 2023, 18 (07):
  • [23] Modular and Parameter-Efficient Multimodal Fusion with Prompting
    Liang, Sheng
    Zhao, Mengjie
    Schuetze, Hinrich
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2976 - 2985
  • [24] Parameter-Efficient Transfer Learning with Diff Pruning
    Guo, Demi
    Rush, Alexander M.
    Kim, Yoon
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4884 - 4896
  • [25] Characterizing Communication in Distributed Parameter-Efficient Fine-Tuning for Large Language Models
    Alnaasan, Nawras
    Huang, Horng-Ruey
    Shafi, Aamir
    Subramoni, Hari
    Panda, Dhabaleswar K.
    2024 IEEE SYMPOSIUM ON HIGH-PERFORMANCE INTERCONNECTS, HOTI 2024, 2024, : 11 - 19
  • [26] Parameter-Efficient Fine-Tuning of Large Pretrained Models for Instance Segmentation Tasks
    Baker, Nermeen Abou
    Rohrschneider, David
    Handmann, Uwe
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (04): : 2783 - 2807
  • [27] Structured segment rescaling with Gaussian processes for parameter efficient ConvNets
    Siddiqui, Bilal
    Alaeddini, Adel
    Zhu, Dakai
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 154
  • [28] PARAMETER-EFFICIENT TRANSFER LEARNING OF AUDIO SPECTROGRAM TRANSFORMERS
    Cappellazzo, Umberto
    Falavigna, Daniele
    Brutti, Alessio
    Ravanelli, Mirco
    2024 IEEE 34TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, MLSP 2024, 2024,
  • [29] Parameter-Efficient Sparse Retrievers and Rerankers Using Adapters
    Pal, Vaishali
    Lassance, Carlos
    Dejean, Herve
    Clinchant, Stephane
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT II, 2023, 13981 : 16 - 31
  • [30] Stochastic Bridges as Effective Regularizers for Parameter-Efficient Tuning
    Chen, Weize
    Han, Xu
    Lin, Yankai
    Liu, Zhiyuan
    Sun, Maosong
    Zhou, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10400 - 10420