PeLK: Parameter-efficient Large Kernel ConvNets with Peripheral Convolution

被引:10
|
作者
Chen, Honghao [1 ,2 ,5 ]
Chu, Xiangxiang [3 ]
Ren, Yongjian [1 ,2 ]
Zhao, Xin [1 ,2 ]
Huang, Kaiqi [1 ,2 ,4 ]
机构
[1] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[3] Meituan, Beijing, Peoples R China
[4] CAS Ctr Excellence Brain Sci & Intelligence Techn, Beijing, Peoples R China
[5] Meituan Inc, Beijing, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52733.2024.00531
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, some large kernel convnets strike back with appealing performance and efficiency. However, given the square complexity of convolution, scaling up kernels can bring about an enormous amount of parameters and the proliferated parameters can induce severe optimization problem. Due to these issues, current CNNs compromise to scale up to 51 x 51 in the form of stripe convolution ( i.e., 51 x 5 + 5 x 51) and start to saturate as the kernel size continues growing. In this paper, we delve into addressing these vital issues and explore whether we can continue scaling up kernels for more performance gains. Inspired by human vision, we propose a human-like peripheral convolution that efficiently reduces over 90% parameter count of dense grid convolution through parameter sharing, and manage to scale up kernel size to extremely large. Our peripheral convolution behaves highly similar to human, reducing the complexity of convolution from O(K-2) to O(log K) without backfiring performance. Built on this, we propose Parameter-efficient Large Kernel Network (PeLK). Our PeLK outperforms modern vision Transformers and ConvNet architectures like Swin, ConvNeXt, RepLKNet and SLaK on various vision tasks including ImageNet classification, semantic segmentation on ADE20K and object detection on MS COCO. For the first time, we successfully scale up the kernel size of CNNs to an unprecedented 101 x 101 and demonstrate consistent improvements.
引用
收藏
页码:5557 / 5567
页数:11
相关论文
共 50 条
  • [1] HAda: Hyper-Adaptive Parameter-Efficient Learning for Multi-View ConvNets
    Wang, Shiye
    Li, Changsheng
    Yan, Zeyu
    Liang, Wanjun
    Yuan, Ye
    Wang, Guoren
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 85 - 99
  • [2] Kernel Modulation: A Parameter-Efficient Method for Training Convolutional Neural Networks
    Hu, Yuhuang
    Liu, Shih-Chii
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2192 - 2198
  • [3] Parameter-Efficient Masking Networks
    Bai, Yue
    Wang, Huan
    Ma, Xu
    Zhang, Yitian
    Tao, Zhiqiang
    Fu, Yun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [4] Parameter-efficient Tuning of Large-scale Multimodal Foundation Model
    Wang, Haixin
    Yang, Xinlong
    Chang, Jianlong
    Jin, Dian
    Sun, Jinan
    Zhang, Shikun
    Luo, Xiao
    Tian, Qi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Parameter-efficient deep probabilistic forecasting
    Sprangers, Olivier
    Schelter, Sebastian
    de Rijke, Maarten
    INTERNATIONAL JOURNAL OF FORECASTING, 2023, 39 (01) : 332 - 345
  • [6] Parameter-Efficient Transfer Learning for NLP
    Houlsby, Neil
    Giurgiu, Andrei
    Jastrzebski, Stanislaw
    Morrone, Bruna
    de laroussilhe, Quentin
    Gesmundo, Andrea
    Attariyan, Mona
    Gelly, Sylvain
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [7] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Luping Wang
    Sheng Chen
    Linnan Jiang
    Shu Pan
    Runze Cai
    Sen Yang
    Fei Yang
    Artificial Intelligence Review, 58 (8)
  • [8] Parameter-Efficient Fine-Tuning Large Speech Model Based on LoRA
    Ou, Ling
    Feng, Gen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 36 - 41
  • [9] Parameter-efficient Tuning for Large Language Model without Calculating Its Gradients
    Jin, Feihu
    Zhang, Jiajun
    Zong, Chengqing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 321 - 330
  • [10] An Efficient Large Kernel Convolution Network Designed for Neural Processing Unit
    Wang, Jiawen
    Liao, Chenfei
    Li, Dewei
    Zhao, Zhongqi
    Chen, Jingchuan
    Yang, Kehu
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 142