Kernel Shape Control for Row-Efficient Convolution on Processing-In-Memory Arrays

被引:2
|
作者
Rhe, Johnny [1 ]
Jeon, Kang Eun [1 ]
Lee, Joo Chan [2 ]
Jeong, Seongmoon [2 ]
Ko, Jong Hwan [3 ]
机构
[1] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon, South Korea
[2] Sungkyunkwan Univ, Dept Artificial Intelligence, Suwon, South Korea
[3] Sungkyunkwan Univ, Coll Informat & Commun Engn, Suwon, South Korea
基金
新加坡国家研究基金会;
关键词
processing-in-memory; shift and duplicate (SDK) weight mapping; weight pruning; neural compression; ARCHITECTURE; PRECISION;
D O I
10.1109/ICCAD57390.2023.10323749
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Processing-in-memory (PIM) architectures have been highlighted as one of the viable solutions for faster and more power-efficient convolutional neural networks (CNNs) inference. Recently, shift and duplicate kernel (SDK) convolutional weight mapping scheme was proposed, achieving up to 50% throughput improvement over the prior arts. However, the traditional pattern-based pruning methods, which were adopted for row-skipping and computing cycle reduction, are not optimal for the latest SDK mapping due to structural irregularity caused by the shifted and duplicated kernels. To address this issue, we propose a method called kernel shape control (KERNTROL) that aims to promote structural regularity for achieving a high row-skipping ratio and model accuracy. Instead of pruning certain weight elements permanently, KERNTROL controls the kernel shapes through the omission of certain weights based on their mapped columns. In comparison to the latest pattern-based pruning approaches, KERNTROL achieves up to 36.4% improvement in the compression rate, and 38.6% in array utilization with maintaining the original model accuracy.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Identification of a convolution kernel in a control problem for the heat equation with a boundary memory term
    Cavaterra, Cecilia
    Guidetti, Davide
    ANNALI DI MATEMATICA PURA ED APPLICATA, 2014, 193 (03) : 779 - 816
  • [32] A Flexible Yet Efficient DNN Pruning Approach for Crossbar-Based Processing-in-Memory Architectures
    Zheng, Long
    Liu, Haifeng
    Huang, Yu
    Chen, Dan
    Liu, Chaoqiang
    He, Haiheng
    Liao, Xiaofei
    Jin, Hai
    Xue, Jingling
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3745 - 3756
  • [33] RIME: A Scalable and Energy-Efficient Processing-In-Memory Architecture for Floating-Point Operations
    Lu, Zhaojun
    Arafin, Md Tanvir
    Qu, Gang
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 120 - 125
  • [34] Efficient Error-Correcting-Code Mechanism for High-Throughput Memristive Processing-in-Memory
    Leitersdorf, Orian
    Perach, Ben
    Ronen, Ronny
    Kvatinsky, Shahar
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 199 - 204
  • [35] Lattice: An ADC/DAC-less ReRAM-based Processing-In-Memory Architecture for Accelerating Deep Convolution Neural Networks
    Zheng, Qilin
    Wang, Zongwei
    Feng, Zishun
    Yan, Bonan
    Cai, Yimao
    Huang, Ru
    Chen, Yiran
    Yang, Chia-Lin
    Li, Hai
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [36] PAIRS: Pruning-AIded Row-Skipping for SDK-Based Convolutional Weight Mapping in Processing-In-Memory Architectures
    Rhe, Johnny
    Jeon, Kang Eun
    Ko, Jong Hwan
    2023 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN, ISLPED, 2023,
  • [37] VW-SDK: Efficient Convolutional Weight Mapping Using Variable Windows for Processing-In-Memory Architectures
    Rhe, Johnny
    Moon, Sungmin
    Ko, Jong Hwan
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 214 - 219
  • [38] Quant-PIM: An Energy-Efficient Processing-in-Memory Accelerator for Layerwise Quantized Neural Networks
    Lee, Young Seo
    Chung, Eui-Young
    Gong, Young-Ho
    Chung, Sung Woo
    IEEE EMBEDDED SYSTEMS LETTERS, 2021, 13 (04) : 162 - 165
  • [39] Design of Processing-in-Memory With Triple Computational Path and Sparsity Handling for Energy-Efficient DNN Training
    Han, Wontak
    Heo, Jaehoon
    Kim, Junsoo
    Lim, Sukbin
    Kim, Joo-Young
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2022, 12 (02) : 354 - 366
  • [40] CRPIM: An efficient compute-reuse scheme for ReRAM-based Processing-in-Memory DNN accelerators
    Hong, Shihao
    Chung, Yeh-Ching
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 153