Bit-width Adaptive Accelerator Design for Convolution Neural Network

被引:1
|
作者
Guo, Jianxin [1 ]
Yin, Shouyi [1 ]
Ouyang, Peng [1 ]
Tu, Fengbin [1 ]
Tang, Shibin [1 ]
Liu, Leibo [1 ]
Wei, Shaojun [1 ]
机构
[1] Tsinghua Univ, Inst Microelect, Beijing 100084, Peoples R China
关键词
D O I
10.1109/ISCAS.2018.8351666
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNNs) have achieved great success in many applications. Recently, various FPGA-based accelerators have been proposed to improve the performance of CNNs. However, current most FPGA-based methods only use the same bit-width selection for all CNN layers which lead to very low resource utilization and difficulty in further performance improvement. In this paper, we propose a bit-width adaptive accelerator design approach which can adapt to the CNN layers with various bit-width requirements in a same network. We construct multiple different bit-width convolutional processors to compute the CNN layers in parallel way. We partition the FPGA DSP resources and use our optimization approach to find the optimal resource allocation. On a Xilinx Virtex-7 FPGA, our design approach achieves higher throughput than the state-of-the-art FPGA-based CNN accelerators from 5.48x to 7.25x and by 6.20x on average, when we evaluate the convolutional layers of AlexNet and deeper VGG CNNs.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Accelerating Low Bit-Width Deep Convolution Neural Network in MRAM
    He, Zhezhi
    Angizi, Shaahin
    Fan, Deliang
    [J]. 2018 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2018, : 533 - 538
  • [2] Bit-width Reduction and Customized Register for Low Cost Convolutional Neural Network Accelerator
    Choi, Kyungrak
    Choi, Woong
    Shin, Kyungho
    Park, Jongsun
    [J]. 2017 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN (ISLPED), 2017,
  • [3] An Efficient Streaming Accelerator for Low Bit-Width Convolutional Neural Networks
    Chen, Qinyu
    Fu, Yuxiang
    Song, Wenqing
    Cheng, Kaifeng
    Lu, Zhonghai
    Zhang, Chuan
    Li, Li
    [J]. ELECTRONICS, 2019, 8 (04)
  • [4] Low Bit-Width Convolutional Neural Network on RRAM
    Cai, Yi
    Tang, Tianqi
    Xia, Lixue
    Li, Boxun
    Wang, Yu
    Yang, Huazhong
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (07) : 1414 - 1427
  • [5] Design Flow of Accelerating Hybrid Extremely Low Bit-width Neural Network in Embedded FPGA
    Wang, Junsong
    Lou, Qiuwen
    Zhang, Xiaofan
    Zhu, Chao
    Lin, Yonghua
    Chen, Deming
    [J]. 2018 28TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2018, : 163 - 169
  • [6] Low Cost Convolutional Neural Network Accelerator Based on Bi-Directional Filtering and Bit-Width Reduction
    Choi, Woong
    Choi, Kyungrak
    Park, Jongsun
    [J]. IEEE ACCESS, 2018, 6 : 14734 - 14746
  • [7] An Energy-Efficient Accelerator for Hybrid Bit-width DNNs
    Liu, Bo
    Ruan, Xing
    Xia, Mengwen
    Gong, Yu
    Yang, Jinjiang
    Ge, Wei
    Yang, Jun
    [J]. 2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 3306 - 3313
  • [8] Adaptive bit-width compression for low-energy frame memory design
    Moshnyaga, VG
    [J]. SIPS 2001: IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS: DESIGN AND IMPLEMENTATION, 2001, : 185 - 192
  • [9] FlexBNN: Fast Private Binary Neural Network Inference With Flexible Bit-Width
    Dong, Ye
    Chen, Xiaojun
    Song, Xiangfu
    Li, Kaiyun
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 2382 - 2397
  • [10] A Low Bit-width Parameter Representation Method for Hardware-oriented Convolution Neural Networks
    Chen, Qiang
    Xin, Chen
    Zou, Chenglong
    Wang, Xinan
    Wang, Bo
    [J]. 2017 IEEE 12TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2017, : 148 - 151