Accelerating CNN Algorithm with Fine-grained Dataflow Architectures

被引:3
|
作者
Xiang, Taoran [1 ,2 ]
Feng, Yujing [1 ]
Ye, Xiaochun [1 ]
Tan, Xu [1 ,2 ]
Li, Wenming [1 ]
Zhu, Yatao [1 ]
Wu, Meng [1 ]
Zhang, Hao [1 ]
Fan, Dongrui [1 ,2 ]
机构
[1] Chinese Acad Sci, ICT, State Key Lab Comp Architecture, Beijing, Peoples R China
[2] UCAS, Sch Comp & Control Engn, Beijing, Peoples R China
来源
IEEE 20TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING AND COMMUNICATIONS / IEEE 16TH INTERNATIONAL CONFERENCE ON SMART CITY / IEEE 4TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND SYSTEMS (HPCC/SMARTCITY/DSS) | 2018年
基金
中国国家自然科学基金;
关键词
fine-grained dataflow; Convolutional Neural Network; general accelerator; data reuse; high parallel;
D O I
10.1109/HPCC/SmartCity/DSS.2018.00063
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional Neural Network(CNN) is a hot and state-of-the-art algorithm which is widely used in applications such as face recognition, intelligent monitoring, image recognition and text recognition. Because of its high computational complexity, many efficient hardware accelerators have been proposed to exploit high degree of parallel processing for CNN. However, accelerators which are implemented on FPGAs and ASICs usually sacrifice generality for higher performance and lower power consumption. Other accelerators, such as GPUs, are general enough, but they lead to higher power consumption. Fine-grained dataflow architectures, which break conventional Von Neumann architectures, show natural advantages in processing CNN-like algorithms with high computational efficiency and low power consumption. At the same time, it remains broadly applicable and adaptable. In this paper, we propose a scheme for implementing and optimizing CNN on fine-grained dataflow architecture based accelerators. The experiment results reveal that by using our scheme, the performance of AlexNet running on the dataflow accelerator is 3.11x higher than that on NVIDIA Tesla K80, and the power consumption of our hardware is 8.52x lower than that of K80.
引用
收藏
页码:243 / 251
页数:9
相关论文
共 50 条
  • [41] Boosting the Performance of CNN Accelerators with Dynamic Fine-Grained Channel Gating
    Hua, Weizhe
    Zhou, Yuan
    De Sa, Christopher
    Zhang, Zhiru
    Suh, G. Edward
    MICRO'52: THE 52ND ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, 2019, : 139 - 150
  • [42] Hybrid ViT-CNN Network for Fine-Grained Image Classification
    Shao, Ran
    Bi, Xiao-Jun
    Chen, Zheng
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1109 - 1113
  • [43] General hardware multicasting for fine-grained message-passing architectures
    Naylor, Matthew
    Moore, Simon W.
    Thomas, David
    Beaumont, Jonathan R.
    Fleming, Shane
    Vousden, Mark
    Markettos, A. Theodore
    Bytheway, Thomas
    Brown, Andrew
    2021 29TH EUROMICRO INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED AND NETWORK-BASED PROCESSING (PDP 2021), 2021, : 126 - 133
  • [44] Fine-Grained Multi-Query Stream Processing on Integrated Architectures
    Zhang, Feng
    Zhang, Chenyang
    Yang, Lin
    Zhang, Shuhao
    He, Bingsheng
    Lu, Wei
    Du, Xiaoyong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (09) : 2303 - 2320
  • [45] Towards fine-grained automated verification of publish-subscribe architectures
    Baresi, Luciano
    Ghezzi, Carlo
    Mottola, Luca
    FORMAL TECHNIQUES FOR NETWORKED AND DISTRIBUTED SYSTEMS - FORTE 2006, 2006, 4229 : 131 - 135
  • [46] A recommendation algorithm based on fine-grained feature analysis
    Lu, Wenjie
    Altenbek, Gulila
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 163
  • [47] An Ideal Fine-Grained GAC Algorithm for Table Constraints
    Qiao, Limeng
    Xu, Zhenhui
    Dong, Jin
    Shao, Yuan
    Tong, Xin
    Li, Zhanshan
    ADVANCES IN SWARM INTELLIGENCE, ICSI 2016, PT I, 2016, 9712 : 86 - 94
  • [48] Fine-Grained Probability Counting: Refined LogLog Algorithm
    Wang, Lun
    Cai, Zekun
    Wang, Hao
    Jiang, Jie
    Yang, Tong
    Cui, Bin
    Li, Xiaoming
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), 2018, : 176 - 183
  • [49] Improve Fine-Grained Feature Learning in Fine-Grained DataSet GAI
    Wang, Hai Peng
    Geng, Zhi Qing
    IEEE ACCESS, 2025, 13 : 12777 - 12788
  • [50] Leveraging Fine-Grained Labels to Regularize Fine-Grained Visual Classification
    Wu, Junfeng
    Yao, Li
    Liu, Bin
    Ding, Zheyuan
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON COMPUTER MODELING AND SIMULATION (ICCMS 2019) AND 8TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND APPLICATIONS (ICICA 2019), 2019, : 133 - 136