DaDianNao: A Machine-Learning Supercomputer

被引:932
|
作者
Chen, Yunji [1 ]
Luo, Tao [1 ,3 ]
Liu, Shaoli [1 ]
Zhang, Shijin [1 ]
He, Liqiang [2 ,4 ]
Wang, Jia [1 ]
Li, Ling [1 ]
Chen, Tianshi [1 ]
Xu, Zhiwei [1 ]
Sun, Ninghui [1 ]
Temam, Olivier [2 ]
机构
[1] Chinese Acad Sci, ICT, SKL Comp Architecture, Beijing, Peoples R China
[2] Inria, Scalay, France
[3] Univ CAS, Beijing, Peoples R China
[4] Inner Mongolia Univ, Hohhot, Peoples R China
关键词
D O I
10.1109/MICRO.2014.58
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Many companies are deploying services, either for consumers or industry, which are largely based on machine-learning algorithms for sophisticated processing of large amounts of data. The state-of-the-art and most popular such machine-learning algorithms are Convolutional and Deep Neural Networks (CNNs and DNNs), which are known to be both computationally and memory intensive. A number of neural network accelerators have been recently proposed which can offer high computational capacity/area ratio, but which remain hampered by memory accesses. However, unlike the memory wall faced by processors on general-purpose workloads, the CNNs and DNNs memory footprint, while large, is not beyond the capability of the on-chip storage of a multi-chip system. This property, combined with the CNN/DNN algorithmic characteristics, can lead to high internal bandwidth and low external communications, which can in turn enable high-degree parallelism at a reasonable area cost. In this article, we introduce a custom multi-chip machine-learning architecture along those lines. We show that, on a subset of the largest known neural network layers, it is possible to achieve a speedup of 450.65x over a GPU, and reduce the energy by 150.31x on average for a 64-chip system. We implement the node down to the place and route at 28nm, containing a combination of custom storage and computational units, with industry-grade interconnects.
引用
收藏
页码:609 / 622
页数:14
相关论文
共 50 条
  • [1] DaDianNao: A Neural Network Supercomputer
    Luo, Tao
    Liu, Shaoli
    Li, Ling
    Wang, Yuqing
    Zhang, Shijin
    Chen, Tianshi
    Xu, Zhiwei
    Temam, Olivier
    Chen, Yunji
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2017, 66 (01) : 73 - 88
  • [2] Machine-learning in astronomy
    Hobson, Michael
    Graff, Philip
    Feroz, Farhan
    Lasenby, Anthony
    [J]. STATISTICAL CHALLENGES IN 21ST CENTURY COSMOLOGY, 2015, 10 (306): : 279 - 287
  • [3] Machine-learning design
    Changjun Zhang
    [J]. Nature Energy, 2018, 3 : 535 - 535
  • [4] Machine-learning design
    Zhang, Changjun
    [J]. NATURE ENERGY, 2018, 3 (07): : 535 - 535
  • [5] Machine-Learning the Landscape
    He, Yang-Hui
    [J]. CALABI-YAU LANDSCAPE: FROM GEOMETRY, TO PHYSICS, TO MACHINE LEARNING, 2021, 2293 : 87 - 130
  • [6] Machine-learning wistfulness
    Davenport, Matt
    [J]. CHEMICAL & ENGINEERING NEWS, 2017, 95 (18) : 40 - 40
  • [7] Machine-learning with Cellular Automata
    Povalej, P
    Kokol, P
    Druzovec, TW
    Stiglic, B
    [J]. ADVANCES IN INTELLIGENT DATA ANALYSIS VI, PROCEEDINGS, 2005, 3646 : 305 - 315
  • [8] Certified Machine-Learning Models
    Damiani, Ernesto
    Ardagna, Claudio A.
    [J]. SOFSEM 2020: THEORY AND PRACTICE OF COMPUTER SCIENCE, 2020, 12011 : 3 - 15
  • [9] Machine-learning competition announced
    Erickson, J
    [J]. DR DOBBS JOURNAL, 1997, 22 (05): : 16 - 16
  • [10] Machine-learning micropattern manufacturing
    Wang, Si
    Shen, Ziao
    Shen, Zhenyu
    Dong, Yuanjun
    Li, Yanran
    Cao, Yuxin
    Zhang, Yanmei
    Guo, Shengshi
    Shuai, Jianwei
    Yang, Yun
    Lin, Changjian
    Chen, Xun
    Zhang, Xingcai
    Huang, Qiaoling
    [J]. NANO TODAY, 2021, 38