A Small-Footprint Accelerator for Large-Scale Neural Networks

被引:3
|
作者
Chen, Tianshi [1 ]
Zhang, Shijin [1 ]
Liu, Shaoli [1 ]
Du, Zidong [1 ]
Luo, Tao [1 ]
Gao, Yuan [2 ]
Liu, Junjie [2 ]
Wang, Dongsheng [2 ]
Wu, Chengyong [1 ]
Sun, Ninghui [1 ]
Chen, Yunji [1 ,4 ]
Temam, Olivier [3 ]
机构
[1] Chinese Acad Sci, ICT, SKLCA, Beijing 100190, Peoples R China
[2] Tsinghua Univ, TNLIST, Beijing 100084, Peoples R China
[3] Inria, Saclay, France
[4] Chinese Acad Sci, Ctr Excellence Brain Sci, Beijing 100190, Peoples R China
来源
ACM TRANSACTIONS ON COMPUTER SYSTEMS | 2015年 / 33卷 / 02期
关键词
Architecture; Processor; Hardware; RECOGNITION;
D O I
10.1145/2701417
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Machine-learning tasks are becoming pervasive in a broad range of domains, and in a broad range of systems (from embedded systems to data centers). At the same time, a small set of machine-learning algorithms (especially Convolutional and Deep Neural Networks, i.e., CNNs and DNNs) are proving to be state-of-the-art across many applications. As architectures evolve toward heterogeneous multicores composed of a mix of cores and accelerators, a machine-learning accelerator can achieve the rare combination of efficiency (due to the small number of target algorithms) and broad application scope. Until now, most machine-learning accelerator designs have been focusing on efficiently implementing the computational part of the algorithms. However, recent state-of-the-art CNNs and DNNs are characterized by their large size. In this study, we design an accelerator for large-scale CNNs and DNNs, with a special emphasis on the impact of memory on accelerator design, performance, and energy. We show that it is possible to design an accelerator with a high throughput, capable of performing 452 GOP/s (key NN operations such as synaptic weight multiplications and neurons outputs additions) in a small footprint of 3.02 mm(2) and 485mW; compared to a 128-bit 2GHz SIMD processor, the accelerator is 117.87x faster, and it can reduce the total energy by 21.08x. The accelerator characteristics are obtained after layout at 65nm. Such a high throughput in a small footprint can open up the usage of state-of-the-art machine-learning algorithms in a broad set of systems and for a broad set of applications.
引用
收藏
页数:27
相关论文
共 50 条
  • [41] Large-scale neural networks and the lateralization of motivation and emotion
    Tops, Mattie
    Quirin, Markus
    Boksem, Maarten A. S.
    Koole, Sander L.
    INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2017, 119 : 41 - 49
  • [42] Large-Scale Plant Classification with Deep Neural Networks
    Heredia, Ignacio
    ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2017, 2017, : 259 - 262
  • [43] Large-scale self-normalizing neural networks
    Zhaodong Chen
    Weiqin Zhao
    Lei Deng
    Yufei Ding
    Qinghao Wen
    Guoqi Li
    Yuan Xie
    Journal of Automation and Intelligence, 2024, 3 (02) : 101 - 110
  • [44] Visualization of density relations in large-scale neural networks
    Nadasdy, Z
    Zaborszky, L
    ANATOMY AND EMBRYOLOGY, 2001, 204 (04): : 303 - 317
  • [45] Neural-Event-Triggered fMRI of large-scale neural networks
    Logothetis, Nikos K.
    CURRENT OPINION IN NEUROBIOLOGY, 2015, 31 : 214 - 222
  • [46] Population networks: A large-scale framework for modelling cortical neural networks
    Mallot, HA
    Giannakopoulos, F
    BIOLOGICAL CYBERNETICS, 1996, 75 (06) : 441 - 452
  • [47] AUTOMATIC GAIN CONTROL AND MULTI-STYLE TRAINING FOR ROBUST SMALL-FOOTPRINT KEYWORD SPOTTING WITH DEEP NEURAL NETWORKS
    Prabhavalkar, Rohit
    Alvarez, Raziel
    Parada, Carolina
    Nakkiran, Preetum
    Sainath, Tara N.
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4704 - 4708
  • [48] PocketSUMMIT: Small-Footprint Continuous Speech Recognition
    Hetherington, I. Lee
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 2173 - 2176
  • [49] Large-Scale Mapping of Small Roads in Lidar Images Using Deep Convolutional Neural Networks
    Salberg, Arnt-Borre
    Trier, Oivind Due
    Kampffmeyer, Michael
    IMAGE ANALYSIS, SCIA 2017, PT II, 2017, 10270 : 193 - 204
  • [50] Node pruning based on Entropy of Weights and Node Activity for Small-footprint Acoustic Model based on Deep Neural Networks
    Takeda, Ryu
    Nakadai, Kazuhiro
    Komatani, Kazunori
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 1636 - 1640