SONIC: A Sparse Neural Network Inference Accelerator with Silicon Photonics for Energy-Efficient Deep Learning

被引:0
|
作者
Sunny, Febin [1 ]
Nikdast, Mandi [1 ]
Pasricha, Sudeep [1 ]
机构
[1] Colorado State Univ, Elect & Comp Engn, Ft Collins, CO 80523 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sparse neural networks can greatly facilitate the deployment of neural networks on resource-constrained platforms as they offer compact model sizes while retaining inference accuracy. Because of the sparsity in parameter matrices, sparse neural networks can, in principle, be exploited in accelerator architectures fur improved energy-efficiency, and latency. However, to realize these improvements in practice, there is a need to explore sparsity-aware hardware-software co-design. In this paper, we propose a novel silicon photonics-based sparse neural network inference accelerator called SONIC. SONIC takes advantage of the high energy-efficiency and low latency of photonic devices along with software co-optimization to accelerate sparse neural networks. Our experimental analysis shows that SONIC can achieve up to 5.8x better performance per-watt and 8.4x lower energy-per-hit than state-of-the-art sparse electronic neural network accelerators; and up to 13.8x better performance-per-watt and 27.6x lower energy-per-bit than the best known photonic neural network accelerators.
引用
收藏
页码:214 / 219
页数:6
相关论文
共 50 条
  • [1] An Energy-Efficient Deep Neural Network Accelerator Design
    Jung, Jueun
    Lee, Kyuho Jason
    [J]. 2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 272 - 276
  • [2] PIE: A Pipeline Energy-efficient Accelerator for Inference Process in Deep Neural Networks
    Zhao, Yangyang
    Yu, Qi
    Zhou, Xuda
    Zhou, Xuehai
    Wang, Chao
    Li, Xi
    [J]. 2016 IEEE 22ND INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2016, : 1067 - 1074
  • [3] An Energy-Efficient Inference Engine for a Configurable ReRAM-Based Neural Network Accelerator
    Zheng, Yang-Lin
    Yang, Wei-Yi
    Chen, Ya-Shu
    Han, Ding-Hung
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 740 - 753
  • [4] Ascend: A Scalable and Energy-Efficient Deep Neural Network Accelerator With Photonic Interconnects
    Li, Yuan
    Wang, Ke
    Zheng, Hao
    Louri, Ahmed
    Karanth, Avinash
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (07) : 2730 - 2741
  • [5] Energy-Efficient Multiply-and-Accumulate using Silicon Photonics for Deep Neural Networks
    Shiflett, Kyle
    Karanth, Avinash
    Louri, Ahmed
    Bunescu, Razvan
    [J]. 2020 IEEE PHOTONICS CONFERENCE (IPC), 2020,
  • [6] Efficient Hardware Accelerator for Compressed Sparse Deep Neural Network
    Xiao, Hao
    Zhao, Kaikai
    Liu, Guangzhu
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (05) : 772 - 775
  • [7] An Energy-Efficient Accelerator with Relative-Indexing Memory for Sparse Compressed Convolutional Neural Network
    Wu, I-Chen
    Huang, Po-Tsang
    Lo, Chin-Yang
    Hwang, Wei
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 42 - 45
  • [8] An Energy-Efficient Silicon Photonic-Assisted Deep Learning Accelerator for Big Data
    Li, Mengkun
    Wang, Yongjian
    [J]. WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2020, 2020
  • [9] An Energy-Efficient Sparse Deep-Neural-Network Learning Accelerator With Fine-Grained Mixed Precision of FP8-FP16
    Lee, Jinsu
    Lee, Juhyoung
    Han, Donghyeon
    Lee, Jinmook
    Park, Gwangtae
    Yoo, Hoi-Jun
    [J]. IEEE SOLID-STATE CIRCUITS LETTERS, 2019, 2 (11): : 232 - 235
  • [10] Implementing a Timing Error-Resilient and Energy-Efficient Near-Threshold Hardware Accelerator for Deep Neural Network Inference
    Gundi, Noel Daniel
    Pandey, Pramesh
    Roy, Sanghamitra
    Chakraborty, Koushik
    [J]. JOURNAL OF LOW POWER ELECTRONICS AND APPLICATIONS, 2022, 12 (02)