A High-Performance and Energy-Efficient Photonic Architecture for Multi-DNN Acceleration

被引:0
|
作者
Li, Yuan [1 ]
Louri, Ahmed [1 ]
Karanth, Avinash [2 ]
机构
[1] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
[2] Ohio Univ, Sch Elect Engn & Comp Sci, Athens, OH 45701 USA
基金
美国国家科学基金会;
关键词
Accelerator; dataflow; deep neural network; silicon photonics;
D O I
10.1109/TPDS.2023.3327535
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Large-scale deep neural network (DNN) accelerators are poised to facilitate the concurrent processing of diverse DNNs, imposing demanding challenges on the interconnection fabric. These challenges encompass overcoming performance degradation and energy increase associated with system scaling while also necessitating flexibility to support dynamic partitioning and adaptable organization of compute resources. Nevertheless, conventional metallic-based interconnects frequently confront inherent limitations in scalability and flexibility. In this paper, we leverage silicon photonic interconnects and adopt an algorithm-architecture co-design approach to develop MDA, a DNN accelerator meticulously crafted to empower high-performance and energy-efficient concurrent processing of diverse DNNs. Specifically, MDA consists of three novel components: 1) a resource allocation algorithm that assigns compute resources to concurrent DNNs based on their computational demands and priorities; 2) a dataflow selection algorithm that determines off-chip and on-chip dataflows for each DNN, with the objectives of minimizing off-chip and on-chip memory accesses, respectively; 3) a flexible silicon photonic network that can be dynamically segmented into sub-networks, each interconnecting the assigned compute resources of a certain DNN while adapting to the communication patterns dictated by the selected on-chip dataflow. Simulation results show that the proposed MDA accelerator outperforms other state-of-the-art multi-DNN accelerators, including PREMA, AI-MT, Planaria, and HDA. MDA accelerator achieves a speedup of 3.6, accompanied by substantial improvements of 7.3x, 12.7x, and 9.2x in energy efficiency, service-level agreement (SLA) satisfaction rate, and fairness, respectively.
引用
收藏
页码:46 / 58
页数:13
相关论文
共 50 条
  • [1] Versa-DNN: A Versatile Architecture Enabling High-Performance and Energy-Efficient Multi-DNN Acceleration
    Yang, Jiaqi
    Zheng, Hao
    Louri, Ahmed
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (02) : 349 - 361
  • [2] Galaxy: A High-Performance Energy-Efficient Multi-Chip Architecture Using Photonic Interconnects
    Demir, Yigit
    Pan, Yan
    Song, Seukwoo
    Hardavellas, Nikos
    Kim, John
    Memik, Gokhan
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON SUPERCOMPUTING, (ICS'14), 2014, : 303 - 312
  • [3] A Silicon Photonic Multi-DNN Accelerator
    Li, Yuan
    Louri, Ahmed
    Karanth, Avinash
    2023 32ND INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT, 2023, : 238 - 249
  • [4] Algorithm/Architecture Co-Design for Energy-Efficient Acceleration of Multi-Task DNN
    Shin, Jaekang
    Choi, Seungkyu
    Ra, Jongwoo
    Kim, Lee -Sup
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 253 - 258
  • [5] Polyform: A Versatile Architecture for Multi-DNN Execution via Spatial and Temporal Acceleration
    Yin, Lingxiang
    Ghazizadeh, Amir
    Tian, Shilin
    Louri, Ahmed
    Zheng, Hao
    2023 IEEE 41ST INTERNATIONAL CONFERENCE ON COMPUTER DESIGN, ICCD, 2023, : 166 - 169
  • [6] A High-Performance, Energy-Efficient Modular DMA Engine Architecture
    Benz, Thomas
    Rogenmoser, Michael
    Scheffler, Paul
    Riedel, Samuel
    Ottaviano, Alessandro
    Kurth, Andreas
    Hoefler, Torsten
    Benini, Luca
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (01) : 263 - 277
  • [7] Morph-GCNX: A Universal Architecture for High-Performance and Energy-Efficient Graph Convolutional Network Acceleration
    Wang, Ke
    Zheng, Hao
    Li, Jiajun
    Louri, Ahmed
    IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING, 2024, 9 (02): : 115 - 127
  • [8] GANPU: An Energy-Efficient Multi-DNN Training Processor for GANs With Speculative Dual-Sparsity Exploitation
    Kang, Sanghoon
    Han, Donghyeon
    Lee, Juhyoung
    Im, Dongseok
    Kim, Sangyeob
    Kim, Soyeon
    Ryu, Junha
    Yoo, Hoi-Jun
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2021, 56 (09) : 2845 - 2857
  • [9] Energy-Efficient and High-Performance Software Architecture for Storage Class Memory
    Baek, Seungjae
    Choi, Jongmoo
    Lee, Donghee
    Noh, Sam H.
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2013, 12 (03)
  • [10] High-performance, energy-efficient IGBTs
    Snyder, Lucy A.
    Electron Prod Garden City NY, 2008, 8