MARS: Exploiting Multi-Level Parallelism for DNN Workloads on Adaptive Multi-Accelerator Systems

被引:0
|
作者
Shen, Guan [1 ]
Zhao, Jieru [1 ]
Wang, Zeke [2 ]
Lin, Zhe [3 ]
Ding, Wenchao [4 ]
Wu, Chentao [1 ]
Chen, Quan [1 ]
Guo, Minyi [1 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Zhejiang Univ, Hangzhou, Peoples R China
[3] Sun Yat Sen Univ, Guangzhou, Peoples R China
[4] Fudan Univ, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/DAC56929.2023.10247992
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Along with the fast evolution of deep neural networks, the hardware system is also developing rapidly. As a promising solution achieving high scalability and low manufacturing cost, multi-accelerator systems widely exist in data centers, cloud platforms, and SoCs. Thus, a challenging problem arises in multi-accelerator systems: selecting a proper combination of accelerators from available designs and searching for efficient DNN mapping strategies. To this end, we propose MARS, a novel mapping framework that can perform computation-aware accelerator selection, and apply communication-aware sharding strategies to maximize parallelism. Experimental results show that MARS can achieve 32.2% latency reduction on average for typical DNN workloads compared to the baseline, and 59.4% latency reduction on heterogeneous models compared to the corresponding state-of-the-art method.
引用
下载
收藏
页数:6
相关论文
共 50 条
  • [21] Multi-accelerator Neural Network Inference in Diversely Heterogeneous Embedded Systems
    Dagli, Ismet
    Belviranli, Mehmet E.
    PROCEEDINGS OF RSDHA 2021: REDEFINING SCALABILITY FOR DIVERSELY HETEROGENEOUS ARCHITECTURES, 2021, : 1 - 7
  • [22] Multi-level parallelism for protein prediction on the parallel computers
    Chen, J.
    Mo, Z. Y.
    Song, L.
    MOLECULAR & CELLULAR PROTEOMICS, 2004, 3 (10) : S248 - S248
  • [23] Exploring multi-level parallelism in cellular automata networks
    Calidonna, CR
    Di Napoli, C
    Giordano, M
    Furnari, MM
    HIGH PERFORMANCE COMPUTING, PROCEEDINGS, 2000, 1940 : 336 - 343
  • [24] The Introduction of Multi-level Parallelism Solvers in Multibody Dynamics
    Andreev, Andrey
    Egunov, Vitaly
    Movchan, Evgenia
    Cherednikov, Nikita
    Kharkov, Egor
    Kohtashvili, Natalia
    CREATIVITY IN INTELLIGENT TECHNOLOGIES AND DATA SCIENCE, PT II, 2019, 1084 : 166 - 180
  • [25] Support for multi-level parallelism on heterogeneous multi-core and performance optimization
    Li, Shi-Gang
    Hu, Chang-Jun
    Wang, Jue
    Li, Jian-Jiang
    Ruan Jian Xue Bao/Journal of Software, 2013, 24 (12): : 2782 - 2796
  • [26] Multi-Variable Agents Decomposition for DCOPs to Exploit Multi-Level Parallelism
    Fioretto, Ferdinando
    Yeoh, William
    Pontelli, Enrico
    PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 1823 - 1824
  • [27] An adaptive multi-level simulation algorithm for stochastic biological systems
    Lester, C.
    Yates, C. A.
    Giles, M. B.
    Baker, R. E.
    JOURNAL OF CHEMICAL PHYSICS, 2015, 142 (02):
  • [28] MULTI-LEVEL ADAPTIVE NETWORKS IN TANDEM AND HYBRID ASR SYSTEMS
    Bell, Peter
    Swietojanski, Pawel
    Renals, Steve
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 6975 - 6979
  • [29] MULTI-LEVEL DIFFUSION ADAPTIVE NETWORKS
    Cattivelli, Federico S.
    Sayed, Ali H.
    2009 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1- 8, PROCEEDINGS, 2009, : 2789 - 2792
  • [30] Multi-level metacognition for adaptive behavior
    Conn, Marvin
    M'Bale, Kenneth
    Josyula, Darsana
    BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, 2018, 26 : 174 - 183