The Case for Domain-Specific Networks

被引:0
|
作者
Abts, Dennis [1 ]
Kim, John [2 ]
机构
[1] NVIDIA, Santa Clara, CA 95050 USA
[2] Korea Adv Inst Sci & Technol, Daejeon, South Korea
关键词
D O I
10.1109/HOTI59126.2023.00021
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Modern parallel computers are dichotomized into capacity or capability systems. Capacity systems cater to a wide range of weak scaling workloads, using distributed parallel systems with message passing while capability systems focus on strong scaling workloads across a significant fraction of the machine's processing units. The interconnection network differs under these regimes, with commodity Ethernet or Infiniband solutions typically deployed for capacity systems, while capabilityclass systems often necessitate tightly-coupled, fine-grained communication. Systems built for AI training and inference embody traits from both classes: tight coupling and strong scaling for model parallelism, and weak scaling for data parallelism in a distributed system. Handling 100-billion-parameter large-language models and trillion-token data sets presents computational challenges for current supercomputing infrastructure. This paper discusses the crucial role of the interconnection network in these large-scale systems, advocating for flexible, low-latency interconnects that can deliver high throughput at large scales with tens of thousands of endpoints. This work also emphasizes the importance of reliability and resilience in enduring long-running training workloads and demanding inference requirements of domain-specific workloads.
引用
下载
收藏
页码:49 / 52
页数:4
相关论文
共 50 条
  • [1] Domain-Specific Networks for Machine Learning
    Abts, Dennis
    2020 14TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON NETWORKS-ON-CHIP (NOCS), 2020,
  • [2] A Case for Domain-Specific Curiosity in Mathematics
    Peterson, Emily Grossnickle
    Cohen, Jana
    EDUCATIONAL PSYCHOLOGY REVIEW, 2019, 31 (04) : 807 - 832
  • [3] A Case for Domain-Specific Curiosity in Mathematics
    Emily Grossnickle Peterson
    Jana Cohen
    Educational Psychology Review, 2019, 31 : 807 - 832
  • [4] NNL:a domain-specific language for neural networks
    王秉睿
    Chen Yunji
    High Technology Letters, 2020, 26 (02) : 160 - 167
  • [5] A domain-specific language approach to programmable networks
    Muller, G
    Lawall, JL
    Thibault, S
    Jensen, REV
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2003, 33 (03): : 370 - 381
  • [6] A Domain-Specific Architecture for Deep Neural Networks
    Jouppi, Norman P.
    Young, Cliff
    Patil, Nishant
    Patterson, David
    COMMUNICATIONS OF THE ACM, 2018, 61 (09) : 50 - 59
  • [7] NNL: a domain-specific language for neural networks
    Wang B.
    Chen Y.
    High Technology Letters, 2020, 26 (02) : 160 - 167
  • [8] Domain-specific model differencing for graphical domain-specific languages
    Jafarlou, Manouchehr Zadahmad
    ACM/IEEE 25TH INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS, MODELS 2022 COMPANION, 2022, : 205 - 208
  • [9] A domain-specific measure of centrality for water distribution networks
    Zarghami, Seyed Ashkan
    Gunawan, Indra
    ENGINEERING CONSTRUCTION AND ARCHITECTURAL MANAGEMENT, 2020, 27 (02) : 341 - 355
  • [10] UTILIZING DOMAIN-SPECIFIC INFORMATION FOR THE OPTIMIZATION OF LOGISTICS NETWORKS
    Rabe, Markus
    Ammouriova, Majsa
    Schmitt, Dominik
    2018 WINTER SIMULATION CONFERENCE (WSC), 2018, : 2873 - 2884