SparCML: High-Performance Sparse Communication for Machine Learning

被引:61
|
作者
Renggli, Cedric [1 ]
Ashkboos, Saleh [2 ]
Aghagolzadeh, Mehdi [3 ]
Alistarh, Dan [2 ]
Hoefler, Torsten [1 ]
机构
[1] Swiss Fed Inst Technol, Zurich, Switzerland
[2] IST Austria, Vienna, Austria
[3] Microsoft, Redmond, WA USA
基金
欧洲研究理事会;
关键词
Sparse AllReduce; Sparse Input Vectors; Sparse AllGather; OPERATIONS; DESCENT;
D O I
10.1145/3295500.3356222
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Applying machine learning techniques to the quickly growing data in science and industry requires highly-scalable algorithms. Large datasets are most commonly processed "data parallel" distributed across many nodes. Each node's contribution to the overall gradient is summed using a global allreduce. This allreduce is the single communication and thus scalability bottleneck for most machine learning workloads. We observe that frequently, many gradient values are (close to) zero, leading to sparse of sparsifyable communications. To exploit this insight, we analyze, design, and implement a set of communication-efficient protocols for sparse input data, in conjunction with efficient machine learning algorithms which can leverage these primitives. Our communication protocols generalize standard collective operations, by allowing processes to contribute arbitrary sparse input data vectors. Our generic communication library, SPARCML(1), extends MPI to support additional features, such as non-blocking (asynchronous) operations and low-precision data representations. As such, SPARCML and its techniques will form the basis of future highly-scalable machine learning frameworks.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Applications of machine learning method in high-performance materials design: a review
    Yuan, Junhao
    Li, Zhen
    Yang, Yujia
    Yin, Anyi
    Li, Wenjie
    Sun, Dan
    Wang, Qing
    JOURNAL OF MATERIALS INFORMATICS, 2024, 4 (03):
  • [22] MACHINE LEARNING AND SIMULATION BASED TEMPERATURE PREDICTION ON HIGH-PERFORMANCE PROCESSORS
    Knox, Carlton
    Yuan, Zihao
    Coskun, Ayse K.
    PROCEEDINGS OF ASME 2022 INTERNATIONAL TECHNICAL CONFERENCE AND EXHIBITION ON PACKAGING AND INTEGRATION OF ELECTRONIC AND PHOTONIC MICROSYSTEMS, INTERPACK2022, 2022,
  • [23] Silas: A high-performance machine learning foundation for logical reasoning and verification
    Bride, Hadrien
    Cai, Cheng-Hao
    Dong, Jie
    Dong, Jin Song
    Hou, Zhe
    Mirjalili, Seyedali
    Sun, Jing
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 176
  • [24] Understanding and Designing a High-Performance Ultrafiltration Membrane Using Machine Learning
    Gao, Haiping
    Zhong, Shifa
    Dangayach, Raghav
    Chen, Yongsheng
    ENVIRONMENTAL SCIENCE & TECHNOLOGY, 2023, 57 (46) : 17831 - 17840
  • [25] High-performance commercial data mining: A multistrategy machine learning application
    Hsu, WH
    Welge, M
    Redman, T
    Clutter, D
    DATA MINING AND KNOWLEDGE DISCOVERY, 2002, 6 (04) : 361 - 391
  • [26] Special-purpose parallel architectures for high-performance machine learning
    Battiti, R
    Lee, P
    Sartori, A
    Tecchiolli, G
    HIGH-PERFORMANCE COMPUTING AND NETWORKING, 1995, 919 : 944 - 944
  • [27] High-Performance Commercial Data Mining: A Multistrategy Machine Learning Application
    William H. Hsu
    Michael Welge
    Tom Redman
    David Clutter
    Data Mining and Knowledge Discovery, 2002, 6 : 361 - 391
  • [28] Interpretable machine learning for developing high-performance organic solar cells
    Abadi, Elyas Abbasi Jannat
    Sahu, Harikrishna
    Javadpour, Seyed Morteza
    Goharimanesh, Masoud
    MATERIALS TODAY ENERGY, 2022, 25
  • [29] Applying machine learning for high-performance named-entity extraction
    Baluja, S
    Mittal, VO
    Sukthankar, R
    COMPUTATIONAL INTELLIGENCE, 2000, 16 (04) : 586 - 595
  • [30] A comprehensive machine learning strategy for designing high-performance photoanode catalysts
    Huang, Meirong
    Wang, Sutong
    Zhu, Hongwei
    JOURNAL OF MATERIALS CHEMISTRY A, 2023, 11 (40) : 21619 - 21627