NEEBS: Nonexpert large-scale environment building system for deep neural network

被引:0
|
作者
Tajima, Yoshiharu [1 ,2 ]
Asaoka, Masahiro [1 ]
Tabuchi, Akihiro [1 ]
Kasagi, Akihiko [1 ]
Tabaru, Tsuguchika [1 ]
机构
[1] Fujitsu Ltd, Fujitsu Labs, Kawasaki, Kanagawa, Japan
[2] Fujitsu Ltd, Fujitsu Labs, 4-1-1 Kamikodanaka,Nakahara ku, Kawasaki, Kanagawa, Japan
来源
关键词
BERT; deep neural network; large-scale clusters; natural language processing;
D O I
10.1002/cpe.7499
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep neural networks (DNNs) have greatly improved the accuracy of various tasks in areas such as natural language processing (NLP). Obtaining a highly accurate DNN model requires multiple repetitions of training on a huge dataset, which requires a large-scale cluster the compute nodes of which are tightly connected by high-speed interconnects to exchange a large amount of intermediate data with very short latency. However, fully using the computational power of a large-scale cluster for training requires knowledge of its components such as a distributed file system, an interconnection, and optimized high-performance libraries. We have developed a Non-Expert large-scale Environment Building System (NEEBS) that aids a user in building a fast-running training environment on a large-scale cluster. It automatically installs and configures the applications and necessary libraries. It also optimally prepares tools to stage both data and executable programs, and launcher scripts suitable for both the applications and job submission systems of the cluster. NEEBS achieves 93.91% throughput scalability in NLP pretraining. We also present an approach to reduce pretraining time of highly accurate DNN model for NLP using a large-scale computation environment built using NEEBS. We trained a Bidirectional Encoder Representations from Transformers (BERT)-3.9b and a BERT-xlarge using a dense masked language model (MLM) on Megatron-LM framework and evaluated the improvement in learning time and learning efficiency for a Japanese language dataset using 768 graphics processing units (GPUs) on the AI Bridging Cloud Infrastructure (ABCI). Our implementation NEEBS improved learning efficiency per iteration by a factor of 10 and completed the pretraining of BERT-xlarge in 4.7 h. This pretraining takes 5 months on a single GPU. To determine if the BERT models are correctly pretrained, we evaluated their accuracy in two tasks, Stanford Natural Language Inference Corpus translated into Japanese (JSNLI) and Twitter reputation analysis (TwitterRA). BERT-3.9b achieved 94.30% accuracy for JSNLI, and BERT-xlarge achieved 90.63% accuracy for TwitterRA. We constructed pretrained models with comparable accuracy to other Japanese BERT models in a shorter time.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] NEEBS: Nonexpert large-scale environment building system for deep neural network
    Tajima, Yoshiharu
    Asaoka, Masahiro
    Tabuchi, Akihiro
    Kasagi, Akihiko
    Tabaru, Tsuguchika
    [J]. Concurrency and Computation: Practice and Experience, 2023, 35 (19):
  • [2] Brain-inspired Large-scale Deep Neural Network System
    Lü J.-C.
    Ye Q.
    Tian Y.-X.
    Han J.-W.
    Wu F.
    [J]. Ruan Jian Xue Bao/Journal of Software, 2022, 33 (04): : 1412 - 1429
  • [3] Large-scale photovoltaic system in green building: MPPT control based on deep neural network and dynamic time-window
    An, Qing
    Tang, Ruoli
    Hu, Qiqi
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (04) : 4769 - 4787
  • [4] Large-scale water quality prediction with integrated deep neural network
    Bi, Jing
    Lin, Yongze
    Dong, Quanxi
    Yuan, Haitao
    Zhou, MengChu
    [J]. INFORMATION SCIENCES, 2021, 571 (571) : 191 - 205
  • [5] Multiscale Deep Alternative Neural Network for Large-Scale Video Classification
    Wang, Jinzhuo
    Wang, Wenmin
    Gao, Wen
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (10) : 2578 - 2592
  • [6] Classification of large-scale stellar spectra based on deep convolutional neural network
    Liu, W.
    Zhu, M.
    Dai, C.
    He, D. Y.
    Yao, Jiawen
    Tian, H. F.
    Wang, B. Y.
    Wu, K.
    Zhan, Y.
    Chen, B. -Q.
    Luo, A-Li
    Wang, R.
    Cao, Y.
    Yu, X. C.
    [J]. MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2019, 483 (04) : 4774 - 4783
  • [7] A large-scale in-memory computing for deep neural network with trained quantization
    Cheng, Yuan
    Wang, Chao
    Chen, Hai-Bao
    Yu, Hao
    [J]. INTEGRATION-THE VLSI JOURNAL, 2019, 69 : 345 - 355
  • [8] GRAPH NEURAL NETWORK FOR LARGE-SCALE NETWORK LOCALIZATION
    Yan, Wenzhong
    Jin, Di
    Lin, Zhidi
    Yin, Feng
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 5250 - 5254
  • [9] Local Self-Adaptation Mechanisms for Large-Scale Neural System Building
    Ortiz, M. Garcia
    Gepperth, A.
    [J]. ADVANCES IN COGNITIVE NEURODYNAMICS (II), 2011, : 543 - 551
  • [10] Temperature and humidity environment of a large-scale wooden building
    Morooka, T
    Norimoto, M
    Isoda, N
    [J]. MOKUZAI GAKKAISHI, 2000, 46 (05): : 421 - 430