Large Scale Recurrent Neural Network on GPU

被引:0
|
作者
Li, Boxun [1 ]
Zhou, Erjin [1 ]
Huang, Bo [1 ]
Duan, Jiayi [1 ]
Wang, Yu [1 ]
Xu, Ningyi [2 ]
Zhang, Jiaxing [2 ]
Yang, Huazhong [1 ]
机构
[1] Tsinghua Univ, Tsinghua Natl Lab Informat Sci & Technol, Dept EE, Beijing 100084, Peoples R China
[2] Microsoft Res Asia, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large scale artificial neural networks (ANNs) have been widely used in data processing applications. The recurrent neural network (RNN) is a special type of neural network equipped with additional recurrent connections. Such a unique architecture enables the recurrent neural network to remember the past processed information and makes it an expressive model for nonlinear sequence processing tasks. However, the large computation complexity makes it difficult to effectively train a recurrent neural network and therefore significantly limits the research on the recurrent neural network in the last 20 years. In recent years, the use of graphics processing units (GPUs) becomes a significant advance to speed up the training process of large scale neural networks by taking advantage of the massive parallelism capabilities of GPUs. In this paper, we propose an efficient GPU implementation of the large scale recurrent neural network and demonstrate the power of scaling up the recurrent neural network with GPUs. We first explore the potential parallelism of the recurrent neural network and propose a fine-grained two-stage pipeline implementation. Experiment results show that the proposed GPU implementation can achieve 2 similar to 11x speed-up compared with the basic CPU implementation with the Intel Math Kernel Library. We then use the proposed GPU implementation to scale up the recurrent neural network and improve its performance. The experiment results of the Microsoft Research Sentence Completion Challenge demonstrate that the large scale recurrent network without class layer is able to beat the traditional class-based modest-size recurrent network and achieve an accuracy of 47%, the best result achieved by a single recurrent neural network on the same dataset.
引用
收藏
页码:4062 / 4069
页数:8
相关论文
共 50 条
  • [1] Forecasting large scale conditional volatility and covariance using neural network on GPU
    Xianggao Cai
    Guoming Lai
    Xiaola Lin
    The Journal of Supercomputing, 2013, 63 : 490 - 507
  • [2] Forecasting large scale conditional volatility and covariance using neural network on GPU
    Cai, Xianggao
    Lai, Guoming
    Lin, Xiaola
    JOURNAL OF SUPERCOMPUTING, 2013, 63 (02): : 490 - 507
  • [3] Reinforcement learning in a large-scale photonic recurrent neural network
    Bueno, J.
    Maktoobi, S.
    Froehly, L.
    Fischer, I.
    Jacquot, M.
    Larger, L.
    Brunner, D.
    OPTICA, 2018, 5 (06): : 756 - 760
  • [4] POSTER: ParGNN: Efficient Training for Large-Scale Graph Neural Network on GPU Clusters
    Li, Shunde
    Gu, Junyu
    Wang, Jue
    Yao, Tiechui
    Liang, Zhiqiang
    Shi, Yumeng
    Li, Shigang
    Xi, Weiting
    Li, Shushen
    Zhou, Chunbao
    Wang, Yangang
    Chi, Xuebin
    PROCEEDINGS OF THE 29TH ACM SIGPLAN ANNUAL SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, PPOPP 2024, 2024, : 469 - 471
  • [5] FontRNN: Generating Large-scale Chinese Fonts via Recurrent Neural Network
    Tang, Shusen
    Xia, Zeqing
    Lian, Zhouhui
    Tang, Yingmin
    Xiao, Jianguo
    COMPUTER GRAPHICS FORUM, 2019, 38 (07) : 567 - 577
  • [6] Data Parallel Large Sparse Deep Neural Network on GPU
    Sattar, Naw Safrin
    Arifuzzaman, Shaikh
    2020 IEEE 34TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW 2020), 2020, : 1006 - 1014
  • [7] Accelerating Maximal Bicliques Enumeration with GPU on large scale network
    Wu, Chunqi
    Li, Jingdong
    Li, Zhao
    Zhang, Ji
    Tang, Pan
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 601 - 613
  • [8] Collective behavior of large-scale neural networks with GPU acceleration
    Jingyi Qu
    Rubin Wang
    Cognitive Neurodynamics, 2017, 11 : 553 - 563
  • [9] Collective behavior of large-scale neural networks with GPU acceleration
    Qu, Jingyi
    Wang, Rubin
    COGNITIVE NEURODYNAMICS, 2017, 11 (06) : 553 - 563
  • [10] Long Short-Term Memory Recurrent Neural Network Architectures for Large Scale Acoustic Modeling
    Sak, Hasim
    Senior, Andrew
    Beaufays, Francoise
    15TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2014), VOLS 1-4, 2014, : 338 - 342