MASSIVELY PARALLEL ARCHITECTURES FOR LARGE-SCALE NEURAL NETWORK SIMULATIONS

被引:13
|
作者
FUJIMOTO, Y
FUKUDA, N
AKABANE, T
机构
[1] SHARP CO LTD,INTEGRATED CIRCUITS GRP,CTR IC DEV,RES STAFF,TENRI,NARA 632,JAPAN
[2] SHARP CO LTD,CORP RES & DEV GRP,CTR INFORMAT SYST RES & DEV,TENRI,NARA 632,JAPAN
来源
关键词
D O I
10.1109/72.165590
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A toroidal lattice architecture (TLA) and a planar lattice architecture (PIA) as massively parallel architectures of neurocomputers for large scale neural network simulations are proposed. The performances of these architectures are almost proportional to the number of node processors and they adopt the most efficient two-dimensional processor connections to be implemented by the wafer scale integration (WSI) technology to date. They also give a solution to the connectivity problem, the performance degradation caused by the data transmission bottleneck, and the load balancing problem for efficient parallel processing in large scale neural network simulations. Furthermore, these architectures will have an even greater expandability of parallelism and exhibit great flexibility for the various configurations of neural networks and varieties of neuron models. First we define the general neuron model that is the basis of these massively parallel architectures. Then, we take a multilayer perceptron (MLP) as a typical example of neural networks and describe the simulation of the MLP using error back propagation learning algorithms on virtual processors (VP's) with the TLA and the PLA. Then, the mapping from the VP's to physical node processors with the same TLA and PLA is presented. This mapping is done by row and column partitions. At the same time, the row and column permutations are carried out for node processor load balancing. The mapping algorithm for the load balancing is given. An equation to estimate the performance of these architectures is also presented. Finally, we describe implementation of the TLA with transputers including a parallel processor configuration, load balance algorithm, and evaluation of its performance. We have implemented a Hopfield neural network and a MLP, and applied them to the traveling salesman problem (TSP) and the identity mapping (IM), respectively. The TLA neurocomputer has achieved 2 MCPS in a feedforward network and 600 KCUPS in a back propagation network using 16 transputers. Actual proof that its performance increases almost in proportion to the number of node processors is given.
引用
收藏
页码:876 / 888
页数:13
相关论文
共 50 条
  • [31] NEURAL NETWORK REASONING ALGORITHM OF LARGE-SCALE GRAGH BASED ON PARALLEL COMPUTING
    Zeng Keqin
    2022 19TH INTERNATIONAL COMPUTER CONFERENCE ON WAVELET ACTIVE MEDIA TECHNOLOGY AND INFORMATION PROCESSING (ICCWAMTIP), 2022,
  • [32] Improving parallel performance of large-scale watershed simulations
    Eller, Paul R.
    Cheng, Jing-Ru C.
    Nguyen, Hung V.
    Maier, Robert S.
    ICCS 2010 - INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE, PROCEEDINGS, 2010, 1 (01): : 801 - 808
  • [33] Parallel computation of large-scale molecular dynamics simulations
    Kwon, Sungjin
    Lee, Youngmin
    Im, Seyoung
    Experimental Mechanics in Nano and Biotechnology, Pts 1 and 2, 2006, 326-328 : 341 - 344
  • [34] Conservative synchronization of large-scale network simulations
    Park, A
    Fujimoto, RM
    Perumalla, KS
    18TH WORKSHOP ON PARALLEL AND DISTRIBUTED SIMULATION, PROCEEDINGS, 2004, : 153 - 161
  • [35] N-BODY SIMULATIONS ON MASSIVELY-PARALLEL ARCHITECTURES
    STILLER, L
    DAEMEN, LL
    GUBERNATIS, JE
    JOURNAL OF COMPUTATIONAL PHYSICS, 1994, 115 (02) : 550 - 552
  • [36] A 3-D Fast Transform-Based Preconditioner for Large-Scale Power Grid Analysis on Massively Parallel Architectures
    Daloukas, Konstantis
    Evmorfopoulos, Nestor
    Tsompanopoulou, Panagiota
    Stamoulis, George I.
    PROCEEDINGS OF THE FIFTEENTH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED 2014), 2015, : 723 - 730
  • [37] New neural network architectures based on fractional Fourier transforms for large-scale optical implementations
    Lee, SY
    Jin, SI
    Bae, YS
    Shin, SG
    Park, CH
    Shin, SY
    PROGRESS IN CONNECTIONIST-BASED INFORMATION SYSTEMS, VOLS 1 AND 2, 1998, : 692 - 695
  • [38] GRAPH NEURAL NETWORK FOR LARGE-SCALE NETWORK LOCALIZATION
    Yan, Wenzhong
    Jin, Di
    Lin, Zhidi
    Yin, Feng
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 5250 - 5254
  • [39] A Large-Scale Study of Activation Functions in Modern Deep Neural Network Architectures for Efficient Convergence
    Rasamoelina, Andrinandrasana David
    Cik, Ivan
    Sincak, Peter
    Mach, Marian
    Hruska, Lukas
    INTELIGENCIA ARTIFICIAL-IBEROAMERICAN JOURNAL OF ARTIFICIAL INTELLIGENCE, 2022, 25 (70): : 95 - 109
  • [40] A Massively Parallel BWP Algorithm for Solving Large-Scale Systems of Nonlinear Equations
    Silva, Bruno
    Lopes, Luiz Guerreiro
    2023 IEEE High Performance Extreme Computing Conference, HPEC 2023, 2023,