MASSIVELY PARALLEL ARCHITECTURES FOR LARGE-SCALE NEURAL NETWORK SIMULATIONS

被引:13
|
作者
FUJIMOTO, Y
FUKUDA, N
AKABANE, T
机构
[1] SHARP CO LTD,INTEGRATED CIRCUITS GRP,CTR IC DEV,RES STAFF,TENRI,NARA 632,JAPAN
[2] SHARP CO LTD,CORP RES & DEV GRP,CTR INFORMAT SYST RES & DEV,TENRI,NARA 632,JAPAN
来源
关键词
D O I
10.1109/72.165590
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A toroidal lattice architecture (TLA) and a planar lattice architecture (PIA) as massively parallel architectures of neurocomputers for large scale neural network simulations are proposed. The performances of these architectures are almost proportional to the number of node processors and they adopt the most efficient two-dimensional processor connections to be implemented by the wafer scale integration (WSI) technology to date. They also give a solution to the connectivity problem, the performance degradation caused by the data transmission bottleneck, and the load balancing problem for efficient parallel processing in large scale neural network simulations. Furthermore, these architectures will have an even greater expandability of parallelism and exhibit great flexibility for the various configurations of neural networks and varieties of neuron models. First we define the general neuron model that is the basis of these massively parallel architectures. Then, we take a multilayer perceptron (MLP) as a typical example of neural networks and describe the simulation of the MLP using error back propagation learning algorithms on virtual processors (VP's) with the TLA and the PLA. Then, the mapping from the VP's to physical node processors with the same TLA and PLA is presented. This mapping is done by row and column partitions. At the same time, the row and column permutations are carried out for node processor load balancing. The mapping algorithm for the load balancing is given. An equation to estimate the performance of these architectures is also presented. Finally, we describe implementation of the TLA with transputers including a parallel processor configuration, load balance algorithm, and evaluation of its performance. We have implemented a Hopfield neural network and a MLP, and applied them to the traveling salesman problem (TSP) and the identity mapping (IM), respectively. The TLA neurocomputer has achieved 2 MCPS in a feedforward network and 600 KCUPS in a back propagation network using 16 transputers. Actual proof that its performance increases almost in proportion to the number of node processors is given.
引用
收藏
页码:876 / 888
页数:13
相关论文
共 50 条
  • [1] Large-scale powder mixer simulations using massively parallel GPU architectures
    Radeke, Charles A.
    Glasser, Benjamin J.
    Khinast, Johannes G.
    CHEMICAL ENGINEERING SCIENCE, 2010, 65 (24) : 6435 - 6442
  • [2] A massively parallel neural network approach to large-scale Euclidean traveling salesman problems
    Wang, Hongjian
    Zhang, Naiyu
    Creput, Jean-Charles
    NEUROCOMPUTING, 2017, 240 : 137 - 151
  • [3] Massively parallel simulation on large-scale carbon nanotubes
    Tejima, S
    Berebr, S
    Minami, K
    Jimbo, N
    Nakamura, H
    Kanada, Y
    Tamanek, D
    NANOTECH 2003, VOL 3, 2003, : 102 - 105
  • [4] Design of large-scale parallel simulations
    Knepley, MG
    Sameh, AH
    Sarin, V
    PARALLEL COMPUTATIONAL FLUID DYNAMICS: TOWARDS TERAFLOPS, OPTIMIZATION, AND NOVEL FORMULATIONS, 2000, : 273 - 279
  • [5] A Simulator for Large-Scale Parallel Computer Architectures
    Janssen, Curtis L.
    Adalsteinsson, Helgi
    Cranford, Scott
    Kenny, Joseph P.
    Pinar, Ali
    Evensky, David A.
    Mayo, Jackson
    INTERNATIONAL JOURNAL OF DISTRIBUTED SYSTEMS AND TECHNOLOGIES, 2010, 1 (02) : 57 - 73
  • [6] Parallel Large-Scale Neural Network Training For Online Advertising
    Qi, Quanchang
    Lu, Guangming
    Zhang, Jun
    Yang, Lichun
    Liu, Haishan
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 343 - 350
  • [7] Massively Parallel Multilevel Fast Multipole Algorithm for Extremely Large-Scale Electromagnetic Simulations: A Review
    He, Wei-Jia
    Huang, Xiao-Wei
    Yang, Ming-Lin
    Sheng, Xin-Qing
    Progress in Electromagnetics Research, 2022, 173 : 37 - 52
  • [8] Fast Transform-Based Preconditioners for Large-Scale Power Grid Analysis on Massively Parallel Architectures
    Daloukas, Konstantis
    Evmorfopoulos, Nestor
    Drasidis, George
    Tsiampas, Michalis
    Tsompanopoulou, Panagiota
    Stamoulis, George I.
    2012 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2012, : 384 - 391
  • [9] Massively Parallel Multilevel Fast Multipole Algorithm for Extremely Large-Scale Electromagnetic Simulations: A Review
    He, Wei-Jia
    Huang, Xiao-Wei
    Yang, Ming-Lin
    Sheng, Xin-Qing
    PROGRESS IN ELECTROMAGNETICS RESEARCH-PIER, 2022, 173 : 37 - 52
  • [10] Visualization of Large-Scale Neural Simulations
    Hernando, Juan B.
    Duelo, Carlos
    Martin, Vicente
    BRAIN-INSPIRED COMPUTING, 2014, 8603 : 184 - 197