A Stochastic Computational Multi-Layer Perceptron with Backward Propagation

被引:66
|
作者
Liu, Yidong [1 ]
Liu, Siting [1 ]
Wang, Yanzhi [2 ]
Lombardi, Fabrizio [3 ]
Han, Jie [1 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 1H9, Canada
[2] Syracuse Univ, Elect Engn & Comp Sci Dept, Syracuse, NY 13244 USA
[3] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02115 USA
基金
加拿大自然科学与工程研究理事会;
关键词
Stochastic computation; binary search; neural network; probability estimator; multi-layer perceptron; HARDWARE IMPLEMENTATION; NEURAL-NETWORK;
D O I
10.1109/TC.2018.2817237
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Stochastic computation has recently been proposed for implementing artificial neural networks with reduced hardware and power consumption, but at a decreased accuracy and processing speed. Most existing implementations are based on pre-training such that the weights are predetermined for neurons at different layers, thus these implementations lack the ability to update the values of the network parameters. In this paper, a stochastic computational multi-layer perceptron (SC-MLP) is proposed by implementing the backward propagation algorithm for updating the layer weights. Using extended stochastic logic (ESL), a reconfigurable stochastic computational activation unit (SCAU) is designed to implement different types of activation functions such as the tanh and the rectifier function. A triple modular redundancy (TMR) technique is employed for reducing the random fluctuations in stochastic computation. A probability estimator (PE) and a divider based on the TMR and a binary search algorithm are further proposed with progressive precision for reducing the required stochastic sequence length. Therefore, the latency and energy consumption of the SC-MLP are significantly reduced. The simulation results show that the proposed design is capable of implementing both the training and inference processes. For the classification of nonlinearly separable patterns, at a slight loss of accuracy by 1.32-1.34 percent, the proposed design requires only 28.5-30.1 percent of the area and 18.9-23.9 percent of the energy consumption incurred by a design using floating point arithmetic. Compared to a fixed-point implementation, the SC-MLP consumes a smaller area (40.7-45.5 percent) and a lower energy consumption (38.0-51.0 percent) with a similar processing speed and a slight drop of accuracy by 0.15-0.33 percent. The area and the energy consumption of the proposed design is from 80.7-87.1 percent and from 71.9-93.1 percent, respectively, of a binarized neural network (BNN), with a similar accuracy.
引用
收藏
页码:1273 / 1286
页数:14
相关论文
共 50 条
  • [31] Geno-mathematical identification of the multi-layer perceptron
    Ostermark, Ralf
    NEURAL COMPUTING & APPLICATIONS, 2009, 18 (04): : 331 - 344
  • [32] Battle royale optimizer for training multi-layer perceptron
    Saeid Agahian
    Taymaz Akan
    Evolving Systems, 2022, 13 : 563 - 575
  • [33] Multi-layer perceptron learning in the domain of attributed graphs
    Jain, BJ
    Wysotzki, F
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 1003 - 1008
  • [34] Many-objective training of a multi-layer perceptron
    Koeppen, Mario
    Yoshida, Kaori
    NEURAL NETWORK WORLD, 2007, 17 (06) : 627 - 637
  • [35] A novel scheme to determine the architecture of a Multi-layer Perceptron
    Chintalapudi, KK
    Pal, NR
    1998 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-5, 1998, : 2297 - 2302
  • [36] Prediction of zenith tropospheric delay by multi-layer perceptron
    Katsougiannopoulos, S.
    Pikridas, C.
    JOURNAL OF APPLIED GEODESY, 2009, 3 (04) : 223 - 229
  • [37] Multi-layer perceptron based modelling of nonlinear systems
    Lightbody, G
    Irwin, GW
    FUZZY SETS AND SYSTEMS, 1996, 79 (01) : 93 - 112
  • [38] Intensified Customer Churn Prediction: Connectivity with Weighted Multi-Layer Perceptron and Enhanced Multipath Back Propagation
    Panimalar, S. Arockia
    Krishnakumar, A.
    Kumar, S. Senthil
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [39] Real-time optimization of Dynamic Stochastic Knapsack Problem using a multi-layer perceptron network
    Ramaraj, S
    Teodorovic, D
    Gracanin, D
    MLMTA'03: INTERNATIONAL CONFERENCE ON MACHINE LEARNING; MODELS, TECHNOLOGIES AND APPLICATIONS, 2003, : 223 - 229
  • [40] PROPAGATION OF CASCADES IN A MULTI-LAYER MEDIUM
    KHRISTOV, KI
    SOVIET PHYSICS JETP-USSR, 1958, 6 (04): : 676 - 680