Memristor-Based Multilayer Neural Networks With Online Gradient Descent Training

被引:185
|
作者
Soudry, Daniel [1 ,2 ]
Di Castro, Dotan [3 ]
Gal, Asaf [4 ]
Kolodny, Avinoam [5 ]
Kvatinsky, Shahar [6 ]
机构
[1] Columbia Univ, Dept Stat, New York, NY 10027 USA
[2] Columbia Univ, Grossman Ctr Stat Mind, Dept Stat, New York, NY 10027 USA
[3] Yahoo Labs, IL-31905 Haifa, Israel
[4] Technion Israel Inst Technol, Dept Elect Engn, Biol Networks Res Labs, IL-32000 Haifa, Israel
[5] Technion Israel Inst Technol, Dept Elect Engn, IL-32000 Haifa, Israel
[6] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
关键词
Backpropagation; hardware; memristive systems; memristor; multilayer neural networks (MNNs); stochastic gradient descent; synapse; ANALOG; PLASTICITY; DEVICES; DESIGN; MODEL;
D O I
10.1109/TNNLS.2014.2383395
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning in multilayer neural networks (MNNs) relies on continuous updating of large matrices of synaptic weights by local rules. Such locality can be exploited for massive parallelism when implementing MNNs in hardware. However, these update rules require a multiply and accumulate operation for each synaptic weight, which is challenging to implement compactly using CMOS. In this paper, a method for performing these update operations simultaneously (incremental outer products) using memristor-based arrays is proposed. The method is based on the fact that, approximately, given a voltage pulse, the conductivity of a memristor will increment proportionally to the pulse duration multiplied by the pulse magnitude if the increment is sufficiently small. The proposed method uses a synaptic circuit composed of a small number of components per synapse: one memristor and two CMOS transistors. This circuit is expected to consume between 2% and 8% of the area and static power of previous CMOS-only hardware alternatives. Such a circuit can compactly implement hardware MNNs trainable by scalable algorithms based on online gradient descent (e.g., backpropagation). The utility and robustness of the proposed memristor-based circuit are demonstrated on standard supervised learning tasks.
引用
收藏
页码:2408 / 2421
页数:14
相关论文
共 50 条
  • [41] On Memristor-Based Impulsive Neural Networks with Time-Delay
    Hu, Bin
    Guan, Zhi-Hong
    Liu, Zhi-Wei
    Jiang, Xiao-Wei
    [J]. 2017 29TH CHINESE CONTROL AND DECISION CONFERENCE (CCDC), 2017, : 4748 - 4753
  • [42] Modeling affections with memristor-based associative memory neural networks
    Hu, Xiaofang
    Duan, Shukai
    Chen, Guanrong
    Chen, Ling
    [J]. NEUROCOMPUTING, 2017, 223 : 129 - 137
  • [43] Adaptive synchronization of memristor-based neural networks with discontinuous activations
    Li, Yueheng
    Luo, Biao
    Liu, Derong
    Yang, Zhanyu
    Zhu, Yunli
    [J]. NEUROCOMPUTING, 2020, 381 : 196 - 206
  • [44] Memristor-based neural networks: Synaptic versus neuronal stochasticity
    Naous, Rawan
    AlShedivat, Maruan
    Neftci, Emre
    Cauwenberghs, Gert
    Salama, Khaled Nabil
    [J]. AIP ADVANCES, 2016, 6 (11):
  • [45] ANALYSIS OF GRADIENT DESCENT LEARNING ALGORITHMS FOR MULTILAYER FEEDFORWARD NEURAL NETWORKS
    GUO, H
    GELFAND, SB
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS, 1991, 38 (08): : 883 - 894
  • [46] Dynamics of on-line gradient descent learning for multilayer neural networks
    Saad, D
    Solla, SA
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 8: PROCEEDINGS OF THE 1995 CONFERENCE, 1996, 8 : 302 - 308
  • [47] Efficient Techniques for Training the Memristor-based Spiking Neural Networks Targeting Better Speed, Energy and Lifetime
    Ma, Yu
    Zhou, Pingqiang
    [J]. 2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 390 - 395
  • [48] Gradient Descent Analysis: On Visualizing the Training of Deep Neural Networks
    Becker, Martin
    Lippel, Jens
    Zielke, Thomas
    [J]. PROCEEDINGS OF THE 14TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS - VOL 3: IVAPP, 2019, : 338 - 345
  • [49] Training Itself: Mixed-signal Training Acceleration for Memristor-based Neural Network
    Li, Boxun
    Wang, Yuzhi
    Wang, Yu
    Chen, Yiran
    Yang, Huazhong
    [J]. 2014 19TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2014, : 361 - 366
  • [50] Training Neural Networks by Time-Fractional Gradient Descent
    Xie, Jingyi
    Li, Sirui
    [J]. AXIOMS, 2022, 11 (10)