In situ training of feed-forward and recurrent convolutional memristor networks

被引:0
|
作者
Zhongrui Wang
Can Li
Peng Lin
Mingyi Rao
Yongyang Nie
Wenhao Song
Qinru Qiu
Yunning Li
Peng Yan
John Paul Strachan
Ning Ge
Nathan McDonald
Qing Wu
Miao Hu
Huaqiang Wu
R. Stanley Williams
Qiangfei Xia
J. Joshua Yang
机构
[1] University of Massachusetts,Department of Electrical and Computer Engineering
[2] Hewlett Packard Labs,Department of Electrical Engineering and Computer Science
[3] Hewlett Packard Enterprise,Information Directorate
[4] Syracuse University,Department of Electrical and Computer Engineering
[5] Air Force Research Laboratory,Institute of Microelectronics
[6] Binghamton University,Department of Electrical and Computer Engineering
[7] Tsinghua University,undefined
[8] Texas A&M University,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
The explosive growth of machine learning is largely due to the recent advancements in hardware and architecture. The engineering of network structures, taking advantage of the spatial or temporal translational isometry of patterns, naturally leads to bio-inspired, shared-weight structures such as convolutional neural networks, which have markedly reduced the number of free parameters. State-of-the-art microarchitectures commonly rely on weight-sharing techniques, but still suffer from the von Neumann bottleneck of transistor-based platforms. Here, we experimentally demonstrate the in situ training of a five-level convolutional neural network that self-adapts to non-idealities of the one-transistor one-memristor array to classify the MNIST dataset, achieving similar accuracy to the memristor-based multilayer perceptron with a reduction in trainable parameters of ~75% owing to the shared weights. In addition, the memristors encoded both spatial and temporal translational invariance simultaneously in a convolutional long short-term memory network—a memristor-based neural network with intrinsic 3D input processing—which was trained in situ to classify a synthetic MNIST sequence dataset using just 850 weights. These proof-of-principle demonstrations combine the architectural advantages of weight sharing and the area/energy efficiency boost of the memristors, paving the way to future edge artificial intelligence.
引用
收藏
页码:434 / 442
页数:8
相关论文
共 50 条
  • [1] In situ training of feed-forward and recurrent convolutional memristor networks
    Wang, Zhongrui
    Li, Can
    Lin, Peng
    Rao, Mingyi
    Nie, Yongyang
    Song, Wenhao
    Qiu, Qinru
    Li, Yunning
    Yan, Peng
    Strachan, John Paul
    Ge, Ning
    McDonald, Nathan
    Wu, Qing
    Hu, Miao
    Wu, Huaqiang
    Williams, R. Stanley
    Xia, Qiangfei
    Yang, J. Joshua
    NATURE MACHINE INTELLIGENCE, 2019, 1 (09) : 434 - 442
  • [2] Evolutionary approach to training feed-forward and recurrent neural networks
    Riley, Jeff
    Ciesielski, Victor B.
    International Conference on Knowledge-Based Intelligent Electronic Systems, Proceedings, KES, 1998, 3 : 596 - 602
  • [3] An evolutionary approach to training feed-forward and recurrent neural networks.
    Riley, J
    Ciesielski, VB
    1998 SECOND INTERNATIONAL CONFERENCE ON KNOWLEDGE-BASED INTELLIGENT ELECTRONIC SYSTEMS, KES '98, PROCEEDINGS, VOL, 3, 1998, : 596 - 602
  • [4] Feed-forward and recurrent neural networks in signal prediction
    Prochazka, Ales
    Pavelka, Ales
    ICCC 2007: 5TH IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL CYBERNETICS, PROCEEDINGS, 2007, : 93 - 96
  • [5] Metaplasticity and memory in multilevel recurrent feed-forward networks
    Zanardi, Gianmarco
    Bettotti, Paolo
    Morand, Jules
    Pavesi, Lorenzo
    Tubiana, Luca
    PHYSICAL REVIEW E, 2024, 110 (05)
  • [6] A greenhouse control with feed-forward and recurrent neural networks
    Fourati, Fathi
    Chtourou, Mohamed
    SIMULATION MODELLING PRACTICE AND THEORY, 2007, 15 (08) : 1016 - 1028
  • [7] An improved training method for feed-forward neural networks
    Lendl, M
    Unbehauen, R
    CLASSIFICATION IN THE INFORMATION AGE, 1999, : 320 - 327
  • [8] A new scheme for training feed-forward neural networks
    AbdelWahhab, O
    SidAhmed, MA
    PATTERN RECOGNITION, 1997, 30 (03) : 519 - 524
  • [9] Training Algorithm with Incomplete Data for Feed-Forward Neural Networks
    Song-Yee Yoon
    Soo-Young Lee
    Neural Processing Letters, 1999, 10 : 171 - 179
  • [10] Differential evolution training algorithm for feed-forward neural networks
    Ilonen, J
    Kamarainen, JK
    Lampinen, J
    NEURAL PROCESSING LETTERS, 2003, 17 (01) : 93 - 105