In situ training of feed-forward and recurrent convolutional memristor networks

被引:0
|
作者
Zhongrui Wang
Can Li
Peng Lin
Mingyi Rao
Yongyang Nie
Wenhao Song
Qinru Qiu
Yunning Li
Peng Yan
John Paul Strachan
Ning Ge
Nathan McDonald
Qing Wu
Miao Hu
Huaqiang Wu
R. Stanley Williams
Qiangfei Xia
J. Joshua Yang
机构
[1] University of Massachusetts,Department of Electrical and Computer Engineering
[2] Hewlett Packard Labs,Department of Electrical Engineering and Computer Science
[3] Hewlett Packard Enterprise,Information Directorate
[4] Syracuse University,Department of Electrical and Computer Engineering
[5] Air Force Research Laboratory,Institute of Microelectronics
[6] Binghamton University,Department of Electrical and Computer Engineering
[7] Tsinghua University,undefined
[8] Texas A&M University,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
The explosive growth of machine learning is largely due to the recent advancements in hardware and architecture. The engineering of network structures, taking advantage of the spatial or temporal translational isometry of patterns, naturally leads to bio-inspired, shared-weight structures such as convolutional neural networks, which have markedly reduced the number of free parameters. State-of-the-art microarchitectures commonly rely on weight-sharing techniques, but still suffer from the von Neumann bottleneck of transistor-based platforms. Here, we experimentally demonstrate the in situ training of a five-level convolutional neural network that self-adapts to non-idealities of the one-transistor one-memristor array to classify the MNIST dataset, achieving similar accuracy to the memristor-based multilayer perceptron with a reduction in trainable parameters of ~75% owing to the shared weights. In addition, the memristors encoded both spatial and temporal translational invariance simultaneously in a convolutional long short-term memory network—a memristor-based neural network with intrinsic 3D input processing—which was trained in situ to classify a synthetic MNIST sequence dataset using just 850 weights. These proof-of-principle demonstrations combine the architectural advantages of weight sharing and the area/energy efficiency boost of the memristors, paving the way to future edge artificial intelligence.
引用
收藏
页码:434 / 442
页数:8
相关论文
共 50 条
  • [21] A Comparison of Feed-forward and Recurrent Neural Networks in Time Series Forecasting
    Brezak, Danko
    Bacek, Tomislav
    Majetic, Dubravko
    Kasac, Josip
    Novakovic, Branko
    2012 IEEE CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR FINANCIAL ENGINEERING & ECONOMICS (CIFER), 2012, : 206 - 211
  • [22] A Comparative Assessment of Feed-Forward and Convolutional Neural Networks for the Classification of Prostate Lesions
    Marnell, Sabrina
    Riley, Patrick
    Olier, Ivan
    Rea, Marc
    Ortega-Martorell, Sandra
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING (IDEAL 2019), PT II, 2019, 11872 : 132 - 138
  • [23] Patterns of synchrony for feed-forward and auto-regulation feed-forward neural networks
    Aguiar, Manuela A. D.
    Dias, Ana Paula S.
    Ferreira, Flora
    CHAOS, 2017, 27 (01)
  • [24] An ensemble of differential evolution and Adam for training feed-forward neural networks
    Xue, Yu
    Tong, Yiling
    Neri, Ferrante
    INFORMATION SCIENCES, 2022, 608 : 453 - 471
  • [25] Unsupervised, smooth training of feed-forward neural networks for mismatch compensation
    Surendran, AC
    Lee, CH
    Rahim, M
    1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, PROCEEDINGS, 1997, : 482 - 489
  • [26] Salp Swarm Algorithm (SSA) for Training Feed-Forward Neural Networks
    Bairathi, Divya
    Gopalani, Dinesh
    SOFT COMPUTING FOR PROBLEM SOLVING, SOCPROS 2017, VOL 1, 2019, 816 : 521 - 534
  • [27] Hybrid learning schemes for fast training of feed-forward neural networks
    Karayiannis, NB
    MATHEMATICS AND COMPUTERS IN SIMULATION, 1996, 41 (1-2) : 13 - 28
  • [28] Hybrid training of feed-forward neural networks with particle swarm optimization
    Carvalho, M.
    Ludermir, T. B.
    NEURAL INFORMATION PROCESSING, PT 2, PROCEEDINGS, 2006, 4233 : 1061 - 1070
  • [30] A training-time analysis of robustness in feed-forward neural networks
    Alippi, C
    Sana, D
    Scotti, F
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 2853 - 2858