Memristive KDG-BNN: Memristive binary neural networks trained via knowledge distillation and generative adversarial networks

被引:5
|
作者
Gao, Tongtong [1 ]
Zhou, Yue [1 ]
Duan, Shukai [1 ,2 ,3 ]
Hu, Xiaofang [1 ,2 ,3 ]
机构
[1] Southwest Univ, Coll Artificial Intelligence, Chongqing 400715, Peoples R China
[2] Brain inspired Comp & Intelligent Control Chongqin, Chongqing 400715, Peoples R China
[3] Southwest Univ, Key Lab Luminescence Anal & Mol Sensing, Minist Educ, Chongqing, Peoples R China
基金
国家重点研发计划;
关键词
Binary neural networks; Knowledge distillation; Generative adversarial networks; Wasserstein generative adversarial networks; Memristive circuit;
D O I
10.1016/j.knosys.2022.108962
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the increasing requirements for the combination of software and hardware, network compression and hardware deployment have become hot research topics. In network compression, binary neural networks (BNNs) are widely applied in artificial intelligence chips because of memory saving, high computational efficiency, and hardware friendliness. However, there is a performance gap between BNNs and full-precision neural networks (FNNs). This paper proposes a BNN training framework called KDG-BNN, consisting of three modules: a full-precision network, a 1-bit binary network, and a discriminator. The full-precision network guides the 1-bit binary network to train through distillation loss in this framework. Meanwhile, the 1-bit binary network acts as a generator and conducts adversarial training with the discriminator. By simultaneously optimizing the adversarial loss and distillation loss, the 1-bit binary network can learn the feature distribution of the full-precision network more accurately. Then, the generative adversarial network (GAN) is replaced by Wasserstein GAN with gradient penalty (WGAN-GP) to deal with gradient disappearance, and KDG-BNN is developed into KDWG-BNN. Experiments show that AdamBNN trained with KDWG-BNN can achieve 85.89% and 70.7% accuracy on CIFAR-10 and ImageNet, respectively, exceeding 0.76% on CIFAR-10 and 0.2% on ImageNet. The memristor has many features for hardware deployment, such as memory functions, continuous input and output, nanoscale size, etc., making it an ideal device for deploying neural networks. Therefore, this paper further proposes a memristor-based KDG-BNN implementation scheme by levering the merits of memristors and the lightweight BNNs in the hope of realizing and promoting end-side intelligent applications. (C) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Lag synchronization for fractional-order memristive neural networks via period intermittent control
    Zhang, Lingzhong
    Yang, Yongqing
    Wang, Fei
    [J]. NONLINEAR DYNAMICS, 2017, 89 (01) : 367 - 381
  • [42] Quasisynchronization of Memristive Neural Networks With Communication Delays via Event-Triggered Impulsive Control
    Zhou, Yufeng
    Zhang, Hao
    Zeng, Zhigang
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (08) : 7682 - 7693
  • [43] Exponential Stabilization of Delayed Inertial Memristive Neural Networks via Aperiodically Intermittent Control Strategy
    Liu, Dan
    Ye, Dan
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (01): : 448 - 458
  • [44] Global synchronization of memristive neural networks subject to random disturbances via distributed pinning control
    Guo, Zhenyuan
    Yang, Shaofu
    Wang, Jun
    [J]. NEURAL NETWORKS, 2016, 84 : 67 - 79
  • [45] Finite-Time Stabilization of Inertial Memristive Neural Networks via Nonreduced Order Method
    Zhang, Jun
    Zhu, Song
    Liu, Xiaoyang
    Wen, Shiping
    Mu, Chaoxu
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [46] Lag synchronization for fractional-order memristive neural networks via period intermittent control
    Lingzhong Zhang
    Yongqing Yang
    Fei wang
    [J]. Nonlinear Dynamics, 2017, 89 : 367 - 381
  • [47] Global Exponential Synchronization of Multiple Memristive Neural Networks With Time Delay via Nonlinear Coupling
    Guo, Zhenyuan
    Yang, Shaofu
    Wang, Jun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (06) : 1300 - 1311
  • [48] Finite-time stabilization of memristive neural networks via two-phase method
    Yu, Tianhu
    Wang, Huamin
    Cao, Jinde
    Xue, Changfeng
    [J]. NEUROCOMPUTING, 2022, 491 : 24 - 33
  • [49] Synchronization of memristive neural networks with unknown parameters via event-triggered adaptive control
    Zhou, Yufeng
    Zhang, Hao
    Zeng, Zhigang
    [J]. NEURAL NETWORKS, 2021, 139 : 255 - 264
  • [50] Synchronization of a class of memristive neural networks with time delays via sampled-data control
    Huaiqin Wu
    Ruoxia Li
    Hongzhi Wei
    Xiaowei Zhang
    Rong Yao
    [J]. International Journal of Machine Learning and Cybernetics, 2015, 6 : 365 - 373