Integrated Optimization in Training Process for Binary Neural Network

被引:0
|
作者
Quang Hieu Vo [1 ]
Hong, Sang Hoon [2 ]
Kim, Lok-Won [1 ]
Hong, Choong Seon [1 ]
机构
[1] Kyung Hee Univ, Dept Comp Sci & Engn, Yongin 17104, South Korea
[2] Kyung Hee Univ, Dept Elect Engn, Yongin 17104, South Korea
基金
新加坡国家研究基金会;
关键词
Binary Neural Network; Deep Neural Network; Deep Learning; Machine Learning;
D O I
10.1109/ICOIN56518.2023.10048969
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have recently become larger and deeper to keep up with more complex applications, resulting in high power and memory consumption. Due to simplicity in computation and storage, Binary Neural Networks (BNNs) have been one of the potential approaches to overcome these challenges. Previous works proposed many techniques to mitigate the accuracy degradation because of less bit-width representation. However, each technique follows different optimization directions, while the combination can gain better results. In addition, the padding value which is an essential factor directly affecting the accuracy and inference implementation has not been touched on in the state-of-the-art solutions. In this paper, based on the previous works, an integrated approach is applied in the training process for BNNs to improve accuracy and training stability. In particular, to increase the probability of changing weights' sign, the ReCU function proposed in related work is used to transform full-precision weight to binary weight, while to make the gradient mismatch of the sign function closer to the real one, the training-aware approximation function is used to replace the sign function. Besides, to make the BNNs compatible with post-XNOR implementation, the padding value for convolution is proposed to change to minus one from the default zero. The integrated method is implemented on the Cifar-10 dataset with VGG-small model shows that the training process is more stable with higher accuracy, compared to the baseline, while the model architecture and training algorithm are preserved.
引用
收藏
页码:545 / 548
页数:4
相关论文
共 50 条
  • [41] Bridging the Gap: Unifying the Training and Evaluation of Neural Network Binary Classifiers
    Tsoi, Nathan
    Candon, Kate
    Li, Deyuan
    Milkessa, Yofti
    Vazquez, Marynel
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [42] Neural network training using genetic algorithm with a novel binary encoding
    Liang, Yong
    Leung, Kwong-Sak
    Xu, Zong-Ben
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2007, PT 2, PROCEEDINGS, 2007, 4492 : 371 - +
  • [43] WELDING PROCESS OPTIMIZATION WITH ARTIFICIAL NEURAL NETWORK APPLICATIONS
    Aktepe, Adnan
    Ersoz, Suleyman
    Luy, Murat
    [J]. NEURAL NETWORK WORLD, 2014, 24 (06) : 655 - 670
  • [44] Integrated Feature and Parameter Optimization for an Evolving Spiking Neural Network
    Schliebs, Stefan
    Defoin-Platel, Michael
    Kasabov, Nikola
    [J]. ADVANCES IN NEURO-INFORMATION PROCESSING, PT I, 2009, 5506 : 1229 - +
  • [45] Robustness of Adaptive Neural Network Optimization Under Training Noise
    Chaudhury, Subhajit
    Yamasaki, Toshihiko
    [J]. IEEE ACCESS, 2021, 9 : 37039 - 37053
  • [46] Neural network training and simulation using a multidimensional optimization system
    Likas, A
    Karras, DA
    Lagaris, IE
    [J]. INTERNATIONAL JOURNAL OF COMPUTER MATHEMATICS, 1998, 67 (1-2) : 33 - 46
  • [47] Training multilayer neural network by gobal chaos optimization algorithms
    Khoa, T. Q. D.
    Nakagawa, Masahiro
    [J]. 2007 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-6, 2007, : 136 - 141
  • [48] Review of Convolutional Neural Network Optimization and Training in Image Processing
    Ren, Yong
    Cheng, Xuemin
    [J]. TENTH INTERNATIONAL SYMPOSIUM ON PRECISION ENGINEERING MEASUREMENTS AND INSTRUMENTATION, 2019, 11053
  • [49] Survival density particle swarm optimization for neural network training
    Liu, HB
    Li, B
    Wang, XK
    Ji, Y
    Tang, YY
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2004, PT 1, 2004, 3173 : 332 - 337
  • [50] Quantum Neural Network With Parallel Training for Wireless Resource Optimization
    Narottama, Bhaskara
    Jamaluddin, Triwidyastuti
    Shin, Soo Young
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (05) : 5835 - 5847