Efficient training algorithms for a class of shunting inhibitory convolutional neural networks

被引:50
|
作者
Tivive, FHC [1 ]
Bouzerdoum, A [1 ]
机构
[1] Univ Wollongong, Sch Elect Comp & Telecommun Engn, Wollongong, NSW 2522, Australia
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2005年 / 16卷 / 03期
关键词
convolutional neural network (CoNN); first- and second-order training methods; shunting inhibitory neuron;
D O I
10.1109/TNN.2005.845144
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article presents some efficient training algorithms, based on first-order, second-order, and conjugate gradient optimization methods, for a class of convolutional neural networks (CoNNs), known as shunting inhibitory convolution neural networks. Furthermore, a new hybrid method is proposed, which is derived from the principles of Quickprop, Rprop, SuperSAB, and least squares (LS). Experimental results show that the new hybrid method can perform as well as the Levenberg-Marquardt (LM) algorithm, but at a much lower computational cost and less memory storage. For comparison sake, the visual pattern recognition task of face/nonface discrimination is chosen as a classification problem to evaluate the performance of the training algorithms. Sixteen training algorithms are implemented for the three different variants of the proposed CoNN architecture: binary-, Toeplitz- and fully connected architectures. All implemented algorithms can train the three network architectures successfully, but their convergence speed vary markedly. In particular, the combination of LS with the new hybrid method and LS with the LM method achieve the best convergence rates in terms of number of training epochs. In addition, the classification accuracies of all three architectures are assessed using ten-fold cross validation. The results show that the binary- and Toeplitz-connected architectures outperform slightly the fully connected architecture: the lowest error rates across all training algorithms are 1.95% for Toeplitz-connected, 2.10% for the binary-connected, and 2.20% for the fully connected network. In general, the modified Broyden-Fletcher-Goldfarb-Shanno (BFGS) methods, the three variants of LM algorithm, and the new hybrid/LS method perform consistently well, achieving error rates of less than 3% averaged across all three architectures.
引用
收藏
页码:541 / 556
页数:16
相关论文
共 50 条
  • [1] A gender recognition system using shunting inhibitory convolutional neural networks
    Tivive, Fok Hing Chi
    Bouzerdoum, Abdesselam
    [J]. 2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 5336 - 5341
  • [2] A face detection system using shunting inhibitory convolutional neural networks
    Tivive, FHC
    Bouzerdoum, A
    [J]. 2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 2571 - 2575
  • [3] ON A CLASS OF EFFICIENT LEARNING ALGORITHMS FOR NEURAL NETWORKS
    BARMANN, F
    BIEGLERKONIG, F
    [J]. NEURAL NETWORKS, 1992, 5 (01) : 139 - 144
  • [4] A compact CMOS face detection architecture based on shunting inhibitory convolutional neural networks
    Zhang, Xiaoxiao
    Bermak, Amine
    Boussaid, Farid
    Bouzerdoum, A.
    [J]. DELTA 2008: FOURTH IEEE INTERNATIONAL SYMPOSIUM ON ELECTRONIC DESIGN, TEST AND APPLICATIONS, PROCEEDINGS, 2008, : 374 - +
  • [5] A shunting inhibitory convolutional neural network for gender classification
    Tivive, Fok Hing Chi
    Bouzerdoum, Abdesselam
    [J]. 18TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 4, PROCEEDINGS, 2006, : 421 - +
  • [6] Efficient Incremental Training for Deep Convolutional Neural Networks
    Tao, Yudong
    Tu, Yuexuan
    Shyu, Mei-Ling
    [J]. 2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 286 - 291
  • [7] A Note on Shunting Inhibitory Neural Networks with Impulses
    Yuan, Doming
    Gui, Zhanji
    Zhang, Jie
    [J]. PROCEEDINGS OF THE 6TH CONFERENCE OF BIOMATHEMATICS, VOLS I AND II: ADVANCES ON BIOMATHEMATICS, 2008, : 88 - 92
  • [8] Depth Dropout: Efficient Training of Residual Convolutional Neural Networks
    Guo, Jian
    Gould, Stephen
    [J]. 2016 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2016, : 343 - 349
  • [9] TNPU: An Efficient Accelerator Architecture for Training Convolutional Neural Networks
    Li, Jiajun
    Yan, Guihai
    Lu, Wenyan
    Jiang, Shuhao
    Gong, Shijun
    Wu, Jingya
    Yan, Junchao
    Li, Xiaowei
    [J]. 24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 450 - 455
  • [10] Enabling Efficient Training of Convolutional Neural Networks for Histopathology Images
    Alali, Mohammed H.
    Roohi, Arman
    Deogun, Jitender S.
    [J]. IMAGE ANALYSIS AND PROCESSING, ICIAP 2022 WORKSHOPS, PT I, 2022, 13373 : 533 - 544