Forward layer-wise learning of convolutional neural networks through separation index maximizing

被引:0
|
作者
Karimi, Ali [1 ]
Kalhor, Ahmad [1 ]
Tabrizi, Melika Sadeghi [1 ]
机构
[1] Univ Tehran, Coll Engn, Sch Elect & Comp Engn, Tehran, Iran
关键词
D O I
10.1038/s41598-024-59176-3
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
This paper proposes a forward layer-wise learning algorithm for CNNs in classification problems. The algorithm utilizes the Separation Index (SI) as a supervised complexity measure to evaluate and train each layer in a forward manner. The proposed method explains that gradually increasing the SI through layers reduces the input data's uncertainties and disturbances, achieving a better feature space representation. Hence, by approximating the SI with a variant of local triplet loss at each layer, a gradient-based learning algorithm is suggested to maximize it. Inspired by the NGRAD (Neural Gradient Representation by Activity Differences) hypothesis, the proposed algorithm operates in a forward manner without explicit error information from the last layer. The algorithm's performance is evaluated on image classification tasks using VGG16, VGG19, AlexNet, and LeNet architectures with CIFAR-10, CIFAR-100, Raabin-WBC, and Fashion-MNIST datasets. Additionally, the experiments are applied to text classification tasks using the DBPedia and AG's News datasets. The results demonstrate that the proposed layer-wise learning algorithm outperforms state-of-the-art methods in accuracy and time complexity.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] A Layer-Wise Theoretical Framework for Deep Learning of Convolutional Neural Networks
    Huu-Thiet Nguyen
    Li, Sitan
    Cheah, Chien Chern
    [J]. IEEE ACCESS, 2022, 10 : 14270 - 14287
  • [2] Layer-Wise Compressive Training for Convolutional Neural Networks
    Grimaldi, Matteo
    Tenace, Valerio
    Calimera, Andrea
    [J]. FUTURE INTERNET, 2019, 11 (01)
  • [3] Layer-Wise Training to Create Efficient Convolutional Neural Networks
    Zeng, Linghua
    Tian, Xinmei
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 631 - 641
  • [4] Interpreting Convolutional Neural Networks via Layer-Wise Relevance Propagation
    Jia, Wohuan
    Zhang, Shaoshuai
    Jiang, Yue
    Xu, Li
    [J]. ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I, 2022, 13338 : 457 - 467
  • [5] Deep Convolutional Neural Networks with Layer-wise Context Expansion and Attention
    Yu, Dong
    Xiong, Wayne
    Droppo, Jasha
    Stolcke, Andreas
    Ye, Guoli
    Li, Jinyu
    Zweig, Geoffrey
    [J]. 17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 17 - 21
  • [6] Collaborative Layer-Wise Discriminative Learning in Deep Neural Networks
    Jin, Xiaojie
    Chen, Yunpeng
    Dong, Jian
    Feng, Jiashi
    Yan, Shuicheng
    [J]. COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 : 733 - 749
  • [7] Implementation of Lightweight Convolutional Neural Networks via Layer-Wise Differentiable Compression
    Diao, Huabin
    Hao, Yuexing
    Xu, Shaoyun
    Li, Gongyan
    [J]. SENSORS, 2021, 21 (10)
  • [8] Activation Distribution-based Layer-wise Quantization for Convolutional Neural Networks
    Ki, Subin
    Kim, Hyun
    [J]. 2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,
  • [9] Optimization Based Layer-Wise Pruning Threshold Method for Accelerating Convolutional Neural Networks
    Ding, Yunlong
    Chen, Di-Rong
    [J]. MATHEMATICS, 2023, 11 (15)
  • [10] Explaining Convolutional Neural Networks using Softmax Gradient Layer-wise Relevance Propagation
    Iwana, Brian Kenji
    Kuroki, Ryohei
    Uchida, Seiichi
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 4176 - 4185