A Layer-Wise Theoretical Framework for Deep Learning of Convolutional Neural Networks

被引:11
|
作者
Huu-Thiet Nguyen [1 ]
Li, Sitan [1 ]
Cheah, Chien Chern [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
关键词
Convergence; Convolution; Deep learning; Convolutional neural networks; 1; f noise; Mathematical models; Analytical models; CNNs; layer-wise learning; explainable AI; trust in AI;
D O I
10.1109/ACCESS.2022.3147869
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As research attention in deep learning has been focusing on pushing empirical results to a higher peak, remarkable progress has been made in the performance race of machine learning applications in the past years. Yet deep learning based on artificial neural networks still remains difficult to understand as it is considered as a black-box approach. A lack of understanding of deep learning networks from the theoretical perspective would not only hinder the employment of them in applications where high-stakes decisions need to be made, but also limit their future development where artificial intelligence is expected to be robust, predictable and trustable. This paper aims to provide a theoretical methodology to investigate and train deep convolutional neural networks so as to ensure convergence. A mathematical model based on matrix representations for convolutional neural networks is first formulated and an analytic layer-wise learning framework for convolutional neural networks is then proposed and tested on several common benchmarking image datasets. The case studies show a reasonable trade-off between accuracy and analytic learning, and also highlight the potential of employing the proposed layer-wise learning method in finding the appropriate number of layers in actual implementations.
引用
收藏
页码:14270 / 14287
页数:18
相关论文
共 50 条
  • [1] Deep Convolutional Neural Networks with Layer-wise Context Expansion and Attention
    Yu, Dong
    Xiong, Wayne
    Droppo, Jasha
    Stolcke, Andreas
    Ye, Guoli
    Li, Jinyu
    Zweig, Geoffrey
    [J]. 17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 17 - 21
  • [2] Collaborative Layer-Wise Discriminative Learning in Deep Neural Networks
    Jin, Xiaojie
    Chen, Yunpeng
    Dong, Jian
    Feng, Jiashi
    Yan, Shuicheng
    [J]. COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 : 733 - 749
  • [3] Layer-Wise Compressive Training for Convolutional Neural Networks
    Grimaldi, Matteo
    Tenace, Valerio
    Calimera, Andrea
    [J]. FUTURE INTERNET, 2019, 11 (01)
  • [4] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    [J]. IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [5] Stochastic Layer-Wise Precision in Deep Neural Networks
    Lacey, Griffin
    Taylor, Graham W.
    Areibi, Shawki
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2018, : 663 - 672
  • [6] Layer-Wise Training to Create Efficient Convolutional Neural Networks
    Zeng, Linghua
    Tian, Xinmei
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 631 - 641
  • [7] Investigating Learning in Deep Neural Networks Using Layer-Wise Weight Change
    Agrawal, Ayush Manish
    Tendle, Atharva
    Sikka, Harshvardhan
    Singh, Sahib
    Kayid, Amr
    [J]. INTELLIGENT COMPUTING, VOL 2, 2021, 284 : 678 - 693
  • [8] Forward layer-wise learning of convolutional neural networks through separation index maximizing
    Karimi, Ali
    Kalhor, Ahmad
    Tabrizi, Melika Sadeghi
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [9] Unsupervised Layer-Wise Model Selection in Deep Neural Networks
    Ludovic, Arnold
    Helene, Paugam-Moisy
    Michele, Sebag
    [J]. ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 : 915 - 920
  • [10] Interpreting Convolutional Neural Networks via Layer-Wise Relevance Propagation
    Jia, Wohuan
    Zhang, Shaoshuai
    Jiang, Yue
    Xu, Li
    [J]. ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I, 2022, 13338 : 457 - 467