Batch Gradient Learning Algorithm with Smoothing L1 Regularization for Feedforward Neural Networks

被引:0
|
作者
Mohamed, Khidir Shaib [1 ,2 ]
机构
[1] Qassim Univ, Coll Sci & Arts Uglat Asugour, Dept Math, Buraydah 51452, Saudi Arabia
[2] Dalanj Univ, Coll Sci, Dept Math & Comp, POB 14, Dilling, Sudan
关键词
convergence; batch gradient learning algorithm; feedforward neural networks; smoothing L-1 regularization; L-1/2; REGULARIZATION; PENALTY; CONVERGENCE; BOUNDEDNESS;
D O I
10.3390/computers12010004
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Regularization techniques are critical in the development of machine learning models. Complex models, such as neural networks, are particularly prone to overfitting and to performing poorly on the training data. L-1 regularization is the most extreme way to enforce sparsity, but, regrettably, it does not result in an NP-hard problem due to the non-differentiability of the 1-norm. However, the L-1 regularization term achieved convergence speed and efficiency optimization solution through a proximal method. In this paper, we propose a batch gradient learning algorithm with smoothing L-1 regularization (BGSL(1)) for learning and pruning a feedforward neural network with hidden nodes. To achieve our study purpose, we propose a smoothing (differentiable) function in order to address the non-differentiability of L-1 regularization at the origin, make the convergence speed faster, improve the network structure ability, and build stronger mapping. Under this condition, the strong and weak convergence theorems are provided. We used N-dimensional parity problems and function approximation problems in our experiments. Preliminary findings indicate that the BGSL(1) has convergence faster and good generalization abilities when compared with BGL(1/2), BGL(1), BGL(2), and BGSL(1/2). As a result, we demonstrate that the error function decreases monotonically and that the norm of the gradient of the error function approaches zero, thereby validating the theoretical finding and the supremacy of the suggested technique.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Learning Optimized Structure of Neural Networks by Hidden Node Pruning With L1 Regularization
    Xie, Xuetao
    Zhang, Huaqing
    Wang, Junze
    Chang, Qin
    Wang, Jian
    Pal, Nikhil R.
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (03) : 1333 - 1346
  • [22] Convergence analysis of the batch gradient-based neuro-fuzzy learning algorithm with smoothing L1/2 regularization for the first-order Takagi-Sugeno system
    Liu, Yan
    Yang, Dakun
    [J]. FUZZY SETS AND SYSTEMS, 2017, 319 : 28 - 49
  • [23] l1 Regularization in Two-Layer Neural Networks
    Li, Gen
    Gu, Yuantao
    Ding, Jie
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 135 - 139
  • [24] Structure Optimization of Neural Networks with L1 Regularization on Gates
    Chang, Qin
    Wang, Junze
    Zhang, Huaqing
    Shi, Lina
    Wang, Jian
    Pal, Nikhil R.
    [J]. 2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 196 - 203
  • [25] Compact Deep Neural Networks with l1,1 and l1,2 Regularization
    Ma, Rongrong
    Niu, Lingfeng
    [J]. 2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 1248 - 1254
  • [26] Boundedness of a batch gradient method with penalty for feedforward neural networks
    Zhang, Huisheng
    Wu, Wei
    Yao, Mingchen
    [J]. APPLIED MATHEMATICS FOR SCIENCE AND ENGINEERING, 2007, : 175 - +
  • [27] A pruning algorithm with relaxed conditions for high-order neural networks based on smoothing group L1/2 regularization and adaptive momentum
    Kang, Qian
    Fan, Qinwei
    Zurada, Jacek M.
    Huang, Tingwen
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 257
  • [28] Structured Pruning of Convolutional Neural Networks via L1 Regularization
    Yang, Chen
    Yang, Zhenghong
    Khattak, Abdul Mateen
    Yang, Liu
    Zhang, Wenxin
    Gao, Wanlin
    Wang, Minjuan
    [J]. IEEE ACCESS, 2019, 7 : 106385 - 106394
  • [29] A parallel algorithm for gradient training of feedforward neural networks
    Hanzalek, Z
    [J]. PARALLEL COMPUTING, 1998, 24 (5-6) : 823 - 839
  • [30] A pruning feedforward small-world neural network by dynamic sparse regularization with smoothing l1/2 norm for nonlinear system modeling
    Li, Wenjing
    Chu, Minghui
    [J]. APPLIED SOFT COMPUTING, 2023, 136