Batch Gradient Learning Algorithm with Smoothing L1 Regularization for Feedforward Neural Networks

被引:0
|
作者
Mohamed, Khidir Shaib [1 ,2 ]
机构
[1] Qassim Univ, Coll Sci & Arts Uglat Asugour, Dept Math, Buraydah 51452, Saudi Arabia
[2] Dalanj Univ, Coll Sci, Dept Math & Comp, POB 14, Dilling, Sudan
关键词
convergence; batch gradient learning algorithm; feedforward neural networks; smoothing L-1 regularization; L-1/2; REGULARIZATION; PENALTY; CONVERGENCE; BOUNDEDNESS;
D O I
10.3390/computers12010004
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Regularization techniques are critical in the development of machine learning models. Complex models, such as neural networks, are particularly prone to overfitting and to performing poorly on the training data. L-1 regularization is the most extreme way to enforce sparsity, but, regrettably, it does not result in an NP-hard problem due to the non-differentiability of the 1-norm. However, the L-1 regularization term achieved convergence speed and efficiency optimization solution through a proximal method. In this paper, we propose a batch gradient learning algorithm with smoothing L-1 regularization (BGSL(1)) for learning and pruning a feedforward neural network with hidden nodes. To achieve our study purpose, we propose a smoothing (differentiable) function in order to address the non-differentiability of L-1 regularization at the origin, make the convergence speed faster, improve the network structure ability, and build stronger mapping. Under this condition, the strong and weak convergence theorems are provided. We used N-dimensional parity problems and function approximation problems in our experiments. Preliminary findings indicate that the BGSL(1) has convergence faster and good generalization abilities when compared with BGL(1/2), BGL(1), BGL(2), and BGSL(1/2). As a result, we demonstrate that the error function decreases monotonically and that the norm of the gradient of the error function approaches zero, thereby validating the theoretical finding and the supremacy of the suggested technique.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Batch gradient method with smoothing L1/2 regularization for training of feedforward neural networks
    Wu, Wei
    Fan, Qinwei
    Zurada, Jacek M.
    Wang, Jian
    Yang, Dakun
    Liu, Yan
    [J]. NEURAL NETWORKS, 2014, 50 : 72 - 78
  • [2] Convergence of batch gradient algorithm with smoothing composition of group l0 and l1/2 regularization for feedforward neural networks
    Ramchoun, Hassan
    Ettaouil, Mohamed
    [J]. PROGRESS IN ARTIFICIAL INTELLIGENCE, 2022, 11 (03) : 269 - 278
  • [3] Convergence of batch gradient learning algorithm with smoothing L1/2 regularization for Sigma-Pi-Sigma neural networks
    Liu, Yan
    Li, Zhengxue
    Yang, Dakun
    Mohamed, Kh. Sh.
    Wang, Jing
    Wu, Wei
    [J]. NEUROCOMPUTING, 2015, 151 : 333 - 341
  • [4] Convergence of online gradient method for feedforward neural networks with smoothing L1/2 regularization penalty
    Fan, Qinwei
    Zurada, Jacek M.
    Wu, Wei
    [J]. NEUROCOMPUTING, 2014, 131 : 208 - 216
  • [5] Batch gradient training method with smoothing regularization for l0 feedforward neural networks
    Zhang, Huisheng
    Tang, Yanli
    Liu, Xiaodong
    [J]. NEURAL COMPUTING & APPLICATIONS, 2015, 26 (02): : 383 - 390
  • [6] Convergence of batch gradient learning with smoothing regularization and adaptive momentum for neural networks
    Fan, Qinwei
    Wu, Wei
    Zurada, Jacek M.
    [J]. SPRINGERPLUS, 2016, 5
  • [7] Online gradient method with smoothing l0 regularization for feedforward neural networks
    Zhang, Huisheng
    Tang, Yanli
    [J]. NEUROCOMPUTING, 2017, 224 : 1 - 8
  • [8] L1/2 regularization learning for smoothing interval neural networks: Algorithms and convergence analysis
    Yang, Dakun
    Liu, Yan
    [J]. NEUROCOMPUTING, 2018, 272 : 122 - 129
  • [9] A modified gradient learning algorithm with smoothing L1/2 regularization for Takagi-Sugeno fuzzy models
    Liu, Yan
    Wu, Wei
    Fan, Qinwei
    Yang, Dakun
    Wang, Jian
    [J]. NEUROCOMPUTING, 2014, 138 : 229 - 237
  • [10] Smoothing L1 regularization for stochastic configuration networks
    基于光滑化L1 正则项的随机配置网络
    [J]. 1600, Northeast University (39):