Sequential Training of Neural Networks With Gradient Boosting

被引:7
|
作者
Emami, Seyedsaman [1 ]
Martinez-Munoz, Gonzalo [1 ]
机构
[1] Univ Autonoma Madrid, Escuela Politecn Super, Madrid 28049, Spain
关键词
Gradient boosting; neural network; CLASSIFIERS;
D O I
10.1109/ACCESS.2023.3271515
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a novel technique based on gradient boosting to train the final layers of a neural network (NN). Gradient boosting is an additive expansion algorithm in which a series of models are trained sequentially to approximate a given function. A neural network can also be seen as an additive expansion where the scalar product of the responses of the last hidden layer and its weights provide the final output of the network. Instead of training the network as a whole, the proposed algorithm trains the network sequentially in T steps. First, the bias term of the network is initialized with a constant approximation that minimizes the average loss of the data. Then, at each step, a portion of the network, composed of J neurons, is trained to approximate the pseudo-residuals on the training data computed from the previous iterations. Finally, the T partial models and bias are integrated as a single NN with T x J neurons in the hidden layer. Extensive experiments in classification and regression tasks, as well as in combination with deep neural networks, are carried out showing a competitive generalization performance with respect to neural networks trained with different standard solvers, such as Adam, L-BFGS, SGD and deep models. Furthermore, we show that the proposed method design permits to switch off a number of hidden units during test (the units that were last trained) without a significant reduction of its generalization ability. This permits the adaptation of the model to different classification speed requirements on the fly.
引用
收藏
页码:42738 / 42750
页数:13
相关论文
共 50 条
  • [1] A Gradient Boosting Approach for Training Convolutional and Deep Neural Networks
    Emami, Seyedsaman
    Martinez-Munoz, Gonzalo
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2023, 4 : 313 - 321
  • [2] Training methods for Adaptive Boosting of neural networks
    Schwenk, H
    Bengio, Y
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 647 - 653
  • [3] Training Neural Networks with Policy Gradient
    Bose, Sourabh
    Huber, Manfred
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 3998 - 4005
  • [4] Boosting the training of neural networks through hybrid metaheuristics
    Mohammed Azmi Al-Betar
    Mohammed A. Awadallah
    Iyad Abu Doush
    Osama Ahmad Alomari
    Ammar Kamal Abasi
    Sharif Naser Makhadmeh
    Zaid Abdi Alkareem Alyasseri
    Cluster Computing, 2023, 26 : 1821 - 1843
  • [5] Boosting the training of neural networks through hybrid metaheuristics
    Al-Betar, Mohammed Azmi
    Awadallah, Mohammed A.
    Abu Doush, Iyad
    Alomari, Osama Ahmad
    Abasi, Ammar Kamal
    Makhadmeh, Sharif Naser
    Alyasseri, Zaid Abdi Alkareem
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2023, 26 (03): : 1821 - 1843
  • [6] Gradient boosting Bayesian neural networks via Langevin MCMC
    Bai, George
    Chandra, Rohitash
    NEUROCOMPUTING, 2023, 558
  • [7] Quantune: Post-training quantization of convolutional neural networks using extreme gradient boosting for fast deployment
    Lee, Jemin
    Yu, Misun
    Kwon, Yongin
    Kim, Taeho
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 132 : 124 - 135
  • [8] SANGRIA: Stacked Autoencoder Neural Networks With Gradient Boosting for Indoor Localization
    Gufran, Danish
    Tiku, Saideep
    Pasricha, Sudeep
    IEEE EMBEDDED SYSTEMS LETTERS, 2024, 16 (02) : 142 - 145
  • [9] A parallel algorithm for gradient training of feedforward neural networks
    Hanzalek, Z
    PARALLEL COMPUTING, 1998, 24 (5-6) : 823 - 839
  • [10] Boosting neural networks
    Schwenk, H
    Bengio, Y
    NEURAL COMPUTATION, 2000, 12 (08) : 1869 - 1887