A Learning Algorithm with a Gradient Normalization and a Learning Rate Adaptation for the Mini-batch Type Learning

被引:0
|
作者
Ito, Daiki [1 ]
Okamoto, Takashi [2 ]
Koakutsu, Seiichi [2 ]
机构
[1] Chiba Univ, Fac Engn, Chiba, Japan
[2] Chiba Univ, Grad Sch Engn, Chiba, Japan
关键词
Neural networks; Convolutional neural networks; Stochastic gradient descent method; Learning algorithm;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of a high-performance optimization algorithm to solve the learning problem of the neural networks is strongly demanded with the advance of the deep learning. The learning algorithms with gradient normalization mechanisms have been investigated, and their effectiveness has been shown. In the learning algorithms, the adaptation of the learning rate is very important issue. The learning algorithms of the neural networks are classified into the batch learning and the mini-batch learning. In the learning with vast training data, the mini-batch type learning is often used due to the limitation of memory size and the computational cost. The mini-batch type learning algorithms with gradient normalization mechanisms have been investigated. However, the adaptation of the learning rate in the mini-batch type learning algorithm with the gradient normalization has not been investigated well. This study proposes to introduce a new learning rate adaptation mechanism based on sign variation of gradient to a mini-batch type learning algorithm with the gradient normalization. The effectiveness of the proposed algorithm is verified through applications to a learning problem of the multi-layered neural networks and a learning problem of the convolutional neural networks.
引用
收藏
页码:811 / 816
页数:6
相关论文
共 50 条
  • [31] Scalable Hardware Accelerator for Mini-Batch Gradient Descent
    Rasoori, Sandeep
    Akella, Venkatesh
    PROCEEDINGS OF THE 2018 GREAT LAKES SYMPOSIUM ON VLSI (GLSVLSI'18), 2018, : 159 - 164
  • [32] A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai–Borwein Stepsize
    Teng-Teng Yu
    Xin-Wei Liu
    Yu-Hong Dai
    Jie Sun
    Journal of the Operations Research Society of China, 2023, 11 : 277 - 307
  • [33] A Modified Stochastic Gradient Descent Optimization Algorithm With Random Learning Rate for Machine Learning and Deep Learning
    Duk-Sun Shim
    Joseph Shim
    International Journal of Control, Automation and Systems, 2023, 21 : 3825 - 3831
  • [34] A Modified Stochastic Gradient Descent Optimization Algorithm With Random Learning Rate for Machine Learning and Deep Learning
    Shim, Duk-Sun
    Shim, Joseph
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2023, 21 (11) : 3825 - 3831
  • [35] DYNAMITE: Dynamic Interplay of Mini-Batch Size and Aggregation Frequency for Federated Learning With Static and Streaming Datasets
    Liu, Weijie
    Zhang, Xiaoxi
    Duan, Jingpu
    Joe-Wong, Carlee
    Zhou, Zhi
    Chen, Xu
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (07) : 7664 - 7679
  • [36] Research on Mini-Batch Affinity Propagation Clustering Algorithm
    Xu, Ziqi
    Lu, Yahui
    Jiang, Yu
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 86 - 95
  • [37] An Asynchronous Mini-batch Algorithm for Regularized Stochastic Optimization
    Feyzmandavian, Hamid Reza
    Aytekin, Arda
    Johansson, Mikael
    2015 54TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2015, : 1384 - 1389
  • [38] Statistical Analysis of Fixed Mini-Batch Gradient Descent Estimator
    Qi, Haobo
    Wang, Feifei
    Wang, Hansheng
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2023, 32 (04) : 1348 - 1360
  • [39] Learning Rate Adaptation for Differentially Private Learning
    Koskela, Antti
    Honkela, Antti
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2465 - 2474
  • [40] An Asynchronous Mini-Batch Algorithm for Regularized Stochastic Optimization
    Feyzmahdavian, Hamid Reza
    Aytekin, Arda
    Johansson, Mikael
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (12) : 3740 - 3754