A Learning Algorithm with a Gradient Normalization and a Learning Rate Adaptation for the Mini-batch Type Learning

被引:0
|
作者
Ito, Daiki [1 ]
Okamoto, Takashi [2 ]
Koakutsu, Seiichi [2 ]
机构
[1] Chiba Univ, Fac Engn, Chiba, Japan
[2] Chiba Univ, Grad Sch Engn, Chiba, Japan
关键词
Neural networks; Convolutional neural networks; Stochastic gradient descent method; Learning algorithm;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of a high-performance optimization algorithm to solve the learning problem of the neural networks is strongly demanded with the advance of the deep learning. The learning algorithms with gradient normalization mechanisms have been investigated, and their effectiveness has been shown. In the learning algorithms, the adaptation of the learning rate is very important issue. The learning algorithms of the neural networks are classified into the batch learning and the mini-batch learning. In the learning with vast training data, the mini-batch type learning is often used due to the limitation of memory size and the computational cost. The mini-batch type learning algorithms with gradient normalization mechanisms have been investigated. However, the adaptation of the learning rate in the mini-batch type learning algorithm with the gradient normalization has not been investigated well. This study proposes to introduce a new learning rate adaptation mechanism based on sign variation of gradient to a mini-batch type learning algorithm with the gradient normalization. The effectiveness of the proposed algorithm is verified through applications to a learning problem of the multi-layered neural networks and a learning problem of the convolutional neural networks.
引用
收藏
页码:811 / 816
页数:6
相关论文
共 50 条
  • [1] A MINI-BATCH STOCHASTIC GRADIENT METHOD FOR SPARSE LEARNING TO RANK
    Cheng, Fan
    Wang, Dongliang
    Zhang, Lei
    Su, Yansen
    Qiu, Jianfeng
    Suo, Yi
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2018, 14 (04): : 1207 - 1221
  • [2] Automatic Setting of Learning Rate and Mini-batch Size in Momentum and AdaM Stochastic Gradient Methods
    Franchini, Giorgia
    Porta, Federica
    INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS 2022, ICNAAM-2022, 2024, 3094
  • [3] Deterministic convergence of complex mini-batch gradient learning algorithm for fully complex-valued neural networks
    Zhang, Huisheng
    Zhang, Ying
    Zhu, Shuai
    Xu, Dongpo
    NEUROCOMPUTING, 2020, 407 : 185 - 193
  • [4] The stability and statistic of domain decomposition algorithm with mini-batch learning for optimal transport
    Lin, Judy Yangjun
    Liu, Huoxia
    NEUROCOMPUTING, 2024, 587
  • [5] Adaptive Natural Gradient Method for Learning of Stochastic Neural Networks in Mini-Batch Mode
    Park, Hyeyoung
    Lee, Kwanyong
    APPLIED SCIENCES-BASEL, 2019, 9 (21):
  • [6] A mini-batch algorithm for large-scale learning problems with adaptive step size
    He, Chongyang
    Zhang, Yiting
    Zhu, Dingyu
    Cao, Mingyuan
    Yang, Yueting
    DIGITAL SIGNAL PROCESSING, 2023, 143
  • [7] Mini-batch learning of exponential family finite mixture models
    Nguyen, Hien D.
    Forbes, Florence
    McLachlan, Geoffrey J.
    STATISTICS AND COMPUTING, 2020, 30 (04) : 731 - 748
  • [8] Mini-batch learning of exponential family finite mixture models
    Hien D. Nguyen
    Florence Forbes
    Geoffrey J. McLachlan
    Statistics and Computing, 2020, 30 : 731 - 748
  • [9] Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD)
    Qi Qian
    Rong Jin
    Jinfeng Yi
    Lijun Zhang
    Shenghuo Zhu
    Machine Learning, 2015, 99 : 353 - 372
  • [10] Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD)
    Qian, Qi
    Jin, Rong
    Yi, Jinfeng
    Zhang, Lijun
    Zhu, Shenghuo
    MACHINE LEARNING, 2015, 99 (03) : 353 - 372