A Learning Algorithm with a Gradient Normalization and a Learning Rate Adaptation for the Mini-batch Type Learning

被引:0
|
作者
Ito, Daiki [1 ]
Okamoto, Takashi [2 ]
Koakutsu, Seiichi [2 ]
机构
[1] Chiba Univ, Fac Engn, Chiba, Japan
[2] Chiba Univ, Grad Sch Engn, Chiba, Japan
关键词
Neural networks; Convolutional neural networks; Stochastic gradient descent method; Learning algorithm;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of a high-performance optimization algorithm to solve the learning problem of the neural networks is strongly demanded with the advance of the deep learning. The learning algorithms with gradient normalization mechanisms have been investigated, and their effectiveness has been shown. In the learning algorithms, the adaptation of the learning rate is very important issue. The learning algorithms of the neural networks are classified into the batch learning and the mini-batch learning. In the learning with vast training data, the mini-batch type learning is often used due to the limitation of memory size and the computational cost. The mini-batch type learning algorithms with gradient normalization mechanisms have been investigated. However, the adaptation of the learning rate in the mini-batch type learning algorithm with the gradient normalization has not been investigated well. This study proposes to introduce a new learning rate adaptation mechanism based on sign variation of gradient to a mini-batch type learning algorithm with the gradient normalization. The effectiveness of the proposed algorithm is verified through applications to a learning problem of the multi-layered neural networks and a learning problem of the convolutional neural networks.
引用
收藏
页码:811 / 816
页数:6
相关论文
共 50 条
  • [41] An adaptive mini-batch stochastic gradient method for AUC maximization
    Cheng, Fan
    Zhang, Xia
    Zhang, Chuang
    Qiu, Jianfeng
    Zhang, Lei
    NEUROCOMPUTING, 2018, 318 : 137 - 150
  • [42] HYPERSPECTRAL UNMIXING VIA PROJECTED MINI-BATCH GRADIENT DESCENT
    Li, Jing
    Li, Xiaorun
    Zhao, Liaoying
    2017 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2017, : 1133 - 1136
  • [43] Mini-batch Dynamic Geometric Embedding for Unsupervised Domain Adaptation
    Khan, Siraj
    Guo, Yuxin
    Ye, Yuzhong
    Li, Chunshan
    Wu, Qingyao
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2063 - 2080
  • [44] Mini-batch Dynamic Geometric Embedding for Unsupervised Domain Adaptation
    Siraj Khan
    Yuxin Guo
    Yuzhong Ye
    Chunshan Li
    Qingyao Wu
    Neural Processing Letters, 2023, 55 : 2063 - 2080
  • [45] A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai-Borwein Stepsize
    Yu, Teng-Teng
    Liu, Xin-Wei
    Dai, Yu-Hong
    Sun, Jie
    JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 277 - 307
  • [46] Test-Time Domain Adaptation by Learning Domain-Aware Batch Normalization
    Wu, Yanan
    Chi, Zhixiang
    Wang, Yang
    Plataniotis, Konstantinos N.
    Feng, Songhe
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 15961 - 15969
  • [47] Formal convergence analysis on deterministic l1-regularization based mini-batch learning for RBF networks
    Liu, Zhaofeng
    Leung, Chi-Sing
    So, Hing Cheung
    NEUROCOMPUTING, 2023, 532 : 77 - 93
  • [48] Properties of the stochastic approximation EM algorithm with mini-batch sampling
    Kuhn, Estelle
    Matias, Catherine
    Rebafka, Tabea
    STATISTICS AND COMPUTING, 2020, 30 (06) : 1725 - 1739
  • [49] A Framework of Convergence Analysis of Mini-batch Stochastic Projected Gradient Methods
    Jian Gu
    Xian-Tao Xiao
    Journal of the Operations Research Society of China, 2023, 11 : 347 - 369
  • [50] A Framework of Convergence Analysis of Mini-batch Stochastic Projected Gradient Methods
    Gu, Jian
    Xiao, Xian-Tao
    JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 347 - 369