A Learning Algorithm with a Gradient Normalization and a Learning Rate Adaptation for the Mini-batch Type Learning

被引:0
|
作者
Ito, Daiki [1 ]
Okamoto, Takashi [2 ]
Koakutsu, Seiichi [2 ]
机构
[1] Chiba Univ, Fac Engn, Chiba, Japan
[2] Chiba Univ, Grad Sch Engn, Chiba, Japan
关键词
Neural networks; Convolutional neural networks; Stochastic gradient descent method; Learning algorithm;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of a high-performance optimization algorithm to solve the learning problem of the neural networks is strongly demanded with the advance of the deep learning. The learning algorithms with gradient normalization mechanisms have been investigated, and their effectiveness has been shown. In the learning algorithms, the adaptation of the learning rate is very important issue. The learning algorithms of the neural networks are classified into the batch learning and the mini-batch learning. In the learning with vast training data, the mini-batch type learning is often used due to the limitation of memory size and the computational cost. The mini-batch type learning algorithms with gradient normalization mechanisms have been investigated. However, the adaptation of the learning rate in the mini-batch type learning algorithm with the gradient normalization has not been investigated well. This study proposes to introduce a new learning rate adaptation mechanism based on sign variation of gradient to a mini-batch type learning algorithm with the gradient normalization. The effectiveness of the proposed algorithm is verified through applications to a learning problem of the multi-layered neural networks and a learning problem of the convolutional neural networks.
引用
收藏
页码:811 / 816
页数:6
相关论文
共 50 条
  • [21] TEACHER-ASSISTED MINI-BATCH SAMPLING FOR BLIND DISTILLATION USING METRIC LEARNING
    Inoue, Nakamasa
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4160 - 4164
  • [22] Convergence of the Mini-Batch SIHT Algorithm
    Damadi, Saeed
    Shen, Jinglai
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, INTELLISYS 2023, 2024, 822 : 223 - 233
  • [23] Accelerating Machine Learning I/O by Overlapping Data Staging and Mini-batch Generations
    Serizawa, Kazuhiro
    Tatebe, Osamu
    BDCAT'19: PROCEEDINGS OF THE 6TH IEEE/ACM INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING, APPLICATIONS AND TECHNOLOGIES, 2019, : 31 - 34
  • [24] An Object Detection Algorithm for Deep Learning Based on Batch Normalization
    Zhou, Yan
    Yuan, Changqing
    Zeng, Fanzhi
    Qian, Jiechang
    Wu, Chen
    SMART COMPUTING AND COMMUNICATION, SMARTCOM 2017, 2018, 10699 : 438 - 448
  • [25] Mini-Batch Learning Strategies for modeling long term temporal dependencies: A study in environmental applications
    Xu, Shaoming
    Khandelwal, Ankush
    Li, Xiang
    Jia, Xiaowei
    Liu, Licheng
    Willard, Jared
    Ghosh, Rahul
    Cutler, Kelly
    Steinbach, Michael
    Duffy, Christopher
    Nieber, John
    Kumar, Vipin
    PROCEEDINGS OF THE 2023 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2023, : 649 - 657
  • [26] Extensible Max-Min Collaborative Retention for Online Mini-Batch Learning Hash Retrieval
    Tan, Junpeng
    Yang, Xiaojun
    Yang, Zhijing
    Chen, Ruihan
    Lu, Yongyi
    Lin, Liang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 6743 - 6758
  • [27] Dynamic Energy Cost Conservation for Distributed Edge Clouds Utilizing Online Mini-Batch Learning
    Jing, Zewei
    Wang, Xianbin
    Yang, Qinghai
    Mei, Muyu
    Wu, Yan
    2023 IEEE 34TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS, PIMRC, 2023,
  • [28] IMPROVING MEMORY BANKS FOR UNSUPERVISED LEARNING WITH LARGE MINI-BATCH, CONSISTENCY AND HARD NEGATIVE MINING
    Bulat, Adrian
    Sanchez-Lozano, Enrique
    Tzimiropoulos, Georgios
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 1695 - 1699
  • [29] Stronger Adversarial Attack: Using Mini-batch Gradient
    Yu, Lin
    Deng, Ting
    Zhang, Wenxiang
    Zeng, Zhigang
    2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 364 - 370
  • [30] Gradient preconditioned mini-batch SGD for ridge regression
    Zhang, Zhuan
    Zhou, Shuisheng
    Li, Dong
    Yang, Ting
    NEUROCOMPUTING, 2020, 413 : 284 - 293