Improving generalization performance of natural gradient learning using optimized regularization by NIC

被引:12
|
作者
Park, H [1 ]
Murata, N
Amari, S
机构
[1] RIKEN, Brain Sci Inst, Wako, Saitama, Japan
[2] Waseda Univ, Tokyo, Japan
关键词
D O I
10.1162/089976604322742065
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural gradient learning is known to be efficient in escaping plateau, which is a main cause of the slow learning speed of neural networks. The adaptive natural gradient learning method for practical implementation also has been developed, and its advantage in real-world problems has been confirmed. In this letter, we deal with the generalization performances of the natural gradient method. Since natural gradient learning makes parameters fit to training data quickly, the overfitting phenomenon may easily occur, which results in poor generalization performance. To solve the problem, we introduce the regularization term in natural gradient learning and propose an efficient optimizing method for the scale of regularization by using a generalized Akaike information criterion (network information criterion). We discuss the properties of the optimized regularization strength by NIC through theoretical analysis as well as computer simulations. We confirm the computational efficiency and generalization performance of the proposed method in real-world applications through computational experiments on benchmark problems.
引用
收藏
页码:355 / 382
页数:28
相关论文
共 50 条
  • [1] Improving Generalization of Deepfake Detectors by Imposing Gradient Regularization
    Guan, Weinan
    Wang, Wei
    Dong, Jing
    Peng, Bo
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 5345 - 5356
  • [2] Improving Generalization in Reinforcement Learning with Mixture Regularization
    Wang, Kaixin
    Kang, Bingyi
    Shao, Jie
    Feng, Jiashi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] A new regularization learning method for improving generalization capability of neural network
    Wu, Y
    Zhang, LM
    [J]. PROCEEDINGS OF THE 4TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-4, 2002, : 2011 - 2015
  • [4] Penalizing Gradient Norm for Efficiently Improving Generalization in Deep Learning
    Zhao, Yang
    Zhang, Hao
    Hu, Xiuyuan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] Improving Generalization of Meta-learning with Inverted Regularization at Inner-level
    Wang, Lianzhe
    Zhou, Shiji
    Zhang, Shanghang
    Chu, Xu
    Chang, Heng
    Zhu, Wenwu
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7826 - 7835
  • [6] Improving the Generalization Performance of Multi-class SVM via Angular Regularization
    Li, Jianxin
    Zhou, Haoyi
    Xie, Pengtao
    Zhang, Yingchun
    [J]. PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2131 - 2137
  • [7] The effect of initial weight, learning rate and regularization on generalization performance and efficiency
    Wu, Y
    Zhang, LM
    [J]. 2002 6TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS, VOLS I AND II, 2002, : 1191 - 1194
  • [8] Improving Generalization Performance in Co-Evolutionary Learning
    Chong, Siang Yew
    Tino, Peter
    Ku, Day Chyi
    Yao, Xin
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2012, 16 (01) : 70 - 85
  • [9] IMPROVING GENERALIZATION PERFORMANCE USING DOUBLE BACKPROPAGATION
    DRUCKER, H
    LECUN, Y
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (06): : 991 - 997
  • [10] Improving Gradient Regularization using Complex-Valued Neural Networks
    Yeats, Eric
    Chen, Yiran
    Li, Hai
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139