Learning to Balance Local Losses via Meta-Learning

被引:0
|
作者
Yoa, Seungdong [1 ]
Jeon, Minkyu [1 ]
Oh, Youngjin [1 ]
Kim, Hyunwoo J. [1 ]
机构
[1] Korea Univ, Dept Comp Sci, Seoul 02841, South Korea
关键词
Training; Neural networks; Loss measurement; Standards; Deep learning; Task analysis; Licenses; image classification; machine learning; meta-learning;
D O I
10.1109/ACCESS.2021.3113934
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The standard training for deep neural networks relies on a global and fixed loss function. For more effective training, dynamic loss functions have been recently proposed. However, the dynamic global loss function is not flexible to differentially train layers in complex deep neural networks. In this paper, we propose a general framework that learns to adaptively train each layer of deep neural networks via meta-learning. Our framework leverages the local error signals from layers and identifies which layer needs to be trained more at every iteration. Also, the proposed method improves the local loss function with our minibatch-wise dropout and cross-validation loop to alleviate meta-overfitting. The experiments show that our method achieved competitive performance compared to state-of-the-art methods on popular benchmark datasets for image classification: CIFAR-10 and CIFAR-100. Surprisingly, our method enables training deep neural networks without skip-connections using dynamically weighted local loss functions.
引用
收藏
页码:130834 / 130844
页数:11
相关论文
共 50 条
  • [1] Contextualizing Meta-Learning via Learning to Decompose
    Ye, Han-Jia
    Zhou, Da-Wei
    Hong, Lanqing
    Li, Zhenguo
    Wei, Xiu-Shen
    Zhan, De-Chuan
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (01) : 117 - 133
  • [2] A Collaborative Learning Framework via Federated Meta-Learning
    Lin, Sen
    Yang, Guang
    Zhang, Junshan
    [J]. 2020 IEEE 40TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS), 2020, : 289 - 299
  • [3] Meta weight learning via model-agnostic meta-learning
    Xu, Zhixiong
    Chen, Xiliang
    Tang, Wei
    Lai, Jun
    Cao, Lei
    [J]. NEUROCOMPUTING, 2021, 432 : 124 - 132
  • [4] Improving progressive sampling via meta-learning on learning curves
    Leite, R
    Brazdil, P
    [J]. MACHINE LEARNING: ECML 2004, PROCEEDINGS, 2004, 3201 : 250 - 261
  • [5] Learning to Transfer: Unsupervised Domain Translation via Meta-Learning
    Lin, Jianxin
    Wang, Yijun
    Chen, Zhibo
    He, Tianyu
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11507 - 11514
  • [6] Learning Meta-Learning (LML) dataset: Survey data of meta-learning parameters
    Corraya, Sonia
    Al Mamun, Shamim
    Kaiser, M. Shamim
    [J]. DATA IN BRIEF, 2023, 51
  • [7] Meta-learning in Reinforcement Learning
    Schweighofer, N
    Doya, K
    [J]. NEURAL NETWORKS, 2003, 16 (01) : 5 - 9
  • [8] Learning to Forget for Meta-Learning
    Baik, Sungyong
    Hong, Seokil
    Lee, Kyoung Mu
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2376 - 2384
  • [9] Edge Sparsification for Graphs via Meta-Learning
    Wan, Guihong
    [J]. 2021 IEEE 37TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2021), 2021, : 2733 - 2738
  • [10] Automatic Modulation Classification via Meta-Learning
    Hao, Xiaoyang
    Feng, Zhixi
    Yang, Shuyuan
    Wang, Min
    Jiao, Licheng
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (14) : 12276 - 12292