Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning

被引:0
|
作者
Zhang, Ce [1 ]
Yao, Xiao [1 ]
Shi, Changfeng [2 ]
Gu, Min [3 ]
机构
[1] Hohai Univ, Coll IoT Engn, Changzhou, Peoples R China
[2] HoHai Univ, Business Sch, ChangZhou, Peoples R China
[3] First Peoples Hosp Changzhou, Changzhou, Peoples R China
关键词
Machine learning; Few-shot learning; K-FAC; Second-order optimization; Adaptive learning rate;
D O I
10.1007/s00530-023-01159-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model-agnostic meta-learning (MAML) highlights the ability to quickly adapt to new tasks with only a small amount of labeled training data among many few-shot learning algorithms. However, the computational complexity is high, because the MAML algorithm generates a large number of second-order parameters in the secondary gradient update. In addition, due to the non-convex nature of the neural network, the loss landscape has many flat areas, leading to slow convergence during training, and excessively long training. In this paper, a second-order optimization method called Kronecker-factored Approximate Curvature (K-FAC) is proposed to approximate Natural Gradient Descent. K-FAC reduces the computational complexity by approximating the large matrix of the Fisher information as the Kronecker product of two much smaller matrices, and the second-order parameter information is fully utilized to accelerate the convergence. Moreover, in order to solve the problem that Natural Gradient Descent is sensitive to the learning rate, this paper proposes Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning (AK-MAML), which automatically adjusts the learning rate according to the curvature and improves the efficiency of training. Experimental results show that AK-MAML has the ability of faster convergence, lower computation, and higher accuracy on few-shot datasets.
引用
收藏
页码:3169 / 3177
页数:9
相关论文
共 50 条
  • [1] Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning
    Zhang, Ce
    Yao, Xiao
    Shi, Changfeng
    Gu, Min
    [J]. MULTIMEDIA SYSTEMS, 2023,
  • [2] Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning
    Ce Zhang
    Xiao Yao
    Changfeng Shi
    Min Gu
    [J]. Multimedia Systems, 2023, 29 (6) : 3169 - 3177
  • [3] Optimizing Neural Networks with Kronecker-factored Approximate Curvature
    Martens, James
    Grosse, Roger
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 2408 - 2417
  • [4] Probabilistic Model-Agnostic Meta-Learning
    Finn, Chelsea
    Xu, Kelvin
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Bayesian Model-Agnostic Meta-Learning
    Yoon, Jaesik
    Kim, Taesup
    Dia, Ousmane
    Kim, Sungwoong
    Bengio, Yoshua
    Ahn, Sungjin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [6] Meta weight learning via model-agnostic meta-learning
    Xu, Zhixiong
    Chen, Xiliang
    Tang, Wei
    Lai, Jun
    Cao, Lei
    [J]. NEUROCOMPUTING, 2021, 432 : 124 - 132
  • [7] Knowledge Distillation for Model-Agnostic Meta-Learning
    Zhang, Min
    Wang, Donglin
    Gai, Sibo
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1355 - 1362
  • [8] Combining Model-Agnostic Meta-Learning and Transfer Learning for Regression
    Satrya, Wahyu Fadli
    Yun, Ji-Hoon
    [J]. SENSORS, 2023, 23 (02)
  • [9] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [10] Learning Symbolic Model-Agnostic Loss Functions via Meta-Learning
    Raymond, Christian
    Chen, Qi
    Xue, Bing
    Zhang, Mengjie
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 13699 - 13714