Online Optimization Method of Learning Process for Meta-Learning

被引:0
|
作者
Xu, Zhixiong [1 ,2 ]
Zhang, Wei [1 ]
Li, Ailin [1 ]
Zhao, Feifei [1 ]
Jing, Yuanyuan [1 ]
Wan, Zheng [1 ]
Cao, Lei [2 ]
Chen, Xiliang [2 ]
机构
[1] Army Acad Border & Coastal Def, Xian 710100, Peoples R China
[2] Army Engn Univ PLA, Nanjing 210001, Peoples R China
来源
COMPUTER JOURNAL | 2023年 / 67卷 / 05期
关键词
Meta-learning; deep reinforcement learning; hyper-parameter; gradient descent;
D O I
10.1093/comjnl/bxad089
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Meta-learning is a pivotal and potentially influential machine learning approach to solve challenging problems in reinforcement learning. However, the costly hyper-parameter tuning for training stability of meta-learning is a known shortcoming and currently a hotspot of research. This paper addresses this shortcoming by introducing an online and easily trainable hyper-parameter optimization approach, called Meta Parameters Learning via Meta-Learning (MPML), to combine online hyper-parameter adjustment scheme into meta-learning algorithm, which reduces the need to tune hyper-parameters. Specifically, a basic learning rate for each training task is put forward. Besides, the proposed algorithm dynamically adapts multiple basic learning rate and a shared meta-learning rate through conducting gradient descent alongside the initial optimization steps. In addition, the sensitivity with respect to hyper-parameter choices in the proposed approach are also discussed compared with model-agnostic meta-learning method. The experimental results on reinforcement learning problems demonstrate MPML algorithm is easy to implement and delivers more highly competitive performance than existing meta-learning methods on a diverse set of challenging control tasks.
引用
收藏
页码:1645 / 1651
页数:7
相关论文
共 50 条
  • [1] Online Optimization Method of Learning Process for Meta-Learning
    Xu, Zhixiong
    Zhang, Wei
    Li, Ailin
    Zhao, Feifei
    Jing, Yuanyuan
    Wan, Zheng
    Cao, Lei
    Chen, Xiliang
    [J]. Computer Journal, 1600, 67 (05): : 1645 - 1651
  • [2] Online Meta-Learning
    Finn, Chelsea
    Rajeswaran, Aravind
    Kakade, Sham
    Levine, Sergey
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [3] Online Structured Meta-learning
    Yao, Huaxiu
    Zhou, Yingbo
    Mahdavi, Mehrdad
    Li, Zhenhui
    Socher, Richard
    Xiong, Caiming
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [4] Online-Within-Online Meta-Learning
    Denevi, Giulia
    Stamos, Dimitris
    Ciliberto, Carlo
    Pontil, Massimiliano
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [5] Online meta-learning for POI recommendation
    Lv, Yao
    Sang, Yu
    Tai, Chong
    Cheng, Wanjun
    Shang, Jedi S.
    Qu, Jianfeng
    Chu, Xiaomin
    Zhang, Ruoqian
    [J]. GEOINFORMATICA, 2023, 27 (01) : 61 - 76
  • [6] Online meta-learning for POI recommendation
    Yao Lv
    Yu Sang
    Chong Tai
    Wanjun Cheng
    Jedi S. Shang
    Jianfeng Qu
    Xiaomin Chu
    Ruoqian Zhang
    [J]. GeoInformatica, 2023, 27 : 61 - 76
  • [7] Meta-Learning with Differentiable Convex Optimization
    Lee, Kwonjoon
    Maji, Subhransu
    Ravichandran, Avinash
    Soatto, Stefano
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10649 - 10657
  • [8] Reconciling meta-learning and continual learning with online mixtures of tasks
    Jerfel, Ghassen
    Grant, Erin
    Griffiths, Thomas L.
    Heller, Katherine
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Constructive meta-learning with machine learning method repositories
    Abe, H
    Yamaguchi, T
    [J]. INNOVATIONS IN APPLIED ARTIFICIAL INTELLIGENCE, 2004, 3029 : 502 - 511
  • [10] Online Meta-Learning by Parallel Algorithm Competition
    Elfwing, Stefan
    Uchibe, Eiji
    Doya, Kenji
    [J]. GECCO'18: PROCEEDINGS OF THE 2018 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2018, : 426 - 433