Online Optimization Method of Learning Process for Meta-Learning

被引:0
|
作者
Xu, Zhixiong [1 ,2 ]
Zhang, Wei [1 ]
Li, Ailin [1 ]
Zhao, Feifei [1 ]
Jing, Yuanyuan [1 ]
Wan, Zheng [1 ]
Cao, Lei [2 ]
Chen, Xiliang [2 ]
机构
[1] Army Acad Border & Coastal Def, Xian 710100, Peoples R China
[2] Army Engn Univ PLA, Nanjing 210001, Peoples R China
来源
COMPUTER JOURNAL | 2023年 / 67卷 / 05期
关键词
Meta-learning; deep reinforcement learning; hyper-parameter; gradient descent;
D O I
10.1093/comjnl/bxad089
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Meta-learning is a pivotal and potentially influential machine learning approach to solve challenging problems in reinforcement learning. However, the costly hyper-parameter tuning for training stability of meta-learning is a known shortcoming and currently a hotspot of research. This paper addresses this shortcoming by introducing an online and easily trainable hyper-parameter optimization approach, called Meta Parameters Learning via Meta-Learning (MPML), to combine online hyper-parameter adjustment scheme into meta-learning algorithm, which reduces the need to tune hyper-parameters. Specifically, a basic learning rate for each training task is put forward. Besides, the proposed algorithm dynamically adapts multiple basic learning rate and a shared meta-learning rate through conducting gradient descent alongside the initial optimization steps. In addition, the sensitivity with respect to hyper-parameter choices in the proposed approach are also discussed compared with model-agnostic meta-learning method. The experimental results on reinforcement learning problems demonstrate MPML algorithm is easy to implement and delivers more highly competitive performance than existing meta-learning methods on a diverse set of challenging control tasks.
引用
收藏
页码:1645 / 1651
页数:7
相关论文
共 50 条
  • [41] Modeling and Optimization Trade-off in Meta-learning
    Gao, Katelyn
    Sener, Ozan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [42] Meta-learning via search combined with parameter optimization
    Duch, W
    Grudzinski, K
    [J]. INTELLIGENT INFORMATION SYSTEMS 2002, PROCEEDINGS, 2002, 17 : 13 - 22
  • [43] Meta-learning digitized-counterdiabatic quantum optimization
    Chandarana, Pranav
    Vieites, Pablo Suarez
    Hegade, Narendra N.
    Solano, Enrique
    Ban, Yue
    Chen, Xi
    [J]. QUANTUM SCIENCE AND TECHNOLOGY, 2023, 8 (04)
  • [44] Submodular Meta-Learning
    Adibi, Arman
    Mokhtari, Aryan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [45] Meta-learning with backpropagation
    Younger, AS
    Hochreiter, S
    Conwell, PR
    [J]. IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 2001 - 2006
  • [46] Competitive Meta-Learning
    Boxi Weng
    Jian Sun
    Gao Huang
    Fang Deng
    Gang Wang
    Jie Chen
    [J]. IEEE/CAA Journal of Automatica Sinica, 2023, 10 (09) : 1902 - 1904
  • [47] TOOLS AND TASKS FOR LEARNING AND META-LEARNING
    Jaworski, Barbara
    [J]. JOURNAL OF MATHEMATICS TEACHER EDUCATION, 2005, 8 (05) : 359 - 361
  • [48] Tools and Tasks for Learning and Meta-learning
    Barbara Jaworski
    [J]. Journal of Mathematics Teacher Education, 2005, 8 (5) : 359 - 361
  • [49] On meta-learning rule learning heuristics
    Janssen, Frederik
    Fuernkranz, Johannes
    [J]. ICDM 2007: PROCEEDINGS OF THE SEVENTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, 2007, : 529 - 534
  • [50] Meta-learning: From classification to forecasting, to optimization, and beyond
    Smith-Miles, Kate
    [J]. 6th IEEE/ACIS International Conference on Computer and Information Science, Proceedings, 2007, : 2 - 2