On the Importance of Hyperparameter Optimization for Model-based Reinforcement Learning

被引:0
|
作者
Zhang, Baohe [1 ]
Rajan, Raghu [1 ]
Pineda, Luis [2 ]
Lambert, Nathan [3 ]
Biedenkapp, Andre [1 ]
Chua, Kurtland [4 ]
Hutter, Frank [1 ,5 ]
Calandra, Roberto [2 ]
机构
[1] Univ Freiburg, Freiburg, Germany
[2] Facebook AI Res, Menlo Pk, CA USA
[3] Univ Calif Berkeley, Berkeley, CA USA
[4] Princeton Univ, Princeton, NJ 08544 USA
[5] Bosch Ctr AI, Tubingen, Germany
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model-based Reinforcement Learning (MBRL) is a promising framework for learning control in a data-efficient manner. MBRL algorithms can be fairly complex due to the separate dynamics modeling and the subsequent planning algorithm, and as a result, they often possess tens of hyper-parameters and architectural choices. For this reason, MBRL typically requires significant human expertise before it can be applied to new problems and domains. To alleviate this problem, we propose to use automatic hyperparameter optimization (HPO). We demonstrate that this problem can be tackled effectively with automated HPO, which we demonstrate to yield significantly improved performance compared to human experts. In addition, we show that tuning of several MBRL hyperparameters dynamically, i.e. during the training itself, further improves the performance compared to using static hyperparameters which are kept fixed for the whole training. Finally, our experiments provide valuable insights into the effects of several hyperparameters, such as plan horizon or learning rate and their influence on the stability of training and resulting rewards.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Efficient hyperparameter optimization through model-based reinforcement learning
    Wu, Jia
    Chen, SenPeng
    Liu, XiYuan
    [J]. NEUROCOMPUTING, 2020, 409 : 381 - 393
  • [2] Deep Reinforcement Learning with Model-based Acceleration for Hyperparameter Optimization
    Chen, SenPeng
    Wu, Jia
    Chen, XiuYun
    [J]. 2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 170 - 177
  • [3] Reinforcement Learning for Model Selection and Hyperparameter Optimization
    Wu, Jia
    Chen, Sen-Peng
    Chen, Xiu-Yun
    Zhou, Rui
    [J]. Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2020, 49 (02): : 255 - 261
  • [4] Model-Based Reinforcement Learning for Quantized Federated Learning Performance Optimization
    Yang, Nuocheng
    Wang, Sihua
    Chen, Mingzhe
    Brinton, Christopher G.
    Yin, Changchuan
    Saad, Walid
    Cui, Shuguang
    [J]. 2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 5063 - 5068
  • [5] Model-Based Reinforcement Learning Method for Microgrid Optimization Scheduling
    Yao, Jinke
    Xu, Jiachen
    Zhang, Ning
    Guan, Yajuan
    [J]. SUSTAINABILITY, 2023, 15 (12)
  • [6] Model-Based Reinforcement Learning via Proximal Policy Optimization
    Sun, Yuewen
    Yuan, Xin
    Liu, Wenzhang
    Sun, Changyin
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 4736 - 4740
  • [7] Hyperparameter Search Space Pruning - A New Component for Sequential Model-Based Hyperparameter Optimization
    Wistuba, Martin
    Schilling, Nicolas
    Schmidt-Thieme, Lars
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2015, PT II, 2015, 9285 : 104 - 119
  • [8] Generalized Population-Based Training for Hyperparameter Optimization in Reinforcement Learning
    Bai, Hui
    Cheng, Ran
    [J]. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [9] Bayesian Optimistic Optimization: Optimistic Exploration for Model-based Reinforcement Learning
    Wu, Chenyang
    Li, Tianci
    Zhang, Zongzhang
    Yu, Yang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] BiES: Adaptive Policy Optimization for Model-Based Offline Reinforcement Learning
    Yang, Yijun
    Jiang, Jing
    Wang, Zhuowei
    Duan, Qiqi
    Shi, Yuhui
    [J]. AI 2021: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13151 : 570 - 581