A survey on model-based reinforcement learning

被引:0
|
作者
Fan-Ming Luo
Tian Xu
Hang Lai
Xiong-Hui Chen
Weinan Zhang
Yang Yu
机构
[1] Nanjing University,National Key Laboratory for Novel Software Technology
[2] Shanghai Jiao Tong University,Department of Computer Science and Engineering
[3] Polixir.ai,undefined
来源
关键词
reinforcement learning; model-based reinforcement learning; planning; model learning; model learning with reduced error; model usage;
D O I
暂无
中图分类号
学科分类号
摘要
Reinforcement learning (RL) interacts with the environment to solve sequential decision-making problems via a trial-and-error approach. Errors are always undesirable in real-world applications, even though RL excels at playing complex video games that permit several trial-and-error attempts. To improve sample efficiency and thus reduce errors, model-based reinforcement learning (MBRL) is believed to be a promising direction, as it constructs environment models in which trial-and-errors can occur without incurring actual costs. In this survey, we investigate MBRL with a particular focus on the recent advancements in deep RL. There is a generalization error between the learned model of a non-tabular environment and the actual environment. Consequently, it is crucial to analyze the disparity between policy training in the environment model and that in the actual environment, guiding algorithm design for improved model learning, model utilization, and policy training. In addition, we discuss the recent developments of model-based techniques in other forms of RL, such as offline RL, goal-conditioned RL, multi-agent RL, and meta-RL. Furthermore, we discuss the applicability and benefits of MBRL for real-world tasks. Finally, this survey concludes with a discussion of the promising future development prospects for MBRL. We believe that MBRL has great unrealized potential and benefits in real-world applications, and we hope this survey will encourage additional research on MBRL.
引用
收藏
相关论文
共 50 条
  • [1] Model-based Reinforcement Learning: A Survey
    Moerland, Thomas M.
    Broekens, Joost
    Plaat, Aske
    Jonker, Catholijn M.
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2023, 16 (01): : 1 - 118
  • [2] A survey on model-based reinforcement learning
    Fan-Ming LUO
    Tian XU
    Hang LAI
    Xiong-Hui CHEN
    Weinan ZHANG
    Yang YU
    [J]. Science China(Information Sciences), 2024, 67 (02) : 59 - 84
  • [3] A survey on model-based reinforcement learning
    Luo, Fan-Ming
    Xu, Tian
    Lai, Hang
    Chen, Xiong-Hui
    Zhang, Weinan
    Yu, Yang
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (02)
  • [4] Model-Based Reinforcement Learning in Robotics: A Survey
    Sun, Shiguang
    Lan, Xuguang
    Zhang, Hanbo
    Zheng, Nanning
    [J]. Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (01): : 1 - 16
  • [5] Survey of Model-Based Reinforcement Learning: Applications on Robotics
    Athanasios S. Polydoros
    Lazaros Nalpantidis
    [J]. Journal of Intelligent & Robotic Systems, 2017, 86 : 153 - 173
  • [6] Survey of Model-Based Reinforcement Learning: Applications on Robotics
    Polydoros, Athanasios S.
    Nalpantidis, Lazaros
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2017, 86 (02) : 153 - 173
  • [7] A Brief Survey of Model-Based Reinforcement Learning Techniques
    Pal, Constantin-Valentin
    Leon, Florin
    [J]. 2020 24TH INTERNATIONAL CONFERENCE ON SYSTEM THEORY, CONTROL AND COMPUTING (ICSTCC), 2020, : 92 - 97
  • [8] High-accuracy model-based reinforcement learning, a survey
    Plaat, Aske
    Kosters, Walter
    Preuss, Mike
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2023, 56 (09) : 9541 - 9573
  • [9] High-accuracy model-based reinforcement learning, a survey
    Aske Plaat
    Walter Kosters
    Mike Preuss
    [J]. Artificial Intelligence Review, 2023, 56 : 9541 - 9573
  • [10] The ubiquity of model-based reinforcement learning
    Doll, Bradley B.
    Simon, Dylan A.
    Daw, Nathaniel D.
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 1075 - 1081