Real-World Reinforcement Learning via Multifidelity Simulators

被引:33
|
作者
Cutler, Mark [1 ]
Walsh, Thomas J. [1 ]
How, Jonathan P. [1 ]
机构
[1] MIT, Lab Informat & Decis Syst, Cambridge, MA 02139 USA
基金
美国国家科学基金会;
关键词
Animation and simulation; autonomous agents; learning and adaptive systems; reinforcement learning (RL); ROBOTICS;
D O I
10.1109/TRO.2015.2419431
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Reinforcement learning (RL) can be a tool for designing policies and controllers for robotic systems. However, the cost of real-world samples remains prohibitive as many RL algorithms require a large number of samples before learning useful policies. Simulators are one way to decrease the number of required real-world samples, but imperfect models make deciding when and how to trust samples from a simulator difficult. We present a framework for efficient RL in a scenario where multiple simulators of a target task are available, each with varying levels of fidelity. The framework is designed to limit the number of samples used in each successively higher-fidelity/cost simulator by allowing a learning agent to choose to run trajectories at the lowest level simulator that will still provide it with useful information. Theoretical proofs of the framework's sample complexity are given and empirical results are demonstrated on a remote-controlled car with multiple simulators. The approach enables RL algorithms to find near-optimal policies in a physical robot domain with fewer expensive real-world samples than previous transfer approaches or learning without simulators.
引用
收藏
页码:655 / 671
页数:17
相关论文
共 50 条
  • [1] STASIS: Reinforcement Learning Simulators for Human-Centric Real-World Environments
    Efstathiadis, Georgios
    Emedom-Nnamdi, Patrick
    Kolbeinsson, Arinbjorn
    Onnela, Jukka-Pekka
    Lu, Junwei
    TRUSTWORTHY MACHINE LEARNING FOR HEALTHCARE, TML4H 2023, 2023, 13932 : 85 - 92
  • [2] Real-world humanoid locomotion with reinforcement learning
    Radosavovic, Ilija
    Xiao, Tete
    Zhang, Bike
    Darrell, Trevor
    Malik, Jitendra
    Sreenath, Koushil
    SCIENCE ROBOTICS, 2024, 9 (89)
  • [3] Reinforcement Learning in Robotics: Applications and Real-World Challenges
    Kormushev, Petar
    Calinon, Sylvain
    Caldwell, Darwin G.
    ROBOTICS, 2013, 2 (03): : 122 - 148
  • [4] Offline Learning of Counterfactual Predictions for Real-World Robotic Reinforcement Learning
    Jin, Jun
    Graves, Daniel
    Haigh, Cameron
    Luo, Jun
    Jagersand, Martin
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 3616 - 3623
  • [5] Validating Robotics Simulators on Real-World Impacts
    Acosta, Brian
    Yang, William
    Posa, Michael
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) : 6471 - 6478
  • [6] Injection optimization at particle accelerators via reinforcement learning: From simulation to real-world application
    Awal, Awal
    Hetzel, Jan
    Gebel, Ralf
    Pretz, Joerg
    PHYSICAL REVIEW ACCELERATORS AND BEAMS, 2025, 28 (03)
  • [7] Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning
    Sharma, Archit
    Ahn, Michael
    Levine, Sergey
    Kumar, Vikash
    Hausmani, Karol
    Gu, Shixiang
    ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [8] Setting up a Reinforcement Learning Task with a Real-World Robot
    Mahmood, A. Rupam
    Korenkevych, Dmytro
    Komer, Brent J.
    Bergstra, James
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 4635 - 4640
  • [9] Real-world reinforcement learning for autonomous humanoid robot docking
    Navarro-Guerrero, Nicolas
    Weber, Cornelius
    Schroeter, Pascal
    Wermter, Stefan
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2012, 60 (11) : 1400 - 1407
  • [10] NeoRL: A Near Real-World Benchmark for Offline Reinforcement Learning
    Qin, Rong-Jun
    Zhang, Xingyuan
    Gao, Songyi
    Chen, Xiong-Hui
    Li, Zewen
    Zhang, Weinan
    Yu, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,