Real-World Reinforcement Learning via Multifidelity Simulators

被引:33
|
作者
Cutler, Mark [1 ]
Walsh, Thomas J. [1 ]
How, Jonathan P. [1 ]
机构
[1] MIT, Lab Informat & Decis Syst, Cambridge, MA 02139 USA
基金
美国国家科学基金会;
关键词
Animation and simulation; autonomous agents; learning and adaptive systems; reinforcement learning (RL); ROBOTICS;
D O I
10.1109/TRO.2015.2419431
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Reinforcement learning (RL) can be a tool for designing policies and controllers for robotic systems. However, the cost of real-world samples remains prohibitive as many RL algorithms require a large number of samples before learning useful policies. Simulators are one way to decrease the number of required real-world samples, but imperfect models make deciding when and how to trust samples from a simulator difficult. We present a framework for efficient RL in a scenario where multiple simulators of a target task are available, each with varying levels of fidelity. The framework is designed to limit the number of samples used in each successively higher-fidelity/cost simulator by allowing a learning agent to choose to run trajectories at the lowest level simulator that will still provide it with useful information. Theoretical proofs of the framework's sample complexity are given and empirical results are demonstrated on a remote-controlled car with multiple simulators. The approach enables RL algorithms to find near-optimal policies in a physical robot domain with fewer expensive real-world samples than previous transfer approaches or learning without simulators.
引用
收藏
页码:655 / 671
页数:17
相关论文
共 50 条
  • [21] Deep reinforcement learning for real-world quadrupedal locomotion: a comprehensive review
    Zhang, Hongyin
    He, Li
    Wang, Donglin
    INTELLIGENCE & ROBOTICS, 2022, 2 (03):
  • [22] Tackling Real-World Autonomous Driving using Deep Reinforcement Learning
    Maramotti, Paolo
    Capasso, Alessandro Paolo
    Bacchiani, Giulio
    Broggi, Alberto
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1274 - 1281
  • [23] Towards Real-World Deployment of Reinforcement Learning for Traffic Signal Control
    Mueller, Arthur
    Rangras, Vishal
    Ferfers, Tobias
    Hufen, Florian
    Schreckenberg, Lukas
    Jasperneite, Juergen
    Schnittker, Georg
    Waldmann, Michael
    Friesen, Maxim
    Wiering, Marco
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 507 - 514
  • [24] Real-world dexterous object manipulation based deep reinforcement learning
    Yao, Qingfeng
    Wang, Jilong
    Yang, Shuyu
    arXiv, 2021,
  • [25] Simulation-Based Reinforcement Learning for Real-World Autonomous Driving
    Osinski, Blazej
    Jakubowski, Adam
    Ziecina, Pawel
    Milos, Piotr
    Galias, Christopher
    Homoceanu, Silviu
    Michalewski, Henryk
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6411 - 6418
  • [26] Real-Sim-Real Transfer for Real-World Robot Control Policy Learning with Deep Reinforcement Learning
    Liu, Naijun
    Cai, Yinghao
    Lu, Tao
    Wang, Rui
    Wang, Shuo
    APPLIED SCIENCES-BASEL, 2020, 10 (05):
  • [27] Intrinsically motivated reinforcement learning for human-robot interaction in the real-world
    Qureshi, Ahmed Hussain
    Nakamura, Yutaka
    Yoshikawa, Yuichiro
    Ishiguro, Hiroshi
    NEURAL NETWORKS, 2018, 107 : 23 - 33
  • [28] Non-blocking Asynchronous Training for Reinforcement Learning in Real-World Environments
    Bohm, Peter
    Pounds, Pauline
    Chapman, Archie C.
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10927 - 10934
  • [29] Domain Adapting Deep Reinforcement Learning for Real-World Speech Emotion Recognition
    Rajapakshe, Thejan
    Rana, Rajib
    Khalifa, Sara
    Schuller, Bjoern W.
    IEEE ACCESS, 2024, 12 : 193101 - 193114
  • [30] Optimizing Reinforcement Learning Control Model in Furuta Pendulum and Transferring it to Real-World
    Hong, Myung Rae
    Kang, Sanghun
    Lee, Jingoo
    Seo, Sungchul
    Han, Seungyong
    Koh, Je-Sung
    Kang, Daeshik
    IEEE ACCESS, 2023, 11 : 95195 - 95200