Real-World Reinforcement Learning via Multifidelity Simulators

被引:33
|
作者
Cutler, Mark [1 ]
Walsh, Thomas J. [1 ]
How, Jonathan P. [1 ]
机构
[1] MIT, Lab Informat & Decis Syst, Cambridge, MA 02139 USA
基金
美国国家科学基金会;
关键词
Animation and simulation; autonomous agents; learning and adaptive systems; reinforcement learning (RL); ROBOTICS;
D O I
10.1109/TRO.2015.2419431
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Reinforcement learning (RL) can be a tool for designing policies and controllers for robotic systems. However, the cost of real-world samples remains prohibitive as many RL algorithms require a large number of samples before learning useful policies. Simulators are one way to decrease the number of required real-world samples, but imperfect models make deciding when and how to trust samples from a simulator difficult. We present a framework for efficient RL in a scenario where multiple simulators of a target task are available, each with varying levels of fidelity. The framework is designed to limit the number of samples used in each successively higher-fidelity/cost simulator by allowing a learning agent to choose to run trajectories at the lowest level simulator that will still provide it with useful information. Theoretical proofs of the framework's sample complexity are given and empirical results are demonstrated on a remote-controlled car with multiple simulators. The approach enables RL algorithms to find near-optimal policies in a physical robot domain with fewer expensive real-world samples than previous transfer approaches or learning without simulators.
引用
收藏
页码:655 / 671
页数:17
相关论文
共 50 条
  • [11] Challenges of real-world reinforcement learning: definitions, benchmarks and analysis
    Dulac-Arnold, Gabriel
    Levine, Nir
    Mankowitz, Daniel J.
    Li, Jerry
    Paduraru, Cosmin
    Gowal, Sven
    Hester, Todd
    MACHINE LEARNING, 2021, 110 (09) : 2419 - 2468
  • [12] Challenges of real-world reinforcement learning: definitions, benchmarks and analysis
    Gabriel Dulac-Arnold
    Nir Levine
    Daniel J. Mankowitz
    Jerry Li
    Cosmin Paduraru
    Sven Gowal
    Todd Hester
    Machine Learning, 2021, 110 : 2419 - 2468
  • [13] Toward the confident deployment of real-world reinforcement learning agents
    Hanna, Josiah P.
    AI MAGAZINE, 2024, 45 (03) : 396 - 403
  • [14] Real-World Human-Robot Collaborative Reinforcement Learning
    Shafti, Ali
    Tjomsland, Jonas
    Dudley, William
    Faisal, A. Aldo
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 11161 - 11166
  • [15] ContainerGym: A Real-World Reinforcement Learning Benchmark for Resource Allocation
    Pendyala, Abhijeet
    Dettmer, Justin
    Glasmachers, Tobias
    Atamna, Asma
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2023, PT I, 2024, 14505 : 78 - 92
  • [16] A Real-World Quadrupedal Locomotion Benchmark for Offline Reinforcement Learning
    Zhang, Hongyin
    Yang, Shuyu
    Wang, Donglin
    2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,
  • [17] Real-world Robot Reaching Skill Learning Based on Deep Reinforcement Learning
    Liu, Naijun
    Lu, Tao
    Cai, Yinghao
    Wang, Rui
    Wang, Shuo
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 4780 - 4784
  • [18] End-to-End Active Object Tracking and Its Real-World Deployment via Reinforcement Learning
    Luo, Wenhan
    Sun, Peng
    Zhong, Fangwei
    Liu, Wei
    Zhang, Tong
    Wang, Yizhou
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (06) : 1317 - 1332
  • [19] Decentralized Circle Formation Control for Fish-like Robots in the Real-world via Reinforcement Learning
    Zhang, Tianhao
    Li, Yueheng
    Li, Shuai
    Ye, Qiwei
    Wang, Chen
    Xie, Guangming
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 8814 - 8820
  • [20] Deep Offline Reinforcement Learning for Real-world Treatment Optimization Applications
    Nambiar, Mila
    Ghosh, Supriyo
    Ong, Priscilla
    Chan, Yu En
    Bee, Yong Mong
    Krishnaswamy, Pavitra
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 4673 - 4684