Offline Reinforcement Learning for Quadrotor Control: Overcoming the Ground Effect

被引:0
|
作者
Sacchetto, Luca [1 ]
Korte, Mathias [1 ]
Gronauer, Sven [1 ]
Diepold, Klaus [1 ]
机构
[1] Tech Univ Munich, Sch Computat Informat & Technol, Arcisstr 21, D-80333 Munich, Germany
关键词
D O I
10.1109/IROS55552.2023.10341599
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying Reinforcement Learning to solve real-world optimization problems presents significant challenges because of the large amount of data normally required. A popular solution is to train the algorithms in a simulation and transfer the weights to the real system. However, sim-to-real approaches are prone to fail when the Reality Gap is too big, e.g. in robotic systems with complex and non-linear dynamics. In this work, we propose the use of Offline Reinforcement Learning as a viable alternative to sim-to-real policy transfer to address such instances. On the example of a small quadrotor, we show that the ground effect causes problems in an otherwise functioning zero-shot sim-to-real framework. Our sim-to-real experiments show that, even with the explicit modelling of the ground effect and the employing of popular transfer techniques, the trained policies fail to capture the physical nuances necessary to perform a real-world take-off maneuver. Contrariwise, we show that state-of-the-art Offline Reinforcement Learning algorithms represent a feasible, reliable and sample efficient alternative in this use case.
引用
收藏
页码:7539 / 7544
页数:6
相关论文
共 50 条
  • [21] Offline Reinforcement Learning with Pseudometric Learning
    Dadashi, Robert
    Rezaeifar, Shideh
    Vieillard, Nino
    Hussenot, Leonard
    Pietquin, Olivier
    Geist, Matthieu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [22] Passive Position Control of a Quadrotor With Ground Effect Interaction
    Davis, Edwin
    Pounds, Paul E. I.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2016, 1 (01): : 539 - 545
  • [23] Offline Model-Based Reinforcement Learning for Tokamak Control
    Char, Ian
    Abbate, Joseph
    Bardoczi, Laszlo
    Boyer, Mark D.
    Chung, Youngseog
    Conlin, Rory
    Erickson, Keith
    Mehta, Viraj
    Richner, Nathan
    Kolemen, Egemen
    Schneider, Jeff
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [24] Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
    Yang, Yiqin
    Hu, Hao
    Li, Wenzhe
    Li, Siyuan
    Yang, Jun
    Zhao, Qianchuan
    Zhang, Chongjie
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10843 - 10851
  • [25] End-to-end offline reinforcement learning for glycemia control
    Beolet, Tristan
    Adenis, Alice
    Huneker, Erik
    Louis, Maxime
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 154
  • [26] Reinforcement learning and model predictive control for robust embedded quadrotor guidance and control
    Colin Greatwood
    Arthur G. Richards
    Autonomous Robots, 2019, 43 : 1681 - 1693
  • [27] Reinforcement learning and model predictive control for robust embedded quadrotor guidance and control
    Greatwood, Colin
    Richards, Arthur G.
    AUTONOMOUS ROBOTS, 2019, 43 (07) : 1681 - 1693
  • [28] Benchmarking Offline Reinforcement Learning
    Tittaferrante, Andrew
    Yassine, Abdulsalam
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 259 - 263
  • [29] Federated Offline Reinforcement Learning
    Zhou, Doudou
    Zhang, Yufeng
    Sonabend-W, Aaron
    Wang, Zhaoran
    Lu, Junwei
    Cai, Tianxi
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024, 119 (548) : 3152 - 3163
  • [30] Distributed Offline Reinforcement Learning
    Heredia, Paulo
    George, Jemin
    Mou, Shaoshuai
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 4621 - 4626