Action-Quantized Offline Reinforcement Learning for Robotic Skill Learning

被引:0
|
作者
Luo, Jianlan [1 ]
Dong, Perry [1 ]
Wu, Jeffrey [1 ]
Kumar, Aviral [1 ]
Geng, Xinyang [1 ]
Levine, Sergey [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
来源
关键词
Offline Reinforcement Learning; Discretization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The offline reinforcement learning (RL) paradigm provides a general recipe to convert static behavior datasets into policies that can perform better than the policy that collected the data. While policy constraints, conservatism, and other methods for mitigating distributional shifts have made offline reinforcement learning more effective, the continuous action setting often necessitates various approximations for applying these techniques. Many of these challenges are greatly alleviated in discrete action settings, where offline RL constraints and regularizers can often be computed more precisely or even exactly. In this paper, we propose an adaptive scheme for action quantization. We use a VQ-VAE to learn state-conditioned action quantization, avoiding the exponential blowup that comes with naive discretization of the action space. We show that several state-of-the-art offline RL methods such as IQL, CQL, and BRAC improve in performance on benchmarks when combined with our proposed discretization scheme. We further validate our approach on a set of challenging long-horizon complex robotic manipulation tasks in the Robomimic environment, where our discretized offline RL algorithms are able to improve upon their continuous counterparts by 2-3x. Our project page is at saqrl.github.io
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Offline reinforcement learning with anderson acceleration for robotic tasks
    Guoyu Zuo
    Shuai Huang
    Jiangeng Li
    Daoxiong Gong
    [J]. Applied Intelligence, 2022, 52 : 9885 - 9898
  • [2] Offline reinforcement learning with anderson acceleration for robotic tasks
    Zuo, Guoyu
    Huang, Shuai
    Li, Jiangeng
    Gong, Daoxiong
    [J]. APPLIED INTELLIGENCE, 2022, 52 (09) : 9885 - 9898
  • [3] Offline Learning of Counterfactual Predictions for Real-World Robotic Reinforcement Learning
    Jin, Jun
    Graves, Daniel
    Haigh, Cameron
    Luo, Jun
    Jagersand, Martin
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 3616 - 3623
  • [4] Learning Pseudometric-based Action Representations for Offline Reinforcement Learning
    Gu, Pengjie
    Zhao, Mengchen
    Chen, Chen
    Li, Dong
    Hao, Jianye
    An, Bo
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] UAC: Offline Reinforcement Learning With Uncertain Action Constraint
    Guan, Jiayi
    Gu, Shangding
    Li, Zhijun
    Hou, Jing
    Yang, Yiqin
    Chen, Guang
    Jiang, Changjun
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (02) : 671 - 680
  • [6] Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills
    Chebotar, Yevgen
    Hausman, Karol
    Lu, Yao
    Xiao, Ted
    Kalashnikov, Dmitry
    Varley, Jake
    Irpan, Alex
    Eysenbach, Benjamin
    Julian, Ryan
    Finn, Chelsea
    Levine, Sergey
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [7] Offline Reinforcement Learning with Pseudometric Learning
    Dadashi, Robert
    Rezaeifar, Shideh
    Vieillard, Nino
    Hussenot, Leonard
    Pietquin, Olivier
    Geist, Matthieu
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [8] Skill Learning for Robotic Insertion Based on One-shot Demonstration and Reinforcement Learning
    Ying Li
    De Xu
    [J]. International Journal of Automation and Computing, 2021, 18 : 457 - 467
  • [9] Skill Learning for Robotic Insertion Based on One-shot Demonstration and Reinforcement Learning
    Li, Ying
    Xu, De
    [J]. INTERNATIONAL JOURNAL OF AUTOMATION AND COMPUTING, 2021, 18 (03) : 457 - 467
  • [10] Skill Learning for Robotic Insertion Based on One-shot Demonstration and Reinforcement Learning
    Ying Li
    De Xu
    [J]. International Journal of Automation and Computing, 2021, 18 (03) : 457 - 467