Invariant Transform Experience Replay: Data Augmentation for Deep Reinforcement Learning

被引:15
|
作者
Lin, Yijiong [1 ]
Huang, Jiancong [1 ]
Zimmer, Matthieu [2 ]
Guan, Yisheng [1 ]
Rojas, Juan [3 ]
Weng, Paul [2 ,4 ]
机构
[1] Guangdong Univ Technol, Sch Electromech Engn, Guangzhou 510000, Peoples R China
[2] Shanghai Jiao Tong Univ, UM SJTU Joint Inst, Shanghai 54000, Peoples R China
[3] Chinese Univ Hong Kong, Sch Mech & Automat Engn, Hong Kong, Peoples R China
[4] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
来源
关键词
AI-based methods; reinforcement learning; deep learning; dexterous manipulation;
D O I
10.1109/LRA.2020.3013937
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Deep Reinforcement Learning (RL) is a promising approach for adaptive robot control, but its current application to robotics is currently hindered by high sample requirements. To alleviate this issue, we propose to exploit the symmetries present in robotic tasks. Intuitively, symmetries from observed trajectories define transformations that leave the space of feasible RL trajectories invariant and can be used to generate new feasible trajectories, which could be used for training. Based on this data augmentation idea, we formulate a general framework, called Invariant Transform Experience Replay that we present with two techniques: (i) Kaleidoscope Experience Replay exploits reflectional symmetries and (ii) Goal-augmented Experience Replay which takes advantage of lax goal definitions. In the Fetch tasks from OpenAI Gym, our experimental results show significant increases in learning rates and success rates. Particularly, we attain a 13, 3, and 5 times speedup in the pushing, sliding, and pick-and-place tasks respectively in the multi-goal setting. Performance gains are also observed in similar tasks with obstacles and we successfully deployed a trained policy on a real Baxter robot. Our work demonstrates that invariant transformations on RL trajectories are a promising methodology to speed up learning in deep RL. Code, video, and supplementary materials are available at [1].
引用
收藏
页码:6615 / 6622
页数:8
相关论文
共 50 条
  • [1] Deep Reinforcement Learning with Experience Replay Based on SARSA
    Zhao, Dongbin
    Wang, Haitao
    Shao, Kun
    Zhu, Yuanheng
    [J]. PROCEEDINGS OF 2016 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2016,
  • [2] Deep Reinforcement Learning With Quantum-Inspired Experience Replay
    Wei, Qing
    Ma, Hailan
    Chen, Chunlin
    Dong, Daoyi
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) : 9326 - 9338
  • [3] Associative Memory Based Experience Replay for Deep Reinforcement Learning
    Li, Mengyuan
    Kazemi, Arman
    Laguna, Ann Franchesca
    Hu, X. Sharon
    [J]. 2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [4] Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay
    Yin, Haiyan
    Pan, Sinno Jialin
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1640 - 1646
  • [5] Trial and Error Experience Replay Based Deep Reinforcement Learning
    Zhang, Cheng
    Ma, Liang
    [J]. 4TH IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2019) / 3RD INTERNATIONAL SYMPOSIUM ON REINFORCEMENT LEARNING (ISRL 2019), 2019, : 221 - 226
  • [6] Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning
    Foerster, Jakob
    Nardelli, Nantas
    Farquhar, Gregory
    Afouras, Triantafyllos
    Torr, Philip H. S.
    Kohli, Pushmeet
    Whiteson, Shimon
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [7] Experience Replay Optimization via ESMM for Stable Deep Reinforcement Learning
    Osei, Richard Sakyi
    Lopez, Daphne
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (01) : 715 - 723
  • [8] Autonomous reinforcement learning with experience replay
    Wawrzynski, Pawel
    Tanwani, Ajay Kumar
    [J]. NEURAL NETWORKS, 2013, 41 : 156 - 167
  • [9] Memory Reduction through Experience Classification for Deep Reinforcement Learning with Prioritized Experience Replay
    Shen, Kai-Huan
    Tsai, Pei-Yun
    [J]. PROCEEDINGS OF THE 2019 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2019), 2019, : 166 - 171
  • [10] Efficient Scheduling of Data Augmentation for Deep Reinforcement Learning
    Ko, Byungchan
    Ok, Jungseul
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,