Reinforcement Learning with Augmented Data

被引:0
|
作者
Laskin, Michael [1 ]
Lee, Kimin [1 ]
Stooke, Adam [1 ]
Pinto, Lerrel [2 ]
Abbeel, Pieter [1 ]
Srinivas, Aravind [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] New York Univ, New York, NY USA
关键词
LEVEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from visual observations is a fundamental yet challenging problem in Reinforcement Learning (RL). Although algorithmic advances combined with convolutional neural networks have proved to be a recipe for success, current methods are still lacking on two fronts: (a) data-efficiency of learning and (b) generalization to new environments. To this end, we present Reinforcement Learning with Augmented Data (RAD), a simple plug-and-play module that can enhance most RL algorithms. We perform the first extensive study of general data augmentations for RL on both pixel-based and state-based inputs, and introduce two new data augmentations - random translate and random amplitude scale. We show that augmentations such as random translate, crop, color jitter, patch cutout, random convolutions, and amplitude scale can enable simple RL algorithms to outperform complex state-of-the-art methods across common benchmarks. RAD sets a new state-of-the-art in terms of data-efficiency and final performance on the DeepMind Control Suite benchmark for pixel-based control as well as OpenAI Gym benchmark for state-based control. We further demonstrate that RAD significantly improves test-time generalization over existing methods on several OpenAI ProcGen benchmarks. Our RAD module and training code are available at https://www.github.com/MishaLaskin/rad.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Retrieval Augmented Reinforcement Learning
    Goyal, Anirudh
    Friesen, Abram L.
    Weber, Theophane
    Banino, Andrea
    Ke, Nan Rosemary
    Badia, Adria Puigdomenech
    Guez, Arthur
    Mirza, Mehdi
    Humphreys, Peter C.
    Konyushkova, Ksenia
    Sifre, Laurent
    Valko, Michal
    Osindero, Simon
    Lillicrap, Timothy
    Heess, Nicolas
    Blundell, Charles
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] A stable data-augmented reinforcement learning method with ensemble exploration and exploitation
    Zuo, Guoyu
    Tian, Zhipeng
    Huang, Gao
    APPLIED INTELLIGENCE, 2023, 53 (21) : 24792 - 24803
  • [3] A stable data-augmented reinforcement learning method with ensemble exploration and exploitation
    Guoyu Zuo
    Zhipeng Tian
    Gao Huang
    Applied Intelligence, 2023, 53 : 24792 - 24803
  • [4] Reinforcement Learning Augmented Optimization for Smart Mobility
    Overko, Roman
    Ordonez-Hurtado, Rodrigo
    Zhuk, Sergiy
    Shorten, Robert
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 1286 - 1292
  • [5] Planning-Augmented Hierarchical Reinforcement Learning
    Gieselmann, Robert
    Pokorny, Florian T.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 5097 - 5104
  • [6] A STOCHASTIC COMPOSITE AUGMENTED LAGRANGIAN METHOD FOR REINFORCEMENT LEARNING
    LI, Yongfeng
    Zhao, Mingming
    Chen, Weijie
    Wen, Zaiwen
    SIAM JOURNAL ON OPTIMIZATION, 2023, 33 (02) : 921 - 949
  • [7] Augmented Memory Replay in Reinforcement Learning With Continuous Control
    Ramicic, Mirza
    Bonarini, Andrea
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (02) : 485 - 496
  • [8] Augmented Proximal Policy Optimization for Safe Reinforcement Learning
    Dai, Juntao
    Ji, Jiaming
    Yang, Long
    Zheng, Qian
    Pan, Gang
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7288 - 7295
  • [9] Predicting investment behavior: An augmented reinforcement learning model
    Shimokawa, Tetsuya
    Suzuki, Kyoko
    Misawa, Tadanobu
    Okano, Yoshitaka
    NEUROCOMPUTING, 2009, 72 (16-18) : 3447 - 3461
  • [10] Imagination-Augmented Agents for Deep Reinforcement Learning
    Racaniere, Sebastien
    Weber, Theophane
    Reichert, David P.
    Buesing, Lars
    Guez, Arthur
    Rezende, Danilo
    Badia, Adria Puigdomenech
    Vinyals, Oriol
    Heess, Nicolas
    Li, Yujia
    Pascanu, Razvan
    Battaglia, Peter
    Hassabis, Demis
    Silver, David
    Wierstra, Daan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30