Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning

被引:0
|
作者
Esteves, Bernardo [1 ,2 ]
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
机构
[1] INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
关键词
Reinforcement learning; Transfer learning; Representation learning;
D O I
10.1007/978-3-031-49008-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work explores pre-training as a strategy to allow reinforcement learning (RL) algorithms to efficiently adapt to new (albeit similar) tasks. We argue for introducing variability during the pre-training phase, in the form of augmentations to the observations of the agent, to improve the sample efficiency of the fine-tuning stage. We categorize such variability in the form of perceptual, dynamic and semantic augmentations, which can be easily employed in standard pre-training methods. We perform extensive evaluations of our proposed augmentation scheme in model-based algorithms, across multiple scenarios of increasing complexity. The results consistently show that our augmentation scheme significantly improves the efficiency of the fine-tuning to novel tasks, outperforming other state-of-the-art pre-training approaches.
引用
收藏
页码:133 / 145
页数:13
相关论文
共 50 条
  • [1] Efficient Conditional Pre-training for Transfer Learning
    Chakraborty, Shuvam
    Uzkent, Burak
    Ayush, Kumar
    Tanmay, Kumar
    Sheehan, Evan
    Ermon, Stefano
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4240 - 4249
  • [2] RePreM: Representation Pre-training with Masked Model for Reinforcement Learning
    Cai, Yuanying
    Zhang, Chuheng
    Shen, Wei
    Zhang, Xuyun
    Ruan, Wenjie
    Huang, Longbo
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 6879 - 6887
  • [3] CyclicFL: Efficient Federated Learning with Cyclic Model Pre-Training
    Zhang, Pengyu
    Zhou, Yingbo
    Hu, Ming
    Wei, Xian
    Chen, Mingsong
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2025,
  • [4] Pre-training with asynchronous supervised learning for reinforcement learning based autonomous driving
    Wang, Yunpeng
    Zheng, Kunxian
    Tian, Daxin
    Duan, Xuting
    Zhou, Jianshan
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2021, 22 (05) : 673 - 686
  • [5] Improving Reinforcement Learning Pre-Training with Variational Dropout
    Blau, Tom
    Ott, Lionel
    Ramos, Fabio
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 4115 - 4122
  • [6] MOTO: Offline Pre-training to Online Fine-tuning for Model-based Robot Learning
    Rafailov, Rafael
    Hatch, Kyle
    Kolev, Victor
    Martin, John D.
    Phielipp, Mariano
    Finn, Chelsea
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [7] Pre-training Framework for Improving Learning Speed of Reinforcement Learning based Autonomous Vehicles
    Kim, Jung-Jae
    Cha, Si-Ho
    Ryu, Minwoo
    Jo, Minho
    2019 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2019, : 321 - 322
  • [8] An Efficient Approach to Model-Based Hierarchical Reinforcement Learning
    Li, Zhuoru
    Narayan, Akshay
    Leong, Tze-Yun
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3583 - 3589
  • [9] Supervised pre-training for improved stability in deep reinforcement learning
    Jang, Sooyoung
    Kim, Hyung-Il
    ICT EXPRESS, 2023, 9 (01): : 51 - 56
  • [10] Efficient reinforcement learning: Model-based acrobot control
    Boone, G
    1997 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION - PROCEEDINGS, VOLS 1-4, 1997, : 229 - 234