Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning

被引:0
|
作者
Esteves, Bernardo [1 ,2 ]
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
机构
[1] INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
关键词
Reinforcement learning; Transfer learning; Representation learning;
D O I
10.1007/978-3-031-49008-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work explores pre-training as a strategy to allow reinforcement learning (RL) algorithms to efficiently adapt to new (albeit similar) tasks. We argue for introducing variability during the pre-training phase, in the form of augmentations to the observations of the agent, to improve the sample efficiency of the fine-tuning stage. We categorize such variability in the form of perceptual, dynamic and semantic augmentations, which can be easily employed in standard pre-training methods. We perform extensive evaluations of our proposed augmentation scheme in model-based algorithms, across multiple scenarios of increasing complexity. The results consistently show that our augmentation scheme significantly improves the efficiency of the fine-tuning to novel tasks, outperforming other state-of-the-art pre-training approaches.
引用
收藏
页码:133 / 145
页数:13
相关论文
共 50 条
  • [11] Model-Based Transfer Reinforcement Learning Based on Graphical Model Representations
    Sun, Yuewen
    Zhang, Kun
    Sun, Changyin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) : 1035 - 1048
  • [12] Length-Based Curriculum Learning for Efficient Pre-training of Language Models
    Nagatsuka, Koichi
    Broni-Bediako, Clifford
    Atsumi, Masayasu
    NEW GENERATION COMPUTING, 2023, 41 (01) : 109 - 134
  • [13] Length-Based Curriculum Learning for Efficient Pre-training of Language Models
    Koichi Nagatsuka
    Clifford Broni-Bediako
    Masayasu Atsumi
    New Generation Computing, 2023, 41 : 109 - 134
  • [14] Model-Based Reinforcement Learning with Adversarial Training for Online Recommendation
    Bai, Xueying
    Guan, Jian
    Wang, Hongning
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [15] PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via Relabeling Experience and Unsupervised Pre-training
    Lee, Kimin
    Smith, Laura
    Abbeel, Pieter
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [16] On the Effect of Pre-training for Transformer in Different Modality on Offline Reinforcement Learning
    Takagi, Shiro
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [17] Efficient hyperparameter optimization through model-based reinforcement learning
    Wu, Jia
    Chen, SenPeng
    Liu, XiYuan
    NEUROCOMPUTING, 2020, 409 : 381 - 393
  • [18] Reinforcement Learning with Action-Free Pre-Training from Videos
    Seo, Younggyo
    Lee, Kimin
    James, Stephen
    Abbeel, Pieter
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 19561 - 19579
  • [19] Knowledge Transfer using Model-Based Deep Reinforcement Learning
    Boloka, Tlou
    Makondo, Ndivhuwo
    Rosman, Benjamin
    2021 SOUTHERN AFRICAN UNIVERSITIES POWER ENGINEERING CONFERENCE/ROBOTICS AND MECHATRONICS/PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA (SAUPEC/ROBMECH/PRASA), 2021,
  • [20] APD: Learning Diverse Behaviors for Reinforcement Learning Through Unsupervised Active Pre-Training
    Zeng, Kailin
    Zhang, QiYuan
    Chen, Bin
    Liang, Bin
    Yang, Jun
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 12251 - 12258