Efficient Scheduling of Data Augmentation for Deep Reinforcement Learning

被引:0
|
作者
Ko, Byungchan [1 ]
Ok, Jungseul [2 ]
机构
[1] NALBI, Seoul, South Korea
[2] POSTECH, GSAI, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
GO;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In deep reinforcement learning (RL), data augmentation is widely considered as a tool to induce a set of useful priors about semantic consistency and to improve sample efficiency and generalization performance. However, even when the prior is useful for generalization, distilling it to RL agent often interferes with RL training and degenerates sample efficiency. Meanwhile, the agent is forgetful of the prior due to the non-stationary nature of RL. These observations suggest two extreme schedules of distillation: (i) over the entire training; or (ii) only at the end. Hence, we devise a stand-alone network distillation method to inject the consistency prior at any time (even after RL), and a simple yet efficient framework to automatically schedule the distillation. Specifically, the proposed framework first focuses on mastering train environments regardless of generalization by adaptively deciding which or no augmentation to be used for the training. After this, we add the distillation to extract the remaining benefits for generalization from all the augmentations, which requires no additional new samples. In our experiments, we demonstrate the utility of the proposed framework, in particular, that considers postponing the augmentation to the end of RL training. https://github.com/kbc6723/es-da
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Opportunistic maintenance scheduling with deep reinforcement learning
    Valet, Alexander
    Altenmueller, Thomas
    Waschneck, Bernd
    May, Marvin Carl
    Kuhnle, Andreas
    Lanza, Gisela
    [J]. JOURNAL OF MANUFACTURING SYSTEMS, 2022, 64 : 518 - 534
  • [22] Efficient deep learning based data augmentation techniques for enhanced learning on inadequate medical imaging data
    Sashank, Madipally Sai Krishna
    Maddila, Vijay Souri
    Boddu, Vikas
    Radhika, Y.
    [J]. ACTA IMEKO, 2022, 11 (01):
  • [23] Data-efficient Deep Reinforcement Learning for Vehicle Trajectory Control
    Frauenknecht, Bernd
    Ehlgen, Tobias
    Trimpe, Sebastian
    [J]. 2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 894 - 901
  • [24] Ensemble and Auxiliary Tasks for Data-Efficient Deep Reinforcement Learning
    Maulana, Muhammad Rizki
    Lee, Wee Sun
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 122 - 138
  • [25] Deep reinforcement learning for dynamic scheduling of energy-efficient automated guided vehicles
    Zhang, Lixiang
    Yan, Yan
    Hu, Yaoguang
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 35 (08) : 3875 - 3888
  • [26] A Data-Efficient Method of Deep Reinforcement Learning for Chinese Chess
    Xu, Changming
    Ding, Hengfeng
    Zhang, Xuejian
    Wang, Cong
    Yang, Hongji
    [J]. 2022 IEEE 22ND INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY, AND SECURITY COMPANION, QRS-C, 2022, : 687 - 693
  • [27] Towards Efficient Workflow Scheduling over Yarn Cluster using Deep Reinforcement Learning
    Xue, Jianguo
    Wang, Ting
    Cai, Puyu
    [J]. IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 473 - 478
  • [28] Efficient Practice for Deep Reinforcement Learning
    Kancharla, Venkata Sai Santosh Ravi Teja
    Lee, Minwoo
    [J]. 2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 78 - 85
  • [29] Scheduling the NASA Deep Space Network with Deep Reinforcement Learning
    Goh, Edwin
    Venkataram, Hamsa Shwetha
    Hoffmann, Mark
    Johnston, Mark D.
    Wilson, Brian
    [J]. 2021 IEEE AEROSPACE CONFERENCE (AEROCONF 2021), 2021,
  • [30] Text Data Augmentation for Deep Learning
    Shorten, Connor
    Khoshgoftaar, Taghi M.
    Furht, Borko
    [J]. JOURNAL OF BIG DATA, 2021, 8 (01)