Efficient Scheduling of Data Augmentation for Deep Reinforcement Learning

被引:0
|
作者
Ko, Byungchan [1 ]
Ok, Jungseul [2 ]
机构
[1] NALBI, Seoul, South Korea
[2] POSTECH, GSAI, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
GO;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In deep reinforcement learning (RL), data augmentation is widely considered as a tool to induce a set of useful priors about semantic consistency and to improve sample efficiency and generalization performance. However, even when the prior is useful for generalization, distilling it to RL agent often interferes with RL training and degenerates sample efficiency. Meanwhile, the agent is forgetful of the prior due to the non-stationary nature of RL. These observations suggest two extreme schedules of distillation: (i) over the entire training; or (ii) only at the end. Hence, we devise a stand-alone network distillation method to inject the consistency prior at any time (even after RL), and a simple yet efficient framework to automatically schedule the distillation. Specifically, the proposed framework first focuses on mastering train environments regardless of generalization by adaptively deciding which or no augmentation to be used for the training. After this, we add the distillation to extract the remaining benefits for generalization from all the augmentations, which requires no additional new samples. In our experiments, we demonstrate the utility of the proposed framework, in particular, that considers postponing the augmentation to the end of RL training. https://github.com/kbc6723/es-da
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Data Centers Job Scheduling with Deep Reinforcement Learning
    Liang, Sisheng
    Yang, Zhou
    Jin, Fang
    Chen, Yong
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2020, PT II, 2020, 12085 : 906 - 917
  • [2] Synthetic Data Augmentation for Deep Reinforcement Learning in Financial Trading
    Liu, Chunli
    Ventre, Carmine
    Polukarov, Maria
    [J]. 3RD ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2022, 2022, : 343 - 351
  • [3] Boosting Deep Reinforcement Learning Agents with Generative Data Augmentation
    Papagiannis, Tasos
    Alexandridis, Georgios
    Stafylopatis, Andreas
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (01):
  • [4] Iot Data Processing and Scheduling Based on Deep Reinforcement Learning
    Jiang, Yuchuan
    Wang, Zhangjun
    Jin, Zhixiong
    [J]. INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2023, 18 (06)
  • [5] Automatic Data Augmentation by Upper Confidence Bounds for Deep Reinforcement Learning
    Gil, Yoonhee
    Baek, Jongchan
    Park, Jonghyuk
    Han, Soohee
    [J]. 2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 1199 - 1203
  • [6] Invariant Transform Experience Replay: Data Augmentation for Deep Reinforcement Learning
    Lin, Yijiong
    Huang, Jiancong
    Zimmer, Matthieu
    Guan, Yisheng
    Rojas, Juan
    Weng, Paul
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 6615 - 6622
  • [7] Energy-efficient VM scheduling based on deep reinforcement learning
    Wang, Bin
    Liu, Fagui
    Lin, Weiwei
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2021, 125 : 616 - 628
  • [8] PROLIFIC: Deep Reinforcement Learning for Efficient EV Fleet Scheduling and Charging
    Ma, Junchi
    Zhang, Yuan
    Duan, Zongtao
    Tang, Lei
    [J]. SUSTAINABILITY, 2023, 15 (18)
  • [9] A Deep Reinforcement Learning Approach to the Optimization of Data Center Task Scheduling
    Che, Haiying
    Bai, Zixing
    Zuo, Rong
    Li, Honglei
    [J]. COMPLEXITY, 2020, 2020
  • [10] A Data-Efficient Training Method for Deep Reinforcement Learning
    Feng, Wenhui
    Han, Chongzhao
    Lian, Feng
    Liu, Xia
    [J]. ELECTRONICS, 2022, 11 (24)