Efficient Scheduling of Data Augmentation for Deep Reinforcement Learning

被引:0
|
作者
Ko, Byungchan [1 ]
Ok, Jungseul [2 ]
机构
[1] NALBI, Seoul, South Korea
[2] POSTECH, GSAI, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
GO;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In deep reinforcement learning (RL), data augmentation is widely considered as a tool to induce a set of useful priors about semantic consistency and to improve sample efficiency and generalization performance. However, even when the prior is useful for generalization, distilling it to RL agent often interferes with RL training and degenerates sample efficiency. Meanwhile, the agent is forgetful of the prior due to the non-stationary nature of RL. These observations suggest two extreme schedules of distillation: (i) over the entire training; or (ii) only at the end. Hence, we devise a stand-alone network distillation method to inject the consistency prior at any time (even after RL), and a simple yet efficient framework to automatically schedule the distillation. Specifically, the proposed framework first focuses on mastering train environments regardless of generalization by adaptively deciding which or no augmentation to be used for the training. After this, we add the distillation to extract the remaining benefits for generalization from all the augmentations, which requires no additional new samples. In our experiments, we demonstrate the utility of the proposed framework, in particular, that considers postponing the augmentation to the end of RL training. https://github.com/kbc6723/es-da
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Data Augmentation for Bayesian Deep Learning
    Wang, Yuexi
    Polson, Nicholas
    Sokolov, Vadim O.
    [J]. BAYESIAN ANALYSIS, 2023, 18 (04): : 1041 - 1069
  • [32] Data Efficient Deep Reinforcement Learning With Action-Ranked Temporal Difference Learning
    Liu, Qi
    Li, Yanjie
    Liu, Yuecheng
    Lin, Ke
    Gao, Jianqi
    Lou, Yunjiang
    [J]. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (04): : 2949 - 2961
  • [33] Text Data Augmentation for Deep Learning
    Shorten, Connor
    Khoshgoftaar, Taghi M.
    Furht, Borko
    [J]. JOURNAL OF BIG DATA, 2021, 8 (01)
  • [34] Cloud Resource Scheduling With Deep Reinforcement Learning and Imitation Learning
    Guo, Wenxia
    Tian, Wenhong
    Ye, Yufei
    Xu, Lingxiao
    Wu, Kui
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (05): : 3576 - 3586
  • [35] Data-Driven Online Energy Scheduling of a Microgrid Based on Deep Reinforcement Learning
    Ji, Ying
    Wang, Jianhui
    Xu, Jiacan
    Li, Donglin
    [J]. ENERGIES, 2021, 14 (08)
  • [36] DeepJS']JS: Job Scheduling Based on Deep Reinforcement Learning in Cloud Data Center
    Li, Fengcun
    Hu, Bo
    [J]. ICBDC 2019: PROCEEDINGS OF 2019 4TH INTERNATIONAL CONFERENCE ON BIG DATA AND COMPUTING, 2019, : 48 - 53
  • [37] Deep Reinforcement Learning with Online Data Augmentation to Improve Sample Efficiency for Intelligent HVAC Control
    Kurte, Kuldeep
    Amasyali, Kadir
    Munk, Jeffrey
    Zandi, Helia
    [J]. PROCEEDINGS OF THE 2022 THE 9TH ACM INTERNATIONAL CONFERENCE ON SYSTEMS FOR ENERGY-EFFICIENT BUILDINGS, CITIES, AND TRANSPORTATION, BUILDSYS 2022, 2022, : 479 - 483
  • [38] Energy efficient task scheduling based on deep reinforcement learning in cloud environment: A specialized review
    Hou, Huanhuan
    Jawaddi, Siti Nuraishah Agos
    Ismail, Azlan
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 151 : 214 - 231
  • [39] A deep reinforcement learning based hybrid algorithm for efficient resource scheduling in edge computing environment
    Xue, Fei
    Hai, Qiuru
    Dong, Tingting
    Cui, Zhihua
    Gong, Yuelu
    [J]. INFORMATION SCIENCES, 2022, 608 : 362 - 374
  • [40] Dynamic VNF Scheduling: A Deep Reinforcement Learning Approach
    Zhang, Zixiao
    He, Fujun
    Oki, Eiji
    [J]. IEICE TRANSACTIONS ON COMMUNICATIONS, 2023, E106B (07) : 557 - 570