Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay

被引:0
|
作者
Yin, Haiyan [1 ]
Pan, Sinno Jialin [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The process for transferring knowledge of multiple reinforcement learning policies into a single multi-task policy via distillation technique is known as policy distillation. When policy distillation is under a deep reinforcement learning setting, due to the giant parameter size and the huge state space for each task domain, it requires extensive computational efforts to train the multi-task policy network. In this paper, we propose a new policy distillation architecture for deep reinforcement learning, where we assume that each task uses its taskspecific high-level convolutional features as the inputs to the multi-task policy network. Furthermore, we propose a new sampling framework termed hierarchical prioritized experience replay to selectively choose experiences from the replay memories of each task domain to perform learning on the network. With the above two attempts, we aim to accelerate the learning of the multi-task policy network while guaranteeing a good performance. We use Atari 2600 games as testing environment to demonstrate the efficiency and effectiveness of our proposed solution for policy distillation.
引用
下载
收藏
页码:1640 / 1646
页数:7
相关论文
共 50 条
  • [41] Multimodal fusion for autonomous navigation via deep reinforcement learning with sparse rewards and hindsight experience replay
    Xiao, Wendong
    Yuan, Liang
    Ran, Teng
    He, Li
    Zhang, Jianbo
    Cui, Jianping
    DISPLAYS, 2023, 78
  • [42] Multi-Input Autonomous Driving Based on Deep Reinforcement Learning With Double Bias Experience Replay
    Cui, Jianping
    Yuan, Liang
    He, Li
    Xiao, Wendong
    Ran, Teng
    Zhang, Jianbo
    IEEE SENSORS JOURNAL, 2023, 23 (11) : 11253 - 11261
  • [43] Unveiling the Effects of Experience Replay on Deep Reinforcement Learning-based Power Allocation in Wireless Networks
    Kopic, Amna
    Perenda, Erma
    Gacanin, Haris
    2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,
  • [44] Knowledge Reasoning Method Based on Deep Transfer Reinforcement Learning: DTRLpath
    Lin, Shiming
    Ye, Ling
    Zhuang, Yijie
    Lu, Lingyun
    Zheng, Shaoqiu
    Huang, Chenxi
    Kwee, Ng Yin
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 80 (01): : 299 - 317
  • [45] Knowledge Transfer using Model-Based Deep Reinforcement Learning
    Boloka, Tlou
    Makondo, Ndivhuwo
    Rosman, Benjamin
    2021 SOUTHERN AFRICAN UNIVERSITIES POWER ENGINEERING CONFERENCE/ROBOTICS AND MECHATRONICS/PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA (SAUPEC/ROBMECH/PRASA), 2021,
  • [46] Deep Reinforcement Learning with Knowledge Transfer for Online Rides Order Dispatching
    Wang, Zhaodong
    Qin, Zhiwei
    Tang, Xiaocheng
    Ye, Jieping
    Zhu, Hongtu
    2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, : 617 - 626
  • [47] In situ compression artifact removal in scientific data using deep transfer learning and experience replay
    Madireddy, Sandeep
    Hwan Park, Ji
    Lee, Sunwoo
    Balaprakash, Prasanna
    Yoo, Shinjae
    Liao, Wei-keng
    Hauck, Cory D.
    Paul Laiu, M.
    Archibald, Richard
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (02):
  • [48] A hierarchical deep reinforcement learning model with expert prior knowledge for intelligent penetration testing
    Li, Qianyu
    Zhang, Min
    Shen, Yi
    Wang, Ruipeng
    Hu, Miao
    Li, Yang
    Hao, Hao
    COMPUTERS & SECURITY, 2023, 132
  • [49] Balanced prioritized experience replay in off-policy reinforcement learning
    Lou Z.
    Wang Y.
    Shan S.
    Zhang K.
    Wei H.
    Neural Computing and Applications, 2024, 36 (25) : 15721 - 15737
  • [50] Enhanced Off-Policy Reinforcement Learning With Focused Experience Replay
    Kong, Seung-Hyun
    Nahrendra, I. Made Aswin
    Paek, Dong-Hee
    IEEE ACCESS, 2021, 9 (09): : 93152 - 93164