Reusing Source Task Knowledge via Transfer Approximator in Reinforcement Transfer Learning

被引:2
|
作者
Cheng, Qiao [1 ]
Wang, Xiangke [1 ]
Niu, Yifeng [1 ]
Shen, Lincheng [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Hunan, Peoples R China
来源
SYMMETRY-BASEL | 2019年 / 11卷 / 01期
基金
中国国家自然科学基金;
关键词
artificial neural networks; probabilistic policy reuse; reinforcement learning; transfer approximator; transfer learning;
D O I
10.3390/sym11010025
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Transfer Learning (TL) has received a great deal of attention because of its ability to speed up Reinforcement Learning (RL) by reusing learned knowledge from other tasks. This paper proposes a new transfer learning framework, referred to as Transfer Learning via Artificial Neural Network Approximator (TL-ANNA). It builds an Artificial Neural Network (ANN) transfer approximator to transfer the related knowledge from the source task into the target task and reuses the transferred knowledge with a Probabilistic Policy Reuse (PPR) scheme. Specifically, the transfer approximator maps the state of the target task symmetrically to states of the source task with a certain mapping rule, and activates the related knowledge (components of the action-value function) of the source task as the input of the ANNs; it then predicts the quality of the actions in the target task with the ANNs. The target learner uses the PPR scheme to bias the RL with the suggested action from the transfer approximator. In this way, the transfer approximator builds a symmetric knowledge path between the target task and the source task. In addition, two mapping rules for the transfer approximator are designed, namely, Full Mapping Rule and Group Mapping Rule. Experiments performed on the RoboCup soccer Keepaway task verified that the proposed transfer learning methods outperform two other transfer learning methods in both jumpstart and time to threshold metrics and are more robust to the quality of source knowledge. In addition, the TL-ANNA with the group mapping rule exhibits slightly worse performance than the one with the full mapping rule, but with less computation and space cost when appropriate grouping method is used.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Transfer Learning by Reusing Structured Knowledge
    Yang, Qiang
    Zheng, Vincent W.
    Li, Bin
    Zhuo, Hankz Hankui
    [J]. AI MAGAZINE, 2011, 32 (02) : 95 - 106
  • [2] Efficient deep reinforcement learning under task variations via knowledge transfer for drone control
    Jang, Sooyoung
    Kim, Hyung-Il
    [J]. ICT EXPRESS, 2024, 10 (03): : 576 - 582
  • [3] Options in Multi-task Reinforcement Learning - Transfer via Reflection
    Denis, Nicholas
    Fraser, Maia
    [J]. ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11489 : 225 - 237
  • [4] Knowledge Transfer in Multi-Task Deep Reinforcement Learning for Continuous Control
    Xu, Zhiyuan
    Wu, Kun
    Che, Zhengping
    Tang, Jian
    Ye, Jieping
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [5] Using advice to transfer knowledge acquired in one reinforcement learning task to another
    Torrey, L
    Walker, T
    Shavlik, J
    Maclin, R
    [J]. MACHINE LEARNING: ECML 2005, PROCEEDINGS, 2005, 3720 : 412 - 424
  • [6] Task similarity measures for transfer in reinforcement learning task libraries
    Carroll, JL
    Seppi, K
    [J]. Proceedings of the International Joint Conference on Neural Networks (IJCNN), Vols 1-5, 2005, : 803 - 808
  • [7] Deep Multi-Task Multi-Agent Reinforcement Learning With Knowledge Transfer
    Mai Y.
    Zang Y.
    Yin Q.
    Ni W.
    Huang K.
    [J]. IEEE Transactions on Games, 2024, 16 (03) : 1 - 11
  • [8] Towards Knowledge Transfer in Deep Reinforcement Learning
    Glatt, Ruben
    da Silva, Felipe Leno
    Reali Costa, Anna Helena
    [J]. PROCEEDINGS OF 2016 5TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2016), 2016, : 91 - 96
  • [9] Improving Deep Reinforcement Learning with Knowledge Transfer
    Glatt, Ruben
    Reali Costa, Anna Helena
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 5036 - 5037
  • [10] REPAINT: Knowledge Transfer in Deep Reinforcement Learning
    Tao, Yunzhe
    Genc, Sahika
    Chung, Jonathan
    Sun, Tao
    Mallya, Sunil
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7145 - 7155