Transfer reinforcement learning via meta-knowledge extraction using auto-pruned decision trees

被引:8
|
作者
Lan, Yixing [1 ]
Xu, Xin [1 ]
Fang, Qiang [1 ]
Zeng, Yujun [1 ]
Liu, Xinwang [2 ]
Zhang, Xianjian [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha, Peoples R China
[2] Natl Univ Def Technol, Coll Comp, Changsha, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Transfer reinforcement learning; Meta-knowledge; Explainable artificial intelligence; POLICIES;
D O I
10.1016/j.knosys.2022.108221
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer reinforcement learning (RL) has recently received increasing attention to make RL agents have better learning performance in target Markov decision problems (MDPs) by using the knowledge learned in source MDPs. However, it is still an open and challenging problem to improve the transfer capability and interpretability of RL algorithms. In this paper, we propose a novel transfer reinforcement learning approach via meta-knowledge extraction using auto-pruned decision trees. In source MDPs, pre-trained policies are firstly learned via RL algorithms using general function approximators. Then, a meta-knowledge extraction algorithm is designed with an auto-pruned decision tree model, where the meta-knowledge is learned by re-training the auto-pruned decision tree based on the data samples generated from the pre-trained policies. The state spaces of meta-knowledge are determined by estimating the uncertainty of state-action pairs in pre-trained policies based on the entropy value of leaf nodes. In target MDPs, according to whether the state is in the state set of meta-knowledge, a hybrid policy is generated by integrating the meta-knowledge and the policies learned on the target MDPs. Based on the proposed transfer RL approach, two meta-knowledge-based transfer reinforcement learning (MKRL) algorithms are developed for MDPs with discrete action spaces and continuous action spaces, respectively. Experimental results in several benchmark tasks show that the MKRL algorithm outperforms other baselines in terms of learning efficiency and interpretability in the target MDPs with generic cases of task similarity. (C)& nbsp;2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Argumentation based reinforcement learning for meta-knowledge extraction
    Xu, Junyi
    Yao, Li
    Li, Le
    Ji, Ming
    Tang, Guoming
    INFORMATION SCIENCES, 2020, 506 : 258 - 272
  • [2] Control of hypothesis space using meta-knowledge in inductive learning
    Inuzuka, Nobuhiro
    Ishida, Hiroyuki
    Nakano, Tomofumi
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2008, 5178 : 911 - 918
  • [3] Learning by Switching Generation and Reasoning Methods-Acquisition of Meta-knowledge for Switching with Reinforcement Learning
    Tomaru, Masahiro
    Umano, Motohide
    Matsumoto, Yuji
    Seta, Kazuhisa
    2008 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-5, 2008, : 1930 - +
  • [4] Knowledge transfer as transformative dialogue: a pedagogical view on learning and meta-knowledge transfer in a leadership development program
    Ohlsson, Jon
    JOURNAL OF ORGANIZATIONAL CHANGE MANAGEMENT, 2023, 36 (08) : 117 - 128
  • [5] Cross-Modal Meta-Knowledge Transfer: A Meta-Learning Framework Adaptable for Multimodal Tasks
    Chen, Yuhe
    Jin, Jingxuan
    Li, De
    Wang, Peng
    PROCEEDINGS OF 2024 INTERNATIONAL CONFERENCE ON COMPUTER AND MULTIMEDIA TECHNOLOGY, ICCMT 2024, 2024, : 558 - 563
  • [6] Reusing Source Task Knowledge via Transfer Approximator in Reinforcement Transfer Learning
    Cheng, Qiao
    Wang, Xiangke
    Niu, Yifeng
    Shen, Lincheng
    SYMMETRY-BASEL, 2019, 11 (01):
  • [7] Integrating scientific knowledge into machine learning using interactive decision trees
    Sarailidis, Georgios
    Wagener, Thorsten
    Pianosi, Francesca
    COMPUTERS & GEOSCIENCES, 2023, 170
  • [8] Task Inference for Offline Meta Reinforcement Learning via Latent Shared Knowledge
    Zhou, Ying
    Cong, Shan
    Yu, Chao
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2023, 2023, 14120 : 356 - 365
  • [9] Speech Enhancement Using Dynamic Learning in Knowledge Distillation via Reinforcement Learning
    Chu, Shih-Chuan
    Wu, Chung-Hsien
    Su, Tsai-Wei
    IEEE ACCESS, 2023, 11 : 144421 - 144434
  • [10] Knowledge Transfer using Model-Based Deep Reinforcement Learning
    Boloka, Tlou
    Makondo, Ndivhuwo
    Rosman, Benjamin
    2021 SOUTHERN AFRICAN UNIVERSITIES POWER ENGINEERING CONFERENCE/ROBOTICS AND MECHATRONICS/PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA (SAUPEC/ROBMECH/PRASA), 2021,