Dynamic fusion for ensemble of deep Q-network

被引:0
|
作者
Patrick P. K. Chan
Meng Xiao
Xinran Qin
Natasha Kees
机构
[1] South China University of Technology,School of Computer Science and Engineering
关键词
Ensemble; Deep reinforcement learning; Dynamic fusion; Deep Q-network;
D O I
暂无
中图分类号
学科分类号
摘要
Ensemble reinforcement learning, which combines the decisions of a set of base agents, is proposed to enhance the decision making process and speed up training time. Many studies indicate that an ensemble model may achieve better results than a single agent because of the complement of base agents, in which the error of an agent may be corrected by others. However, the fusion method is a fundamental issue in ensemble. Currently, existing studies mainly focus on static fusion which either assumes all agents have the same ability or ignores the ones with poor average performance. This assumption causes current static fusion methods to overlook base agents with poor overall performance, but excellent results in select scenarios, which results in the ability of some agents not being fully utilized. This study aims to propose a dynamic fusion method which utilizes each base agent according to its local competence on test states. The performance of a base agent on the validation set is measured in terms of the rewards achieved by the agent in next n steps. The similarity between a validation state and a new state is quantified by Euclidian distance in the latent space and the weights of each base agent are updated according to its performance on validation states and their similarity to a new state. The experimental studies confirm that the proposed dynamic fusion method outperforms its base agents and also the static fusion methods. This is the first dynamic fusion method proposed for deep reinforcement learning, which extends the study on dynamic fusion from classification to reinforcement learning.
引用
收藏
页码:1031 / 1040
页数:9
相关论文
共 50 条
  • [31] Influence on Learning of Various Conditions in Deep Q-Network
    Niitsuma, Jun
    Osana, Yuko
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 1932 - 1935
  • [32] A Framework of Hierarchical Deep Q-Network for Portfolio Management
    Gao, Yuan
    Gao, Ziming
    Hu, Yi
    Song, Sifan
    Jiang, Zhengyong
    Su, Jionglong
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2021, : 132 - 140
  • [33] An Improved Deep Q-Network with Convolution Block Attention
    Li, Shilin
    Qu, Junsuo
    Yang, Dan
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 2921 - 2929
  • [34] Manufacturing Resource Scheduling Based on Deep Q-Network
    ZHANG Yufei
    ZOU Yuanhao
    ZHAO Xiaodong
    Wuhan University Journal of Natural Sciences, 2022, 27 (06) : 531 - 538
  • [35] Inhomogeneous deep Q-network for time sensitive applications
    Chen, Xu
    Wang, Jun
    ARTIFICIAL INTELLIGENCE, 2022, 312
  • [36] Multiagent Learning and Coordination with Clustered Deep Q-Network
    Pageaud, Simon
    Deslandres, Veronique
    Lehoux, Vassilissa
    Hassas, Salima
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2156 - 2158
  • [37] Deep Reinforcement Learning. Case Study: Deep Q-Network
    Vrejoiu, Mihnea Horia
    ROMANIAN JOURNAL OF INFORMATION TECHNOLOGY AND AUTOMATIC CONTROL-REVISTA ROMANA DE INFORMATICA SI AUTOMATICA, 2019, 29 (03): : 65 - 78
  • [38] Deep Reinforcement Learning Pairs Trading with a Double Deep Q-Network
    Brim, Andrew
    2020 10TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2020, : 222 - 227
  • [39] A Dueling Deep Recurrent Q-Network Framework for Dynamic Multichannel Access in Heterogeneous Wireless Networks
    Chen, Haitao
    Zhao, Haitao
    Zhou, Li
    Zhang, Jiao
    Liu, Yan
    Pan, Xiaoqian
    Liu, Xingguang
    Wei, Jibo
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [40] Learning the Dynamic Treatment Regimes from Medical Registry Data through Deep Q-network
    Ning Liu
    Ying Liu
    Brent Logan
    Zhiyuan Xu
    Jian Tang
    Yanzhi Wang
    Scientific Reports, 9