Faster Deep Q-learning using Neural Episodic Control

被引:4
|
作者
Nishio, Daichi [1 ]
Yamane, Satoshi [1 ]
机构
[1] Kanazawa Univ, Inst Sci & Engn, Kanazawa, Ishikawa, Japan
关键词
Deep reinforcement learning; DQN; Neural Episodic Control; Sample efficiency;
D O I
10.1109/COMPSAC.2018.00075
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The research on deep reinforcement learning which estimates Q-value by deep learning has been attracted the interest of researchers recently. In deep reinforcement learning, it is important to efficiently learn the experiences that an agent has collected by exploring environment. We propose NEC2DQN that improves learning speed of a poor sample efficiency algorithm such as DQN by using good one such as NEC at the beginning of learning. We show it is able to learn faster than Double DQN or N-step DQN in the experiments of Pong.
引用
收藏
页码:486 / 491
页数:6
相关论文
共 50 条
  • [21] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [22] Comparison of Deep Q-Learning, Q-Learning and SARSA Reinforced Learning for Robot Local Navigation
    Anas, Hafiq
    Ong, Wee Hong
    Malik, Owais Ahmed
    [J]. ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 443 - 454
  • [23] Neural Q-learning for solving PDEs
    Cohen, Samuel N.
    Jiang, Deqing
    Sirignano, Justin
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [24] Deep Q-Learning from Demonstrations
    Hester, Todd
    Vecerik, Matej
    Pietquin, Olivier
    Lanctot, Marc
    Schaul, Tom
    Piot, Bilal
    Horgan, Dan
    Quan, John
    Sendonaris, Andrew
    Osband, Ian
    Dulac-Arnold, Gabriel
    Agapiou, John
    Leibo, Joel Z.
    Gruslys, Audrunas
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3223 - 3230
  • [25] Hierarchical clustering with deep Q-learning
    Forster, Richard
    Fulop, Agnes
    [J]. ACTA UNIVERSITATIS SAPIENTIAE INFORMATICA, 2018, 10 (01) : 86 - 109
  • [26] Active deep Q-learning with demonstration
    Chen, Si-An
    Tangkaratt, Voot
    Lin, Hsuan-Tien
    Sugiyama, Masashi
    [J]. MACHINE LEARNING, 2020, 109 (9-10) : 1699 - 1725
  • [27] Active deep Q-learning with demonstration
    Si-An Chen
    Voot Tangkaratt
    Hsuan-Tien Lin
    Masashi Sugiyama
    [J]. Machine Learning, 2020, 109 : 1699 - 1725
  • [28] A Theoretical Analysis of Deep Q-Learning
    Fan, Jianqing
    Wang, Zhaoran
    Xie, Yuchen
    Yang, Zhuoran
    [J]. LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 486 - 489
  • [29] ADAPTIVE CONTENTION WINDOW DESIGN USING DEEP Q-LEARNING
    Kumar, Abhishek
    Verma, Gunjan
    Rao, Chirag
    Swami, Ananthram
    Segarra, Santiago
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4950 - 4954
  • [30] Deep Q-Learning with Prioritized Sampling
    Zhai, Jianwei
    Liu, Quan
    Zhang, Zongzhang
    Zhong, Shan
    Zhu, Haijun
    Zhang, Peng
    Sun, Cijia
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 13 - 22