Gated multi-attention representation in reinforcement learning

被引:8
|
作者
Liang, Dayang [1 ]
Chen, Qihang [1 ]
Liu, Yunlong [1 ]
机构
[1] Xiamen Univ, Dept Automat, Xiamen 361005, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Gated multi-attention module; Deep Q-learning network; Atari; 2600; games;
D O I
10.1016/j.knosys.2021.107535
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) has achieved great success in recent years by combining the feature extraction power of deep learning and the decision power of reinforcement learning techniques. In the literature, Convolutional Neural Networks (CNN) is usually used as the feature extraction method and recent studies have shown that the performances of the DRL algorithms can be greatly improved with the utilization of the attention mechanism, where the raw attentions are directly used for the decision-making. However, as is well-known, reinforcement learning is a trial-and-error process and it is almost impossible to learn an optimal policy in the beginning of the learning, especially in environments with sparse rewards, which in turn will cause the raw attention-based models can only remember and utilize the attention information indiscriminately for different areas and may focus on some task -irrelevant regions, but the focusing on such task-irrelevant regions is usually helpless and ineffective for the agent to find the optimal policy. To address this issue, we propose a gated multi-attention mechanism, which is then combined with the Deep Q-learning network (GMAQN). The gated multi -attention representation module (GMA) in GMAQN can effectively eliminate task-irrelevant attention information in the early phase of the trial-and-error process and improve the stability of the model. The proposed method has been demonstrated on the challenging domain of classic Atari 2600 games and experimental results show that compared with the baselines, our method can achieve better performance in terms of both the scores and the effect of focusing in the key regions. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Dynamic Graph Combinatorial Optimization with Multi-Attention Deep Reinforcement Learning
    Gunarathna, Udesh
    Borovica-Gajic, Renata
    Karunasekera, Shanika
    Tanin, Egemen
    30TH ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS, ACM SIGSPATIAL GIS 2022, 2022, : 153 - 164
  • [2] Multi-attention multiple instance learning
    Andrei V. Konstantinov
    Lev V. Utkin
    Neural Computing and Applications, 2022, 34 : 14029 - 14051
  • [3] Multi-attention multiple instance learning
    Konstantinov, Andrei, V
    Utkin, Lev, V
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (16): : 14029 - 14051
  • [4] Multi-attention Network for One Shot Learning
    Wang, Peng
    Liu, Lingqiao
    Shen, Chunhua
    Huang, Zi
    van den Hengel, Anton
    Shen, Heng Tao
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6212 - 6220
  • [5] Multi-Group Multi-Attention: Towards Discriminative Spatiotemporal Representation
    Shi, Zhensheng
    Cao, Liangjie
    Guan, Cheng
    Liang, Ju
    Li, Qianqian
    Gu, Zhaorui
    Zheng, Haiyong
    Zheng, Bing
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 2057 - 2066
  • [6] Multi-attention deep reinforcement learning and re-ranking for vehicle re-identification
    Liu, Yu
    Shen, Jianbing
    He, Haibo
    NEUROCOMPUTING, 2020, 414 : 27 - 35
  • [7] Question-Answering Aspect Classification with Multi-attention Representation
    Wu, Hanqian
    Liu, Mumu
    Wang, Jingjing
    Xie, Jue
    Li, Shoushan
    INFORMATION RETRIEVAL, CCIR 2018, 2018, 11168 : 78 - 89
  • [8] A novel multi-attention reinforcement learning for the scheduling of unmanned shipment vessels (USV) in automated container terminals
    Zhu, Jianxin
    Zhang, Weidan
    Yu, Lean
    Guo, Xinghai
    OMEGA-INTERNATIONAL JOURNAL OF MANAGEMENT SCIENCE, 2024, 129
  • [9] Gated Multi-Attention Feedback Network for Medical Image Super-Resolution
    Shang, Jianrun
    Zhang, Xue
    Zhang, Guisheng
    Song, Wenhao
    Chen, Jinyong
    Li, Qilei
    Gao, Mingliang
    ELECTRONICS, 2022, 11 (21)
  • [10] Event Extraction with Deep Contextualized Word Representation and Multi-attention Layer
    Ding, Ruixue
    Li, Zhoujun
    ADVANCED DATA MINING AND APPLICATIONS, ADMA 2018, 2018, 11323 : 189 - 201