Gated multi-attention representation in reinforcement learning

被引:8
|
作者
Liang, Dayang [1 ]
Chen, Qihang [1 ]
Liu, Yunlong [1 ]
机构
[1] Xiamen Univ, Dept Automat, Xiamen 361005, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Gated multi-attention module; Deep Q-learning network; Atari; 2600; games;
D O I
10.1016/j.knosys.2021.107535
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) has achieved great success in recent years by combining the feature extraction power of deep learning and the decision power of reinforcement learning techniques. In the literature, Convolutional Neural Networks (CNN) is usually used as the feature extraction method and recent studies have shown that the performances of the DRL algorithms can be greatly improved with the utilization of the attention mechanism, where the raw attentions are directly used for the decision-making. However, as is well-known, reinforcement learning is a trial-and-error process and it is almost impossible to learn an optimal policy in the beginning of the learning, especially in environments with sparse rewards, which in turn will cause the raw attention-based models can only remember and utilize the attention information indiscriminately for different areas and may focus on some task -irrelevant regions, but the focusing on such task-irrelevant regions is usually helpless and ineffective for the agent to find the optimal policy. To address this issue, we propose a gated multi-attention mechanism, which is then combined with the Deep Q-learning network (GMAQN). The gated multi -attention representation module (GMA) in GMAQN can effectively eliminate task-irrelevant attention information in the early phase of the trial-and-error process and improve the stability of the model. The proposed method has been demonstrated on the challenging domain of classic Atari 2600 games and experimental results show that compared with the baselines, our method can achieve better performance in terms of both the scores and the effect of focusing in the key regions. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Multi-attention Fusion for Multimodal Sentiment Classification
    Li, Guangmin
    Zeng, Xin
    Chen, Chi
    Zhou, Long
    PROCEEDINGS OF 2024 ACM ICMR WORKSHOP ON MULTIMODAL VIDEO RETRIEVAL, ICMR-MVR 2024, 2024, : 1 - 7
  • [42] Deliberate Multi-Attention Network for Image Captioning
    Dan, Zedong
    Fang, Yanmei
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, PRCV 2022, 2022, 13534 : 475 - 487
  • [43] Spatial multi-attention conditional neural processes
    Bao, Li-Li
    Zhang, Jiang-She
    Zhang, Chun-Xia
    NEURAL NETWORKS, 2024, 173
  • [44] Multi-Attention Network for Aspect Sentiment Analysis
    Han, Huiyu
    Li, Xiaoge
    Zhi, Shuting
    Wang, Haoyue
    2019 8TH INTERNATIONAL CONFERENCE ON SOFTWARE AND COMPUTER APPLICATIONS (ICSCA 2019), 2019, : 22 - 26
  • [45] Attention-Gated Reinforcement Learning in Neural Networks-A Unified View
    Brosch, Tobias
    Schwenker, Friedhelm
    Neumann, Heiko
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2013, 2013, 8131 : 272 - 279
  • [46] Learning Multi-Attention Context Graph for Group-Based Re-Identification
    Yan, Yichao
    Qin, Jie
    Ni, Bingbing
    Chen, Jiaxin
    Liu, Li
    Zhu, Fan
    Zheng, Wei-Shi
    Yang, Xiaokang
    Shao, Ling
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) : 7001 - 7018
  • [47] Hierarchical multi-attention networks for document classification
    Yingren Huang
    Jiaojiao Chen
    Shaomin Zheng
    Yun Xue
    Xiaohui Hu
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 1639 - 1647
  • [48] Recurrent Networks for Guided Multi-Attention Classification
    Dai, Xin
    Kong, Xiangnan
    Guo, Tian
    Lee, John Boaz
    Liu, Xinyue
    Moore, Constance
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 412 - 420
  • [49] Learning From Demonstrations Via Multi-Level and Multi-Attention Domain-Adaptive Meta-Learning
    Hu, Ziye
    Gan, Zhongxue
    Li, Wei
    Guo, Weikun
    Gao, Xiang
    Zhu, Jiwei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 11910 - 11917
  • [50] Multi-Timescale Memory Dynamics Extend Task Repertoire in a Reinforcement Learning Network With Attention-Gated Memory
    Martinolli, Marco
    Gerstner, Wulfram
    Gilra, Aditya
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2018, 12