Gated multi-attention representation in reinforcement learning

被引:8
|
作者
Liang, Dayang [1 ]
Chen, Qihang [1 ]
Liu, Yunlong [1 ]
机构
[1] Xiamen Univ, Dept Automat, Xiamen 361005, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Gated multi-attention module; Deep Q-learning network; Atari; 2600; games;
D O I
10.1016/j.knosys.2021.107535
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) has achieved great success in recent years by combining the feature extraction power of deep learning and the decision power of reinforcement learning techniques. In the literature, Convolutional Neural Networks (CNN) is usually used as the feature extraction method and recent studies have shown that the performances of the DRL algorithms can be greatly improved with the utilization of the attention mechanism, where the raw attentions are directly used for the decision-making. However, as is well-known, reinforcement learning is a trial-and-error process and it is almost impossible to learn an optimal policy in the beginning of the learning, especially in environments with sparse rewards, which in turn will cause the raw attention-based models can only remember and utilize the attention information indiscriminately for different areas and may focus on some task -irrelevant regions, but the focusing on such task-irrelevant regions is usually helpless and ineffective for the agent to find the optimal policy. To address this issue, we propose a gated multi-attention mechanism, which is then combined with the Deep Q-learning network (GMAQN). The gated multi -attention representation module (GMA) in GMAQN can effectively eliminate task-irrelevant attention information in the early phase of the trial-and-error process and improve the stability of the model. The proposed method has been demonstrated on the challenging domain of classic Atari 2600 games and experimental results show that compared with the baselines, our method can achieve better performance in terms of both the scores and the effect of focusing in the key regions. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] A Multi-attention Collaborative Deep Learning Approach for Blood Pressure Prediction
    He, Luo
    Liu, Hongyan
    Yang, Yinghui
    Wang, Bei
    ACM TRANSACTIONS ON MANAGEMENT INFORMATION SYSTEMS, 2022, 13 (02)
  • [22] Multi-attention representation network partial domain adaptation for COVID-19 diagnosis
    He, Chunmei
    Zheng, Lanqing
    Tan, Taifeng
    Fan, Xianjun
    Ye, Zhengchun
    APPLIED SOFT COMPUTING, 2022, 125
  • [23] Multi-Attention Network for Sentiment Analysis
    Du, Tingting
    Huang, Yunyin
    Wu, Xian
    Chang, Huiyou
    PROCEEDINGS OF THE 2018 2ND INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL (NLPIR 2018), 2018, : 49 - 54
  • [24] Multi-Depth Learning with Multi-Attention for fine-grained image classification
    Dai, Zuhua
    Li, Hongyi
    Li, Kelong
    Zhou, Anwei
    2020 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND HUMAN-COMPUTER INTERACTION (ICHCI 2020), 2020, : 206 - 212
  • [25] Multi-Layer Attention-based State Representation for the Reinforcement Learning of Visual Servoing
    Kitajima, Hiromu
    Bounyong, Souksakhone
    Yoshioka, Mototaka
    2023 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS, ICCE, 2023,
  • [26] A Shared Multi-Attention Framework for Multi-Label Zero-Shot Learning
    Huynh, Dat
    Elhamifar, Ehsan
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 8773 - 8783
  • [27] Multi-Attention Network for Stereo Matching
    Yang, Xiaowei
    He, Lin
    Zhao, Yong
    Sang, Haiwei
    Yang, Zuliu
    Cheng, Xianjing
    IEEE ACCESS, 2020, 8 : 113371 - 113382
  • [28] Interpretable MOOC recommendation: a multi-attention network for personalized learning behavior analysis
    Fan, Ju
    Jiang, Yuanchun
    Liu, Yezheng
    Zhou, Yonghang
    INTERNET RESEARCH, 2022, 32 (02) : 588 - 605
  • [29] Multi-attention mutual information distributed framework for few-shot learning
    Wang, Zhe
    Ma, Pingchuan
    Chi, Ziqiu
    Li, Dongdong
    Yang, Hai
    Du, Wenli
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [30] Multi-Object Tracking Via Multi-Attention
    Wang, Xianrui
    Ling, Hefei
    Chen, Jiazhong
    Li, Ping
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,