Reinforcement Learning with Attention that Works: A Self-Supervised Approach

被引:28
|
作者
Manchin, Anthony [1 ]
Abbasnejad, Ehsan [1 ]
van den Hengel, Anton [1 ]
机构
[1] Univ Adelaide, Australian Inst Machine Learning, Adelaide, SA, Australia
关键词
Reinforcement learning; Attention; Deep learning;
D O I
10.1007/978-3-030-36802-9_25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Attention models have had a significant positive impact on deep learning across a range of tasks. However previous attempts at integrating attention with reinforcement learning have failed to produce significant improvements. Unlike the selective attention models used in previous attempts, which constrain the attention via preconceived notions of importance, our implementation utilises the Markovian properties inherent in the state input. We propose the first combination of self attention and reinforcement learning that is capable of producing significant improvements, including new state of the art results in the Arcade Learning Environment.
引用
收藏
页码:223 / 230
页数:8
相关论文
共 50 条
  • [1] Self-Supervised Attention-Aware Reinforcement Learning
    Wu, Haiping
    Khetarpa, Khimya
    Precup, Doina
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10311 - 10319
  • [2] Intrinsically Motivated Self-supervised Learning in Reinforcement Learning
    Zhao, Yue
    Du, Chenzhuang
    Zhao, Hang
    Li, Tiejun
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 3605 - 3615
  • [3] Self-Supervised Reinforcement Learning for Recommender Systems
    Xin, Xin
    Karatzoglou, Alexandros
    Arapakis, Ioannis
    Jose, Joemon M.
    [J]. PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 931 - 940
  • [4] Self-supervised Attention Learning for Robot Control
    Cong, Lin
    Shi, Yunlei
    Zhang, Jianwei
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 1153 - 1158
  • [5] Guiding Attention for Self-Supervised Learning with Transformers
    Deshpande, Ameet
    Narasimhan, Karthik
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4676 - 4686
  • [6] There Is No Turning Back: A Self-Supervised Approach for Reversibility-Aware Reinforcement Learning
    Grinsztajn, Nathan
    Ferret, Johan
    Pietquin, Olivier
    Preux, Philippe
    Geist, Matthieu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [7] Self-Supervised Discovering of Interpretable Features for Reinforcement Learning
    Shi, Wenjie
    Huang, Gao
    Song, Shiji
    Wang, Zhuoyuan
    Lin, Tingyu
    Wu, Cheng
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (05) : 2712 - 2724
  • [8] Self-Supervised Reinforcement Learning for Active Object Detection
    Fang, Fen
    Liang, Wenyu
    Wu, Yan
    Xu, Qianli
    Lim, Joo-Hwee
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 10224 - 10231
  • [9] Heuristic Attention Representation Learning for Self-Supervised Pretraining
    Van Nhiem Tran
    Liu, Shen-Hsuan
    Li, Yung-Hui
    Wang, Jia-Ching
    [J]. SENSORS, 2022, 22 (14)
  • [10] Graph Multihead Attention Pooling with Self-Supervised Learning
    Wang, Yu
    Hu, Liang
    Wu, Yang
    Gao, Wanfu
    [J]. ENTROPY, 2022, 24 (12)