Learning task-relevant representations via rewards and real actions for reinforcement learning

被引:0
|
作者
Yuan, Linghui [1 ]
Lu, Xiaowei [1 ]
Liu, Yunlong [1 ]
机构
[1] Xiamen Univ, Dept Automat, Xiamen, Peoples R China
关键词
Visual reinforcement learning; Task-relevant representations; Representation learning method; Conditional mutual information maximization;
D O I
10.1016/j.knosys.2024.111788
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The input of visual reinforcement learning often contains redundant information, which will reduce the decision efficiency and decrease the performance of the agent. To address this issue, task-relevant representations of the input are usually learned, where only task-related information is preserved for the decision making. While in the literature, auxiliary tasks that are constructed by using reward signals, optimal policy or by extracting controllable elements in the input are commonly adopted to learn the task-relevant representations, the methods based on reward signals do not work well in sparse reward environments, the effectiveness of methods using optimal policy relies heavily on the optimality degree of the given policy, and the methods by extracting the controllable elements will ignore the uncontrollable task-relevant information in the input. To alleviate these problems and to learn better task-relevant representations, in this paper, we first encourage the encoder to encode controllable parts in the input by maximizing the conditional mutual information between the representations and agent's real actions. And then as reward signals are directly related to the underlying tasks , they are used to make more task-related information encoded irrespective of whether such information is controllable or not. Finally, a temporal coherence constraint is incorporated into the whole framework to reduce the task-irrelevant information in the representations. Experiments on Distracting DeepMind Control Suite and autonomous driving simulator CARLA show that our proposed approach can achieve better performances than some state-of-the-art (SOTA) baselines, which demonstrates the method's effectiveness in enhancing the agent's decision efficiency and overall performances. Code is available at https://github.com/DMU-XMU/Learning-Task-relevant-Representations-via-Rewards-and-Real-Actions-forReinforcement-Learning.git.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Learning to Discover Task-Relevant Features for Interpretable Reinforcement Learning
    Zhang, Qiyuan
    Ma, Xiaoteng
    Yang, Yiqin
    Li, Chenghao
    Yang, Jun
    Liu, Yu
    Liang, Bin
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 6601 - 6607
  • [2] Learning Task-relevant Representations for Generalization via Characteristic Functions of Reward Sequence Distributions
    Yang, Rui
    Wang, Jie
    Geng, Zijie
    Ye, Mingxuan
    Ji, Shuiwang
    Li, Bin
    Wu, Feng
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 2242 - 2252
  • [3] Task-Relevant Chunking in Sequence Learning
    Perlman, Amotz
    Pothos, Emmanuel M.
    Edwards, Darren J.
    Tzelgov, Joseph
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 2010, 36 (03) : 649 - 661
  • [4] Learning explainable task-relevant state representation for model-free deep reinforcement learning
    Zhao, Tingting
    Li, Guixi
    Zhao, Tuo
    Chen, Yarui
    Xie, Ning
    Niu, Gang
    Sugiyama, Masashi
    NEURAL NETWORKS, 2024, 180
  • [5] Learning Multi-Task Transferable Rewards via Variational Inverse Reinforcement Learning
    Yoo, Se-Wook
    Seo, Seung-Woo
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022,
  • [6] Effective Robot Task Learning by Focusing on Task-relevant Objects
    Lee, Kyu Hwa
    Lee, Jinhan
    Thomaz, Andrea L.
    Bobick, Aaron F.
    2009 IEEE-RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2009, : 2551 - 2556
  • [7] Offline reinforcement learning with representations for actions
    Lou, Xingzhou
    Yin, Qiyue
    Zhang, Junge
    Yu, Chao
    He, Zhaofeng
    Cheng, Nengjie
    Huang, Kaiqi
    INFORMATION SCIENCES, 2022, 610 : 746 - 758
  • [8] Learning task-relevant features from robot data
    Vlassis, N
    Bunschoten, R
    Kröse, B
    2001 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS I-IV, PROCEEDINGS, 2001, : 499 - 504
  • [9] Robust Task Representations for Offline Meta-Reinforcement Learning via Contrastive Learning
    Yuan, Haoqi
    Lu, Zongqing
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [10] Passive exposure to task-relevant stimuli enhances categorization learning
    Schmid, Christian
    Haziq, Muhammad
    Baese-Berk, Melissa M.
    Murray, James M.
    Jaramillo, Santiago
    ELIFE, 2024, 12