Multi-step Prediction for Learning Invariant Representations in Reinforcement Learning

被引:0
|
作者
Xu, Xinyue [1 ,2 ]
Lv, Kai [1 ,2 ]
Dong, Xingye [1 ,2 ]
Han, Sheng [1 ,2 ]
Lin, Youfang [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] CAAC, Key Lab Intelligent Passenger Serv Civil Aviat, Beijing, Peoples R China
关键词
multi-step prediction; bisimulation metrics; representation learning; reinforcement learning;
D O I
10.1109/HPBDIS53214.2021.9658436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on how to achieve task-relevant feature representations in reinforcement learning from image observations without relying either on domain knowledge or pixel-reconstruction. Although the existing algorithms based on reconstruction and contrastive learning have achieved excellent success, the sample efficiency and robustness of the algorithm are limited due to task-irrelevant information. In this paper, we utilize bisimulation metrics to construct an invariant representation learning method and extract task-relevant information. The research shows that the multi-step prediction environment model can retain longer-term state-transition information. In addition, we propose a multi-step prediction method to collect cumulative loss and update the extractor for representing learning, thereby improving the relevance of the extracted information of the task. Experiments on tasks with or without distractors show that the proposed method achieves better results.
引用
收藏
页码:202 / 206
页数:5
相关论文
共 50 条
  • [1] A Multi-Step Reinforcement Learning Algorithm
    Zhang, Zhicong
    Hu, Kaishun
    Huang, Huiyu
    Li, Shuai
    Zhao, Shaoyong
    FRONTIERS OF MANUFACTURING AND DESIGN SCIENCE, PTS 1-4, 2011, 44-47 : 3611 - 3615
  • [2] Multi-Step Reinforcement Learning: A Unifying Algorithm
    De Asis, Kristopher
    Hernandez-Garcia, J. Fernando
    Holland, G. Zacharias
    Sutton, Richard S.
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2902 - 2909
  • [3] Reinforcement Learning for Multi-Step Expert Advice
    Philipp, Patrick
    Rettinger, Achim
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 962 - 971
  • [4] Multi-agent reinforcement learning for vehicular task offloading with multi-step trajectory prediction
    Zhang, Xinyi
    Zhu, Yanmin
    Wang, Chunyang
    Cao, Jian
    Chen, Yirong
    Wang, Jie
    CCF TRANSACTIONS ON PERVASIVE COMPUTING AND INTERACTION, 2024, 6 (02) : 101 - 114
  • [5] Multi-Agent Reinforcement Learning with Multi-Step Generative Models
    Krupnik, Orr
    Mordatch, Igor
    Tamar, Aviv
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [6] Multi-step medical image segmentation based on reinforcement learning
    Zhiqiang Tian
    Xiangyu Si
    Yaoyue Zheng
    Zhang Chen
    Xiaojian Li
    Journal of Ambient Intelligence and Humanized Computing, 2022, 13 : 5011 - 5022
  • [7] Speeding-up reinforcement learning with multi-step actions
    Schoknecht, R
    Riedmiller, M
    ARTIFICIAL NEURAL NETWORKS - ICANN 2002, 2002, 2415 : 813 - 818
  • [8] Multi-step medical image segmentation based on reinforcement learning
    Tian, Zhiqiang
    Si, Xiangyu
    Zheng, Yaoyue
    Chen, Zhang
    Li, Xiaojian
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2020, 13 (11) : 5011 - 5022
  • [9] The Effect of Multi-step Methods on Overestimation in Deep Reinforcement Learning
    Meng, Lingheng
    Gorbet, Rob
    Kulic, Dana
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 347 - 353
  • [10] Transfer Learning for Multi-Step Resource Utilization Prediction
    Parera, Claudia
    Liao, Qi
    Malanchini, Ilaria
    Wellington, Dan
    Redondi, Alessandro E. C.
    Cesana, Matteo
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,