Multi-step Prediction for Learning Invariant Representations in Reinforcement Learning

被引:0
|
作者
Xu, Xinyue [1 ,2 ]
Lv, Kai [1 ,2 ]
Dong, Xingye [1 ,2 ]
Han, Sheng [1 ,2 ]
Lin, Youfang [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] CAAC, Key Lab Intelligent Passenger Serv Civil Aviat, Beijing, Peoples R China
关键词
multi-step prediction; bisimulation metrics; representation learning; reinforcement learning;
D O I
10.1109/HPBDIS53214.2021.9658436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on how to achieve task-relevant feature representations in reinforcement learning from image observations without relying either on domain knowledge or pixel-reconstruction. Although the existing algorithms based on reconstruction and contrastive learning have achieved excellent success, the sample efficiency and robustness of the algorithm are limited due to task-irrelevant information. In this paper, we utilize bisimulation metrics to construct an invariant representation learning method and extract task-relevant information. The research shows that the multi-step prediction environment model can retain longer-term state-transition information. In addition, we propose a multi-step prediction method to collect cumulative loss and update the extractor for representing learning, thereby improving the relevance of the extracted information of the task. Experiments on tasks with or without distractors show that the proposed method achieves better results.
引用
收藏
页码:202 / 206
页数:5
相关论文
共 50 条
  • [21] Learning multi-step prediction models for receding horizon control
    Terzi, Enrico
    Fagiano, Lorenzo
    Farina, Marcello
    Scattolini, Riccardo
    2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, : 1335 - 1340
  • [22] A Unified Approach for Multi-step Temporal-Difference Learning with Eligibility Traces in Reinforcement Learning
    Yang, Long
    Shi, Minhao
    Zheng, Qian
    Meng, Wenjia
    Pan, Gang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2984 - 2990
  • [23] Learn multi-step object sorting tasks through deep reinforcement learning
    Bao, Jiatong
    Zhang, Guoqing
    Peng, Yi
    Shao, Zhiyu
    Song, Aiguo
    ROBOTICA, 2022, 40 (11) : 3878 - 3894
  • [24] Reinforcement Learning based Multi-Step Look-Ahead Bayesian Optimization
    Cheon, Mujin
    Byun, Haeun
    Lee, Jay H.
    IFAC PAPERSONLINE, 2022, 55 (07): : 100 - 105
  • [25] Multi-step reinforcement learning-based offloading for vehicle edge computing
    Han Shaodong
    Chen Yingqun
    Chen Guihong
    Yin, Jiao
    Hang, Hua
    Cao, Jinli
    2023 15TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE, ICACI, 2023,
  • [26] Recurrent neural networks that learn multi-step visual routines with reinforcement learning
    Mollard, Sami
    Wacongne, Catherine
    Bohte, Sander M.
    Roelfsema, Pieter R.
    PLOS COMPUTATIONAL BIOLOGY, 2024, 20 (04)
  • [27] Designing Internal Reward of Reinforcement Learning Agents in Multi-Step Dilemma Problem
    Ichikawa, Yoshihiro
    Takadama, Keiki
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2013, 17 (06) : 926 - 931
  • [28] MULTI-STEP ACTOR-CRITIC FRAMEWORK FOR REINFORCEMENT LEARNING IN CONTINUOUS CONTROL
    Huang T.
    Chen G.
    Journal of Applied and Numerical Optimization, 2023, 5 (02): : 189 - 200
  • [29] Deep Learning for Multi-Step Performance Prediction in Operational Optical Networks
    Mezni, Ameni
    Charlton, Douglas W.
    Tremblay, Christine
    Desrosiers, Christian
    2020 CONFERENCE ON LASERS AND ELECTRO-OPTICS (CLEO), 2020,
  • [30] A novel multi-step Q-learning method to improve data efficiency for deep reinforcement learning
    Yuan, Yinlong
    Yu, Zhu Liang
    Gu, Zhenghui
    Yeboah, Yao
    Wei, Wu
    Deng, Xiaoyan
    Li, Jingcong
    Li, Yuanqing
    KNOWLEDGE-BASED SYSTEMS, 2019, 175 : 107 - 117