Multi-step Prediction for Learning Invariant Representations in Reinforcement Learning

被引:0
|
作者
Xu, Xinyue [1 ,2 ]
Lv, Kai [1 ,2 ]
Dong, Xingye [1 ,2 ]
Han, Sheng [1 ,2 ]
Lin, Youfang [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] CAAC, Key Lab Intelligent Passenger Serv Civil Aviat, Beijing, Peoples R China
关键词
multi-step prediction; bisimulation metrics; representation learning; reinforcement learning;
D O I
10.1109/HPBDIS53214.2021.9658436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on how to achieve task-relevant feature representations in reinforcement learning from image observations without relying either on domain knowledge or pixel-reconstruction. Although the existing algorithms based on reconstruction and contrastive learning have achieved excellent success, the sample efficiency and robustness of the algorithm are limited due to task-irrelevant information. In this paper, we utilize bisimulation metrics to construct an invariant representation learning method and extract task-relevant information. The research shows that the multi-step prediction environment model can retain longer-term state-transition information. In addition, we propose a multi-step prediction method to collect cumulative loss and update the extractor for representing learning, thereby improving the relevance of the extracted information of the task. Experiments on tasks with or without distractors show that the proposed method achieves better results.
引用
收藏
页码:202 / 206
页数:5
相关论文
共 50 条
  • [31] Learning Efficient Coordination Strategy for Multi-step Tasks in Multi-agent Systems using Deep Reinforcement Learning
    Zhu, Zean
    Diallo, Elhadji Amadou Oury
    Sugawara, Toshiharu
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 1, 2020, : 287 - 294
  • [32] DensePhysNet: Learning Dense Physical Object Representations via Multi-step Dynamic Interactions
    Xu, Zhenjia
    Wu, Jiajun
    Zeng, Andy
    Tenenbaum, Joshua B.
    Song, Shuran
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [33] Automatic Generation Control in a Distributed Power Grid Based on Multi-Step Reinforcement Learning
    Zhao, Wenmeng
    Zeng, Tuo
    Liu, Zhihong
    Xie, Lihui
    Xi, Lei
    Ma, Hui
    PROTECTION AND CONTROL OF MODERN POWER SYSTEMS, 2024, 9 (04) : 39 - 50
  • [34] A Multi-step Neural Control for Motor Brain-Machine Interface by Reinforcement Learning
    Wang, Fang
    Xu, Kai
    Zhang, Qiaosheng
    Wang, Yiwen
    Zheng, Xiaoxiang
    ADVANCES IN BIONIC ENGINEERING, 2014, 461 : 565 - 569
  • [35] Understanding world models through multi-step pruning policy via reinforcement learning
    He, Zhiqiang
    Qiu, Wen
    Zhao, Wei
    Shao, Xun
    Liu, Zhi
    INFORMATION SCIENCES, 2025, 686
  • [36] Ensemble Reinforcement Learning in Continuous Spaces - A Hierarchical Multi-Step Approach for Policy Training
    Chen, Gang
    Huang, Victoria
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 3514 - 3522
  • [37] Discrete-Time Nonlinear Optimal Control Using Multi-Step Reinforcement Learning
    An, Ningbo
    Wang, Qishao
    Zhao, Xiaochuan
    Wang, Qingyun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) : 2279 - 2283
  • [38] Evaluation of Deep Learning Models for Multi-Step Ahead Time Series Prediction
    Chandra, Rohitash
    Goyal, Shaurya
    Gupta, Rishabh
    IEEE ACCESS, 2021, 9 : 83105 - 83123
  • [39] Enhancing a Multi-Step Discharge Prediction with Deep Learning and a Response Time Parameter
    Thaisiam, Wandee
    Saelo, Warintra
    Wongchaisuwat, Papis
    WATER, 2022, 14 (18)
  • [40] Prediction of daily leaf wetness duration using multi-step machine learning
    Alsafadi, Karam
    Alatrach, Basma
    Sammen, Saad Sh
    Cao, Wenzhi
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 224