Multi-step Prediction for Learning Invariant Representations in Reinforcement Learning

被引:0
|
作者
Xu, Xinyue [1 ,2 ]
Lv, Kai [1 ,2 ]
Dong, Xingye [1 ,2 ]
Han, Sheng [1 ,2 ]
Lin, Youfang [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] CAAC, Key Lab Intelligent Passenger Serv Civil Aviat, Beijing, Peoples R China
关键词
multi-step prediction; bisimulation metrics; representation learning; reinforcement learning;
D O I
10.1109/HPBDIS53214.2021.9658436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on how to achieve task-relevant feature representations in reinforcement learning from image observations without relying either on domain knowledge or pixel-reconstruction. Although the existing algorithms based on reconstruction and contrastive learning have achieved excellent success, the sample efficiency and robustness of the algorithm are limited due to task-irrelevant information. In this paper, we utilize bisimulation metrics to construct an invariant representation learning method and extract task-relevant information. The research shows that the multi-step prediction environment model can retain longer-term state-transition information. In addition, we propose a multi-step prediction method to collect cumulative loss and update the extractor for representing learning, thereby improving the relevance of the extracted information of the task. Experiments on tasks with or without distractors show that the proposed method achieves better results.
引用
收藏
页码:202 / 206
页数:5
相关论文
共 50 条
  • [41] A Stacked Machine Learning Algorithm for Multi-Step Ahead Prediction of Soil Moisture
    Granata, Francesco
    Di Nunno, Fabio
    Najafzadeh, Mohammad
    Demir, Ibrahim
    HYDROLOGY, 2023, 10 (01)
  • [42] Multi-step truncated Q learning algorithm
    Chen, SL
    Wu, HZ
    Han, XL
    Xiao, L
    PROCEEDINGS OF 2005 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-9, 2005, : 194 - 198
  • [43] Incremental multi-step Q-learning
    Peng, J
    Williams, RJ
    MACHINE LEARNING, 1996, 22 (1-3) : 283 - 290
  • [44] Learning-assisted multi-step planning
    Hauser, K
    Bretl, T
    Latombe, JC
    2005 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-4, 2005, : 4575 - 4580
  • [45] Multi-Step Transfer Learning for Sentiment Analysis
    Golubev, Anton
    Loukachevitch, Natalia
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2021), 2021, 12801 : 209 - 217
  • [46] Emergence of Multi-step Discrete State Transition through Reinforcement Learning with a Recurrent Neural Network
    Samsudin, Mohamad Faizal
    Sawatsubashi, Yoshito
    Shibata, Katsunari
    NEURAL INFORMATION PROCESSING, ICONIP 2012, PT II, 2012, 7664 : 583 - 590
  • [47] MRL-Seg: Overcoming Imbalance in Medical Image Segmentation With Multi-Step Reinforcement Learning
    Yang, Feiyang
    Li, Xiongfei
    Duan, Haoran
    Xu, Feilong
    Huang, Yawen
    Zhang, Xiaoli
    Long, Yang
    Zheng, Yefeng
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (02) : 858 - 869
  • [48] A Deep Reinforcement Learning Based Multi-Step Coarse to Fine Question Answering (MSCQA) System
    Wang, Yu
    Jin, Hongxia
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 7224 - 7232
  • [49] A multi-step predictive deep reinforcement learning algorithm for HVAC control systems in smart buildings
    Liu, Xiangfei
    Ren, Mifeng
    Yang, Zhile
    Yan, Gaowei
    Guo, Yuanjun
    Cheng, Lan
    Wu, Chengke
    ENERGY, 2022, 259
  • [50] "Good Robot!": Efficient Reinforcement Learning for Multi-Step Visual Tasks with Sim to Real Transfer
    Hundt, Andrew
    Killeen, Benjamin
    Greene, Nicholas
    Wu, Hongtao
    Kwon, Heeyeon
    Paxton, Chris
    Hager, Gregory D.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6724 - 6731