Enhancing visual reinforcement learning with State-Action Representation

被引:0
|
作者
Yan, Mengbei [1 ]
Lyu, Jiafei [1 ]
Li, Xiu [1 ]
机构
[1] Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Lishui Rd, Shenzhen 518055, Peoples R China
关键词
Visual reinforcement learning; State-action representation; Sample efficiency;
D O I
10.1016/j.knosys.2024.112487
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the remarkable progress made in visual reinforcement learning (RL) in recent years, sample inefficiency remains a major challenge. Many existing approaches attempt to address this by extracting better representations from raw images using techniques like data augmentation or introducing some auxiliary tasks. However, these methods overlook the environmental dynamic information embedded in the collected transitions, which can be crucial for efficient control. In this paper, we present STAR: State-Action Action Representation Learning, a simple yet effective approach for visual continuous control. STAR learns a joint state-action representation by modeling the dynamics of the environment in the latent space. By incorporating the learned joint state- action representation into the critic, STAR enhances the value estimation with latent dynamics information. We theoretically show that the value function can still converge to the optima when involving additional representation inputs. On various challenging visual continuous control tasks from DeepMind Control Suite, STAR achieves significant improvements in sample efficiency compared to strong baseline algorithms.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Using Memory-Based Learning to Solve Tasks with State-Action Constraints
    Verghese, Mrinal
    Atkeson, Christopher
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9558 - 9565
  • [42] SA-Net: Robust State-Action Recognition for Learning from Observations
    Soans, Nihal
    Asali, Ehsan
    Hong, Yi
    Doshi, Prashant
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2153 - 2159
  • [43] Rebalancing Shared Mobility Systems by User Incentive Schemes: State-Action Representation Design and Analysis
    Schofield, Matthew
    Wang, Ning
    Ho, Shen-Shyang
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 1045 - 1047
  • [44] Enhancing Visual Generalization in Reinforcement Learning with Cycling Augmentation
    Sun, Shengjie
    Lyu, Jiafei
    Li, Lu
    Guo, Jiazhe
    Yan, Mengbei
    Liu, Runze
    Li, Xiu
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT IV, 2024, 15019 : 397 - 411
  • [45] Action-driven contrastive representation for reinforcement learning
    Kim, Minbeom
    Rho, Kyeongha
    Kim, Yong-duk
    Jung, Kyomin
    PLOS ONE, 2022, 17 (03):
  • [46] Autonomous control of a snake-like robot using reinforcement learning -Discussion of the role of the mechanical body in abstraction of state-action space
    Takayama, Akihiro
    Ito, Kazuyuki
    Minamino, Tomoko
    IECON 2008: 34TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, VOLS 1-5, PROCEEDINGS, 2008, : 1525 - 1530
  • [47] Multi-Layer Attention-based State Representation for the Reinforcement Learning of Visual Servoing
    Kitajima, Hiromu
    Bounyong, Souksakhone
    Yoshioka, Mototaka
    2023 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS, ICCE, 2023,
  • [48] Constrained Visual Representation Learning With Bisimulation Metrics for Safe Reinforcement Learning
    Wang, Rongrong
    Cheng, Yuhu
    Wang, Xuesong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 379 - 393
  • [49] State Representation Learning for Goal-Conditioned Reinforcement Learning
    Steccanella, Lorenzo
    Jonsson, Anders
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT IV, 2023, 13716 : 84 - 99
  • [50] Vector quantization for state-action map compression
    Ueda, R
    Fukase, T
    Kobayashi, Y
    Arai, T
    2003 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-3, PROCEEDINGS, 2003, : 2356 - 2361