Enhancing visual reinforcement learning with State-Action Representation

被引:0
|
作者
Yan, Mengbei [1 ]
Lyu, Jiafei [1 ]
Li, Xiu [1 ]
机构
[1] Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Lishui Rd, Shenzhen 518055, Peoples R China
关键词
Visual reinforcement learning; State-action representation; Sample efficiency;
D O I
10.1016/j.knosys.2024.112487
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the remarkable progress made in visual reinforcement learning (RL) in recent years, sample inefficiency remains a major challenge. Many existing approaches attempt to address this by extracting better representations from raw images using techniques like data augmentation or introducing some auxiliary tasks. However, these methods overlook the environmental dynamic information embedded in the collected transitions, which can be crucial for efficient control. In this paper, we present STAR: State-Action Action Representation Learning, a simple yet effective approach for visual continuous control. STAR learns a joint state-action representation by modeling the dynamics of the environment in the latent space. By incorporating the learned joint state- action representation into the critic, STAR enhances the value estimation with latent dynamics information. We theoretically show that the value function can still converge to the optima when involving additional representation inputs. On various challenging visual continuous control tasks from DeepMind Control Suite, STAR achieves significant improvements in sample efficiency compared to strong baseline algorithms.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] For SALE: State-Action Representation Learning for Deep Reinforcement Learning
    Fujimoto, Scott
    Chang, Wei-Di
    Smith, Edward J.
    Gu, Shixiang Shane
    Precup, Doina
    Meger, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [2] PROJECTED STATE-ACTION BALANCING WEIGHTS FOR OFFLINE REINFORCEMENT LEARNING
    Wang, Jiayi
    Qi, Zhengling
    Wong, Raymond K. W.
    ANNALS OF STATISTICS, 2023, 51 (04): : 1639 - 1665
  • [3] A REINFORCEMENT LEARNING MODEL USING DETERMINISTIC STATE-ACTION SEQUENCES
    Murata, Makoto
    Ozawa, Seiichi
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2010, 6 (02): : 577 - 590
  • [4] Learning cooperative assembly with the graph representation of a state-action space
    Ferch, M
    Höchsmann, M
    Zhang, JW
    2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, 2002, : 990 - 995
  • [5] Learning cooperative grasping with the graph representation of a state-action space
    Ferch, M
    Zhang, JW
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2002, 38 (3-4) : 183 - 195
  • [6] Efficient Reinforcement Learning Using State-Action Uncertainty with Multiple Heads
    Aizu, Tomoharu
    Oba, Takeru
    Ukita, Norimichi
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VIII, 2023, 14261 : 184 - 196
  • [7] Speeding up Tabular Reinforcement Learning Using State-Action Similarities
    Rosenfeld, Ariel
    Taylor, Matthew E.
    Kraus, Sarit
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 1722 - 1724
  • [8] Jointly-Learned State-Action Embedding for Efficient Reinforcement Learning
    Pritz, Paul J.
    Ma, Liang
    Leung, Kin K.
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 1447 - 1456
  • [9] Model-Based Reinforcement Learning Exploiting State-Action Equivalence
    Asadi, Mahsa
    Talebi, Mohammad Sadegh
    Bourel, Hippolyte
    Maillard, Odalric-Ambrym
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 204 - 219
  • [10] Swarm Reinforcement Learning Methods for Problems with Continuous State-Action Space
    Iima, Hitoshi
    Kuroe, Yasuaki
    Emoto, Kazuo
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2173 - 2180