Zero-shot sim-to-real transfer using Siamese-Q-Based reinforcement learning

被引:0
|
作者
Zhang, Zhenyu [1 ]
Xie, Shaorong [1 ]
Zhang, Han [1 ]
Luo, Xiangfeng [1 ]
Yu, Hang [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, 99 Shangda Rd, Shanghai 200444, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Representation learning; Simulation to real; Contrastive learning; NETWORK;
D O I
10.1016/j.inffus.2024.102664
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To address real world decision problems in reinforcement learning, it is common to train a policy in a simulator first for safety. Unfortunately, the sim-real gap hinders effective simulation-to-real transfer without substantial training data. However, collecting real samples of complex tasks is often impractical, and the sample inefficiency of reinforcement learning exacerbates the simulation-to-real problem, even with online interaction or data. Representation learning can improve sample efficiency while keeping generalization by projecting high-dimensional inputs into low-dimensional representations. However, whether trained independently or simultaneously with reinforcement learning, representation learning remains a separate auxiliary task, lacking task-related features and generalization for simulation-to-real transfer. This paper proposes Siamese-Q, a new representation learning method employing Siamese networks and zero-shot simulation-to-real transfer, which narrows the distance between inputs with the same semantics in the latent space with respect to Q values. This allows us to fuse task-related information into the representation and improve the generalization of the policy. Evaluation in virtual and real autonomous vehicle scenarios demonstrates substantial improvements of 19.5% and 94.2% respectively over conventional representation learning, without requiring any real-world observations or on-policy interaction, and enabling reinforcement learning policies trained in simulations transfer to reality.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Sim-to-Real Control of Trifinger Robot by Deep Reinforcement Learning
    Wan, Qiang
    Wu, Tianyang
    Ye, Jiawei
    Wan, Lipeng
    Lau, Xuguang
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT VI, 2025, 15206 : 300 - 314
  • [42] Dynamic Bipedal Turning through Sim-to-Real Reinforcement Learning
    Yu, Fangzhou
    Batke, Ryan
    Dao, Jeremy
    Hurst, Jonathan
    Green, Kevin
    Fern, Alan
    2022 IEEE-RAS 21ST INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2022, : 903 - 910
  • [43] Zero-Shot Policy Transfer in Autonomous Racing: Reinforcement Learning vs Imitation Learning
    Hamilton, Nathaniel
    Musau, Patrick
    Lopez, Diego Manzanas
    Johnson, Taylor T.
    2022 IEEE INTERNATIONAL CONFERENCE ON ASSURED AUTONOMY (ICAA 2022), 2022, : 11 - 20
  • [44] Zero-Shot Transfer with Deictic Object-Oriented Representation in Reinforcement Learning
    Marom, Ofir
    Rosman, Benjamin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [45] Sim-to-real transfer in reinforcement learning-based, non-steady-state control for chemical plants
    Kubosawa S.
    Onishi T.
    Tsuruoka Y.
    SICE Journal of Control, Measurement, and System Integration, 2022, 15 (01) : 10 - 23
  • [46] Sim-to-real transfer reinforcement learning for control of thermal effects of an atmospheric pressure plasma jet
    Witman, Matthew
    Gidon, Dogan
    Graves, David B.
    Smit, Berend
    Mesbah, Ali
    PLASMA SOURCES SCIENCE & TECHNOLOGY, 2019, 28 (09):
  • [47] Reinforcement Learning-based Sim-to-Real Impedance Parameter Tuning for Robotic Assembly
    Kim, Yong-Geon
    Na, Minwoo
    Song, Jae-Bok
    2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 833 - 836
  • [48] Sim-to-Real Application of Reinforcement Learning Agents for Autonomous, Real Vehicle Drifting
    Toth, Szilard Hunor
    Viharos, Zsolt Janos
    Bardos, Adam
    Szalay, Zsolt
    VEHICLES, 2024, 6 (02): : 781 - 798
  • [49] A Survey of Zero-shot Generalisation in Deep Reinforcement Learning
    Kirk R.
    Zhang A.
    Grefenstette E.
    Rocktäschel T.
    Journal of Artificial Intelligence Research, 2023, 76 : 201 - 264
  • [50] Zero-Shot Transfer Learning Based on Visual and Textual Resemblance
    Yang, Gang
    Xu, Jieping
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT III, 2019, 11955 : 353 - 362