Zero-shot sim-to-real transfer using Siamese-Q-Based reinforcement learning

被引:0
|
作者
Zhang, Zhenyu [1 ]
Xie, Shaorong [1 ]
Zhang, Han [1 ]
Luo, Xiangfeng [1 ]
Yu, Hang [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, 99 Shangda Rd, Shanghai 200444, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Representation learning; Simulation to real; Contrastive learning; NETWORK;
D O I
10.1016/j.inffus.2024.102664
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To address real world decision problems in reinforcement learning, it is common to train a policy in a simulator first for safety. Unfortunately, the sim-real gap hinders effective simulation-to-real transfer without substantial training data. However, collecting real samples of complex tasks is often impractical, and the sample inefficiency of reinforcement learning exacerbates the simulation-to-real problem, even with online interaction or data. Representation learning can improve sample efficiency while keeping generalization by projecting high-dimensional inputs into low-dimensional representations. However, whether trained independently or simultaneously with reinforcement learning, representation learning remains a separate auxiliary task, lacking task-related features and generalization for simulation-to-real transfer. This paper proposes Siamese-Q, a new representation learning method employing Siamese networks and zero-shot simulation-to-real transfer, which narrows the distance between inputs with the same semantics in the latent space with respect to Q values. This allows us to fuse task-related information into the representation and improve the generalization of the policy. Evaluation in virtual and real autonomous vehicle scenarios demonstrates substantial improvements of 19.5% and 94.2% respectively over conventional representation learning, without requiring any real-world observations or on-policy interaction, and enabling reinforcement learning policies trained in simulations transfer to reality.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Model-based Reinforcement Learning for Sim-to-Real Transfer in Robotics using HTM neural networks
    Diprasetya, M. R.
    Pullani, A. N.
    Schwung, D.
    Schwung, A.
    2024 10TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES, CODIT 2024, 2024, : 43 - 48
  • [22] Using sim-to-real transfer learning to close gaps between simulation and real environments through reinforcement learning
    Ushida, Yuto
    Razan, Hafiyanda
    Ishizuya, Shunta
    Sakuma, Takuto
    Kato, Shohei
    ARTIFICIAL LIFE AND ROBOTICS, 2022, 27 (01) : 130 - 136
  • [23] Using sim-to-real transfer learning to close gaps between simulation and real environments through reinforcement learning
    Yuto Ushida
    Hafiyanda Razan
    Shunta Ishizuya
    Takuto Sakuma
    Shohei Kato
    Artificial Life and Robotics, 2022, 27 : 130 - 136
  • [24] Pre- and post-contact policy decomposition for non-prehensile manipulation with zero-shot sim-to-real transfer
    Kim, Minchan
    Han, Junhyek
    Kim, Jaehyung
    Kim, Beomjoon
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 10644 - 10651
  • [25] One-shot sim-to-real transfer policy for robotic assembly via reinforcement learning with visual demonstration
    Xiao, Ruihong
    Yang, Chenguang
    Jiang, Yiming
    Zhang, Hui
    ROBOTICA, 2024, 42 (04) : 1074 - 1093
  • [26] Human-Guided Reinforcement Learning With Sim-to-Real Transfer for Autonomous Navigation
    Wu, Jingda
    Zhou, Yanxin
    Yang, Haohan
    Huang, Zhiyu
    Lv, Chen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 14745 - 14759
  • [27] A Domain Data Pattern Randomization based Deep Reinforcement Learning method for Sim-to-Real transfer
    Gong, Peng
    Shi, Dianxi
    Xue, Chao
    Chen, Xucan
    2021 5TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE (ICIAI 2021), 2021, : 1 - 7
  • [28] A Survey of Sim-to-Real Transfer Techniques Applied to Reinforcement Learning for Bioinspired Robots
    Zhu, Wei
    Guo, Xian
    Owaki, Dai
    Kutsuzawa, Kyo
    Hayashibe, Mitsuhiro
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3444 - 3459
  • [29] Sim-to-Real: Mapless Navigation for USVs Using Deep Reinforcement Learning
    Wang, Ning
    Wang, Yabiao
    Zhao, Yuming
    Wang, Yong
    Li, Zhigang
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2022, 10 (07)
  • [30] Sim-to-Real Robotic Sketching using Behavior Cloning and Reinforcement Learning
    Jia, Biao (biao@umd.edu), 1600, Institute of Electrical and Electronics Engineers Inc.