Contrastive Learning Methods for Deep Reinforcement Learning

被引:2
|
作者
Wang, Di [1 ]
Hu, Mengqi [1 ]
机构
[1] Univ Illinois, Dept Mech & Ind Engn, Chicago, IL 60609 USA
基金
美国国家科学基金会;
关键词
Contrastive learning; deep reinforcement learning; different-age experience; experience replay buffer; parallel learning; BUFFER;
D O I
10.1109/ACCESS.2023.3312383
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning (DRL) has shown promising performance in various application areas (e.g., games and autonomous vehicles). Experience replay buffer strategy and parallel learning strategy are widely used to boost the performances of offline and online deep reinforcement learning algorithms. However, state-action distribution shifts lead to bootstrap errors. Experience replay buffer learns policies with elder experience trajectories, limiting its application to off-policy algorithms. Balancing the new and the old experience is challenging. Parallel learning strategies can train policies with online experiences. However, parallel environmental instances organize the agent pool inefficiently with higher simulation or physical costs. To overcome these shortcomings, we develop four lightweight and effective DRL algorithms, instance-actor, parallel-actor, instance-critic, and parallel-critic methods, to contrast different-age trajectory experiences. We train the contrast DRL according to the received rewards and proposed contrast loss, which is calculated by designed positive/negative keys. Our benchmark experiments using PyBullet robotics environments show that our proposed algorithm matches or is better than the state-of-the-art DRL algorithms.
引用
收藏
页码:97107 / 97117
页数:11
相关论文
共 50 条
  • [31] Learning Deep Representations via Contrastive Learning for Instance Retrieval
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C., II
    [J]. 2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 1501 - 1506
  • [32] Deep Reinforcement Learning for Adaptive Learning Systems
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    [J]. JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2023, 48 (02) : 220 - 243
  • [33] Learning to Break Rocks With Deep Reinforcement Learning
    Samtani, Pavan
    Leiva, Francisco
    Ruiz-del-Solar, Javier
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (02) : 1077 - 1084
  • [34] Learning to Walk via Deep Reinforcement Learning
    Haarnoja, Tuomas
    Ha, Sehoon
    Zhou, Aurick
    Tan, Jie
    Tucker, George
    Levine, Sergey
    [J]. ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [35] Learning an Index Advisor with Deep Reinforcement Learning
    Lai, Sichao
    Wu, Xiaoying
    Wang, Senyang
    Peng, Yuwei
    Peng, Zhiyong
    [J]. WEB AND BIG DATA, APWEB-WAIM 2021, PT II, 2021, 12859 : 178 - 185
  • [36] Transfer Learning in Deep Reinforcement Learning: A Survey
    Zhu, Zhuangdi
    Lin, Kaixiang
    Jain, Anil K.
    Zhou, Jiayu
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 13344 - 13362
  • [37] Class Incremental Learning With Deep Contrastive Learning and Attention Distillation
    Zhu, Jitao
    Luo, Guibo
    Duan, Baishan
    Zhu, Yuesheng
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1224 - 1228
  • [38] The Difficulty of Passive Learning in Deep Reinforcement Learning
    Ostrovski, Georg
    Castro, Pablo Samuel
    Dabney, Will
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [39] Deep learning and reinforcement learning approach on microgrid
    Chandrasekaran, Kumar
    Kandasamy, Prabaakaran
    Ramanathan, Srividhya
    [J]. INTERNATIONAL TRANSACTIONS ON ELECTRICAL ENERGY SYSTEMS, 2020, 30 (10):
  • [40] Deep learning, reinforcement learning, and world models
    Matsuo, Yutaka
    LeCun, Yann
    Sahani, Maneesh
    Precup, Doina
    Silver, David
    Sugiyama, Masashi
    Uchibe, Eiji
    Morimoto, Jun
    [J]. NEURAL NETWORKS, 2022, 152 : 267 - 275