Data-Efficient Offline Reinforcement Learning with Approximate Symmetries

被引:0
|
作者
Angelotti, Giorgio [1 ,2 ]
Drougard, Nicolas [1 ,2 ]
Chanel, Caroline P. C. [1 ,2 ]
机构
[1] Univ Toulouse, ANITI, Toulouse, France
[2] Univ Toulouse, ISAE Supaero, Toulouse, France
关键词
Offline reinforcement learning; Approximate symmetries; Data augmentation;
D O I
10.1007/978-3-031-55326-4_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of Offline Reinforcement Learning (ORL) models in Markov Decision Processes (MDPs) is heavily contingent upon the quality and diversity of the training data. This research furthers the exploration of expert-guided symmetry detection and data augmentation techniques by considering approximate symmetries in discrete MDPs, providing a fresh perspective on data efficiency in the domain of ORL. We scrutinize the adaptability and resilience of these established methodologies in varied stochastic environments, featuring alterations in transition probabilities with respect to the already tested stochastic environments. Key findings from these investigations elucidate the potential of approximate symmetries for the data augmentation process and confirm the robustness of the existing methods under altered stochastic conditions. Our analysis reinforces the applicability of the established symmetry detection techniques in diverse scenarios while opening new horizons for enhancing the efficiency of ORL models.
引用
收藏
页码:164 / 186
页数:23
相关论文
共 50 条
  • [1] Data-Efficient Pipeline for Offline Reinforcement Learning with Limited Data
    Nie, Allen
    Flet-Berliac, Yannis
    Jordan, Deon R.
    Steenbergen, William
    Brunskill, Emma
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Data-Efficient Hierarchical Reinforcement Learning
    Nachum, Ofir
    Gu, Shixiang
    Lee, Honglak
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] PerSim: Data-efficient Offline Reinforcement Learning with Heterogeneous Agents via Personalized Simulators
    Agarwal, Anish
    Alomar, Abdullah
    Alumootil, Varkey
    Shah, Devavrat
    Shen, Dennis
    Xu, Zhi
    Yang, Cindy
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] Data-Efficient Reinforcement Learning for Malaria Control
    Zou, Lixin
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 507 - 513
  • [5] Pretraining Representations for Data-Efficient Reinforcement Learning
    Schwarzer, Max
    Rajkumar, Nitarshan
    Noukhovitch, Michael
    Anand, Ankesh
    Charlin, Laurent
    Hjelm, Devon
    Bachman, Philip
    Courville, Aaron
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] EqR: Equivariant Representations for Data-Efficient Reinforcement Learning
    Mondal, Arnab Kumar
    Jain, Vineet
    Siddiqi, Kaleem
    Ravanbakhsh, Siamak
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] Efficient Online Reinforcement Learning with Offline Data
    Ball, Philip J.
    Smith, Laura
    Kostrikov, Ilya
    Levine, Sergey
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [8] Data-Efficient Reinforcement Learning for Variable Impedance Control
    Anand, Akhil S.
    Kaushik, Rituraj
    Gravdahl, Jan Tommy
    Abu-Dakka, Fares J.
    [J]. IEEE ACCESS, 2024, 12 : 15631 - 15641
  • [9] BarlowRL: Barlow Twins for Data-Efficient Reinforcement Learning
    Cagatan, Omer Veysel
    Akgun, Baris
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [10] Concurrent Credit Assignment for Data-efficient Reinforcement Learning
    Dauce, Emmanuel
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,