Data-Efficient Offline Reinforcement Learning with Approximate Symmetries

被引:0
|
作者
Angelotti, Giorgio [1 ,2 ]
Drougard, Nicolas [1 ,2 ]
Chanel, Caroline P. C. [1 ,2 ]
机构
[1] Univ Toulouse, ANITI, Toulouse, France
[2] Univ Toulouse, ISAE Supaero, Toulouse, France
关键词
Offline reinforcement learning; Approximate symmetries; Data augmentation;
D O I
10.1007/978-3-031-55326-4_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of Offline Reinforcement Learning (ORL) models in Markov Decision Processes (MDPs) is heavily contingent upon the quality and diversity of the training data. This research furthers the exploration of expert-guided symmetry detection and data augmentation techniques by considering approximate symmetries in discrete MDPs, providing a fresh perspective on data efficiency in the domain of ORL. We scrutinize the adaptability and resilience of these established methodologies in varied stochastic environments, featuring alterations in transition probabilities with respect to the already tested stochastic environments. Key findings from these investigations elucidate the potential of approximate symmetries for the data augmentation process and confirm the robustness of the existing methods under altered stochastic conditions. Our analysis reinforces the applicability of the established symmetry detection techniques in diverse scenarios while opening new horizons for enhancing the efficiency of ORL models.
引用
收藏
页码:164 / 186
页数:23
相关论文
共 50 条
  • [41] Uniform Priors for Data-Efficient Learning
    Sinha, Samarth
    Roth, Karsten
    Goyal, Anirudh
    Ghassemi, Marzyeh
    Akata, Zeynep
    Larochelle, Hugo
    Garg, Animesh
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4026 - 4037
  • [42] DATA-EFFICIENT DEEP REINFORCEMENT LEARNING WITH CONVOLUTION-BASED STATE ENCODER NETWORKS
    Fang, Qiang
    Xu, Xin
    Lan, Yixin
    Zhang, Yichuan
    Zeng, Yujun
    Tang, Tao
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2021, 36
  • [43] Data-Efficient Reinforcement Learning in Continuous State-Action Gaussian-POMDPs
    McAllister, Rowan Thomas
    Rasmussen, Carl Edward
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [44] Koopman Q-learning: Offline Reinforcement Learning via Symmetries of Dynamics
    Weissenbacher, Matthias
    Sinha, Samarth
    Garg, Animesh
    Kawahara, Yoshinobu
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [45] Efficient Diffusion Policies for Offline Reinforcement Learning
    Kang, Bingyi
    Ma, Xiao
    Du, Chao
    Pang, Tianyu
    Yan, Shuicheng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [46] Efficient Offline Reinforcement Learning With Relaxed Conservatism
    Huang, Longyang
    Dong, Botao
    Zhang, Weidong
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (08) : 5260 - 5272
  • [47] Deep reinforcement learning for data-efficient weakly supervised business process anomaly detection
    Elaziz, Eman Abd
    Fathalla, Radwa
    Shaheen, Mohamed
    [J]. JOURNAL OF BIG DATA, 2023, 10 (01)
  • [48] Deep reinforcement learning for data-efficient weakly supervised business process anomaly detection
    Eman Abd Elaziz
    Radwa Fathalla
    Mohamed Shaheen
    [J]. Journal of Big Data, 10
  • [49] Data-efficient Learning of Morphology and Controller for a Microrobot
    Liao, Thomas
    Wang, Grant
    Yang, Brian
    Lee, Rene
    Pister, Kristofer
    Levine, Sergey
    Calandra, Roberto
    [J]. 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 2488 - 2494
  • [50] Robust Reinforcement Learning using Offline Data
    Panaganti, Kishan
    Xu, Zaiyan
    Kalathil, Dileep
    Ghavamzadeh, Mohammad
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,