ORLEP: an efficient offline reinforcement learning evaluation platform

被引:0
|
作者
Mao, Keming [1 ]
Chen, Chen [1 ]
Zhang, Jinkai [1 ]
Li, Yiyang [1 ]
机构
[1] Northeastern Univ, Software Coll, Shenyang, Peoples R China
关键词
Offline reinforcement learning; Model training and evaluation; Platform design and implementation;
D O I
10.1007/s11042-023-16906-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Developing offline reinforcement learning evaluation applications faces challenges such as heterogeneous data and algorithm integration, user-friendly interface, and flexible resource management. This paper designs and implements ORLEP, an efficient platform to provide high-level services for offline reinforcement learning evaluation. Besides integrating underlying infrastructure with highly concurrency and reliability, core components with distributed deployment and 3rd party libs and benchmarks incorporation, ORLEP supplies high-level abstractions for (1) data management, (2) model training and evaluation, (3) result visualization, and (4) resource configuration and supervision. Moreover, this paper verifies specific cases and the results demonstrate the performance and scalability of the proposed ORLEP.
引用
收藏
页码:37073 / 37087
页数:15
相关论文
共 50 条
  • [1] ORLEP: an efficient offline reinforcement learning evaluation platform
    Keming Mao
    Chen Chen
    Jinkai Zhang
    Yiyang Li
    [J]. Multimedia Tools and Applications, 2024, 83 : 37073 - 37087
  • [2] Efficient Diffusion Policies for Offline Reinforcement Learning
    Kang, Bingyi
    Ma, Xiao
    Du, Chao
    Pang, Tianyu
    Yan, Shuicheng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Efficient Online Reinforcement Learning with Offline Data
    Ball, Philip J.
    Smith, Laura
    Kostrikov, Ilya
    Levine, Sergey
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [4] Efficient Offline Reinforcement Learning With Relaxed Conservatism
    Huang, Longyang
    Dong, Botao
    Zhang, Weidong
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (08) : 5260 - 5272
  • [5] Offline Evaluation of Online Reinforcement Learning Algorithms
    Mandel, Travis
    Liu, Yun-En
    Brunskill, Emma
    Popovic, Zoran
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1926 - 1933
  • [6] Sample Efficient Offline-to-Online Reinforcement Learning
    Guo, Siyuan
    Zou, Lixin
    Chen, Hechang
    Qu, Bohao
    Chi, Haotian
    Yu, Philip S.
    Chang, Yi
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (03) : 1299 - 1310
  • [7] Efficient experience replay architecture for offline reinforcement learning
    Zhang, Longfei
    Feng, Yanghe
    Wang, Rongxiao
    Xu, Yue
    Xu, Naifu
    Liu, Zeyi
    Du, Hang
    [J]. ROBOTIC INTELLIGENCE AND AUTOMATION, 2023, 43 (01): : 35 - 43
  • [8] Statistically Efficient Advantage Learning for Offline Reinforcement Learning in Infinite Horizons
    Shi, Chengchun
    Luo, Shikai
    Le, Yuan
    Zhu, Hongtu
    Song, Rui
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024, 119 (545) : 232 - 245
  • [9] Federated Offline Reinforcement Learning with Proximal Policy Evaluation
    Yue, Sheng
    Deng, Yongheng
    Wang, Guanbo
    Ren, Ju
    Zhang, Yaoxue
    [J]. Chinese Journal of Electronics, 2024, 33 (06) : 1360 - 1372
  • [10] Federated Offline Reinforcement Learning with Proximal Policy Evaluation
    Sheng YUE
    Yongheng DENG
    Guanbo WANG
    Ju REN
    Yaoxue ZHANG
    [J]. Chinese Journal of Electronics, 2024, 33 (06) : 1360 - 1372