STASIS: Reinforcement Learning Simulators for Human-Centric Real-World Environments

被引:0
|
作者
Efstathiadis, Georgios [1 ]
Emedom-Nnamdi, Patrick [1 ]
Kolbeinsson, Arinbjorn [2 ]
Onnela, Jukka-Pekka [1 ]
Lu, Junwei [1 ]
机构
[1] Harvard TH Chan Sch Publ Hlth, Dept Biostat, Boston, MA 02115 USA
[2] Evidat Hlth, London, England
关键词
reinforcement learning; health care; real-world simulators;
D O I
10.1007/978-3-031-39539-0_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present on-going work toward building Stasis, a suite of reinforcement learning (RL) environments that aim to maintain realism for human-centric agents operating in real-world settings. Through representation learning and alignment with real-world offline data, Stasis allows for the evaluation of RL algorithms in offline environments with adjustable characteristics, such as observability, heterogeneity and levels of missing data. We aim to introduce environments the encourage training RL agents that are capable of maintaining a level of performance and robustness comparable to agents trained in real-world online environments, while avoiding the high cost and risks associated with making mistakes during online training. We provide examples of two environments that will be part of Stasis and discuss its implications for the deployment of RL-based systems in sensitive and high-risk areas of application.
引用
收藏
页码:85 / 92
页数:8
相关论文
共 50 条
  • [31] Intention awareness: improving upon situation awareness in human-centric environments
    Howard, Newton
    Cambria, Erik
    HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2013, 3 : 1 - 17
  • [32] Information Fusion in Challenging Environments for Human-Centric Cyber Physical Systems
    Grewe, Lynne
    Lagali, Christopher
    Overell, William
    SIGNAL PROCESSING, SENSOR/INFORMATION FUSION, AND TARGET RECOGNITION XXV, 2016, 9842
  • [33] Deep Offline Reinforcement Learning for Real-world Treatment Optimization Applications
    Nambiar, Mila
    Ghosh, Supriyo
    Ong, Priscilla
    Chan, Yu En
    Bee, Yong Mong
    Krishnaswamy, Pavitra
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 4673 - 4684
  • [34] Deep reinforcement learning for real-world quadrupedal locomotion: a comprehensive review
    Zhang, Hongyin
    He, Li
    Wang, Donglin
    INTELLIGENCE & ROBOTICS, 2022, 2 (03):
  • [35] Tackling Real-World Autonomous Driving using Deep Reinforcement Learning
    Maramotti, Paolo
    Capasso, Alessandro Paolo
    Bacchiani, Giulio
    Broggi, Alberto
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1274 - 1281
  • [36] Towards Real-World Deployment of Reinforcement Learning for Traffic Signal Control
    Mueller, Arthur
    Rangras, Vishal
    Ferfers, Tobias
    Hufen, Florian
    Schreckenberg, Lukas
    Jasperneite, Juergen
    Schnittker, Georg
    Waldmann, Michael
    Friesen, Maxim
    Wiering, Marco
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 507 - 514
  • [37] Real-world dexterous object manipulation based deep reinforcement learning
    Yao, Qingfeng
    Wang, Jilong
    Yang, Shuyu
    arXiv, 2021,
  • [38] Simulation-Based Reinforcement Learning for Real-World Autonomous Driving
    Osinski, Blazej
    Jakubowski, Adam
    Ziecina, Pawel
    Milos, Piotr
    Galias, Christopher
    Homoceanu, Silviu
    Michalewski, Henryk
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6411 - 6418
  • [39] Real-Sim-Real Transfer for Real-World Robot Control Policy Learning with Deep Reinforcement Learning
    Liu, Naijun
    Cai, Yinghao
    Lu, Tao
    Wang, Rui
    Wang, Shuo
    APPLIED SCIENCES-BASEL, 2020, 10 (05):
  • [40] Active Visual Object Search Using Affordance-Map in Real World : A Human-Centric Approach
    Piyathilaka, Lasitha
    Kodagoda, Sarath
    2014 13TH INTERNATIONAL CONFERENCE ON CONTROL AUTOMATION ROBOTICS & VISION (ICARCV), 2014, : 1427 - 1432