STASIS: Reinforcement Learning Simulators for Human-Centric Real-World Environments

被引:0
|
作者
Efstathiadis, Georgios [1 ]
Emedom-Nnamdi, Patrick [1 ]
Kolbeinsson, Arinbjorn [2 ]
Onnela, Jukka-Pekka [1 ]
Lu, Junwei [1 ]
机构
[1] Harvard TH Chan Sch Publ Hlth, Dept Biostat, Boston, MA 02115 USA
[2] Evidat Hlth, London, England
关键词
reinforcement learning; health care; real-world simulators;
D O I
10.1007/978-3-031-39539-0_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present on-going work toward building Stasis, a suite of reinforcement learning (RL) environments that aim to maintain realism for human-centric agents operating in real-world settings. Through representation learning and alignment with real-world offline data, Stasis allows for the evaluation of RL algorithms in offline environments with adjustable characteristics, such as observability, heterogeneity and levels of missing data. We aim to introduce environments the encourage training RL agents that are capable of maintaining a level of performance and robustness comparable to agents trained in real-world online environments, while avoiding the high cost and risks associated with making mistakes during online training. We provide examples of two environments that will be part of Stasis and discuss its implications for the deployment of RL-based systems in sensitive and high-risk areas of application.
引用
收藏
页码:85 / 92
页数:8
相关论文
共 50 条
  • [1] Real-World Reinforcement Learning via Multifidelity Simulators
    Cutler, Mark
    Walsh, Thomas J.
    How, Jonathan P.
    IEEE TRANSACTIONS ON ROBOTICS, 2015, 31 (03) : 655 - 671
  • [2] Human-centric dialog training via offline reinforcement learning
    Jaques, Natasha
    Shen, Judy Hanwen
    Ghandehariouni, Asma
    Ferguson, Craig
    Lapedriza, Agata
    Jones, Noah
    Gu, Shixiang Shane
    Picard, Rosalind
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3985 - 4003
  • [3] Real-World Human-Robot Collaborative Reinforcement Learning
    Shafti, Ali
    Tjomsland, Jonas
    Dudley, William
    Faisal, A. Aldo
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 11161 - 11166
  • [4] Non-blocking Asynchronous Training for Reinforcement Learning in Real-World Environments
    Bohm, Peter
    Pounds, Pauline
    Chapman, Archie C.
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10927 - 10934
  • [5] Human-centric computing and intelligent environments Preface
    Gomez, Carles
    Prati, Andrea
    Aghajan, Hamid
    Augusto, Juan Carlos
    JOURNAL OF AMBIENT INTELLIGENCE AND SMART ENVIRONMENTS, 2016, 8 (04) : 377 - 377
  • [6] Real-world humanoid locomotion with reinforcement learning
    Radosavovic, Ilija
    Xiao, Tete
    Zhang, Bike
    Darrell, Trevor
    Malik, Jitendra
    Sreenath, Koushil
    SCIENCE ROBOTICS, 2024, 9 (89)
  • [7] Intrinsically motivated reinforcement learning for human-robot interaction in the real-world
    Qureshi, Ahmed Hussain
    Nakamura, Yutaka
    Yoshikawa, Yuichiro
    Ishiguro, Hiroshi
    NEURAL NETWORKS, 2018, 107 : 23 - 33
  • [8] Reinforcement Learning in Robotics: Applications and Real-World Challenges
    Kormushev, Petar
    Calinon, Sylvain
    Caldwell, Darwin G.
    ROBOTICS, 2013, 2 (03): : 122 - 148
  • [9] Human-Centric Justification of Machine Learning Predictions
    Biran, Or
    McKeown, Kathleen
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1461 - 1467
  • [10] Learning by Exploration: New Challenges in Real-World Environments
    Wu, Qingyun
    Wang, Huazheng
    Wang, Hongning
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3575 - 3576