Projection based inverse reinforcement learning for the analysis of dynamic treatment regimes

被引:0
|
作者
Syed Ihtesham Hussain Shah
Giuseppe De Pietro
Giovanni Paragliola
Antonio Coronato
机构
[1] Parthenope University,Department of ICT and Engineering
[2] National Research Council,ICAR
[3] Università Telematica Giustino Fortunato,undefined
来源
Applied Intelligence | 2023年 / 53卷
关键词
Inverse Reinforcement Learning (IRL); Dynamic Treatment Regime (DTR); Reinforcement Learning (RL); Decision Support System (DSS);
D O I
暂无
中图分类号
学科分类号
摘要
Dynamic Treatment Regimes (DTRs) are adaptive treatment strategies that allow clinicians to personalize dynamically the treatment for each patient based on their step-by-step response to their treatment. There are a series of predefined alternative treatments for each disease and any patient may associate with one of these treatments according to his/her demographics. DTRs for a certain disease are studied and evaluated by means of statistical approaches where patients are randomized at each step of the treatment and their responses are observed. Recently, the Reinforcement Learning (RL) paradigm has also been applied to determine DTRs. However, such approaches may be limited by the need to design a true reward function, which may be difficult to formalize when the expert knowledge is not well assessed, as when the DTR is in the design phase. To address this limitation, an extension of the RL paradigm, namely Inverse Reinforcement Learning (IRL), has been adopted to learn the reward function from data, such as those derived from DTR trials. In this paper, we define a Projection Based Inverse Reinforcement Learning (PB-IRL) approach to learn the true underlying reward function for given demonstrations (DTR trials). Such a reward function can be used both to evaluate the set of DTRs determined for a certain disease, as well as to enable an RL-based intelligent agent to self-learn the best way and then act as a decision support system for the clinician.
引用
收藏
页码:14072 / 14084
页数:12
相关论文
共 50 条
  • [31] Inverse Reinforcement Learning in Tracking Control Based on Inverse Optimal Control
    Xue, Wenqian
    Kolaric, Patrik
    Fan, Jialu
    Lian, Bosen
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (10) : 10570 - 10581
  • [32] Efficient Potential-based Exploration in Reinforcement Learning using Inverse Dynamic Bisimulation Metric
    Wang, Yiming
    Yang, Ming
    Dong, Renzhi
    Sun, Binbin
    Liu, Furui
    Hou, Leong U.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [33] A behavior fusion method based on inverse reinforcement learning
    Shi, Haobin
    Li, Jingchen
    Chen, Shicong
    Hwang, Kao-Shing
    INFORMATION SCIENCES, 2022, 609 : 429 - 444
  • [34] Online Observer-Based Inverse Reinforcement Learning
    Self, Ryan
    Coleman, Kevin
    Bai, He
    Kamalapurkar, Rushikesh
    2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 1959 - 1964
  • [35] Online Observer-Based Inverse Reinforcement Learning
    Self, Ryan
    Coleman, Kevin
    Bai, He
    Kamalapurkar, Rushikesh
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (06): : 1922 - 1927
  • [36] Inverse Reinforcement Learning for Identification in Linear-Quadratic Dynamic Games
    Koepf, Florian
    Inga, Jairo
    Rothfufss, Simon
    Flad, Michael
    Hohmann, Soeren
    IFAC PAPERSONLINE, 2017, 50 (01): : 14902 - 14908
  • [37] Dynamic Regimes for Corporate Human Capital Development Used Reinforcement Learning Methods
    Orlova, Ekaterina V.
    MATHEMATICS, 2023, 11 (18)
  • [38] Adversarial Cooperative Imitation Learning for Dynamic Treatment Regimes
    Wang, Lu
    Yu, Wenchao
    Cheng, Wei
    Min, Martin Renqiang
    Zong, Bo
    He, Xiaofeng
    Zha, Hongyuan
    Wang, Wei
    Chen, Haifeng
    WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 1785 - 1795
  • [39] Preference-learning based Inverse Reinforcement Learning for Dialog Control
    Sugiyama, Hiroaki
    Meguro, Toyomi
    Minami, Yasuhiro
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 222 - 225
  • [40] Stroke-Based Stylization Learning and Rendering with Inverse Reinforcement Learning
    Xie, Ning
    Zhao, Tingting
    Tian, Feng
    Zhang, Xiaohua
    Sugiyama, Masashi
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 2531 - 2537