Projection based inverse reinforcement learning for the analysis of dynamic treatment regimes

被引:0
|
作者
Syed Ihtesham Hussain Shah
Giuseppe De Pietro
Giovanni Paragliola
Antonio Coronato
机构
[1] Parthenope University,Department of ICT and Engineering
[2] National Research Council,ICAR
[3] Università Telematica Giustino Fortunato,undefined
来源
Applied Intelligence | 2023年 / 53卷
关键词
Inverse Reinforcement Learning (IRL); Dynamic Treatment Regime (DTR); Reinforcement Learning (RL); Decision Support System (DSS);
D O I
暂无
中图分类号
学科分类号
摘要
Dynamic Treatment Regimes (DTRs) are adaptive treatment strategies that allow clinicians to personalize dynamically the treatment for each patient based on their step-by-step response to their treatment. There are a series of predefined alternative treatments for each disease and any patient may associate with one of these treatments according to his/her demographics. DTRs for a certain disease are studied and evaluated by means of statistical approaches where patients are randomized at each step of the treatment and their responses are observed. Recently, the Reinforcement Learning (RL) paradigm has also been applied to determine DTRs. However, such approaches may be limited by the need to design a true reward function, which may be difficult to formalize when the expert knowledge is not well assessed, as when the DTR is in the design phase. To address this limitation, an extension of the RL paradigm, namely Inverse Reinforcement Learning (IRL), has been adopted to learn the reward function from data, such as those derived from DTR trials. In this paper, we define a Projection Based Inverse Reinforcement Learning (PB-IRL) approach to learn the true underlying reward function for given demonstrations (DTR trials). Such a reward function can be used both to evaluate the set of DTRs determined for a certain disease, as well as to enable an RL-based intelligent agent to self-learn the best way and then act as a decision support system for the clinician.
引用
收藏
页码:14072 / 14084
页数:12
相关论文
共 50 条
  • [41] Bidding strategy evolution analysis based on multi-task inverse reinforcement learning
    Tang, Qinghu
    Guo, Hongye
    Chen, Qixin
    Electric Power Systems Research, 2022, 212
  • [42] Bidding strategy evolution analysis based on multi-task inverse reinforcement learning
    Tang, Qinghu
    Guo, Hongye
    Chen, Qixin
    ELECTRIC POWER SYSTEMS RESEARCH, 2022, 212
  • [43] Dynamic Service Composition Method Based on Zero-Sum Game Integrated Inverse Reinforcement Learning
    Yuan, Yuan
    Guo, Yuhan
    Ma, Wanqing
    IEEE ACCESS, 2023, 11 : 111897 - 111908
  • [44] Repeated Inverse Reinforcement Learning
    Amin, Kareem
    Jiang, Nan
    Singh, Satinder
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [45] Cooperative Inverse Reinforcement Learning
    Hadfield-Menell, Dylan
    Dragan, Anca
    Abbeel, Pieter
    Russell, Stuart
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [46] Misspecification in Inverse Reinforcement Learning
    Skalse, Joar
    Abate, Alessandro
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 12, 2023, : 15136 - 15143
  • [47] Lifelong Inverse Reinforcement Learning
    Mendez, Jorge A.
    Shivkumar, Shashank
    Eaton, Eric
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [48] Bayesian Inverse Reinforcement Learning
    Ramachandran, Deepak
    Amir, Eyal
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 2586 - 2591
  • [49] Inverse Constrained Reinforcement Learning
    Malik, Shehryar
    Anwar, Usman
    Aghasi, Alireza
    Ahmed, Ali
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [50] Inverse reinforcement learning with evaluation
    da Silva, Valdinei Freire
    Reali Costa, Anna Helena
    Lima, Pedro
    2006 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-10, 2006, : 4246 - +