Safe Reinforcement Learning for Sepsis Treatment

被引:5
|
作者
Jia, Yan [1 ]
Burden, John [1 ]
Lawton, Tom [2 ,3 ]
Habli, Ibrahim [1 ]
机构
[1] Univ York, Dept Comp Sci, York, England
[2] Bradford Royal Infirm, Bradford, England
[3] Bradford Inst Hlth Res, Bradford, England
关键词
Sepsis treatment; Reinforcement learning; Safe policy; SEPTIC SHOCK; FLUIDS;
D O I
10.1109/ICHI48887.2020.9374367
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sepsis, a life-threatening illness, is estimated to be the primary cause of death for 50,000 people a year in the UK and many more worldwide. Managing the treatment of sepsis is very challenging as it is frequently missed at an early stage and the optimal treatment is not yet clear. There are promising attempts to use Reinforcement Learning (RL) to learn optimal strategies to treat sepsis patients, especially for the administration of intravenous fluids and vasopressors. However, some RL agents only take the current state of patients into account when recommending the dosage of vasopressors. This is inconsistent with clinical safety practice in which the dosage of vasopressors is increased or decreased gradually. A sudden major change of the dosage might cause significant harm to patients and as such is considered unsafe in sepsis treatment. In this paper, we have adapted one of the deep RL methods published previously and evaluated whether the learned policy contains these sudden, major changes when recommending the vasopressor dosage. Then, we have modified this method to address the above safety constraint and learnt a safer policy by incorporating current clinical knowledge and practice.
引用
收藏
页码:108 / 114
页数:7
相关论文
共 50 条
  • [1] Individualized sepsis treatment using reinforcement learning
    Suchi Saria
    [J]. Nature Medicine, 2018, 24 : 1641 - 1642
  • [2] Individualized sepsis treatment using reinforcement learning
    Saria, Suchi
    [J]. NATURE MEDICINE, 2018, 24 (11) : 1641 - 1642
  • [3] Deep Inverse Reinforcement Learning for Sepsis Treatment
    Yu, Chao
    Ren, Guoqi
    Liu, Jiming
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2019, : 373 - 375
  • [4] Reinforcement Learning with Balanced Clinical Reward for Sepsis Treatment
    Lu, Zhilin
    Liu, Jingming
    Luo, Ruihong
    Li, Chunping
    [J]. ARTIFICIAL INTELLIGENCE IN MEDICINE, PT I, AIME 2024, 2024, 14844 : 161 - 171
  • [5] On Normative Reinforcement Learning via Safe Reinforcement Learning
    Neufeld, Emery A.
    Bartocci, Ezio
    Ciabattoni, Agata
    [J]. PRIMA 2022: PRINCIPLES AND PRACTICE OF MULTI-AGENT SYSTEMS, 2023, 13753 : 72 - 89
  • [6] Reinforcement Learning For Sepsis Treatment: A Continuous Action Space Solution
    Huang, Yong
    Cao, Rui
    Rahmani, Amir
    [J]. MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 182, 2022, 182 : 631 - 647
  • [7] Optimizing sepsis treatment strategies via a reinforcement learning model
    Zhang, Tianyi
    Qu, Yimeng
    Wang, Deyong
    Zhong, Ming
    Cheng, Yunzhang
    Zhang, Mingwei
    [J]. BIOMEDICAL ENGINEERING LETTERS, 2024, 14 (02) : 279 - 289
  • [8] Optimizing sepsis treatment strategies via a reinforcement learning model
    Tianyi Zhang
    Yimeng Qu
    Deyong wang
    Ming Zhong
    Yunzhang Cheng
    Mingwei Zhang
    [J]. Biomedical Engineering Letters, 2024, 14 : 279 - 289
  • [9] Reinforcement Learning With Safe Exploration for Adaptive Plasma Cancer Treatment
    Hou, Zichao
    Lee, Taeyoung
    Keidar, Michael
    [J]. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2022, 6 (04) : 482 - 492
  • [10] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Horie, Naoto
    Matsui, Tohgoroh
    Moriyama, Koichi
    Mutoh, Atsuko
    Inuzuka, Nobuhiro
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2019, 24 (03) : 352 - 359