Towards Real-World Applications of Personalized Anesthesia Using Policy Constraint Q Learning for Propofol Infusion Control

被引:0
|
作者
Cai, Xiuding [1 ,2 ]
Chen, Jiao [3 ,4 ,5 ]
Zhu, Yaoyao [1 ,2 ]
Wang, Beimin [1 ,2 ]
Yao, Yu [1 ,2 ]
机构
[1] Chinese Acad Sci, Chengdu Inst Comp applicat, Chengdu 610093, Peoples R China
[2] Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 101408, Peoples R China
[3] Sichuan Univ, West China Hosp, Dept Anesthesiol, Chengdu 610041, Peoples R China
[4] Res Units West China, Chengdu 610041, Peoples R China
[5] Chinese Acad Med Sci, Beijing 100006, Peoples R China
基金
中国国家自然科学基金;
关键词
anesthetic administration; offline reinforcement learning (ORL); propofol; Anesthesia; CLOSED-LOOP; SYSTEMS; METAANALYSIS; HYPNOSIS;
D O I
10.1109/JBHI.2023.3321099
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automated anesthesia promises to enable more personalized and precise anesthetic administration and free anesthesiologists from repetitive tasks, allowing them to focus on the most critical aspects of surgical care for patients. Current research has typically focused on creating simulated environments from which agents can learn. These approaches have demonstrated good experimental results, but are still far from clinical application. In this paper, Policy Constraint Q-Learning (PCQL), a data-driven reinforcement learning algorithm for solving the problem of learning strategies on real world anesthesia data, is proposed. Conservative Q-Learning is first introduced to alleviate the problem of Q function overestimation in an offline context. A policy constraint term is then added to agent training to keep the action distribution of the agent and the anesthesiologist consistent, ensuring the agent makes safer decisions in anesthesia scenarios. The effectiveness of PCQL was validated by extensive experiments on a real clinical anesthesia dataset we collected. Experimental results show that PCQL is predicted to achieve higher gains than the baseline approaches while maintaining good agreement with the reference dose given by the anesthesiologist, using less total dose, and being more responsive to the patient's vital signs. In addition, the confidence intervals of the agent were investigated and were founded to cover most of the clinical decisions of the anesthesiologist. Finally, an interpretable method, SHAP, was used to analyze the contributing components of the model predictions, increasing the transparency of the model.
引用
收藏
页码:459 / 469
页数:11
相关论文
共 20 条
  • [1] Towards Efficient, Personalized Anesthesia using Continuous Reinforcement Learning for Propofol Infusion Control
    Lowery, Cristobal
    Faisal, A. Aldo
    [J]. 2013 6TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING (NER), 2013, : 1414 - 1417
  • [2] Towards Real-World Deployment of Reinforcement Learning for Traffic Signal Control
    Mueller, Arthur
    Rangras, Vishal
    Ferfers, Tobias
    Hufen, Florian
    Schreckenberg, Lukas
    Jasperneite, Juergen
    Schnittker, Georg
    Waldmann, Michael
    Friesen, Maxim
    Wiering, Marco
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 507 - 514
  • [3] Real-Sim-Real Transfer for Real-World Robot Control Policy Learning with Deep Reinforcement Learning
    Liu, Naijun
    Cai, Yinghao
    Lu, Tao
    Wang, Rui
    Wang, Shuo
    [J]. APPLIED SCIENCES-BASEL, 2020, 10 (05):
  • [4] Perceptual Learning: Policy Insights From Basic Research to Real-World Applications
    Seitz, Aaron R.
    Sekuler, Allison
    Dosher, Barbara
    Wright, Beverly A.
    Huang, Chang-Bing
    Green, C. Shawn
    Pack, Christopher C.
    Sagi, Dov
    Levi, Dennis
    Tadin, Duje
    Quinlan, Elizabeth
    Jiang, Fang
    Diaz, Gabriel J.
    Ghose, Geoffrey
    Fiser, Jozsef
    Banai, Karen
    Visscher, Kristina
    Huxlin, Krystel
    Shams, Ladan
    Battelli, Lorella
    Carrasco, Marisa
    Herzog, Michael
    Webster, Michael
    Eckstein, Miguel
    Turk-Browne, Nicholas B.
    Censor, Nitzan
    De Weerd, Peter
    Vogels, Rufin
    Hochstein, Shaul
    Watanabe, Takeo
    Sasaki, Yuka
    Polat, Uri
    Lu, Zhong-Lin
    Kourtzi, Zoe
    [J]. POLICY INSIGHTS FROM THE BEHAVIORAL AND BRAIN SCIENCES, 2023, 10 (02) : 324 - 332
  • [5] machine learning applications using real-world data: A literature review
    Adair, Nicholas
    Icten, Zeynep
    Friedman, Mark
    Menzin, Joseph
    [J]. PHARMACOEPIDEMIOLOGY AND DRUG SAFETY, 2020, 29 : 339 - 339
  • [6] Towards Real-World Federated Learning in Medical Image Analysis Using Kaapana
    Kades, Klaus
    Scherer, Jonas
    Zenk, Maximilian
    Kempf, Marius
    Maier-Hein, Klaus
    [J]. DISTRIBUTED, COLLABORATIVE, AND FEDERATED LEARNING, AND AFFORDABLE AI AND HEALTHCARE FOR RESOURCE DIVERSE GLOBAL HEALTH, DECAF 2022, FAIR 2022, 2022, 13573 : 130 - 140
  • [7] First steps towards real-world traffic signal control optimisation by reinforcement learning
    Meess, Henri
    Gerner, Jeremias
    Hein, Daniel
    Schmidtner, Stefanie
    Elger, Gordon
    Bogenberger, Klaus
    [J]. JOURNAL OF SIMULATION, 2024,
  • [8] Towards Distributed Communication and Control in Real-World Multi-Agent Reinforcement Learning
    Liu, Jieyan
    Liu, Yi
    Du, Zhekai
    Lu, Ke
    [J]. IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 4974 - 4979
  • [9] Teaching Real-World Applications of Business Statistics Using Communication to Scaffold Learning
    Green, Gareth P.
    Jones, Stacey
    Bean, John C.
    [J]. BUSINESS AND PROFESSIONAL COMMUNICATION QUARTERLY, 2015, 78 (03) : 314 - 335
  • [10] A Review of Machine Learning Classification Using Quantum Annealing for Real-World Applications
    Nath R.K.
    Thapliyal H.
    Humble T.S.
    [J]. SN Computer Science, 2021, 2 (5)