Selective experience replay compression using coresets for lifelong deep reinforcement learning in medical imaging

被引:0
|
作者
Zheng, Guangyao [1 ]
Zhou, Samson [1 ]
Braverman, Vladimir [1 ]
Jacobs, Michael A. [2 ,3 ]
Parekh, Vishwa S. [4 ]
机构
[1] Rice Univ, Dept Comp Sci, Houston, TX 77005 USA
[2] UTHealth Houston, McGovern Med Sch, Dept Diagnost & Intervent Imaging, Houston, TX USA
[3] Johns Hopkins Univ, Russell H Morgan Dept Radiol & Radiol Sci, Sch Med, Baltimore, MD 21205 USA
[4] Univ Maryland, Univ Maryland Med Intelligent Imaging UM2ii, Dept Diagnost Radiol & Nucl Med, Sch Med, Baltimore, MD 21201 USA
关键词
Deep reinforcement learning; lifelong learning; continual learning; medical imaging; coresets; clustering;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Selective experience replay is a popular strategy for integrating lifelong learning with deep reinforcement learning. Selective experience replay aims to recount selected experiences from previous tasks to avoid catastrophic forgetting. Furthermore, selective experience replay based techniques are model agnostic and allow experiences to be shared across different models. However, storing experiences from all previous tasks make lifelong learning using selective experience replay computationally very expensive and impractical as the number of tasks increase. To that end, we propose a reward distribution-preserving coreset compression technique for compressing experience replay buffers stored for selective experience replay. We evaluated the coreset lifelong deep reinforcement learning technique on the brain tumor segmentation (BRATS) dataset for the task of ventricle localization and on the whole-body MRI for localization of left knee cap, left kidney, right trochanter, left lung, and spleen. The coreset lifelong learning models trained on a sequence of 10 different brain MR imaging environments demonstrated excellent performance localizing the ventricle with a mean pixel error distance of 12.93, 13.46, 17.75, and 18.55 for the compression ratios of 10x, 20x, 30x, and 40x, respectively. In comparison, the conventional lifelong learning model localized the ventricle with a mean pixel distance of 10.87. Similarly, the coreset lifelong learning models trained on whole-body MRI demonstrated no significant difference (p=0.28) between the 10x compressed coreset lifelong learning models and conventional lifelong learning models for all the landmarks. The mean pixel distance for the 10x compressed models across all the landmarks was 25.30, compared to 19.24 for the conventional lifelong learning models. Our results demonstrate that the potential of the coreset-based ERB compression method for compressing experiences without a significant drop in performance.
引用
收藏
页码:1751 / 1764
页数:14
相关论文
共 50 条
  • [1] Selective Experience Replay for Lifelong Learning
    Isele, David
    Cosgun, Akansel
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3302 - 3309
  • [2] SELECTIVE EXPERIENCE REPLAY IN REINFORCEMENT LEARNING FOR REIDENTIFICATION
    Thakoor, Ninad
    Bhanu, Bir
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 4250 - 4254
  • [3] Deep Reinforcement Learning with Experience Replay Based on SARSA
    Zhao, Dongbin
    Wang, Haitao
    Shao, Kun
    Zhu, Yuanheng
    [J]. PROCEEDINGS OF 2016 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2016,
  • [4] Deep Reinforcement Learning With Quantum-Inspired Experience Replay
    Wei, Qing
    Ma, Hailan
    Chen, Chunlin
    Dong, Daoyi
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) : 9326 - 9338
  • [5] Associative Memory Based Experience Replay for Deep Reinforcement Learning
    Li, Mengyuan
    Kazemi, Arman
    Laguna, Ann Franchesca
    Hu, X. Sharon
    [J]. 2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [6] Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay
    Yin, Haiyan
    Pan, Sinno Jialin
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1640 - 1646
  • [7] Trial and Error Experience Replay Based Deep Reinforcement Learning
    Zhang, Cheng
    Ma, Liang
    [J]. 4TH IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2019) / 3RD INTERNATIONAL SYMPOSIUM ON REINFORCEMENT LEARNING (ISRL 2019), 2019, : 221 - 226
  • [8] Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning
    Foerster, Jakob
    Nardelli, Nantas
    Farquhar, Gregory
    Afouras, Triantafyllos
    Torr, Philip H. S.
    Kohli, Pushmeet
    Whiteson, Shimon
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [9] Experience Replay Optimization via ESMM for Stable Deep Reinforcement Learning
    Osei, Richard Sakyi
    Lopez, Daphne
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (01) : 715 - 723
  • [10] Invariant Transform Experience Replay: Data Augmentation for Deep Reinforcement Learning
    Lin, Yijiong
    Huang, Jiancong
    Zimmer, Matthieu
    Guan, Yisheng
    Rojas, Juan
    Weng, Paul
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 6615 - 6622