A multimodal deep learning model for detecting endoscopic images of near-infrared fluorescence capsules

被引:0
|
作者
Wang, Junhao [1 ,2 ]
Zhou, Cheng [1 ]
Wang, Wei [1 ]
Zhang, Hanxiao [2 ]
Zhang, Amin [4 ]
Cui, Daxiang [1 ,3 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Inst & Med Robot, Shanghai 200240, Peoples R China
[3] Henan Univ, Med & Engn Cross Res Inst, Sch Med, Kaifeng 475004, Peoples R China
[4] Shanghai Jiao Tong Univ, Sch Agr & Biol, Dept Food Sci & Technol, Shanghai 200240, Peoples R China
来源
基金
上海市自然科学基金; 中国国家自然科学基金; 国家自然科学基金国际合作与交流项目; 中国博士后科学基金;
关键词
Fluorescence endoscopy; Multimodal deep learning; Disease detection; CANCER;
D O I
10.1016/j.bios.2025.117251
中图分类号
Q6 [生物物理学];
学科分类号
071011 ;
摘要
Early screening for gastrointestinal (GI) diseases is critical for preventing cancer development. With the rapid advancement of deep learning technology, artificial intelligence (AI) has become increasingly prominent in the early detection of GI diseases. Capsule endoscopy is a non-invasive medical imaging technique used to examine the gastrointestinal tract. In our previous work, we developed a near-infrared fluorescence capsule endoscope (NIRF-CE) capable of exciting and capturing near-infrared (NIR) fluorescence images to specifically identify subtle mucosal microlesions and submucosal abnormalities while simultaneously capturing conventional white- light images to detect lesions with significant morphological changes. However, limitations such as low camera resolution and poor lighting within the gastrointestinal tract may lead to misdiagnosis and other medical errors. Manually reviewing and interpreting large volumes of capsule endoscopy images is time-consuming and prone to errors. Deep learning models have shown potential in automatically detecting abnormalities in NIRF-CE images. This study focuses on an improved deep learning model called Retinex-Attention-YOLO (RAY), which is based on single-modality image data and built on the YOLO series of object detection models. RAY enhances the accuracy and efficiency of anomaly detection, especially under low-light conditions. To further improve detection performance, we also propose a multimodal deep learning model, Multimodal-Retinex-Attention-YOLO (MRAY), which combines both white-light and fluorescence image data. The dataset used in this study consists of images of pig stomachs captured by our NIRF-CE system, simulating the human GI tract. In conjunction with a targeted fluorescent probe, which accumulates at lesion sites and releases fluorescent signals for imaging when abnormalities are present, a bright spot indicates a lesion. The MRAY model achieved an impressive precision of 96.3%, outperforming similar object detection models. To further validate the model's performance, ablation experiments were conducted, and comparisons were made with publicly available datasets. MRAY shows great promise for the automated detection of GI cancers, ulcers, inflammations, and other medical conditions in clinical practice.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Analysis and dehazing of near-infrared images
    Yu, Jae Taeg
    Ra, Sung Woong
    JOURNAL OF THE KOREAN SOCIETY FOR AERONAUTICAL AND SPACE SCIENCES, 2016, 44 (01) : 33 - 39
  • [42] A deep learning multimodal fusion framework for wood species identification using near-infrared spectroscopy GADF and RGB image
    Pan, Xi
    Yu, Zhiming
    Yang, Zhong
    HOLZFORSCHUNG, 2023, 77 (11-12) : 816 - 827
  • [43] Unsupervised deep learning registration model for multimodal brain images
    Abbasi, Samaneh
    Mehdizadeh, Alireza
    Boveiri, Hamid Reza
    Mosleh Shirazi, Mohammad Amin
    Javidan, Reza
    Khayami, Raouf
    Tavakoli, Meysam
    JOURNAL OF APPLIED CLINICAL MEDICAL PHYSICS, 2023, 24 (11):
  • [44] Endoscopic near-infrared photoimmunotherapy in an orthotopic head and neck cancer model
    Okada, Ryuhei
    Furusawa, Aki
    Inagaki, Fuyuki
    Wakiyama, Hiroaki
    Kato, Takuya
    Okuyama, Shuhei
    Furumoto, Hideyuki
    Fukushima, Hiroshi
    Choyke, Peter L.
    Kobayashi, Hisataka
    CANCER SCIENCE, 2021, 112 (08) : 3041 - 3049
  • [45] Complementary chemometrics and deep learning for semantic segmentation of tall and wide visible and near-infrared spectral images of plants
    Mishra, Puneet
    Sadeh, Roy
    Bino, Ehud
    Polder, Gerrit
    Boer, Martin P.
    Rutledge, Douglas N.
    Herrmann, Ittai
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2021, 186
  • [46] Retraction Note: Deep learning architectures for land cover classification using red and near-infrared satellite images
    Anju Unnikrishnan
    V. Sowmya
    K. P. Soman
    Multimedia Tools and Applications, 2024, 83 (34) : 81899 - 81899
  • [47] Transfer Learning Based Deep Neural Network for Detecting Artefacts in Endoscopic Images
    Natarajan, Kirthika
    Balusamy, Sargunam
    INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2022, 13 (08) : 633 - 641
  • [48] Ultraluminous infrared galaxies: Atlas of near-infrared images
    Bushouse, HA
    Borne, KD
    Colina, L
    Lucas, RA
    Rowan-Robinson, M
    Baker, AC
    Clements, DL
    Lawrence, A
    Oliver, S
    ASTROPHYSICAL JOURNAL SUPPLEMENT SERIES, 2002, 138 (01): : 1 - 18
  • [49] Near-Infrared Fluorescence Imaging of Carotid Plaques in an Atherosclerotic Murine Model
    Wu, Xiaotian
    Daniel Ulumben, Amy
    Long, Steven
    Katagiri, Wataru
    Wilks, Moses Q.
    Yuan, Hushan
    Cortese, Brian
    Yang, Chengeng
    Kashiwagi, Satoshi
    Choi, Hak Soo
    Normandin, Marc D.
    El Fakhri, Georges
    Zaman, Raiyan T.
    BIOMOLECULES, 2021, 11 (12)
  • [50] Near-infrared fluorescence imaging of microcalcification in an animal model of breast cancer
    Lenkinski, RE
    Ahmed, M
    Zaheer, A
    Frangioni, JV
    Goldberg, SN
    ACADEMIC RADIOLOGY, 2003, 10 (10) : 1159 - 1164