Deep learning framework for interpretable quality control of echocardiography video

被引:0
|
作者
Du, Liwei [1 ]
Xue, Wufeng [1 ]
Qi, Zhanru [2 ]
Shi, Zhongqing [2 ]
Guo, Guanjun [2 ]
Yang, Xin [1 ]
Ni, Dong [1 ]
Yao, Jing [2 ,3 ,4 ]
机构
[1] Shenzhen Univ, Med Sch, Sch Biomed Engn, Shenzhen, Peoples R China
[2] Nanjing Univ, Affiliated Hosp, Dept Ultrasound Med, Med Sch, Nanjing, Peoples R China
[3] Nanjing Univ, Affiliated Hosp, Med Imaging Ctr, Med Sch, Nanjing, Peoples R China
[4] Yizheng Hosp, Nanjing Drum Tower Hosp Grp, Yangzhou, Peoples R China
关键词
echocardiography video; multitask network; quality control; real-time; visualized explanation;
D O I
10.1002/mp.17722
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Background: Echocardiography (echo) has become an indispensable tool in modern cardiology, offering real-time imaging that helps clinicians evaluate heart function and identify abnormalities. Despite these advantages, the acquisition of high-quality echo is time-consuming, labor-intensive, and highly subjective. Purpose: The objective of this study is to introduce a comprehensive system for the automated quality control (QC) of echo videos. This system focuses on real-time monitoring of key imaging parameters, reducing the variability associated with manual QC processes. Methods: Our multitask network analyzes cardiac cycle integrity, anatomical structures (AS), depth, cardiac axis angle (CAA), and gain. The network consists of a shared convolutional neural network (CNN) backbone for spatial feature extraction, along with three additional modules: (1) a bidirectional long short-term memory (Bi-LSTM) phase analysis (PA) module for detecting cardiac cycles and QC targets; (2) an oriented object detection head for AS analysis and depth/CAA quantification; and (3) a classification head for gain analysis. The model was trained and tested on a dataset of 1331 echo videos. Through model inference, a comprehensive score is generated, offering easily interpretable insights. Results: The model achieved a mean average precision of 0.962 for AS detection, with PA yielding average frame errors of 1.603 +/-+/- 1.181 (end-diastolic) and 1.681 +/-+/- 1.332 (end-systolic). The gain classification model demonstrated robust performance (Area Under the Curve > 0.98), and the overall processing speed reached 112.4 frames per second. On 203 randomly collected echo videos, the model achieved a kappa coefficient of 0.79 for rating consistency compared to expert evaluations CONCLUSIONS: Given the model's performance on the clinical dataset and its consistency with expert evaluations, our results indicate that the model not only delivers real-time, interpretable quality scores but also demonstrates strong clinical reliability.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Improving Alzheimer Diagnoses With An Interpretable Deep Learning Framework: Including Neuropsychiatric Symptoms
    Liu, Shujuan
    Zheng, Yuanjie
    Li, Hongzhuang
    Pan, Minmin
    Fang, Zhicong
    Liu, Mengting
    Qiao, Yuchuan
    Pan, Ningning
    Jia, Weikuan
    Ge, Xinting
    NEUROSCIENCE, 2023, 531 : 86 - 98
  • [22] gaBERT: An Interpretable Pretrained Deep Learning Framework for Cancer Gene Marker Discovery
    Hou, Jiale
    Wang, Zikai
    Lu, Haoran
    Pang, Xinzhe
    Cai, Yunpeng
    BIOINFORMATICS RESEARCH AND APPLICATIONS, PT I, ISBRA 2024, 2024, 14954 : 395 - 407
  • [23] Development and validation of an interpretable deep learning framework for Alzheimer's disease classification
    Qiu, Shangran
    Joshi, Prajakta S.
    Miller, Matthew, I
    Xue, Chonghua
    Zhou, Xiao
    Karjadi, Cody
    Chang, Gary H.
    Joshi, Anant S.
    Dwyer, Brigid
    Zhu, Shuhan
    Kaku, Michelle
    Zhou, Yan
    Alderazi, Yazan J.
    Swaminathan, Arun
    Kedar, Sachin
    Saint-Hilaire, Marie-Helene
    Auerbach, Sanford H.
    Yuan, Jing
    Sartor, E. Alton
    Au, Rhoda
    Kolachalama, Vijaya B.
    BRAIN, 2020, 143 : 1920 - 1933
  • [24] GenNet framework: interpretable deep learning for predicting phenotypes from genetic data
    Arno van Hilten
    Steven A. Kushner
    Manfred Kayser
    M. Arfan Ikram
    Hieab H. H. Adams
    Caroline C. W. Klaver
    Wiro J. Niessen
    Gennady V. Roshchupkin
    Communications Biology, 4
  • [25] An interpretable deep learning framework identifies proteomic drivers of Alzheimer's disease
    Panizza, Elena
    Cerione, Richard A.
    FRONTIERS IN CELL AND DEVELOPMENTAL BIOLOGY, 2024, 12
  • [26] DEEP LEARNING BASED POWER CONTROL FOR QUALITY-DRIVEN WIRELESS VIDEO TRANSMISSIONS
    Ye, Chuang
    Gursoy, M. Cenk
    Velipasalar, Senem
    2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 574 - 578
  • [27] ENCODING COMPLEXITY CONTROL FOR LIVE VIDEO APPLICATIONS: AN INTERPRETABLE MACHINE LEARNING APPROACH
    Huang, Chao
    Peng, Zongju
    Chen, Fen
    Jiang, Qiuping
    Cui, Xin
    Jiang, Gangyi
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1456 - 1461
  • [28] Interpretable deep learning methods for multiview learning
    Wang, Hengkang
    Lu, Han
    Sun, Ju
    Safo, Sandra E.
    BMC BIOINFORMATICS, 2024, 25 (01)
  • [29] Framework for biometric iris recognition in video, by deep learning and quality assessment of the iris-pupil region
    Eduardo Garea-Llano
    Annette Morales-Gonzalez
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 : 6517 - 6529
  • [30] Framework for biometric iris recognition in video, by deep learning and quality assessment of the iris-pupil region
    Garea-Llano, Eduardo
    Morales-Gonzalez, Annette
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 14 (6) : 6517 - 6529