Spatio-temporal context based recurrent visual attention model for lymph node detection

被引:2
|
作者
Peng, Haixin [1 ]
Peng, Yinjun [1 ,2 ]
机构
[1] Shandong Univ Sci & Technol, Coll Comp Sci & Engn, Qingdao 266590, Peoples R China
[2] Shandong Univ Sci & Technol, Shandong Prov Key Lab Wisdom Min Informat Technol, Qingdao, Peoples R China
基金
中国国家自然科学基金;
关键词
biomedical image classification; false-positive reduction; mixture density networks; recurrent visual attention; CONVOLUTIONAL NEURAL-NETWORKS; AUTOMATIC DETECTION; SEGMENTATION; CNN;
D O I
10.1002/ima.22430
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
False-positive reduction is one of the most crucial components in an automated lymph nodes (LNs) detection task in volumetric computed tomography (CT) scans, which is a highly sought goal for cancer diagnosis and early treatment. In this article, treating the three-dimensional (3D) LN detection task as object detection on sequence problem, we propose a novel spatio-temporal context-based recurrent visual attention model (STRAM) for the LNs false positive reduction. We firstly extract the deep spatial features maps for two-dimensional LN patches from pre-trained Inception-V3 model. A new Gaussian kernel-based spatial attention method is then presented to extract the most discriminating spatial features for the corresponding center slices. Additionally, to combine the temporal information between 3D CT slices, we devise a novel "Siamese" mixture density networks which can learn to adaptively focus on the most relevant parts of the CT slices. Considering the lesion areas always locate around the centroid of the 3D CT scans, a hard constraint is imposed on the predicted attention locations with batch normalization technique and the Siamese architecture. The proposed model is a fully differentiable unit that can be optimized end-to-end by using stochastic gradient descent. The effectiveness of our method is verified on LN dataset: 388 mediastinal LNs labeled by radiologists in 90 patient CT scans, and 595 abdominal LNs in 86 patient CT scans. Our method demonstrates sensitivities of about 87%/82% at 3 FP/vol. and 93%/89% at 6 FP/vol. for mediastinum and abdomen, respectively, which compares favorably to previous methods.
引用
收藏
页码:1220 / 1242
页数:23
相关论文
共 50 条
  • [31] Combining Spatio-Temporal Context and Kalman Filtering for Visual Tracking
    Yang, Haoran
    Wang, Juanjuan
    Miao, Yi
    Yang, Yulu
    Zhao, Zengshun
    Wang, Zhigang
    Sun, Qian
    Wu, Dapeng Oliver
    MATHEMATICS, 2019, 7 (11)
  • [32] Spatio-temporal attention model for video content analysis
    Guironnet, M
    Guyader, N
    Pellerin, D
    Ladret, P
    2005 International Conference on Image Processing (ICIP), Vols 1-5, 2005, : 2989 - 2992
  • [33] Adaptive Spatio-Temporal Context Learning for Visual Target Tracking
    Marvasti-Zadeh, Seyed Mojtaba
    Ghanei-Yakhdan, Hossein
    Kasaei, Shohreh
    2017 10TH IRANIAN CONFERENCE ON MACHINE VISION AND IMAGE PROCESSING (MVIP), 2017, : 10 - 14
  • [34] A Characterization of Interactive Visual Data Stories With a Spatio-Temporal Context
    Mayer, Benedikt
    Steinhauer, Nastasja
    Preim, Bernhard
    Meuschke, Monique
    COMPUTER GRAPHICS FORUM, 2023, 42 (06)
  • [35] Robust Visual Tracking with Dual Spatio-Temporal Context Trackers
    Sun, Shiyan
    Zhang, Hong
    Yuan, Ding
    SEVENTH INTERNATIONAL CONFERENCE ON GRAPHIC AND IMAGE PROCESSING (ICGIP 2015), 2015, 9817
  • [36] Spatio-Temporal Attention LSTM Model for Flood Forecasting
    Ding, Yukai
    Zhu, Yuelong
    Wu, Yirui
    Feng, Jun
    Cheng, Zirun
    2019 INTERNATIONAL CONFERENCE ON INTERNET OF THINGS (ITHINGS) AND IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) AND IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) AND IEEE SMART DATA (SMARTDATA), 2019, : 458 - 465
  • [37] Spatio-temporal Attention Model for Tactile Texture Recognition
    Cao, Guanqun
    Zhou, Yi
    Bollegala, Danushka
    Luo, Shan
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 9896 - 9902
  • [38] Video anomaly detection based on attention and efficient spatio-temporal feature extraction
    Rahimpour, Seyed Mohammad
    Kazemi, Mohammad
    Moallem, Payman
    Safayani, Mehran
    VISUAL COMPUTER, 2024, 40 (10): : 6825 - 6841
  • [39] Spatio-Temporal Attention Deep Recurrent Q-Network for POMDPs
    Etchart, Mariano
    Ladosz, Pawel
    Mulvaney, David
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2019, PT I, 2019, 11804 : 98 - 105
  • [40] Oversaturated part-based visual tracking via spatio-temporal context learning
    Liu, Wei
    Li, Jicheng
    Shi, Zhiguang
    Chen, Xiaotian
    Chen, Xiao
    APPLIED OPTICS, 2016, 55 (25) : 6960 - 6968