Learning laparoscopic video shot classification for gynecological surgery

被引:43
|
作者
Petscharnig, Stefan [1 ]
Schoeffmann, Klaus [1 ]
机构
[1] Alpen Adria Univ Klagenfurt, Fak Tech Wissensch, Univ Str 65-67, A-9020 Klagenfurt, Austria
关键词
Video classification; Deep learning; Convolutional Neural Network; CONVOLUTIONAL NEURAL-NETWORKS; DEEP; RECOGNITION; TASKS;
D O I
10.1007/s11042-017-4699-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Videos of endoscopic surgery are used for education of medical experts, analysis in medical research, and documentation for everyday clinical life. Hand-crafted image descriptors lack the capabilities of a semantic classification of surgical actions and video shots of anatomical structures. In this work, we investigate how well single-frame convolutional neural networks (CNN) for semantic shot classification in gynecologic surgery work. Together with medical experts, we manually annotate hours of raw endoscopic gynecologic surgery videos showing endometriosis treatment and myoma resection of over 100 patients. The cleaned ground truth dataset comprises 9 h of annotated video material (from 111 different recordings). We use the well-known CNN architectures AlexNet and GoogLeNet and train these architectures for both, surgical actions and anatomy, from scratch. Furthermore, we extract high-level features from AlexNet with weights from a pre-trained model from the Caffe model zoo and feed them to an SVM classifier. Our evaluation shows that we reach an average recall of .697 and .515 for classification of anatomical structures and surgical actions respectively using off-the-shelf CNN features. Using GoogLeNet, we achieve a mean recall of .782 and .617 for classification of anatomical structures and surgical actions respectively. With AlexNet the achieved recall is .615 for anatomical structures and .469 for surgical action classification respectively. The main conclusion of our work is that advances in general image classification methods transfer to the domain of endoscopic surgery videos in gynecology. This is relevant as this domain is different from natural images, e.g. it is distinguished by smoke, reflections, or a limited amount of colors.
引用
收藏
页码:8061 / 8079
页数:19
相关论文
共 50 条
  • [1] Learning laparoscopic video shot classification for gynecological surgery
    Stefan Petscharnig
    Klaus Schöffmann
    Multimedia Tools and Applications, 2018, 77 : 8061 - 8079
  • [2] GYNECOLOGICAL LAPAROSCOPIC SURGERY: LEARNING CURVE
    Khanum, Zohra
    Khanum, Amna
    Aman-ur-Rehman
    ANNALS OF KING EDWARD MEDICAL UNIVERSITY LAHORE PAKISTAN, 2015, 21 (04): : 253 - 256
  • [3] Learning curve in Colorectal Laparoscopic Video Surgery
    不详
    REVISTA ARGENTINA DE RESIDENTES DE CIRUGIA, 2012, 17 : 11 - 12
  • [4] Advances in Gynecological Laparoscopic Surgery
    Chang, Wen-Chun
    Huang, Su-Cheng
    Sheu, Bor-Ching
    JOURNAL OF THE FORMOSAN MEDICAL ASSOCIATION, 2010, 109 (04) : 245 - 247
  • [5] Learning to Model Relationships for Zero-Shot Video Classification
    Gao, Junyu
    Zhang, Tianzhu
    Xu, Changsheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (10) : 3476 - 3491
  • [6] Deep Learning for Shot Classification in Gynecologic Surgery Videos
    Petscharnig, Stefan
    Schoeffmann, Klaus
    MULTIMEDIA MODELING (MMM 2017), PT I, 2017, 10132 : 702 - 713
  • [7] Video shot segmentation and classification
    Gong, YH
    Liu, X
    15TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 1, PROCEEDINGS: COMPUTER VISION AND IMAGE ANALYSIS, 2000, : 860 - 863
  • [8] Shot classification in sports video
    Tong, XF
    Liu, QS
    Lu, HQ
    Jin, HL
    2004 7TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS, VOLS 1-3, 2004, : 1364 - 1367
  • [9] Laparoscopic surgery as standard in gynecological emergencies
    Schindelhauer, A.
    Wimberger, P.
    GYNAKOLOGE, 2014, 47 (03): : 179 - 183
  • [10] Learning Implicit Temporal Alignment for Few-shot Video Classification
    Zhang, Songyang
    Zhou, Jiale
    He, Xuming
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1309 - 1315