Learning Invariant Representation of Tasks for Robust Surgical State Estimation

被引:4
|
作者
Qin, Yidan [1 ,2 ]
Allan, Max [1 ]
Yue, Yisong [2 ]
Burdick, Joel W. [2 ]
Azizian, Mahdi [1 ]
机构
[1] Intuit Surg Inc, Sunnyvale, CA 94086 USA
[2] CALTECH, Dept Mech & Civil Engn, Pasadena, CA 91125 USA
关键词
AI-Based Methods; deep learning methods; laparoscopy; medical robots and systems; surgical robotics;
D O I
10.1109/LRA.2021.3063014
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Surgical state estimators in robot-assisted surgery (RAS)-especially those trained via learning techniques-rely heavily on datasets that capture surgeon actions in laboratory or real-world surgical tasks. Real-world RAS datasets are costly to acquire, are obtained from multiple surgeons who may use different surgical strategies, and are recorded under uncontrolled conditions in highly complex environments. The combination of high diversity and limited data calls for new learning methods that are robust and invariant to operating conditions and surgical techniques. We propose StiseNet, a Surgical Task Invariance State Estimation Network with an invariance induction framework that minimizes the effects of variations in surgical technique and operating environments inherent to RAS datasets. StiseNet's adversarial architecture learns to separate nuisance factors from information needed for surgical state estimation. StiseNet is shown to outperform state-of-the-art state estimation methods on three datasets (including a new real-world RAS dataset: HERNIA-20).
引用
收藏
页码:3208 / 3215
页数:8
相关论文
共 50 条
  • [21] Learning Degradation-Invariant Representation for Robust Real-World Person Re-Identification
    Huang, Yukun
    Fu, Xueyang
    Li, Liang
    Zha, Zheng-Jun
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (11) : 2770 - 2796
  • [22] Deep contrastive representation learning for supervised tasks
    Duan, Chenguang
    Jiao, Yuling
    Kang, Lican
    Yang, Jerry Zhijian
    Zhou, Fusheng
    PATTERN RECOGNITION, 2025, 161
  • [23] Pareto Invariant Representation Learning for Multimedia Recommendation
    Huang, Shanshan
    Li, Haoxuan
    Li, Qingsong
    Zheng, Chunyuan
    Liu, Li
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 6410 - 6419
  • [24] Scale-invariant representation of machine learning
    Lee, Sungyeop
    Jo, Junghyo
    PHYSICAL REVIEW E, 2022, 105 (04)
  • [25] Learning discriminative and invariant representation for fingerprint retrieval
    Dehua SONG
    Ruilin LI
    Fandong ZHANG
    Jufu FENG
    ScienceChina(InformationSciences), 2019, 62 (01) : 220 - 222
  • [26] Unbiased Recommendation Through Invariant Representation Learning
    Tang, Min
    Zou, Lixin
    Cui, Shujie
    Liang, Shiuan-ni
    Jin, Zhe
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES-APPLIED DATA SCIENCE TRACK, PT X, ECML PKDD 2024, 2024, 14950 : 280 - 296
  • [27] Variational Invariant Representation Learning for Multimodal Recommendation
    Yang, Wei
    Zhang, Haoran
    Zhang, Li
    PROCEEDINGS OF THE 2024 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2024, : 752 - 760
  • [28] Fundamental Limits and Tradeoffs in Invariant Representation Learning
    Zhao, Han
    Dan, Chen
    Aragam, Bryon
    Jaakkola, Tommi S.
    Gordon, Geoffrey J.
    Ravikumar, Pradeep
    Journal of Machine Learning Research, 2022, 23
  • [29] Learning discriminative and invariant representation for fingerprint retrieval
    Song, Dehua
    Li, Ruilin
    Zhang, Fandong
    Feng, Jufu
    SCIENCE CHINA-INFORMATION SCIENCES, 2019, 62 (01)
  • [30] LEARNING A TEMPORALLY INVARIANT REPRESENTATION FOR VISUAL TRACKING
    Ma, Chao
    Yang, Xiaokang
    Zhang, Chongyang
    Yang, Ming-Hsuan
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 857 - 861