Multi-view knowledge distillation for efficient semantic segmentation

被引:4
|
作者
Wang, Chen [1 ]
Zhong, Jiang [1 ]
Dai, Qizhu [1 ]
Qi, Yafei [2 ]
Shi, Fengyuan [3 ]
Fang, Bin [1 ]
Li, Xue [4 ]
机构
[1] Chongqing Univ, Sch Comp Sci, Chongqing 400044, Peoples R China
[2] Cent South Univ, Sch Comp Sci & Engn, Changsha 410083, Peoples R China
[3] Northeastern Univ, Sch Informat Sci & Engn, Shenyang 110819, Peoples R China
[4] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia
基金
中国国家自然科学基金;
关键词
Multi-view learning; Knowledge distillation; Knowledge aggregation; Semantic segmentation; ENSEMBLE;
D O I
10.1007/s11554-023-01296-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current state-of-the-art semantic segmentation models achieve remarkable success in segmentation accuracy. However, the huge model size and computing cost restrict their applications on low-latency online systems or devices. Knowledge distillation has been one popular solution for compressing large-scale segmentation models, which train a small segmentation model from a large teacher model. However, one teacher model's knowledge may be insufficiently diverse to train an accurate student model. Meanwhile, the student model may inherit bias from the teacher model. This paper proposes a multi-view knowledge distillation framework called MVKD for efficient semantic segmentation. MVKD could aggregate the multi-view knowledge from multiple teacher models and transfer the multi-view knowledge to the student model. In MVKD, we introduce one multi-view co-tuning strategy to acquire uniformity among the multi-view knowledge in features from different teachers. In addition, we propose a multi-view feature distillation loss and a multi-view output distillation loss to transfer the multi-view knowledge in the features and outputs from multiple teachers to the student. We evaluate the proposed MVKD on three benchmark datasets, Cityscapes, CamVid, and Pascal VOC 2012. Experimental results demonstrate the effectiveness of the proposed MVKD in compressing semantic segmentation models.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Multi-view knowledge distillation for efficient semantic segmentation
    Chen Wang
    Jiang Zhong
    Qizhu Dai
    Yafei Qi
    Fengyuan Shi
    Bin Fang
    Xue Li
    Journal of Real-Time Image Processing, 2023, 20
  • [2] Multi-View Radar Semantic Segmentation
    Ouaknine, Arthur
    Newson, Alasdair
    Perez, Patrick
    Tupin, Florence
    Rebut, Julien
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15651 - 15660
  • [3] MULTI-VIEW SEMANTIC TEMPORAL VIDEO SEGMENTATION
    Theodoridis, Thomas
    Tefas, Anastasios
    Pitas, Ioannis
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 3947 - 3951
  • [4] TransKD: Transformer Knowledge Distillation for Efficient Semantic Segmentation
    Liu, Ruiping
    Yang, Kailun
    Roitberg, Alina
    Zhang, Jiaming
    Peng, Kunyu
    Liu, Huayao
    Wang, Yaonan
    Stiefelhagen, Rainer
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (12) : 20933 - 20949
  • [5] MULTI-VIEW CONTRASTIVE LEARNING FOR ONLINE KNOWLEDGE DISTILLATION
    Yang, Chuanguang
    An, Zhulin
    Xu, Yongjun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3750 - 3754
  • [6] Learning Where to Classify in Multi-view Semantic Segmentation
    Riemenschneider, Hayko
    Bodis-Szomoru, Andras
    Weissenberg, Julien
    Van Gool, Luc
    COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 : 516 - 532
  • [7] Channel-spatial knowledge distillation for efficient semantic segmentation
    Karine, Ayoub
    Napoleon, Thibault
    Jridi, Maher
    PATTERN RECOGNITION LETTERS, 2024, 180 : 48 - 54
  • [8] Knowledge Distillation for Efficient Panoptic Semantic Segmentation: applied to agriculture
    Li, Maohui
    Hasltead, Michael
    McCool, Chris
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 4204 - 4211
  • [9] Video Event Extraction with Multi-View Interaction Knowledge Distillation
    Wei, Kaiwen
    Du, Runyan
    Jin, Li
    Liu, Jian
    Yin, Jianhua
    Zhang, Linhao
    Liu, Jintao
    Liu, Nayu
    Zhang, Jingyuan
    Guo, Zhi
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19224 - 19233
  • [10] Collaborative knowledge distillation for incomplete multi-view action prediction
    Kumar, Deepak
    Kumar, Chetan
    Shao, Ming
    IMAGE AND VISION COMPUTING, 2021, 107