DEPTH REMOVAL DISTILLATION FOR RGB-D SEMANTIC SEGMENTATION

被引:7
|
作者
Fang, Tiyu [1 ]
Liang, Zhen [1 ]
Shao, Xiuli [2 ]
Dong, Zihao [1 ]
Li, Jinping [1 ]
机构
[1] Univ Jinan, Sch Informat Sci & Engn, Jinan 250022, Peoples R China
[2] Nankai Univ, Coll Comp Sci, Tianjin 300350, Peoples R China
关键词
RGB-D semantic segmentation; convolutional neural networks; knowledge distillation;
D O I
10.1109/ICASSP43922.2022.9747767
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
RGB-D semantic segmentation is attracting wide attention due to its better performance than conventional RGB methods. However, most of RGB-D semantic segmentation methods need to acquire the real depth information for segmenting RGB images effectively. Therefore, it is extremely challenging to take full advantage of RGB-D semantic segmentation methods for segmenting RGB images without the depth input. To address this challenge, a general depth removal distillation method is proposed to remove depth dependence from RGB-D semantic segmentation model by knowledge distillation, which can be employed to any CNN-based segmentation network structure. Specifically, a depth-aware convolution is adopted to construct the teacher network for getting sufficient knowledge from RGB-D images. Then according to the structure consistency between depth-aware convolution and general convolution, the teacher network is used to transfer the learned knowledge to the student network with general convolutions by sharing parameters. Next, the student network makes up for the lack of depth in manner of learning by RGB images. Meantime, a Variable Temperature Cross Entropy (VTCE) loss function is proposed to further increase the accuracy of the student model by soft target distillation. Extensive experiments on NYUv2 and SUN RGB-D datasets demonstrate the superiority of our proposed approach.
引用
收藏
页码:2405 / 2409
页数:5
相关论文
共 50 条
  • [1] RGB×D: Learning depth-weighted RGB patches for RGB-D indoor semantic segmentation
    Cao, Jinming
    Leng, Hanchao
    Cohen-Or, Daniel
    Lischinski, Dani
    Chen, Ying
    Tu, Changhe
    Li, Yangyan
    [J]. Neurocomputing, 2021, 462 : 568 - 580
  • [2] RGB-D SEMANTIC SEGMENTATION: A REVIEW
    Hu, Yaosi
    Chen, Zhenzhong
    Lin, Weiyao
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW 2018), 2018,
  • [3] RGB-D Gate-guided edge distillation for indoor semantic segmentation
    Wenbin Zou
    Yingqing Peng
    Zhengyu Zhang
    Shishun Tian
    Xia Li
    [J]. Multimedia Tools and Applications, 2022, 81 : 35815 - 35830
  • [4] RGB-D Gate-guided edge distillation for indoor semantic segmentation
    Zou, Wenbin
    Peng, Yingqing
    Zhang, Zhengyu
    Tian, Shishun
    Li, Xia
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (25) : 35815 - 35830
  • [5] RGBxD: Learning depth-weighted RGB patches for RGB-D indoor semantic segmentation
    Cao, Jinming
    Leng, Hanchao
    Cohen-Or, Daniel
    Lischinski, Dani
    Chen, Ying
    Tu, Changhe
    Li, Yangyan
    [J]. NEUROCOMPUTING, 2021, 462 : 568 - 580
  • [6] Joining geometric and RGB features for RGB-D semantic segmentation
    Zhang, Shaopeng
    Zhong, Min
    Zeng, Gang
    Gan, Rui
    [J]. 2019 INTERNATIONAL CONFERENCE ON IMAGE AND VIDEO PROCESSING, AND ARTIFICIAL INTELLIGENCE, 2019, 11321
  • [7] 3D Neighborhood Convolution: Learning Depth-Aware Features for RGB-D and RGB Semantic Segmentation
    Chen, Yunlu
    Mensink, Thomas
    Gavves, Efstratios
    [J]. 2019 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2019), 2019, : 173 - 182
  • [8] 2.5D CONVOLUTION FOR RGB-D SEMANTIC SEGMENTATION
    Xing, Yajie
    Wang, Jingbo
    Chen, Xiaokang
    Zeng, Gang
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 1410 - 1414
  • [9] Depth-Aware CNN for RGB-D Segmentation
    Wang, Weiyue
    Neumann, Ulrich
    [J]. COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 144 - 161
  • [10] Learning Depth-Sensitive Conditional Random Fields for Semantic Segmentation of RGB-D Images
    Mueller, Andreas C.
    Behnke, Sven
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2014, : 6232 - 6237