DEPTH REMOVAL DISTILLATION FOR RGB-D SEMANTIC SEGMENTATION

被引:7
|
作者
Fang, Tiyu [1 ]
Liang, Zhen [1 ]
Shao, Xiuli [2 ]
Dong, Zihao [1 ]
Li, Jinping [1 ]
机构
[1] Univ Jinan, Sch Informat Sci & Engn, Jinan 250022, Peoples R China
[2] Nankai Univ, Coll Comp Sci, Tianjin 300350, Peoples R China
关键词
RGB-D semantic segmentation; convolutional neural networks; knowledge distillation;
D O I
10.1109/ICASSP43922.2022.9747767
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
RGB-D semantic segmentation is attracting wide attention due to its better performance than conventional RGB methods. However, most of RGB-D semantic segmentation methods need to acquire the real depth information for segmenting RGB images effectively. Therefore, it is extremely challenging to take full advantage of RGB-D semantic segmentation methods for segmenting RGB images without the depth input. To address this challenge, a general depth removal distillation method is proposed to remove depth dependence from RGB-D semantic segmentation model by knowledge distillation, which can be employed to any CNN-based segmentation network structure. Specifically, a depth-aware convolution is adopted to construct the teacher network for getting sufficient knowledge from RGB-D images. Then according to the structure consistency between depth-aware convolution and general convolution, the teacher network is used to transfer the learned knowledge to the student network with general convolutions by sharing parameters. Next, the student network makes up for the lack of depth in manner of learning by RGB images. Meantime, a Variable Temperature Cross Entropy (VTCE) loss function is proposed to further increase the accuracy of the student model by soft target distillation. Extensive experiments on NYUv2 and SUN RGB-D datasets demonstrate the superiority of our proposed approach.
引用
收藏
页码:2405 / 2409
页数:5
相关论文
共 50 条
  • [41] SCN: Switchable Context Network for Semantic Segmentation of RGB-D Images
    Lin, Di
    Zhang, Ruimao
    Ji, Yuanfeng
    Li, Ping
    Huang, Hui
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (03) : 1120 - 1131
  • [42] Zig-Zag Network for Semantic Segmentation of RGB-D Images
    Lin, Di
    Huang, Hui
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (10) : 2642 - 2655
  • [43] ADRNet-S*: Asymmetric depth registration network via contrastive knowledge distillation for RGB-D mirror segmentation
    Zhou, Wujie
    Cai, Yuqi
    Dong, Xiena
    Qiang, Fangfang
    Qiu, Weiwei
    [J]. INFORMATION FUSION, 2024, 108
  • [44] Semantic RGB-D Image Synthesis
    Li, Shijie
    Li, Rong
    Gall, Juergen
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 944 - 952
  • [45] Learning Common and Specific Features for RGB-D Semantic Segmentation with Deconvolutional Networks
    Wang, Jinghua
    Wang, Zhenhua
    Tao, Dacheng
    See, Simon
    Wang, Gang
    [J]. COMPUTER VISION - ECCV 2016, PT V, 2016, 9909 : 664 - 679
  • [46] Attention-Aware and Semantic-Aware Network for RGB-D Indoor Semantic Segmentation
    Duan, Li-Juan
    Sun, Qi-Chao
    Qiao, Yuan-Hua
    Chen, Jun-Cheng
    Cui, Guo-Qin
    [J]. Jisuanji Xuebao/Chinese Journal of Computers, 2021, 44 (02): : 275 - 291
  • [47] Cross-modal attention fusion network for RGB-D semantic segmentation
    Zhao, Qiankun
    Wan, Yingcai
    Xu, Jiqian
    Fang, Lijin
    [J]. NEUROCOMPUTING, 2023, 548
  • [48] PYRAMID-CONTEXT GUIDED FEATURE FUSION FOR RGB-D SEMANTIC SEGMENTATION
    Liu, Haoming
    Guo, Li
    Zhou, Zhongwen
    Zhang, Hanyuan
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (IEEE ICMEW 2022), 2022,
  • [49] RGB-D Semantic Segmentation for Indoor Modeling Using Deep Learning: A Review
    Rached, Ishraq
    Hajji, Rafika
    Landes, Tania
    [J]. RECENT ADVANCES IN 3D GEOINFORMATION SCIENCE, 3D GEOINFO 2023, 2024, : 587 - 604
  • [50] RGB-D joint modelling with scene geometric information for indoor semantic segmentation
    Liu, Hong
    Wu, Wenshan
    Wang, Xiangdong
    Qian, Yueliang
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (17) : 22475 - 22488