Layer-Specific Knowledge Distillation for Class Incremental Semantic Segmentation

被引:3
|
作者
Wang, Qilong [1 ]
Wu, Yiwen [1 ]
Yang, Liu [1 ]
Zuo, Wangmeng [2 ]
Hu, Qinghua [1 ,3 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
[3] Minist Educ Peoples Republ China, Engn Res Ctr City Intelligence & Digital Governanc, Beijing 100816, Peoples R China
关键词
Knowledge distillation; incremental learning; semantic segmentation;
D O I
10.1109/TIP.2024.3372448
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, class incremental semantic segmentation (CISS) towards the practical open-world setting has attracted increasing research interest, which is mainly challenged by the well-known issue of catastrophic forgetting. Particularly, knowledge distillation (KD) techniques have been widely studied to alleviate catastrophic forgetting. Despite the promising performance, existing KD-based methods generally use the same distillation schemes for different intermediate layers to transfer old knowledge, while employing manually tuned and fixed trade-off weights to control the effect of KD. These KD-based methods take no consideration of feature characteristics from different intermediate layers, limiting the effectiveness of KD for CISS. In this paper, we propose a layer-specific knowledge distillation (LSKD) method to assign appropriate knowledge schemes and weights for various intermediate layers by considering feature characteristics, aiming to further explore the potential of KD in improving the performance of CISS. Specifically, we present a mask-guided distillation (MD) to alleviate the background shift on semantic features, which performs distillation by masking the features affected by the background. Furthermore, a mask-guided context distillation (MCD) is presented to explore global context information lying in high-level semantic features. Based on them, our LSKD assigns different distillation schemes according to feature characteristics. To adjust the effect of layer-specific distillation adaptively, LSKD introduces a regularized gradient equilibrium method to learn dynamic trade-off weights. Additionally, our LSKD makes an attempt to simultaneously learn distillation schemes and trade-off weights of different layers by developing a bi-level optimization method. Extensive experiments on widely used Pascal VOC 12 and ADE20K show our LSKD clearly outperforms its counterparts while achieving state-of-the-art results.
引用
收藏
页码:1977 / 1989
页数:13
相关论文
共 50 条
  • [41] Semantic Segmentation of Medical Images Based on Knowledge Distillation Algorithm
    Liu, Hanqing
    Li, Fang
    Yang, Jingyi
    Wang, Xiaotian
    Han, Junling
    Wei, Jin
    Kang, Xiaodong
    12TH ASIAN-PACIFIC CONFERENCE ON MEDICAL AND BIOLOGICAL ENGINEERING, VOL 1, APCMBE 2023, 2024, 103 : 180 - 196
  • [42] Domain Adaptive Knowledge Distillation for Driving Scene Semantic Segmentation
    Kothandaraman, Divya
    Nambiar, Athira
    Mittal, Anurag
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS (WACVW 2021), 2021, : 134 - 143
  • [43] Dual attention-guided distillation for class incremental semantic segmentationDual attention-guided distillation for class incremental semantic segmentationP. Xu et al.
    Pengju Xu
    Yan Wang
    Bingye Wang
    Haiying Zhao
    Applied Intelligence, 2025, 55 (7)
  • [44] SemCKD: Semantic Calibration for Cross-Layer Knowledge Distillation
    Wang, Can
    Chen, Defang
    Mei, Jian-Ping
    Zhang, Yuan
    Feng, Yan
    Chen, Chun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (06) : 6305 - 6319
  • [45] Class similarity weighted knowledge distillation for few shot incremental learning
    Akmel, Feidu
    Meng, Fanman
    Wu, Qingbo
    Chen, Shuai
    Zhang, Runtong
    Assefa, Maregu
    NEUROCOMPUTING, 2024, 584
  • [46] Weakly-supervised Incremental learning for Semantic segmentation with Class Hierarchy
    Kim, Hyoseo
    Choe, Junsuk
    PATTERN RECOGNITION LETTERS, 2024, 182 : 31 - 38
  • [47] Continual Learning for Class- and Domain-Incremental Semantic Segmentation
    Kalb, Tobias
    Roschani, Masoud
    Ruf, Miriam
    Beyerer, Juergen
    2021 32ND IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2021, : 1345 - 1351
  • [48] Discriminative Gradient Adjustment with Coupled Knowledge Distillation for Class Incremental Learning
    Zhang, Hao
    Hu, Yanxu
    Peng, Jiawen
    Ma, Andy J.
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 822 - 827
  • [49] Semantic Segmentation Optimization Algorithm Based on Knowledge Distillation and Model Pruning
    Yao, Weiwei
    Zhang, Jie
    Li, Chen
    Li, Shiyun
    He, Li
    Zhang, Bo
    2019 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2019), 2019, : 261 - 265
  • [50] Region-aware mutual relational knowledge distillation for semantic segmentation
    Zheng, Haowen
    Lin, Xuxin
    Liang, Hailun
    Zhou, Benjia
    Liang, Yanyan
    PATTERN RECOGNITION, 2025, 161