Layer-Specific Knowledge Distillation for Class Incremental Semantic Segmentation

被引:3
|
作者
Wang, Qilong [1 ]
Wu, Yiwen [1 ]
Yang, Liu [1 ]
Zuo, Wangmeng [2 ]
Hu, Qinghua [1 ,3 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
[3] Minist Educ Peoples Republ China, Engn Res Ctr City Intelligence & Digital Governanc, Beijing 100816, Peoples R China
关键词
Knowledge distillation; incremental learning; semantic segmentation;
D O I
10.1109/TIP.2024.3372448
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, class incremental semantic segmentation (CISS) towards the practical open-world setting has attracted increasing research interest, which is mainly challenged by the well-known issue of catastrophic forgetting. Particularly, knowledge distillation (KD) techniques have been widely studied to alleviate catastrophic forgetting. Despite the promising performance, existing KD-based methods generally use the same distillation schemes for different intermediate layers to transfer old knowledge, while employing manually tuned and fixed trade-off weights to control the effect of KD. These KD-based methods take no consideration of feature characteristics from different intermediate layers, limiting the effectiveness of KD for CISS. In this paper, we propose a layer-specific knowledge distillation (LSKD) method to assign appropriate knowledge schemes and weights for various intermediate layers by considering feature characteristics, aiming to further explore the potential of KD in improving the performance of CISS. Specifically, we present a mask-guided distillation (MD) to alleviate the background shift on semantic features, which performs distillation by masking the features affected by the background. Furthermore, a mask-guided context distillation (MCD) is presented to explore global context information lying in high-level semantic features. Based on them, our LSKD assigns different distillation schemes according to feature characteristics. To adjust the effect of layer-specific distillation adaptively, LSKD introduces a regularized gradient equilibrium method to learn dynamic trade-off weights. Additionally, our LSKD makes an attempt to simultaneously learn distillation schemes and trade-off weights of different layers by developing a bi-level optimization method. Extensive experiments on widely used Pascal VOC 12 and ADE20K show our LSKD clearly outperforms its counterparts while achieving state-of-the-art results.
引用
收藏
页码:1977 / 1989
页数:13
相关论文
共 50 条
  • [1] Decomposed Knowledge Distillation for Class-Incremental Semantic Segmentation
    Baek, Donghyeon
    Oh, Youngmin
    Lee, Sanghoon
    Lee, Junghyup
    Ham, Bumsub
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Knowledge distillation for incremental learning in semantic segmentation
    Michieli, Umberto
    Zanuttigh, Pietro
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 205
  • [3] Incrementer: Transformer for Class-Incremental Semantic Segmentation with Knowledge Distillation Focusing on Old Class
    Shang, Chao
    Li, Hongliang
    Meng, Fanman
    Wu, Qingbo
    Qiu, Heqian
    Wang, Lanxiao
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7214 - 7224
  • [4] Dual attention-guided distillation for class incremental semantic segmentation
    Xu, Pengju
    Wang, Yan
    Wang, Bingye
    Zhao, Haiying
    APPLIED INTELLIGENCE, 2025, 55 (07)
  • [5] Class Similarity Weighted Knowledge Distillation for Continual Semantic Segmentation
    Minh Hieu Phan
    The-Anh Ta
    Son Lam Phung
    Long Tran-Thanh
    Bouzerdoum, Abdesselam
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16845 - 16854
  • [6] Cs2K: Class-Specific and Class-Shared Knowledge Guidance for Incremental Semantic Segmentation
    Cong, Wei
    Cong, Yang
    Liu, Yuyang
    Sun, Gan
    COMPUTER VISION - ECCV 2024, PT V, 2025, 15063 : 244 - 261
  • [7] Structured Knowledge Distillation for Semantic Segmentation
    Liu, Yifan
    Chen, Ke
    Liu, Chris
    Qin, Zengchang
    Luo, Zhenbo
    Wang, Jingdong
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2599 - 2608
  • [8] Robust Semantic Segmentation by Redundant Networks With a Layer-Specific Loss Contribution and Majority Vote
    Baer, Andreas
    Klingner, Marvin
    Varghese, Serin
    Hueger, Fabian
    Schlicht, Peter
    Fingscheidt, Tim
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 1348 - 1358
  • [9] Class-Incremental Learning for Semantic Segmentation in Aerial Imagery via Distillation in All Aspects
    Shan, Lianlei
    Wang, Weiqiang
    Lv, Ke
    Luo, Bin
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [10] Inherit With Distillation and Evolve With Contrast: Exploring Class Incremental Semantic Segmentation Without Exemplar Memory
    Zhao, Danpei
    Yuan, Bo
    Shi, Zhenwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 11932 - 11947