Improving Semantic Segmentation via Efficient Self-Training

被引:38
|
作者
Zhu, Yi [1 ]
Zhang, Zhongyue [2 ]
Wu, Chongruo [3 ]
Zhang, Zhi [1 ]
He, Tong [1 ]
Zhang, Hang [4 ]
Manmatha, R. [1 ]
Li, Mu [1 ]
Smola, Alexander [1 ]
机构
[1] Amazon Web Serv, Santa Clara, CA 95054 USA
[2] Snapchat, Sunnyvale, CA 94085 USA
[3] Univ Calif Davis, Davis, CA 95616 USA
[4] Facebook, Menlo Pk, CA 94025 USA
基金
澳大利亚研究理事会;
关键词
Training; Semantics; Computational modeling; Image segmentation; Data models; Schedules; Predictive models; Semantic segmentation; semi-supervised learning; self-training; fast training schedule; cross-domain generalization;
D O I
10.1109/TPAMI.2021.3138337
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Starting from the seminal work of Fully Convolutional Networks (FCN), there has been significant progress on semantic segmentation. However, deep learning models often require large amounts of pixelwise annotations to train accurate and robust models. Given the prohibitively expensive annotation cost of segmentation masks, we introduce a self-training framework in this paper to leverage pseudo labels generated from unlabeled data. In order to handle the data imbalance problem of semantic segmentation, we propose a centroid sampling strategy to uniformly select training samples from every class within each epoch. We also introduce a fast training schedule to alleviate the computational burden. This enables us to explore the usage of large amounts of pseudo labels. Our Centroid Sampling based Self-Training framework (CSST) achieves state-of-the-art results on Cityscapes and CamVid datasets. On PASCAL VOC 2012 test set, our models trained with the original train set even outperform the same models trained on the much bigger augmented train set. This indicates the effectiveness of CSST when there are fewer annotations. We also demonstrate promising few-shot generalization capability from Cityscapes to BDD100K and from Cityscapes to Mapillary datasets.
引用
收藏
页码:1589 / 1602
页数:14
相关论文
共 50 条
  • [11] Learning from Future: A Novel Self-Training Framework for Semantic Segmentation
    Du, Ye
    Shen, Yujun
    Wang, Haochen
    Fei, Jingjing
    Li, Wei
    Wu, Liwei
    Zhao, Rui
    Fu, Zehua
    Liu, Qingjie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [12] Bidirectional Self-Training with Multiple Anisotropic Prototypes for Domain Adaptive Semantic Segmentation
    Lu, Yulei
    Luo, Yawei
    Zhang, Li
    Li, Zheyang
    Yang, Yi
    Xiao, Jun
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1405 - 1415
  • [13] Domain Adaptive Semantic Segmentation via Entropy-Ranking and Uncertain Learning-Based Self-Training
    Peng, Chengli
    Ma, Jiayi
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2022, 9 (08) : 1524 - 1527
  • [14] Domain Adaptive Semantic Segmentation via Entropy-Ranking and Uncertain Learning-Based Self-Training
    Chengli Peng
    Jiayi Ma
    IEEE/CAA Journal of Automatica Sinica, 2022, 9 (08) : 1524 - 1527
  • [15] Self-training for Cell Segmentation and Counting
    Luo, J.
    Oore, S.
    Hollensen, P.
    Fine, A.
    Trappenberg, T.
    ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11489 : 406 - 412
  • [16] Multilevel Self-Training Approach for Cross-Domain Semantic Segmentation in Intelligent Vehicles
    Chen, Yung-Yao
    Jhong, Sin-Ye
    IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2024, 16 (01) : 148 - 161
  • [17] Active self-training for weakly supervised 3D scene semantic segmentation
    Liu, Gengxin
    van Kaick, Oliver
    Huang, Hui
    Hu, Ruizhen
    COMPUTATIONAL VISUAL MEDIA, 2024, 10 (06) : 1063 - 1078
  • [18] A Three-Stage Self-Training Framework for Semi-Supervised Semantic Segmentation
    Ke, Rihuan
    Aviles-Rivero, Angelica, I
    Pandey, Saurabh
    Reddy, Saikumar
    Schonlieb, Carola-Bibiane
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1805 - 1815
  • [19] Active self-training for weakly supervised 3D scene semantic segmentation
    Gengxin Liu
    Oliver van Kaick
    Hui Huang
    Ruizhen Hu
    Computational Visual Media, 2024, 10 : 425 - 438
  • [20] DAST: Unsupervised Domain Adaptation in Semantic Segmentation Based on Discriminator Attention and Self-Training
    Yu, Fei
    Zhang, Mo
    Dong, Hexin
    Hu, Sheng
    Dong, Bin
    Zhang, Li
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10754 - 10762