Teacher-Student Mutual Training for Semi-Supervised Object Detection Based on PPYOLOE

被引:0
|
作者
Zhang G. [1 ]
Wei J. [1 ]
机构
[1] School of Electrical and Information Engineering, Tianjin University, Tianjin
基金
中国国家自然科学基金;
关键词
object detection; PPYOLOE; semi-supervised learning; teacher-student mutual training;
D O I
10.11784/tdxbz202302035
中图分类号
学科分类号
摘要
With the continuous advancements in deep learning,object-detection technology based on convolutional neural network has become a research hotspot in the field of computer vision. Currently,mainstream object-detection algorithms rely on supervised learning and training models on extensive labeled data. However,unlabeled data are easy to obtain,while labeled data are usually challenging,time-consuming,and labor-intensive to collect. This study proposed a semi-supervised object-detection(PPYOLOE-SSOD)algorithm based on teacher-student mutual training to easily obtain data annotations. First,the student and gradually improved teacher models were trained simultaneously. The teacher model was then used to filter high-quality pseudo labels,which guided students during model training and extracted information from unlabeled images. Further,the exponential average method was used in each iteration to update the teacher model parameters to reduce the instability of parameter transfer. In addition,different data-augmentation methods were introduced to enhance the anti-interference ability of the network. Finally,the unsupervised learning branch was added for the learning of unlabeled data,and the features predicted by the model were processed using an intensive learning method. By sorting the classification features predicted by the teacher model,high-quality features were automatically selected as the pseudo labels generated by the teacher model,thus avoiding the tedious post-processing of pseudo labels and improving the accuracy and training speed of the network. On the MS COCO dataset,the accuracy of the PPYOLOE is improved by 1.4%,1.6%,and 2.1% on 1%,5%,and 10% labeled datasets,respectively,using the semi-supervised learning method. Compared with other SSOD algorithms,PPYOLOE-SSOD achieves the highest accuracy. The source code is at https://github.com/ wjm202/PPYYOLOE-SSOD. © 2024 Beijing Institute of Technology. All rights reserved.
引用
收藏
页码:415 / 423
页数:8
相关论文
共 35 条
  • [1] Deng J,, Xuan X J, Wang W F,, Et al., A review of research on object detection based on deep learning [C], Journal of Physics Conference Series, pp. 12028-12067, (2020)
  • [2] Park H J, Kang J W,, Kim B G., SSFPN:Scale sequence(S2)feature-based feature pyramid network for object detection[J], Sensors, 23, 9, pp. 4432-4440, (2023)
  • [3] End-to-end object detection with transformers[C], Computer Vision-ECCV 2020:16th European Conference, pp. 213-229, (2020)
  • [4] Sohn K,, Berthelot D,, Carlini N,, Et al., Fixmatch:Simplifying semi-supervised learning with consistency and confidence[C], Advances in Neural Information Processing Systems, pp. 596-608, (2020)
  • [5] Zhou H Y, Et al., Dense teacher:Dense pseudo-labels for semi-supervised object detection[C], Computer Vision-ECCV 2022:17th European Conference, pp. 35-50, (2022)
  • [6] Zang Y H, Zhou K Y, Huang C, Et al., Semi-supervised and long-tailed object detection with CascadeMatch[J], International Journal of Computer Vision, 131, 3, pp. 1-15, (2023)
  • [7] Chen G B,, Choi W G, Et al., Learning efficient object detection models with knowledge distillation[C], Advances in Neural Information Processing Systems, pp. 1010-1022, (2017)
  • [8] Berthelot D, Carlini N, Goodfellow I, Et al., Mixmatch : A holistic approach to semi-supervised learning[C], Advances in Neural Information Processing Systems, pp. 980-994, (2019)
  • [9] Zhang X F,, Dai L W., Image enhancement based on rough set and fractional order differentiator[J], Fractal and Fractional, 6, 4, pp. 214-215, (2022)
  • [10] Yan H, Zhang J X,, Zhang X F., Injected infrared and visible image fusion via L<sub>1</sub> decomposition model and guided filtering[J], IEEE Transactions on Computational Imaging, 8, 3, pp. 162-173, (2022)