Reducing Training Data Using Pre-Trained Foundation Models: A Case Study on Traffic Sign Segmentation Using the Segment Anything Model

被引:0
|
作者
Henninger, Sofia [1 ]
Kellner, Maximilian [1 ,2 ]
Rombach, Benedikt [1 ]
Reiterer, Alexander [1 ,2 ]
机构
[1] Fraunhofer Inst Phys Measurement Tech IPM, D-79110 Freiburg, Germany
[2] Albert Ludwigs Univ Freiburg, Dept Sustainable Syst Engn INATECH, D-79110 Freiburg, Germany
关键词
semantic segmentation; segment anything model; Mask R-CNN; training data reduction; traffic signs;
D O I
10.3390/jimaging10090220
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The utilization of robust, pre-trained foundation models enables simple adaptation to specific ongoing tasks. In particular, the recently developed Segment Anything Model (SAM) has demonstrated impressive results in the context of semantic segmentation. Recognizing that data collection is generally time-consuming and costly, this research aims to determine whether the use of these foundation models can reduce the need for training data. To assess the models' behavior under conditions of reduced training data, five test datasets for semantic segmentation will be utilized. This study will concentrate on traffic sign segmentation to analyze the results in comparison to Mask R-CNN: the field's leading model. The findings indicate that SAM does not surpass the leading model for this specific task, regardless of the quantity of training data. Nevertheless, a knowledge-distilled student architecture derived from SAM exhibits no reduction in accuracy when trained on data that have been reduced by 95%.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Using pre-trained models and graph convolution networks to find the causal relations among events in the Chinese financial text data
    Kai Hu
    Qing Li
    Jie Xie
    Yingyan Pu
    Ya Guo
    Multimedia Tools and Applications, 2024, 83 : 18699 - 18720
  • [42] LETS: A Label-Efficient Training Scheme for Aspect-Based Sentiment Analysis by Using a Pre-Trained Language Model
    Shim, Heereen
    Lowet, Dietwig
    Luca, Stijn
    Vanrumste, Bart
    IEEE ACCESS, 2021, 9 : 115563 - 115578
  • [43] Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging
    Hans Christian
    Derwin Suhartono
    Andry Chowanda
    Kamal Z. Zamli
    Journal of Big Data, 8
  • [44] A study on effective data preprocessing and augmentation method in diabetic retinopathy classification using pre-trained deep learning approaches
    Incir, Ramazan
    Bozkurt, Ferhat
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 12185 - 12208
  • [45] Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging
    Christian, Hans
    Suhartono, Derwin
    Chowanda, Andry
    Zamli, Kamal Z.
    JOURNAL OF BIG DATA, 2021, 8 (01)
  • [46] A study on effective data preprocessing and augmentation method in diabetic retinopathy classification using pre-trained deep learning approaches
    Ramazan İncir
    Ferhat Bozkurt
    Multimedia Tools and Applications, 2024, 83 : 12185 - 12208
  • [47] Feature selection of pre-trained shallow CNN using the QLESCA optimizer: COVID-19 detection as a case study
    Hamad, Qusay Shihab
    Samma, Hussein
    Suandi, Shahrel Azmin
    APPLIED INTELLIGENCE, 2023, 53 (15) : 18630 - 18652
  • [48] Feature selection of pre-trained shallow CNN using the QLESCA optimizer: COVID-19 detection as a case study
    Qusay Shihab Hamad
    Hussein Samma
    Shahrel Azmin Suandi
    Applied Intelligence, 2023, 53 : 18630 - 18652
  • [49] Deepening into the suitability of using pre-trained models of ImageNet against a lightweight convolutional neural network in medical imaging: an experimental study
    Alzubaidi, Laith
    Duan, Ye
    Al-Dujaili, Ayad
    Ibraheem, Ibraheem Kasim
    Alkenani, Ahmed H.
    Santamaria, Jose
    Fadhel, Mohammed A.
    Al-Shamma, Omran
    Zhang, Jinglan
    PEERJ COMPUTER SCIENCE, 2021, 7 : 1 - 27
  • [50] Crop monitoring using remote sensing land use and land change data: Comparative analysis of deep learning methods using pre-trained CNN models
    Peng, Min
    Liu, Yunxiang
    Khan, Asad
    Ahmed, Bilal
    Sarker, Subrata K.
    Ghadi, Yazeed Yasin
    Bhatti, Uzair Aslam
    Al-Razgan, Muna
    Ali, Yasser A.
    BIG DATA RESEARCH, 2024, 36