Reducing Training Data Using Pre-Trained Foundation Models: A Case Study on Traffic Sign Segmentation Using the Segment Anything Model

被引:0
|
作者
Henninger, Sofia [1 ]
Kellner, Maximilian [1 ,2 ]
Rombach, Benedikt [1 ]
Reiterer, Alexander [1 ,2 ]
机构
[1] Fraunhofer Inst Phys Measurement Tech IPM, D-79110 Freiburg, Germany
[2] Albert Ludwigs Univ Freiburg, Dept Sustainable Syst Engn INATECH, D-79110 Freiburg, Germany
关键词
semantic segmentation; segment anything model; Mask R-CNN; training data reduction; traffic signs;
D O I
10.3390/jimaging10090220
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The utilization of robust, pre-trained foundation models enables simple adaptation to specific ongoing tasks. In particular, the recently developed Segment Anything Model (SAM) has demonstrated impressive results in the context of semantic segmentation. Recognizing that data collection is generally time-consuming and costly, this research aims to determine whether the use of these foundation models can reduce the need for training data. To assess the models' behavior under conditions of reduced training data, five test datasets for semantic segmentation will be utilized. This study will concentrate on traffic sign segmentation to analyze the results in comparison to Mask R-CNN: the field's leading model. The findings indicate that SAM does not surpass the leading model for this specific task, regardless of the quantity of training data. Nevertheless, a knowledge-distilled student architecture derived from SAM exhibits no reduction in accuracy when trained on data that have been reduced by 95%.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] STORM DRAIN DETECTION AND LOCALISATION ON MOBILE LIDAR DATA USING A PRE-TRAINED RANDLA-NET SEMANTIC SEGMENTATION NETWORK
    Mattheuwsen, L.
    Bassier, M.
    Vergauwen, M.
    XXIV ISPRS CONGRESS IMAGING TODAY, FORESEEING TOMORROW, COMMISSION II, 2022, 43-B2 : 237 - 244
  • [32] Landslide Hazard Assessment in Highway Areas of Guangxi Using Remote Sensing Data and a Pre-Trained XGBoost Model
    Zhang, Yuze
    Deng, Lei
    Han, Ying
    Sun, Yunhua
    Zang, Yu
    Zhou, Minlu
    REMOTE SENSING, 2023, 15 (13)
  • [33] Detection of driver distraction in the Australian naturalistic driving study videos using pre-trained models and transfer learning
    Elhenawy, Mohammed
    Masoud, Mahmoud
    Haworth, Narelle
    Young, Kristie
    Rakotonirainy, Andry
    Grzebieta, Raphael
    Williamson, Ann
    TRANSPORTATION RESEARCH PART F-TRAFFIC PSYCHOLOGY AND BEHAVIOUR, 2023, 97 : 31 - 43
  • [34] What does the language system look like in pre-trained language models? A study using complex networks
    Zheng, Jianyu
    KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [35] A pre-trained language model for emergency department intervention prediction using routine physiological data and clinical narratives
    Huang, Ting-Yun
    Chong, Chee-Fah
    Lin, Heng-Yu
    Chen, Tzu-Ying
    Chang, Yung-Chun
    Lin, Ming-Chin
    INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2024, 191
  • [36] Mining Medication-Effect Relations from Twitter Data Using Pre-trained Transformer Language Model
    Jiang, Keyuan
    Zhang, Dingkai
    Bernard, Gordon R.
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, PT II, 2021, 1525 : 468 - 478
  • [37] A Framework for Effective Knowledge Extraction from A Data Space Formed by Unstructured Technical Reports using Pre-trained Models
    Zhang, Jian
    Qin, Bo
    Zhang, Yufei
    Zhou, Junhua
    Wang, Hongwei
    2021 IEEE INTERNATIONAL CONFERENCE ON E-BUSINESS ENGINEERING (ICEBE 2021), 2021, : 120 - 125
  • [38] Heterogeneous data-based information retrieval using a fine-tuned pre-trained BERT language model
    Shaik, Amjan
    Saxena, Surabhi
    Gupta, Manisha
    Parveen, Nikhat
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (21) : 59537 - 59559
  • [39] Improving seismic fault mapping through data conditioning using a pre-trained deep convolutional neural network: A case study on Groningen field
    Otchere, Daniel Asante
    Tackie-Otoo, Bennet Nii
    Mohammad, Mohammad Abdalla Ayoub
    Ganat, Tarek Omar Arbi
    Kuvakin, Nikita
    Miftakhov, Ruslan
    Efremov, Igor
    Bazanov, Andrey
    JOURNAL OF PETROLEUM SCIENCE AND ENGINEERING, 2022, 213
  • [40] Using pre-trained models and graph convolution networks to find the causal relations among events in the Chinese financial text data
    Hu, Kai
    Li, Qing
    Xie, Jie
    Pu, Yingyan
    Guo, Ya
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (07) : 18699 - 18720