Reducing Training Data Using Pre-Trained Foundation Models: A Case Study on Traffic Sign Segmentation Using the Segment Anything Model

被引:0
|
作者
Henninger, Sofia [1 ]
Kellner, Maximilian [1 ,2 ]
Rombach, Benedikt [1 ]
Reiterer, Alexander [1 ,2 ]
机构
[1] Fraunhofer Inst Phys Measurement Tech IPM, D-79110 Freiburg, Germany
[2] Albert Ludwigs Univ Freiburg, Dept Sustainable Syst Engn INATECH, D-79110 Freiburg, Germany
关键词
semantic segmentation; segment anything model; Mask R-CNN; training data reduction; traffic signs;
D O I
10.3390/jimaging10090220
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The utilization of robust, pre-trained foundation models enables simple adaptation to specific ongoing tasks. In particular, the recently developed Segment Anything Model (SAM) has demonstrated impressive results in the context of semantic segmentation. Recognizing that data collection is generally time-consuming and costly, this research aims to determine whether the use of these foundation models can reduce the need for training data. To assess the models' behavior under conditions of reduced training data, five test datasets for semantic segmentation will be utilized. This study will concentrate on traffic sign segmentation to analyze the results in comparison to Mask R-CNN: the field's leading model. The findings indicate that SAM does not surpass the leading model for this specific task, regardless of the quantity of training data. Nevertheless, a knowledge-distilled student architecture derived from SAM exhibits no reduction in accuracy when trained on data that have been reduced by 95%.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Pre-trained artificial intelligence-aided analysis of nanoparticles using the segment anything model
    Gabriel A. A. Monteiro
    Bruno A. A. Monteiro
    Jefersson A. dos Santos
    Alexander Wittemann
    Scientific Reports, 15 (1)
  • [2] Kurdish Sign Language Recognition Using Pre-Trained Deep Learning Models
    Alsaud, Ali A.
    Yousif, Raghad Z.
    Aziz, Marwan. M.
    Kareem, Shahab W.
    Maho, Amer J.
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (06) : 1334 - 1344
  • [3] Hippocampus segmentation and classification for dementia analysis using pre-trained neural network models
    Priyanka, Ahana
    Ganesan, Kavitha
    BIOMEDICAL ENGINEERING-BIOMEDIZINISCHE TECHNIK, 2021, 66 (06): : 581 - 592
  • [4] A novel fusion framework for sequential data using pre-trained model
    Ruan, Tao
    Jin, Canghong
    Xu, Lei
    Ding, Jianchao
    Ying, Shengyu
    Wu, Minghui
    Li, Huanqiang
    IAENG International Journal of Computer Science, 2020, 47 (03) : 593 - 598
  • [5] Training Compact Models for Low Resource Entity Tagging using Pre-trained Language Models
    Izsak, Peter
    Guskin, Shira
    Wasserblat, Moshe
    FIFTH WORKSHOP ON ENERGY EFFICIENT MACHINE LEARNING AND COGNITIVE COMPUTING - NEURIPS EDITION (EMC2-NIPS 2019), 2019, : 44 - 47
  • [6] Chinese Grammatical Error Correction Using Pre-trained Models and Pseudo Data
    Wang, Hongfei
    Kurosawa, Michiki
    Katsumata, Satoru
    Mita, Masato
    Komachi, Mamoru
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (03)
  • [7] Generalist Vision Foundation Models for Medical Imaging: A Case Study of Segment Anything Model on Zero-Shot Medical Segmentation
    Shi, Peilun
    Qiu, Jianing
    Abaxi, Sai Mu Dalike
    Wei, Hao
    Lo, Frank P. -W.
    Yuan, Wu
    DIAGNOSTICS, 2023, 13 (11)
  • [8] A Comparative Study of Using Pre-trained Language Models for Toxic Comment Classification
    Zhao, Zhixue
    Zhang, Ziqi
    Hopfgartner, Frank
    WEB CONFERENCE 2021: COMPANION OF THE WORLD WIDE WEB CONFERENCE (WWW 2021), 2021, : 500 - 507
  • [9] Using Pre-Trained Language Models for Abstractive DBPEDIA Summarization: A Comparative Study
    Zahera, Hamada M.
    Vitiugin, Fedor
    Sherif, Mohamed Ahmed
    Castillo, Carlos
    Ngomo, Axel-Cyrille Ngonga
    KNOWLEDGE GRAPHS: SEMANTICS, MACHINE LEARNING, AND LANGUAGES, 2023, 56 : 19 - 37
  • [10] An Empirical Study on Robustness to Spurious Correlations using Pre-trained Language Models
    Tu, Lifu
    Lalwani, Garima
    Gella, Spandana
    He, He
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 621 - 633