Reducing Training Data Using Pre-Trained Foundation Models: A Case Study on Traffic Sign Segmentation Using the Segment Anything Model

被引:0
|
作者
Henninger, Sofia [1 ]
Kellner, Maximilian [1 ,2 ]
Rombach, Benedikt [1 ]
Reiterer, Alexander [1 ,2 ]
机构
[1] Fraunhofer Inst Phys Measurement Tech IPM, D-79110 Freiburg, Germany
[2] Albert Ludwigs Univ Freiburg, Dept Sustainable Syst Engn INATECH, D-79110 Freiburg, Germany
关键词
semantic segmentation; segment anything model; Mask R-CNN; training data reduction; traffic signs;
D O I
10.3390/jimaging10090220
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The utilization of robust, pre-trained foundation models enables simple adaptation to specific ongoing tasks. In particular, the recently developed Segment Anything Model (SAM) has demonstrated impressive results in the context of semantic segmentation. Recognizing that data collection is generally time-consuming and costly, this research aims to determine whether the use of these foundation models can reduce the need for training data. To assess the models' behavior under conditions of reduced training data, five test datasets for semantic segmentation will be utilized. This study will concentrate on traffic sign segmentation to analyze the results in comparison to Mask R-CNN: the field's leading model. The findings indicate that SAM does not surpass the leading model for this specific task, regardless of the quantity of training data. Nevertheless, a knowledge-distilled student architecture derived from SAM exhibits no reduction in accuracy when trained on data that have been reduced by 95%.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study
    Tekiroglu, Serra Sinem
    Bonaldi, Helena
    Fanton, Margherita
    Guerini, Marco
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3099 - 3114
  • [32] Fake Listing or Truth? Using Pre-trained Deep Learning Model with Data Augmentation to Detect the Imposter
    Nuntachit, Nontakan
    Sugannasil, Prompong
    Sukhahuta, Rattasit
    ADVANCES IN INTELLIGENT NETWORKING AND COLLABORATIVE SYSTEMS, INCOS-2022, 2022, 527 : 105 - 116
  • [33] Downstream Task Performance of BERT Models Pre-Trained Using Automatically De-Identified Clinical Data
    Vakili, Thomas
    Lamproudis, Anastasios
    Henriksson, Aron
    Dalianis, Hercules
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 4245 - 4252
  • [34] Optical flow-based segmentation of moving objects for mobile robot navigation using pre-trained deep learning models
    Karoly, Artur I.
    Elek, Renata Nagyne
    Haidegger, Tamas
    Szell, Karoly
    Galambos, Peter
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 3080 - 3086
  • [35] Detection and Segmentation of Anomalous Traffic Signs Using Yolov8 and Segment Anything Model (SAM) for Indian Specific Conditions
    Varshney, Divya
    Pal, Tirtharaj
    Ghosh, Indrajit
    10TH INTERNATIONAL CONFERENCE ON ELECTRONICS, COMPUTING AND COMMUNICATION TECHNOLOGIES, CONECCT 2024, 2024,
  • [36] STORM DRAIN DETECTION AND LOCALISATION ON MOBILE LIDAR DATA USING A PRE-TRAINED RANDLA-NET SEMANTIC SEGMENTATION NETWORK
    Mattheuwsen, L.
    Bassier, M.
    Vergauwen, M.
    XXIV ISPRS CONGRESS IMAGING TODAY, FORESEEING TOMORROW, COMMISSION II, 2022, 43-B2 : 237 - 244
  • [37] Landslide Hazard Assessment in Highway Areas of Guangxi Using Remote Sensing Data and a Pre-Trained XGBoost Model
    Zhang, Yuze
    Deng, Lei
    Han, Ying
    Sun, Yunhua
    Zang, Yu
    Zhou, Minlu
    REMOTE SENSING, 2023, 15 (13)
  • [38] Detection of driver distraction in the Australian naturalistic driving study videos using pre-trained models and transfer learning
    Elhenawy, Mohammed
    Masoud, Mahmoud
    Haworth, Narelle
    Young, Kristie
    Rakotonirainy, Andry
    Grzebieta, Raphael
    Williamson, Ann
    TRANSPORTATION RESEARCH PART F-TRAFFIC PSYCHOLOGY AND BEHAVIOUR, 2023, 97 : 31 - 43
  • [39] What does the language system look like in pre-trained language models? A study using complex networks
    Zheng, Jianyu
    KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [40] A pre-trained language model for emergency department intervention prediction using routine physiological data and clinical narratives
    Huang, Ting-Yun
    Chong, Chee-Fah
    Lin, Heng-Yu
    Chen, Tzu-Ying
    Chang, Yung-Chun
    Lin, Ming-Chin
    INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2024, 191