Segment Anything Model Can Not Segment Anything: Assessing AI Foundation Model's Generalizability in Permafrost Mapping

被引:7
|
作者
Li, Wenwen [1 ]
Hsu, Chia-Yu [1 ]
Wang, Sizhe [1 ,2 ]
Yang, Yezhou [2 ]
Lee, Hyunho [1 ]
Liljedahl, Anna [3 ]
Witharana, Chandi [4 ]
Yang, Yili [3 ]
Rogers, Brendan M. [3 ]
Arundel, Samantha T. [5 ]
Jones, Matthew B. [6 ]
McHenry, Kenton [7 ]
Solis, Patricia [1 ]
机构
[1] Arizona State Univ, Sch Geog Sci & Urban Planning, Tempe, AZ 85287 USA
[2] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ 85287 USA
[3] Woodwell Climate Res Ctr, Falmouth, MA 02540 USA
[4] Univ Connecticut, Dept Nat Resources & Environm, Storrs, CT 06269 USA
[5] US Geol Survey, Ctr Excellence Geospatial Informat Sci, Rolla, MO 65401 USA
[6] Univ Calif Santa Barbara, Natl Ctr Ecol Anal & Synth, Santa Barbara, CA 93106 USA
[7] Univ Illinois, Natl Ctr Supercomp Applicat NCSA, Champaign, IL 61820 USA
基金
美国国家科学基金会;
关键词
foundation model; artificial intelligence; mapping; zero-shot; segmentation; GeoAI; GEOAI;
D O I
10.3390/rs16050797
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
This paper assesses trending AI foundation models, especially emerging computer vision foundation models and their performance in natural landscape feature segmentation. While the term foundation model has quickly garnered interest from the geospatial domain, its definition remains vague. Hence, this paper will first introduce AI foundation models and their defining characteristics. Built upon the tremendous success achieved by Large Language Models (LLMs) as the foundation models for language tasks, this paper discusses the challenges of building foundation models for geospatial artificial intelligence (GeoAI) vision tasks. To evaluate the performance of large AI vision models, especially Meta's Segment Anything Model (SAM), we implemented different instance segmentation pipelines that minimize the changes to SAM to leverage its power as a foundation model. A series of prompt strategies were developed to test SAM's performance regarding its theoretical upper bound of predictive accuracy, zero-shot performance, and domain adaptability through fine-tuning. The analysis used two permafrost feature datasets, ice-wedge polygons and retrogressive thaw slumps because (1) these landform features are more challenging to segment than man-made features due to their complicated formation mechanisms, diverse forms, and vague boundaries; (2) their presence and changes are important indicators for Arctic warming and climate change. The results show that although promising, SAM still has room for improvement to support AI-augmented terrain mapping. The spatial and domain generalizability of this finding is further validated using a more general dataset EuroCrops for agricultural field mapping. Finally, we discuss future research directions that strengthen SAM's applicability in challenging geospatial domains.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Plug-and-play segment anything model improves nnUNet performance
    Li, Yunxiang
    Jing, Bowen
    Li, Zihan
    Wang, Jing
    Zhang, You
    MEDICAL PHYSICS, 2025, 52 (02) : 899 - 912
  • [32] Enhanced Segment Anything Model for Accurate White Blood Cell Segmentation
    Zang, Yu
    Su, Yang
    Hu, Jun
    ELECTRONICS LETTERS, 2025, 61 (01)
  • [33] GazeSAM: Interactive Image Segmentation with Eye Gaze and Segment Anything Model
    Wang, Bin
    Aboah, Armstrong
    Zhang, Zheyuan
    Pan, Hongyi
    Bagci, Ulas
    GAZE MEETS MACHINE LEARNING WORKSHOP, 2023, 226 : 254 - 264
  • [34] Automated galaxy sizes in Euclid images using the Segment Anything Model
    Vega-Ferrero, J.
    Buitrago, F.
    Fernandez-Iglesias, J.
    Raji, S.
    Sahelices, B.
    Sanchez, H. Dominguez
    ASTRONOMY & ASTROPHYSICS, 2025, 693
  • [35] EyeSAM: Unveiling the Potential of Segment Anything Model in Ophthalmic Image Segmentation
    da Silva, Alan Sousa
    Naik, Gunjan
    Bagga, Pallavi
    Soornro, Taha
    Reis, Ana P. Ribeiro
    Zhang, Gongyu
    Waisberg, Ethan
    Kandakji, Lynn
    Liu, Siyin
    Fu, Dun Jack
    Woof, Wiliam
    Moghul, Ismail
    Balaskas, Konstantinos
    Pontikos, Nikolas
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2024, 65 (07)
  • [36] Tooth segmentation on multimodal images using adapted segment anything model
    Peijuan Wang
    Hanjie Gu
    Yuliang Sun
    Scientific Reports, 15 (1)
  • [37] An efficient fine tuning strategy of segment anything model for polyp segmentation
    Mingyan Wang
    Cun Xu
    Kefeng Fan
    Scientific Reports, 15 (1)
  • [38] SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model
    Zhou, Shili
    He, Ruian
    Tan, Weimin
    Yan, Bo
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 7695 - 7703
  • [39] Evaluation and Improvement of Segment Anything Model for Interactive Histopathology Image Segmentation
    Kim, SeungKyu
    Oh, Hyun-Jic
    Min, Seonghui
    Jeong, Won-Ki
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023 WORKSHOPS, 2023, 14393 : 245 - 255
  • [40] Data Efficiency of Segment Anything Model for Optic Disc and Cup Segmentation
    Yii, Fabian
    MacGillivray, Tom
    Bernabeu, Miguel O.
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2023 WORKSHOPS, 2023, 14394 : 336 - 346