SAMFlow: Eliminating Any Fragmentation in Optical Flow with Segment Anything Model

被引:0
|
作者
Zhou, Shili [1 ]
He, Ruian [1 ]
Tan, Weimin [1 ]
Yan, Bo [1 ]
机构
[1] Fudan Univ, Shanghai Key Lab Intelligent Informat Proc, Sch Comp Sci, Shanghai, Peoples R China
基金
上海市自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optical Flow Estimation aims to find the 2D dense motion field between two frames. Due to the limitation of model structures and training datasets, existing methods often rely too much on local clues and ignore the integrity of objects, resulting in fragmented motion estimation. Through theoretical analysis, we find the pre-trained large vision models are helpful in optical flow estimation, and we notice that the recently famous Segment Anything Model (SAM) demonstrates a strong ability to segment complete objects, which is suitable for solving the fragmentation problem. We thus propose a solution to embed the frozen SAM image encoder into FlowFormer to enhance object perception. To address the challenge of in-depth utilizing SAM in non-segmentation tasks like optical flow estimation, we propose an Optical Flow Task-Specific Adaption scheme, including a Context Fusion Module to fuse the SAM encoder with the optical flow context encoder, and a Context Adaption Module to adapt the SAM features for optical flow task with Learned Task-Specific Embedding. Our proposed SAMFlow model reaches 0.86/2.10 clean/final EPE and 3.55/12.32 EPE/F1-all on Sintel and KITTI-15 training set, surpassing Flowformer by 8.5%/9.9% and 13.2%/16.3%. Furthermore, our model achieves state-of-the-art performance on the Sintel and KITTI-15 benchmarks, ranking #1 among all two-frame methods on Sintel clean pass.
引用
收藏
页码:7695 / 7703
页数:9
相关论文
共 50 条
  • [31] Progressive Self-Prompting Segment Anything Model for Salient Object Detection in Optical Remote Sensing Images
    Zhang, Xiaoning
    Yu, Yi
    Li, Daqun
    Wang, Yuqing
    REMOTE SENSING, 2025, 17 (02)
  • [32] Superpixel-Guided Segment Anything Model for Liver Tumor Segmentation with Couinaud Segment Prompt
    Lyu, Fei
    Xu, Jingwen
    Zhu, Ye
    Wong, Grace Lai-Hung
    Yuen, Pong C.
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT VIII, 2024, 15008 : 678 - 688
  • [33] Evaluating segment anything model (SAM) on MRI scans of brain tumors
    Ali, Luqman
    Alnajjar, Fady
    Swavaf, Muhammad
    Elharrouss, Omar
    Abd-alrazaq, Alaa
    Damseh, Rafat
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [34] IRSAM: Advancing Segment Anything Model for Infrared Small Target Detection
    Zhang, Mingjin
    Wang, Yuchun
    Guo, Jie
    Li, Yunsong
    Gao, Xinbo
    Zhang, Jing
    COMPUTER VISION - ECCV 2024, PT LXVII, 2025, 15125 : 233 - 249
  • [35] Boosting Deep Unsupervised Edge Detection via Segment Anything Model
    Yang, Wenya
    Chen, Xiao-Diao
    Wu, Wen
    Qin, Hongshuai
    Yan, Kangming
    Mao, Xiaoyang
    Song, Haichuan
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (06) : 8961 - 8971
  • [36] EFFICIENT CUTTING TOOL WEAR SEGMENTATION BASED ON SEGMENT ANYTHING MODEL
    Li, Zongshuo
    Huo, Ding
    Meurer, Markus
    Bergs, Thomas
    PROCEEDINGS OF ASME 2024 19TH INTERNATIONAL MANUFACTURING SCIENCE AND ENGINEERING CONFERENCE, MSEC2024, VOL 2, 2024,
  • [37] Drilling rock image segmentation and analysis using segment anything model
    Shan, Liqun
    Liu, Yanchang
    Du, Ke
    Paul, Shovon
    Zhang, Xingli
    Hei, Xiali
    ADVANCES IN GEO-ENERGY RESEARCH, 2024, 12 (02): : 89 - 101
  • [38] Plug-and-play segment anything model improves nnUNet performance
    Li, Yunxiang
    Jing, Bowen
    Li, Zihan
    Wang, Jing
    Zhang, You
    MEDICAL PHYSICS, 2025, 52 (02) : 899 - 912
  • [39] Enhanced Segment Anything Model for Accurate White Blood Cell Segmentation
    Zang, Yu
    Su, Yang
    Hu, Jun
    ELECTRONICS LETTERS, 2025, 61 (01)
  • [40] GazeSAM: Interactive Image Segmentation with Eye Gaze and Segment Anything Model
    Wang, Bin
    Aboah, Armstrong
    Zhang, Zheyuan
    Pan, Hongyi
    Bagci, Ulas
    GAZE MEETS MACHINE LEARNING WORKSHOP, 2023, 226 : 254 - 264