Temporal Transductive Inference for Few-Shot Video Object Segmentation

被引:0
|
作者
Siam, Mennatullah [1 ]
机构
[1] Univ British Columbia, Comp Sci, Vancouver, BC, Canada
关键词
Few-shot learning; Transductive inference; Video object segmentation;
D O I
10.1007/s11263-025-02390-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot video object segmentation (FS-VOS) aims at segmenting video frames using a few labelled examples of classes not seen during initial training. In this paper, we present a simple but effective temporal transductive inference (TTI) approach that leverages temporal consistency in the unlabelled video frames during few-shot inference without episodic training. Key to our approach is the use of a video-level temporal constraint that augments frame-level constraints. The objective of the video-level constraint is to learn consistent linear classifiers for novel classes across the image sequence. It acts as a spatiotemporal regularizer during the transductive inference to increase temporal coherence and reduce overfitting on the few-shot support set. Empirically, our approach outperforms state-of-the-art meta-learning approaches in terms of mean intersection over union on YouTube-VIS by 2.5%. In addition, we introduce an improved benchmark dataset that is exhaustively labelled (i.e., all object occurrences are labelled, unlike the currently available). Our empirical results and temporal consistency analysis confirm the added benefits of the proposed spatiotemporal regularizer to improve temporal coherence. Our code and benchmark dataset is publicly available at, https://github.com/MSiam/tti_fsvos/.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Few-shot video object segmentation with prototype evolution
    Mao, Binjie
    Liu, Xiyan
    Shi, Linsu
    Yu, Jiazhong
    Li, Fei
    Xiang, Shiming
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (10): : 5367 - 5382
  • [2] Few-shot video object segmentation with prototype evolution
    Binjie Mao
    Xiyan Liu
    Linsu Shi
    Jiazhong Yu
    Fei Li
    Shiming Xiang
    Neural Computing and Applications, 2024, 36 : 5367 - 5382
  • [3] Multi-grained Temporal Prototype Learning for Few-shot Video Object Segmentation
    Liu, Nian
    Nan, Kepan
    Zhao, Wangbo
    Liu, Yuanwei
    Yao, Xiwen
    Khan, Salman
    Cholakkal, Hisham
    Anwer, Rao Muhammad
    Han, Junwei
    Khan, Fahad Shahbaz
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 18816 - 18825
  • [4] Inductive and Transductive Few-Shot Video Classification via Appearance and Temporal Alignments
    Nguyen, Khoi D.
    Quoc-Huy Tran
    Khoi Nguyen
    Binh-Son Hua
    Rang Nguyen
    COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 471 - 487
  • [5] Exploring the Better Correlation for Few-Shot Video Object Segmentation
    Luo, Naisong
    Wang, Yuan
    Sun, Rui
    Xiong, Guoxin
    Zhang, Tianzhu
    Wu, Feng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2133 - 2146
  • [6] Few-Shot Video Object Detection
    Fan, Qi
    Tang, Chi-Keung
    Tai, Yu-Wing
    COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 76 - 98
  • [7] Holistic Prototype Attention Network for Few-Shot Video Object Segmentation
    Tang, Yin
    Chen, Tao
    Jiang, Xiruo
    Yao, Yazhou
    Xie, Guo-Sen
    Shen, Heng-Tao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (08) : 6699 - 6709
  • [8] STTMC: A Few-Shot Spatial Temporal Transductive Modulation Classifier
    Shi, Yunhao
    Xu, Hua
    Qi, Zisen
    Zhang, Yue
    Wang, Dan
    Jiang, Lei
    IEEE Transactions on Machine Learning in Communications and Networking, 2024, 2 : 546 - 559
  • [9] Attentional prototype inference for few-shot segmentation
    Sun, Haoliang
    Lu, Xiankai
    Wang, Haochen
    Yin, Yilong
    Zhen, Xiantong
    Snoek, Cees G. M.
    Shao, Ling
    PATTERN RECOGNITION, 2023, 142
  • [10] Temporal Aggregation with Context Focusing for Few-Shot Video Object Detection
    Han, Wentao
    Lei, Jie
    Wang, Fahong
    Feng, Zunlei
    Liang, Ronghua
    Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics, 2023, : 2196 - 2201