Towards Real-World Visual Tracking With Temporal Contexts

被引:23
|
作者
Cao, Ziang [1 ]
Huang, Ziyuan [2 ]
Pan, Liang [1 ]
Zhang, Shiwei [3 ]
Liu, Ziwei [1 ]
Fu, Changhong [4 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Natl Univ Singapore, Dept Mech Engn, Singapore 119077, Singapore
[3] DAMO Acad, Alibaba Grp, Hangzhou 310052, Zhejiang, Peoples R China
[4] Tongji Univ, Sch Mech Engn, Shanghai 201804, Peoples R China
基金
上海市自然科学基金; 中国国家自然科学基金;
关键词
Latency-aware evaluations; real-world tests; temporal contexts; two-level framework; visual tracking; PLUS PLUS; NETWORK;
D O I
10.1109/TPAMI.2023.3307174
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual tracking has made significant improvements in the past few decades. Most existing state-of-the-art trackers 1) merely aim for performance in ideal conditions while overlooking the real-world conditions; 2) adopt the tracking-by-detection paradigm, neglecting rich temporal contexts; 3) only integrate the temporal information into the template, where temporal contexts among consecutive frames are far from being fully utilized. To handle those problems, we propose a two-level framework (TCTrack) that can exploit temporal contexts efficiently. Based on it, we propose a stronger version for real-world visual tracking, i.e., TCTrack++. It boils down to two levels: features and similarity maps. Specifically, for feature extraction, we propose an attention-based temporally adaptive convolution to enhance the spatial features using temporal information, which is achieved by dynamically calibrating the convolution weights. For similarity map refinement, we introduce an adaptive temporal transformer to encode the temporal knowledge efficiently and decode it for the accurate refinement of the similarity map. To further improve the performance, we additionally introduce a curriculum learning strategy. Also, we adopt online evaluation to measure performance in real-world conditions. Exhaustive experiments on 8 well-known benchmarks demonstrate the superiority of TCTrack++. Real-world tests directly verify that TCTrack++ can be readily used in real-world applications.
引用
收藏
页码:15834 / 15849
页数:16
相关论文
共 50 条
  • [11] Towards Interactive Search: Investigating Visual Search in a Novel Real-World Paradigm
    Sauter, Marian
    Stefani, Maximilian
    Mack, Wolfgang
    BRAIN SCIENCES, 2020, 10 (12) : 1 - 13
  • [12] Towards Theory for Real-World Data
    Martens, Wim
    PROCEEDINGS OF THE 41ST ACM SIGMOD-SIGACT-SIGAI SYMPOSIUM ON PRINCIPLES OF DATABASE SYSTEMS (PODS '22), 2022, : 261 - 276
  • [13] Towards a real-World Wide Web
    Kindberg, Tim
    Barton, John
    HP Laboratories Technical Report, 2000, (47):
  • [14] Evaluating the quality of medical evidence in real-world contexts
    Jones, Andrew
    Steel, Daniel
    JOURNAL OF EVALUATION IN CLINICAL PRACTICE, 2018, 24 (05) : 950 - 956
  • [15] Materialising contexts: virtual soundscapes for real-world exploration
    Laurence Cliffe
    James Mansell
    Chris Greenhalgh
    Adrian Hazzard
    Personal and Ubiquitous Computing, 2021, 25 : 623 - 636
  • [16] Materialising contexts: virtual soundscapes for real-world exploration
    Cliffe, Laurence
    Mansell, James
    Greenhalgh, Chris
    Hazzard, Adrian
    PERSONAL AND UBIQUITOUS COMPUTING, 2021, 25 (04) : 623 - 636
  • [17] Tracking real-world phenomena with smart dust
    Römer, K
    WIRELESS SENSOR NETWORKS, PROCEEDINGS, 2004, 2920 : 28 - 43
  • [18] Real-world tracking system performance evaluation
    Blackman, SS
    Dempster, RJ
    SIGNAL AND DATA PROCESSING OF SMALL TARGETS 2002, 2002, 4728 : 470 - 476
  • [19] Real-World Field Snail Detection and Tracking
    Wang, Zhiyan
    Lee, Ivan
    Tie, Yun
    Cai, Jinhai
    Qi, Lin
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 1703 - 1708
  • [20] Towards visual neural prosthetics: Real-world indoor mobility with simulated phosphene vision
    van Wezel, Richard R. J. A.
    van Gerven, Marcel M.
    van Steveninck, Jaap J. de Ruyter
    PERCEPTION, 2021, 50 (1_SUPPL) : 58 - 59