VITAL: VIsual Tracking via Adversarial Learning

被引:450
|
作者
Song, Yibing [1 ]
Ma, Chao [2 ]
Wu, Xiaohe [3 ]
Gong, Lijun [4 ]
Bao, Linchao [1 ]
Zuo, Wangmeng [3 ]
Shen, Chunhua [2 ]
Lau, Rynson W. H. [5 ]
Yang, Ming-Hsuan [6 ]
机构
[1] Tencent AI Lab, Shenzhen, Peoples R China
[2] Univ Adelaide, Adelaide, SA, Australia
[3] Harbin Inst Technol, Harbin, Heilongjiang, Peoples R China
[4] Tencent, Shenzhen, Peoples R China
[5] City Univ Hong Kong, Hong Kong, Peoples R China
[6] Univ Calif Merced, Merced, CA USA
关键词
D O I
10.1109/CVPR.2018.00937
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The tracking-by-detection framework consists of two stages, i.e., drawing samples around the target object in the first stage and classifying each sample as the target object or as background in the second stage. The performance of existing trackers using deep classification networks is limited by two aspects. First, the positive samples in each frame are highly spatially overlapped, and they fail to capture rich appearance variations. Second, there exists extreme class imbalance between positive and negative samples. This paper presents the VITAL algorithm to address these two problems via adversarial learning. To augment positive samples, we use a generative network to randomly generate masks, which are applied to adaptively dropout input features to capture a variety of appearance changes. With the use of adversarial learning, our network identifies the mask that maintains the most robust features of the target objects over a long temporal span. In addition, to handle the issue of class imbalance, we propose a high-order cost sensitive loss to decrease the effect of easy negative samples to facilitate training the classification network. Extensive experiments on benchmark datasets demonstrate that the proposed tracker performs favorably against state-of-the-art approaches.
引用
收藏
页码:8990 / 8999
页数:10
相关论文
共 50 条
  • [1] I-VITAL: Information aided visual tracking with adversarial learning
    Dasari, Mohana Murali
    Kuchibhotla, Hari Chandana
    Rajiv, Aravind
    Gorthi, Rama Krishna
    DISPLAYS, 2023, 77
  • [2] IMPROVED REAL-TIME VISUAL TRACKING VIA ADVERSARIAL LEARNING
    Zhong, Haoxiang
    Yan, Xiyu
    Jiang, Yong
    Xia, Shu-Tao
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1853 - 1857
  • [3] Multilayer Perceptron Generative Model via Adversarial Learning for Robust Visual Tracking
    AlBasiouny, Eman R. R.
    Heliel, Abdel-Fattah Attia
    Abdelmunim, Hossam E. E.
    Abbas, Hazem M. M.
    IEEE ACCESS, 2022, 10 : 121230 - 121248
  • [4] Adversarial Feature Sampling Learning for Efficient Visual Tracking
    Yin, Yingjie
    Xu, De
    Wang, Xingang
    Zhang, Lei
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2020, 17 (02) : 847 - 857
  • [5] Visual Tracking via Sparsity Pattern Learning
    Wang, Yuxi
    Liu, Yue
    Li, Zhuwen
    Cheong, Loong-Fah
    Ling, Haibin
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2716 - 2721
  • [6] Visual Tracking via Nonlocal Similarity Learning
    Liu, Qingshan
    Fan, Jiaqing
    Song, Huihui
    Chen, Wei
    Zhang, Kaihua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (10) : 2826 - 2835
  • [7] ROBUST VISUAL TRACKING VIA TRANSFER LEARNING
    Luo, Wenhan
    Li, Xi
    Li, Wei
    Hu, Weiming
    2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2011, : 485 - 488
  • [8] Visual Tracking via Dynamic Graph Learning
    Li, Chenglong
    Lin, Liang
    Zuo, Wangmeng
    Tang, Jin
    Yang, Ming-Hsuan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (11) : 2770 - 2782
  • [9] EEG classification based on visual stimuli via adversarial learning
    Mishra, Rahul
    Bhavsar, Arnav
    COGNITIVE NEURODYNAMICS, 2024, 18 (03) : 1135 - 1151
  • [10] Adversarial Transfer Networks for Visual Tracking
    Liu, Lijie
    Lu, Jiwen
    Zhou, Jie
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 75 - 81