VITAL: VIsual Tracking via Adversarial Learning

被引:450
|
作者
Song, Yibing [1 ]
Ma, Chao [2 ]
Wu, Xiaohe [3 ]
Gong, Lijun [4 ]
Bao, Linchao [1 ]
Zuo, Wangmeng [3 ]
Shen, Chunhua [2 ]
Lau, Rynson W. H. [5 ]
Yang, Ming-Hsuan [6 ]
机构
[1] Tencent AI Lab, Shenzhen, Peoples R China
[2] Univ Adelaide, Adelaide, SA, Australia
[3] Harbin Inst Technol, Harbin, Heilongjiang, Peoples R China
[4] Tencent, Shenzhen, Peoples R China
[5] City Univ Hong Kong, Hong Kong, Peoples R China
[6] Univ Calif Merced, Merced, CA USA
关键词
D O I
10.1109/CVPR.2018.00937
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The tracking-by-detection framework consists of two stages, i.e., drawing samples around the target object in the first stage and classifying each sample as the target object or as background in the second stage. The performance of existing trackers using deep classification networks is limited by two aspects. First, the positive samples in each frame are highly spatially overlapped, and they fail to capture rich appearance variations. Second, there exists extreme class imbalance between positive and negative samples. This paper presents the VITAL algorithm to address these two problems via adversarial learning. To augment positive samples, we use a generative network to randomly generate masks, which are applied to adaptively dropout input features to capture a variety of appearance changes. With the use of adversarial learning, our network identifies the mask that maintains the most robust features of the target objects over a long temporal span. In addition, to handle the issue of class imbalance, we propose a high-order cost sensitive loss to decrease the effect of easy negative samples to facilitate training the classification network. Extensive experiments on benchmark datasets demonstrate that the proposed tracker performs favorably against state-of-the-art approaches.
引用
收藏
页码:8990 / 8999
页数:10
相关论文
共 50 条
  • [21] VTAAN: Visual Tracking with Attentive Adversarial Network
    Futian Wang
    Xiaoping Wang
    Jin Tang
    Bin Luo
    Chenglong Li
    Cognitive Computation, 2021, 13 : 646 - 656
  • [22] Adversarial attack can help visual tracking
    Sungmin Cho
    Hyeseong Kim
    Ji Soo Kim
    Hyomin Kim
    Junseok Kwon
    Multimedia Tools and Applications, 2022, 81 : 35283 - 35292
  • [23] VTAAN: Visual Tracking with Attentive Adversarial Network
    Wang, Futian
    Wang, Xiaoping
    Tang, Jin
    Luo, Bin
    Li, Chenglong
    COGNITIVE COMPUTATION, 2021, 13 (03) : 646 - 656
  • [24] Learning Discriminative Features for Visual Tracking via Scenario Decoupling
    Ma, Yinchao
    Yu, Qianjin
    Yang, Wenfei
    Zhang, Tianzhu
    Zhang, Jinpeng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, : 2950 - 2966
  • [25] Visual object tracking via online sparse instance learning
    Yan, Jia
    Chen, Xi
    Deng, Dexiang
    Zhu, Qiuping
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2015, 26 : 231 - 246
  • [26] Adaptive Visual Tracking via Learning Detector of Specific Landmarks
    Hwang, Chih-Lyang
    Chang, Kuo-Ching
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND VIRTUAL ENVIRONMENTS FOR MEASUREMENT SYSTEMS AND APPLICATIONS (CIVEMSA), 2013, : 66 - 71
  • [27] Robust Visual Tracking via Binocular Consistent Sparse Learning
    Ziang Ma
    Zhiyu Xiang
    Neural Processing Letters, 2017, 46 : 627 - 642
  • [28] Visual Tracking via Sparse Representation and Online Dictionary Learning
    Cheng, Xu
    Li, Nijun
    Zhou, Tongchi
    Zhou, Lin
    Wu, Zhenyang
    ACTIVITY MONITORING BY MULTIPLE DISTRIBUTED SENSING, 2014, 8703 : 87 - 103
  • [29] Robust visual tracking via self-similarity learning
    Song, Huihui
    Zheng, Yuhui
    Zhang, Kaihua
    ELECTRONICS LETTERS, 2017, 53 (01) : 20 - 21
  • [30] Robust visual tracking via randomly projected instance learning
    Cheng F.
    Liu K.
    Gong M.-G.
    Fu K.
    Xi J.
    Gong, Mao-Guo (mggong@mail.xidian.edu.cn), 1600, Emerald Group Holdings Ltd. (10): : 258 - 271