Disturbed Augmentation Invariance for Unsupervised Visual Representation Learning

被引:3
|
作者
Cheng, Haoyang [1 ]
Li, Hongliang [1 ]
Wu, Qingbo [1 ]
Qiu, Heqian [1 ]
Zhang, Xiaoliang [1 ]
Meng, Fanman [1 ]
Zhao, Taijin [1 ]
机构
[1] Univ Elect Sci & Technol China UESTC, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Unsupervised learning; self-supervised learning; representation learning; contrastive learning; convolutional neural network;
D O I
10.1109/TCSVT.2023.3272741
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Contrastive learning has gained great prominence recently, which achieves excellent performance by simple augmentation invariance. However, the simple contrastive pairs suffer from lacking of diversity due to the mechanical augmentation strategies. In this paper, we propose Disturbed Augmentation Invariance (DAI for abbreviation), which constructs disturbed contrastive pairs by generating appropriate disturbed views for each augmented view in the feature space to increase the diversity. In practice, we establish a multivariate normal distribution for each augmented view, whose mean is corresponding augmented view and covariance matrix is estimated from its nearest neighbors in the dataset. Then we sample random vectors from this distribution as the disturbed views to construct disturbed contrastive pairs. In order to avoid extra computational cost with the increase of disturbed contrastive pairs, we utilize an upper bound of the trivial disturbed augmentation invariance loss to construct the DAI loss. In addition, we propose Bottleneck version of Disturbed Augmentation Invariance (BDAI for abbreviation) inspired by the Information Bottleneck principle, which further refines the extracted information and learns a compact representation by additionally increasing the variance of the original contrastive pair. In order to make BDAI work effectively, we design a statistical strategy to control the balance between the amount of the information shared by all disturbed contrastive pairs and the compactness of the representation. Our approach gets a consistent improvement over the popular contrastive learning methods on a variety of downstream tasks, e.g. image classification, object detection and instance segmentation.
引用
收藏
页码:6924 / 6938
页数:15
相关论文
共 50 条
  • [11] Temporal Knowledge Consistency for Unsupervised Visual Representation Learning
    Feng, Weixin
    Wang, Yuanjiang
    Ma, Lihua
    Yuan, Ye
    Zhang, Chi
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10150 - 10160
  • [12] Unsupervised Visual Representation Learning by Tracking Patches in Video
    Wang, Guangting
    Zhou, Yizhou
    Luo, Chong
    Xie, Wenxuan
    Zeng, Wenjun
    Xiong, Zhiwei
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2563 - 2572
  • [13] Fixed-Rank Representation for Unsupervised Visual Learning
    Liu, Risheng
    Lin, Zhouchen
    De la Torre, Fernando
    Su, Zhixun
    2012 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012, : 598 - 605
  • [14] Transitive Invariance for Self-supervised Visual Representation Learning
    Wang, Xiaolong
    He, Kaiming
    Gupta, Abhinav
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1338 - 1347
  • [15] Fast and Unsupervised Neural Architecture Evolution for Visual Representation Learning
    Xue, Song
    Chen, Hanlin
    Xie, Chunyu
    Zhang, Baochang
    Gong, Xuan
    Doermann, David
    IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2021, 16 (03) : 22 - 32
  • [16] Collaborative Unsupervised Visual Representation Learning from Decentralized Data
    Zhuang, Weiming
    Gan, Xin
    Wen, Yonggang
    Zhang, Shuai
    Yi, Shuai
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4892 - 4901
  • [17] Two momentum contrast in triplet for unsupervised visual representation learning
    Long, Xianzhong
    Du, Han
    Li, Yun
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 10467 - 10480
  • [18] Unsupervised Galaxy Morphological Visual Representation with Deep Contrastive Learning
    Wei, Shoulin
    Li, Yadi
    Lu, Wei
    Li, Nan
    Liang, Bo
    Dai, Wei
    Zhang, Zhijian
    PUBLICATIONS OF THE ASTRONOMICAL SOCIETY OF THE PACIFIC, 2022, 134 (1041)
  • [19] Efficient Unsupervised Visual Representation Learning with Explicit Cluster Balancing
    Metaxas, Ioannis Maniadis
    Tzimiropoulos, Georgios
    Patras, Ioannis
    COMPUTER VISION - ECCV 2024, PT XXXII, 2025, 15090 : 436 - 454
  • [20] Two momentum contrast in triplet for unsupervised visual representation learning
    Xianzhong Long
    Han Du
    Yun Li
    Multimedia Tools and Applications, 2024, 83 : 10467 - 10480