Disturbed Augmentation Invariance for Unsupervised Visual Representation Learning

被引:3
|
作者
Cheng, Haoyang [1 ]
Li, Hongliang [1 ]
Wu, Qingbo [1 ]
Qiu, Heqian [1 ]
Zhang, Xiaoliang [1 ]
Meng, Fanman [1 ]
Zhao, Taijin [1 ]
机构
[1] Univ Elect Sci & Technol China UESTC, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Unsupervised learning; self-supervised learning; representation learning; contrastive learning; convolutional neural network;
D O I
10.1109/TCSVT.2023.3272741
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Contrastive learning has gained great prominence recently, which achieves excellent performance by simple augmentation invariance. However, the simple contrastive pairs suffer from lacking of diversity due to the mechanical augmentation strategies. In this paper, we propose Disturbed Augmentation Invariance (DAI for abbreviation), which constructs disturbed contrastive pairs by generating appropriate disturbed views for each augmented view in the feature space to increase the diversity. In practice, we establish a multivariate normal distribution for each augmented view, whose mean is corresponding augmented view and covariance matrix is estimated from its nearest neighbors in the dataset. Then we sample random vectors from this distribution as the disturbed views to construct disturbed contrastive pairs. In order to avoid extra computational cost with the increase of disturbed contrastive pairs, we utilize an upper bound of the trivial disturbed augmentation invariance loss to construct the DAI loss. In addition, we propose Bottleneck version of Disturbed Augmentation Invariance (BDAI for abbreviation) inspired by the Information Bottleneck principle, which further refines the extracted information and learns a compact representation by additionally increasing the variance of the original contrastive pair. In order to make BDAI work effectively, we design a statistical strategy to control the balance between the amount of the information shared by all disturbed contrastive pairs and the compactness of the representation. Our approach gets a consistent improvement over the popular contrastive learning methods on a variety of downstream tasks, e.g. image classification, object detection and instance segmentation.
引用
收藏
页码:6924 / 6938
页数:15
相关论文
共 50 条
  • [31] Federated unsupervised representation learning
    Zhang, Fengda
    Kuang, Kun
    Chen, Long
    You, Zhaoyang
    Shen, Tao
    Xiao, Jun
    Zhang, Yin
    Wu, Chao
    Wu, Fei
    Zhuang, Yueting
    Li, Xiaolin
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2023, 24 (08) : 1181 - 1193
  • [32] Visual Graph Memory with Unsupervised Representation for Visual Navigation
    Kwon, Obin
    Kim, Nuri
    Choi, Yunho
    Yoo, Hwiyeon
    Park, Jeongho
    Oh, Songhwai
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15870 - 15879
  • [33] Continual Unsupervised Representation Learning
    Rao, Dushyant
    Visin, Francesco
    Rusu, Andrei A.
    Teh, Yee Whye
    Pascanu, Razvan
    Hadsell, Raia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [34] MULTI-AUGMENTATION FOR EFFICIENT SELF-SUPERVISED VISUAL REPRESENTATION LEARNING
    Tran, Van Nhiem
    Huang, Chi-En
    Liu, Shen-Hsuan
    Yang, Kai-Lin
    Ko, Timothy
    Li, Yung-Hui
    2022 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (IEEE ICMEW 2022), 2022,
  • [35] Rethinking the Augmentation Module in Contrastive Learning: Learning Hierarchical Augmentation Invariance with Expanded Views
    Zhang, Junbo
    Ma, Kaisheng
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16629 - 16638
  • [36] Unsupervised learning of visual structure
    Edelman, S
    Intrator, N
    Jacobson, JS
    BIOLOGICALLY MOTIVATED COMPUTER VISION, PROCEEDINGS, 2002, 2525 : 629 - 642
  • [37] Train a One-Million-Way Instance Classifier for Unsupervised Visual Representation Learning
    Liu, Yu
    Huang, Lianghua
    Pan, Pan
    Wang, Bin
    Xu, Yinghui
    Jin, Rong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8706 - 8714
  • [38] Unsupervised learning of visual taxonomies
    Bart, Evgeniy
    Porteous, Ian
    Perona, Pietro
    Welling, Max
    2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 2166 - +
  • [39] Propagate Yourself: Exploring Pixel-Level Consistency for Unsupervised Visual Representation Learning
    Xie, Zhenda
    Lin, Yutong
    Zhang, Zheng
    Cao, Yue
    Lin, Stephen
    Hu, Han
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16679 - 16688
  • [40] Unsupervised learning using topological data augmentation
    Balabanov, Oleksandr
    Granath, Mats
    PHYSICAL REVIEW RESEARCH, 2020, 2 (01):