EMCRL: EM-Enhanced Negative Sampling Strategy for Contrastive Representation Learning

被引:0
|
作者
Zhang, Kun [1 ]
Lv, Guangyi [2 ]
Wu, Le [1 ]
Hong, Richang [1 ]
Wang, Meng [1 ]
机构
[1] Hefei Univ Technol, Sch Comp & Informat, Hefei 230029, Anhui, Peoples R China
[2] Lenovo Res, AI Lab, Beijing 100094, Peoples R China
基金
中国国家自然科学基金;
关键词
Representation learning; Data augmentation; Data models; Semantics; Optimization; Estimation; Sampling methods; Robustness; Natural languages; Crops; Contrastive learning (CL); expectation maximization (EM); negative examples; representation learning;
D O I
10.1109/TCSS.2024.3454056
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
As one representative framework of self-supervised learning (SSL), contrastive learning (CL) has drawn enormous attention in the representation learning area. By pulling together a "positive" example and an anchor, as well as pushing away many "negative" examples from the anchor, CL is able to generate high-quality representations for the data of different modalities. Therefore, the qualities of selected positive and negative examples are critical for the performance of CL-based models. However, due to the assumption of label unavailability, most existing work follows the paradigm of contrastive instance discrimination, which treats each input instance as an individual category. Therefore, they focused more on positive example generation and designed plenty of data augmentation strategies. For negative examples, they just leverage the in-batch negative sampling strategy. We argue that this negative sampling strategy will easily select false negatives and inhibit the capability of CL, which we also believe is one of the reasons why a large size of negatives is needed in CL. Apart from using annotated labels, we try to tackle this problem in an unsupervised manner. We propose to integrate expectation maximization (EM) into the selection of negative examples and develop a novel EM-enhanced negative sampling strategy (EMCRL) to distinguish false negatives from true ones for CL performance improvement. Specifically, EMCRL employs EM to estimate the distribution of ground-truth relations between each sample and corresponding in-batch negatives and then optimizes model parameters with the estimations. Considering the sensitivity of EM algorithm to the parameter initialization, we propose to add a random flip into the distribution estimation to enhance the robustness of the learning process. Extensive experiments over several advanced models on sentence representation and image representation tasks demonstrate the effectiveness of EMCRL. Our method is easy to implement, and the code is publicly available at https://github.com/zhangkunzk/EMCRL_pytorch.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Contrastive Learning with Positive-Negative Frame Mask for Music Representation
    Yao, Dong
    Zhao, Zhou
    Zhang, Shengyu
    Zhu, Jieming
    Zhu, Yudong
    Zhang, Rui
    He, Xiuqiang
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 2906 - 2915
  • [22] HNS: Hierarchical negative sampling for network representation learning
    Chen, Junyang
    Gong, Zhiguo
    Wang, Wei
    Liu, Weiwen
    INFORMATION SCIENCES, 2021, 542 : 343 - 356
  • [23] Unsupervised Path Representation Learning with Curriculum Negative Sampling
    Bin Yang, Sean
    Guo, Chenjuan
    Hu, Jilin
    Tang, Jian
    Yang, Bin
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3286 - 3292
  • [24] Incorporating GAN for Negative Sampling in Knowledge Representation Learning
    Wang, Peifeng
    Li, Shuangyin
    Pan, Rong
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2005 - 2012
  • [25] Cosine Mixup: A Mixup Strategy for Point Cloud Contrastive Representation Learning
    Li, Guangping
    Gao, Xiang
    Liu, Chenghui
    Liang, Dingkai
    Ling, Bingo Wing-Kuen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (02) : 967 - 971
  • [26] Drug repositioning by multi-aspect heterogeneous graph contrastive learning and positive-fusion negative sampling strategy
    Liu, Junkai
    Hu, Fuyuan
    Zou, Quan
    Tiwari, Prayag
    Wu, Hongjie
    Ding, Yijie
    INFORMATION FUSION, 2024, 112
  • [27] Community-Enhanced Contrastive Siamese Networks for Graph Representation Learning
    Li, Yafang
    Wang, Wenbo
    Ma, Guixiang
    Zu, Baokai
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2023, 2023, 14117 : 300 - 314
  • [28] Mining negative samples on contrastive learning via curricular weighting strategy
    Zhuang, Jin
    Jing, Xiao-Yuan
    Jia, Xiaodong
    INFORMATION SCIENCES, 2024, 668
  • [29] NFGCL: A negative-sampling-free graph contrastive learning framework for recommendation
    Xiao, Yuxi
    Ma, Rui
    Sang, Jun
    INFORMATION SCIENCES, 2025, 695
  • [30] AFANS: Augmentation-Free Graph Contrastive Learning with Adversarial Negative Sampling
    Wang, Shihao
    Wang, Chenxu
    Meng, Panpan
    Wang, Zhanggong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT XII, ICIC 2024, 2024, 14873 : 376 - 387