EMCRL: EM-Enhanced Negative Sampling Strategy for Contrastive Representation Learning

被引:0
|
作者
Zhang, Kun [1 ]
Lv, Guangyi [2 ]
Wu, Le [1 ]
Hong, Richang [1 ]
Wang, Meng [1 ]
机构
[1] Hefei Univ Technol, Sch Comp & Informat, Hefei 230029, Anhui, Peoples R China
[2] Lenovo Res, AI Lab, Beijing 100094, Peoples R China
基金
中国国家自然科学基金;
关键词
Representation learning; Data augmentation; Data models; Semantics; Optimization; Estimation; Sampling methods; Robustness; Natural languages; Crops; Contrastive learning (CL); expectation maximization (EM); negative examples; representation learning;
D O I
10.1109/TCSS.2024.3454056
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
As one representative framework of self-supervised learning (SSL), contrastive learning (CL) has drawn enormous attention in the representation learning area. By pulling together a "positive" example and an anchor, as well as pushing away many "negative" examples from the anchor, CL is able to generate high-quality representations for the data of different modalities. Therefore, the qualities of selected positive and negative examples are critical for the performance of CL-based models. However, due to the assumption of label unavailability, most existing work follows the paradigm of contrastive instance discrimination, which treats each input instance as an individual category. Therefore, they focused more on positive example generation and designed plenty of data augmentation strategies. For negative examples, they just leverage the in-batch negative sampling strategy. We argue that this negative sampling strategy will easily select false negatives and inhibit the capability of CL, which we also believe is one of the reasons why a large size of negatives is needed in CL. Apart from using annotated labels, we try to tackle this problem in an unsupervised manner. We propose to integrate expectation maximization (EM) into the selection of negative examples and develop a novel EM-enhanced negative sampling strategy (EMCRL) to distinguish false negatives from true ones for CL performance improvement. Specifically, EMCRL employs EM to estimate the distribution of ground-truth relations between each sample and corresponding in-batch negatives and then optimizes model parameters with the estimations. Considering the sensitivity of EM algorithm to the parameter initialization, we propose to add a random flip into the distribution estimation to enhance the robustness of the learning process. Extensive experiments over several advanced models on sentence representation and image representation tasks demonstrate the effectiveness of EMCRL. Our method is easy to implement, and the code is publicly available at https://github.com/zhangkunzk/EMCRL_pytorch.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Negative sampling strategies for contrastive self-supervised learning of graph representations
    Hafidi, Hakim
    Ghogho, Mounir
    Ciblat, Philippe
    Swami, Ananthram
    SIGNAL PROCESSING, 2022, 190
  • [32] AdaNS: Adaptive negative sampling for unsupervised graph representation learning
    Wang, Yu
    Hu, Liang
    Gao, Wanfu
    Cao, Xiaofeng
    Chang, Yi
    PATTERN RECOGNITION, 2023, 136
  • [33] A novel negative sampling based on TFIDF for learning word representation
    Qin, Pengda
    Xu, Weiran
    Guo, Jun
    NEUROCOMPUTING, 2016, 177 : 257 - 265
  • [34] Comprehensive Analysis of Negative Sampling in Knowledge Graph Representation Learning
    Kamigaito, Hidetaka
    Hayashi, Katsuhiko
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10661 - 10675
  • [35] ConFunc: Enhanced Binary Function-Level Representation through Contrastive Learning
    Li, Longfei
    Yin, Xiaokang
    Li, Xiao
    Zhu, Xiaoya
    Liu, Shengli
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 1241 - 1248
  • [36] ReCLR: Reference-Enhanced Contrastive Learning of Audio Representation for Depression Detection
    Zhang, Pingyue
    Wu, Mengyue
    Yu, Kai
    INTERSPEECH 2023, 2023, : 2998 - 3002
  • [37] Domain generalization by class-aware negative sampling-based contrastive learning
    Xie, Mengwei
    Zhao, Suyun
    Chen, Hong
    Li, Cuiping
    AI OPEN, 2022, 3 : 200 - 207
  • [38] Self-Contrastive Learning with Hard Negative Sampling for Self-supervised Point Cloud Learning
    Du, Bi'an
    Gao, Xiang
    Hu, Wei
    Li, Xin
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3133 - 3142
  • [39] Contrastive learning enhanced by graph neural networks for Universal Multivariate Time Series Representation
    Wang, Xinghao
    Xing, Qiang
    Xiao, Huimin
    Ye, Ming
    INFORMATION SYSTEMS, 2024, 125
  • [40] Contrastive learning enhanced by graph neural networks for Universal Multivariate Time Series Representation
    College of Artificial Intelligence, Southwest University, Chongqing
    400715, China
    Inf. Syst.,