SNCSE: Contrastive Learning for Unsupervised Sentence Embedding with Soft Negative Samples

被引:10
|
作者
Wang, Hao [1 ]
Dou, Yong [1 ]
机构
[1] Natl Univ Def Technol, Changsha 410073, Peoples R China
关键词
Unsupervised Sentence Embedding; Contrastive Learning; Feature Suppression; Soft Negative Samples; Bidirectional Margin Loss;
D O I
10.1007/978-981-99-4752-2_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised sentence embedding aims to obtain the most appropriate embedding for a sentence to reflect its semantics. Contrastive learning has been attracting developing attention. For a sentence, current models utilize diverse data augmentation methods to generate positive samples, while consider other independent sentences as negative samples. Then they adopt InfoNCE loss to pull the embeddings of positive pairs gathered, and push those of negative pairs scattered. Although these models have made great progress, we argue that they may suffer from feature suppression, where the models fail to distinguish and decouple textual similarity and semantic similarity. They may overestimate the semantic similarity of any sentence pairs with similar text regardless of the actual semantic difference between them, and vice versa. Herein, we propose contrastive learning for unsupervised sentence embedding with soft negative samples (SNCSE). Soft negative samples share highly similar text but have surely and apparently different semantics with the original samples. Specifically, we take the negation of original sentences as soft negative samples, and propose BidirectionalMargin Loss (BML) to introduce them into traditional contrastive learning framework. Our experimental results on semantic textual similarity (STS) task show that SNCSE can obtain state-of-the-art performance with different encoders, indicating its strength on unsupervised sentence embedding. Our code and models are released at https:// github.com/Sense-GVT/SNCSE.
引用
收藏
页码:419 / 431
页数:13
相关论文
共 50 条
  • [31] Importance-aware contrastive learning via semantically augmented instances for unsupervised sentence embeddings
    Ma, Xin
    Li, Hong
    Shi, Jiawen
    Zhang, Yi
    Long, Zhigao
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (09) : 2979 - 2990
  • [32] Contrastive Learning Models for Sentence Representations
    Xu, Lingling
    Xie, Haoran
    Li, Zongxi
    Wang, Fu Lee
    Wang, Weiming
    Li, Qing
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (04)
  • [33] SimRE: Simple contrastive learning with soft logical rule for knowledge graph embedding
    Zhang, Dong
    Rong, Zhe
    Xue, Chengyuan
    Li, Guanyu
    INFORMATION SCIENCES, 2024, 661
  • [34] An Unsupervised Sentence Embedding Method by Mutual Information Maximization
    Yan Zhang
    He, Ruidan
    Liu, Zuozhu
    Lim, Kwan Hui
    Bing, Lidong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1601 - 1610
  • [35] HSimCSE: Improving Contrastive Learning of Unsupervised Sentence Representation with Adversarial Hard Positives and Dual Hard Negatives
    Xu, Bo
    Wei, Shouang
    Cheng, Luyi
    Huang, Shizhou
    Song, Hui
    Du, Ming
    Wang, Hongya
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [36] Noise-Contrastive Estimation Based on Relative Neighbour Sampling for Unsupervised Image Embedding Learning
    Lv, Haoyu
    Cao, Jiangzhong
    Dai, Qingyun
    Liang, Guanshu
    Zhao, Qi
    2019 4TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS (ICBDA 2019), 2019, : 308 - 311
  • [37] Heterogeneous data augmentation in graph contrastive learning for effective negative samples
    Ali, Adnan
    Li, Jinlong
    Chen, Huanhuan
    COMPUTERS & ELECTRICAL ENGINEERING, 2024, 118
  • [38] Mining negative samples on contrastive learning via curricular weighting strategy
    Zhuang, Jin
    Jing, Xiao-Yuan
    Jia, Xiaodong
    INFORMATION SCIENCES, 2024, 668
  • [39] Heterogeneous data augmentation in graph contrastive learning for effective negative samples
    Ali, Adnan
    Li, Jinlong
    Chen, Huanhuan
    COMPUTERS & ELECTRICAL ENGINEERING, 2024, 118
  • [40] Kalman contrastive unsupervised representation learning
    Yekta, Mohammad Mahdi Jahani
    SCIENTIFIC REPORTS, 2024, 14 (01):