SNCSE: Contrastive Learning for Unsupervised Sentence Embedding with Soft Negative Samples

被引:10
|
作者
Wang, Hao [1 ]
Dou, Yong [1 ]
机构
[1] Natl Univ Def Technol, Changsha 410073, Peoples R China
关键词
Unsupervised Sentence Embedding; Contrastive Learning; Feature Suppression; Soft Negative Samples; Bidirectional Margin Loss;
D O I
10.1007/978-981-99-4752-2_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised sentence embedding aims to obtain the most appropriate embedding for a sentence to reflect its semantics. Contrastive learning has been attracting developing attention. For a sentence, current models utilize diverse data augmentation methods to generate positive samples, while consider other independent sentences as negative samples. Then they adopt InfoNCE loss to pull the embeddings of positive pairs gathered, and push those of negative pairs scattered. Although these models have made great progress, we argue that they may suffer from feature suppression, where the models fail to distinguish and decouple textual similarity and semantic similarity. They may overestimate the semantic similarity of any sentence pairs with similar text regardless of the actual semantic difference between them, and vice versa. Herein, we propose contrastive learning for unsupervised sentence embedding with soft negative samples (SNCSE). Soft negative samples share highly similar text but have surely and apparently different semantics with the original samples. Specifically, we take the negation of original sentences as soft negative samples, and propose BidirectionalMargin Loss (BML) to introduce them into traditional contrastive learning framework. Our experimental results on semantic textual similarity (STS) task show that SNCSE can obtain state-of-the-art performance with different encoders, indicating its strength on unsupervised sentence embedding. Our code and models are released at https:// github.com/Sense-GVT/SNCSE.
引用
收藏
页码:419 / 431
页数:13
相关论文
共 50 条
  • [21] Synthetic Hard Negative Samples for Contrastive Learning
    Hengkui Dong
    Xianzhong Long
    Yun Li
    Neural Processing Letters, 56
  • [22] Contrastive sentence representation learning with adaptive false negative cancellation
    Xu, Lingling
    Xie, Haoran
    Wang, Fu Lee
    Tao, Xiaohui
    Wang, Weiming
    Li, Qing
    INFORMATION FUSION, 2024, 102
  • [23] Negative samples selecting strategy for graph contrastive learning
    Miao, Rui
    Yang, Yintao
    Ma, Yao
    Juan, Xin
    Xue, Haotian
    Tang, Jiliang
    Wang, Ying
    Wang, Xin
    INFORMATION SCIENCES, 2022, 613 : 667 - 681
  • [24] Contrastive Unsupervised Representation Learning With Optimize-Selected Training Samples
    Cheng, Yujun
    Zhang, Zhewei
    Li, Xuejing
    Wang, Shengjin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [25] WhiteningBERT: An Easy Unsupervised Sentence Embedding Approach
    Huang, Junjie
    Tang, Duyu
    Zhong, Wanjun
    Lu, Shuai
    Shou, Linjun
    Gong, Ming
    Jiang, Daxin
    Duan, Nan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 238 - 244
  • [26] UNSEE: Unsupervised Non-contrastive Sentence Embeddings
    Cagatan, Omer Veysel
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 384 - 393
  • [27] Robust Contrastive Learning Using Negative Samples with Diminished Semantics
    Ge, Songwei
    Mishra, Shlok
    Wang, Haohan
    Li, Chun-Liang
    Jacobs, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [28] DropMix: Better Graph Contrastive Learning with Harder Negative Samples
    Ma, Yueqi
    Chen, Minjie
    Li, Xiang
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1105 - 1112
  • [29] Do More Negative Samples Necessarily Hurt In Contrastive Learning?
    Awasthi, Pranjal
    Dikkala, Nishanth
    Kamath, Pritish
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [30] Importance-aware contrastive learning via semantically augmented instances for unsupervised sentence embeddings
    Xin Ma
    Hong Li
    Jiawen Shi
    Yi Zhang
    Zhigao Long
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 2979 - 2990