Do More Negative Samples Necessarily Hurt In Contrastive Learning?

被引:0
|
作者
Awasthi, Pranjal [1 ]
Dikkala, Nishanth [1 ]
Kamath, Pritish [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
关键词
MATRIX;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent investigations in noise contrastive estimation suggest, both empirically as well as theoretically, that while having more "negative samples" in the contrastive loss improves downstream classification performance initially, beyond a threshold, it hurts downstream performance due to a "collision-coverage" trade-off. But is such a phenomenon inherent in contrastive learning? We show in a simple theoretical setting, where positive pairs are generated by sampling from the underlying latent class (introduced by Saunshi et al. (ICML 2019)), that the downstream performance of the representation optimizing the (population) contrastive loss in fact does not degrade with the number of negative samples. Along the way, we give a structural characterization of the optimal representation in our framework, for noise contrastive estimation. We also provide empirical support for our theoretical results on CIFAR-10 and CIFAR-100 datasets.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Synthetic Hard Negative Samples for Contrastive Learning
    Dong, Hengkui
    Long, Xianzhong
    Li, Yun
    NEURAL PROCESSING LETTERS, 2024, 56 (01)
  • [2] Synthetic Hard Negative Samples for Contrastive Learning
    Hengkui Dong
    Xianzhong Long
    Yun Li
    Neural Processing Letters, 56
  • [3] Negative samples selecting strategy for graph contrastive learning
    Miao, Rui
    Yang, Yintao
    Ma, Yao
    Juan, Xin
    Xue, Haotian
    Tang, Jiliang
    Wang, Ying
    Wang, Xin
    INFORMATION SCIENCES, 2022, 613 : 667 - 681
  • [4] Robust Contrastive Learning Using Negative Samples with Diminished Semantics
    Ge, Songwei
    Mishra, Shlok
    Wang, Haohan
    Li, Chun-Liang
    Jacobs, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] DropMix: Better Graph Contrastive Learning with Harder Negative Samples
    Ma, Yueqi
    Chen, Minjie
    Li, Xiang
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1105 - 1112
  • [6] SNCSE: Contrastive Learning for Unsupervised Sentence Embedding with Soft Negative Samples
    Wang, Hao
    Dou, Yong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT IV, 2023, 14089 : 419 - 431
  • [7] Heterogeneous data augmentation in graph contrastive learning for effective negative samples
    Ali, Adnan
    Li, Jinlong
    Chen, Huanhuan
    COMPUTERS & ELECTRICAL ENGINEERING, 2024, 118
  • [8] Mining negative samples on contrastive learning via curricular weighting strategy
    Zhuang, Jin
    Jing, Xiao-Yuan
    Jia, Xiaodong
    INFORMATION SCIENCES, 2024, 668
  • [9] Heterogeneous data augmentation in graph contrastive learning for effective negative samples
    Ali, Adnan
    Li, Jinlong
    Chen, Huanhuan
    COMPUTERS & ELECTRICAL ENGINEERING, 2024, 118
  • [10] Contrastive learning for unsupervised sentence embeddings using negative samples with diminished semantics
    Yu, Zhiyi
    Li, Hong
    Feng, Jialin
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (04): : 5428 - 5445