Hard Negative Sampling via Regularized Optimal Transport for Contrastive Representation Learning

被引:1
|
作者
Jiang, Ruijie [1 ]
Ishwar, Prakash [2 ]
Aeron, Shuchin [1 ]
机构
[1] Tufts Univ, Dept ECE, Medford, MA 02155 USA
[2] Boston Univ, Dept ECE, Boston, MA USA
关键词
contrastive representation learning; hard negative sampling; optimal transport (OT);
D O I
10.1109/IJCNN54540.2023.10191650
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of designing hard negative sampling distributions for unsupervised contrastive representation learning. We propose and analyze a novel min-max framework that seeks a representation which minimizes the maximum (worst-case) generalized contrastive learning loss over all couplings (joint distributions between positive and negative samples subject to marginal constraints) and prove that the resulting min-max optimum representation will be degenerate. This provides the first theoretical justification for incorporating additional regularization constraints on the couplings. We re-interpret the min-max problem through the lens of Optimal Transport (OT) theory and utilize regularized transport couplings to control the degree of hardness of negative examples. Through experiments we demonstrate that the negative samples generated from our designed negative distribution are more similar to the anchor than those generated from the baseline negative distribution. We also demonstrate that entropic regularization yields negative sampling distributions with parametric form similar to that in a recent state-of-the-art negative sampling design and has similar performance in multiple datasets. Utilizing the uncovered connection with OT, we propose a new ground cost for designing the negative distribution and show improved performance of the learned representation on downstream tasks compared to the representation learned when using squared Euclidean cost.(1)
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Label-aware Hard Negative Sampling Strategies with Momentum Contrastive Learning for Implicit Hate Speech Detection
    Kim, Jaehoon
    Jin, Seungwan
    Park, Sohyun
    Park, Someen
    Han, Kyungsik
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 16177 - 16188
  • [22] Adversarial Hard Negative Generation for Complementary Graph Contrastive Learning
    Wang, Senzhang
    Yan, Hao
    Du, Jinlong
    Yin, Jun
    Zhu, Junxing
    Li, Chaozhuo
    Wang, Jianxin
    PROCEEDINGS OF THE 2023 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2023, : 163 - 171
  • [23] Learning Common Semantics via Optimal Transport for Contrastive Multi-View Clustering
    Zhang, Qian
    Zhang, Lin
    Song, Ran
    Cong, Runmin
    Liu, Yonghuai
    Zhang, Wei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4501 - 4515
  • [24] Boosting Graph Contrastive Learning via Adaptive Sampling
    Wan, Sheng
    Zhan, Yibing
    Chen, Shuo
    Pan, Shirui
    Yang, Jian
    Tao, Dacheng
    Gong, Chen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15971 - 15983
  • [25] An effective negative sampling approach for contrastive learning of sentence embedding
    Tan, Qitao
    Song, Xiaoying
    Ye, Guanghui
    Wu, Chuan
    MACHINE LEARNING, 2023, 112 (12) : 4837 - 4861
  • [26] An effective negative sampling approach for contrastive learning of sentence embedding
    Qitao Tan
    Xiaoying Song
    Guanghui Ye
    Chuan Wu
    Machine Learning, 2023, 112 : 4837 - 4861
  • [27] Probing Negative Sampling for Contrastive Learning to Learn Graph Representations
    Chen, Shiyi
    Wang, Ziao
    Zhang, Xinni
    Zhang, Xiaofeng
    Peng, Dan
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT II, 2021, 12976 : 434 - 449
  • [28] Contrastive Learning and Multi-Choice Negative Sampling Recommendation
    Xue, Yun
    Cai, Xiaodong
    Fang, Sheng
    Zhou, Li
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (05) : 905 - 912
  • [29] JPEG Artifacts Removal via Contrastive Representation Learning
    Wang, Xi
    Fu, Xueyang
    Zhu, Yurui
    Zha, Zheng-Jun
    COMPUTER VISION - ECCV 2022, PT XVII, 2022, 13677 : 615 - 631
  • [30] Learning Fair Representation via Distributional Contrastive Disentanglement
    Oh, Changdae
    Won, Heeji
    So, Junhyuk
    Kim, Taero
    Kim, Yewon
    Choi, Hosik
    Song, Kyungwoo
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 1295 - 1305