Hard Negative Sampling via Regularized Optimal Transport for Contrastive Representation Learning

被引:1
|
作者
Jiang, Ruijie [1 ]
Ishwar, Prakash [2 ]
Aeron, Shuchin [1 ]
机构
[1] Tufts Univ, Dept ECE, Medford, MA 02155 USA
[2] Boston Univ, Dept ECE, Boston, MA USA
关键词
contrastive representation learning; hard negative sampling; optimal transport (OT);
D O I
10.1109/IJCNN54540.2023.10191650
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of designing hard negative sampling distributions for unsupervised contrastive representation learning. We propose and analyze a novel min-max framework that seeks a representation which minimizes the maximum (worst-case) generalized contrastive learning loss over all couplings (joint distributions between positive and negative samples subject to marginal constraints) and prove that the resulting min-max optimum representation will be degenerate. This provides the first theoretical justification for incorporating additional regularization constraints on the couplings. We re-interpret the min-max problem through the lens of Optimal Transport (OT) theory and utilize regularized transport couplings to control the degree of hardness of negative examples. Through experiments we demonstrate that the negative samples generated from our designed negative distribution are more similar to the anchor than those generated from the baseline negative distribution. We also demonstrate that entropic regularization yields negative sampling distributions with parametric form similar to that in a recent state-of-the-art negative sampling design and has similar performance in multiple datasets. Utilizing the uncovered connection with OT, we propose a new ground cost for designing the negative distribution and show improved performance of the learned representation on downstream tasks compared to the representation learned when using squared Euclidean cost.(1)
引用
收藏
页数:8
相关论文
共 50 条
  • [41] Unsupervised Path Representation Learning with Curriculum Negative Sampling
    Bin Yang, Sean
    Guo, Chenjuan
    Hu, Jilin
    Tang, Jian
    Yang, Bin
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3286 - 3292
  • [42] Incorporating GAN for Negative Sampling in Knowledge Representation Learning
    Wang, Peifeng
    Li, Shuangyin
    Pan, Rong
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2005 - 2012
  • [43] Scalable Thompson Sampling via Optimal Transport
    Zhang, Ruiyi
    Wen, Zheng
    Chen, Changyou
    Fang, Chen
    Yu, Tong
    Carin, Lawrence
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89 : 87 - 96
  • [44] Convergence Rates for Regularized Optimal Transport via Quantization br
    Eckstein, Stephan
    Nutz, Marcel
    MATHEMATICS OF OPERATIONS RESEARCH, 2024, 49 (02) : 1223 - 1240
  • [45] Optimal learning rates for least squares regularized regression with unbounded sampling
    Wang, Cheng
    Zhou, Ding-Xuan
    JOURNAL OF COMPLEXITY, 2011, 27 (01) : 55 - 67
  • [46] Multimodal Fake News Detection with Contrastive Learning and Optimal Transport
    Shen, Xiaorong
    Huang, Maowei
    Hu, Zheng
    Cai, Shimin
    Zhou, Tao
    FRONTIERS IN COMPUTER SCIENCE, 2024, 6
  • [47] PolyCL: contrastive learning for polymer representation learning via explicit and implicit augmentations
    Zhou, Jiajun
    Yang, Yijie
    Mroz, Austin M.
    Jelfs, Kim E.
    DIGITAL DISCOVERY, 2025, 4 (01): : 149 - 160
  • [48] HSimCSE: Improving Contrastive Learning of Unsupervised Sentence Representation with Adversarial Hard Positives and Dual Hard Negatives
    Xu, Bo
    Wei, Shouang
    Cheng, Luyi
    Huang, Shizhou
    Song, Hui
    Du, Ming
    Wang, Hongya
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [49] Toward Interpretable Semantic Textual Similarity via Optimal Transport-based Contrastive Sentence Learning
    Lee, Seonghyeon
    Lee, Dongha
    Jang, Seongbo
    Yu, Hwanjo
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5969 - 5979
  • [50] MocGCL: Molecular Graph Contrastive Learning via Negative Selection
    Cui, Jinhao
    Chai, Heyan
    Gong, Yanbin
    Ding, Ye
    Hua, Zhongyun
    Gao, Cuiyun
    Liao, Qing
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,