Joint unsupervised contrastive learning and robust GMM for text clustering

被引:6
|
作者
Hu, Chenxi [1 ]
Wu, Tao [1 ,2 ]
Liu, Shuaiqi [2 ]
Liu, Chunsheng [1 ]
Ma, Tao [1 ]
Yang, Fang [1 ]
机构
[1] Natl Univ Def Technol, Coll Elect Engn, Hefei 230031, Peoples R China
[2] Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
关键词
Text clustering; Contrastive learning; Negative sampling; Gaussian mixture model; Expectation maximization;
D O I
10.1016/j.ipm.2023.103529
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text clustering aims to organize a vast collection of documents into meaningful and coherent clusters, thereby facilitating the extraction of valuable insights. While current frameworks for text clustering try to minimize the anisotropy of pre-trained language models through contrastive learning of text embeddings, the approach of treating in-batch samples as negatives is suboptimal. The K-means algorithm offers a way to sample both hard negatives and false negatives. However, relying solely on a single measure of semantic similarity between distribu-tions and using coarse-grained weighting for negative pairs may potentially limit performance. Furthermore, considering the very similar distribution in text clusters due to rich semantics, the Mahalanobis distance-based Gaussian Mixture Model (GMM) is prone to falling into local optima due to one Gaussian model, having a smaller weight, may gradually merging into another during the parameter evaluation by the EM algorithm. To tackle these challenges, we propose a model named JourTC: Joint unsupervised contrastive learning and robust GMM for Text Clustering. In the contrastive learning phase, hard negatives, potential false negatives, and their corresponding global similarity-aware weights are determined through posterior probabilities derived from a Robust GMM (RGMM). This RGMM utilizes the entropy of each individual Gaussian model as a metric and adaptively adjusts the posterior probabilities of samples based on the Gaussian models with both maximum and minimum entropy to diminish the influence of low-entropy Gaussian models. Extensive experiments have shown that JourTC can be seamlessly integrated into existing text clustering frameworks, leading to a notable improvement in accuracy. Our code is publicly available.1
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Joint Reinforcement and Contrastive Learning for Unsupervised Video Summarization
    Zhang, Yunzuo
    Liu, Yameng
    Zhu, Pengfei
    Kang, Weili
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2587 - 2591
  • [2] MoCoUTRL: a momentum contrastive framework for unsupervised text representation learning
    Zou, Ao
    Hao, Wenning
    Jin, Dawei
    Chen, Gang
    Sun, Feiyan
    CONNECTION SCIENCE, 2023, 35 (01)
  • [3] Unsupervised Clustering of Microseismic Signals Using a Contrastive Learning Model
    Yang, Zhen
    Li, Huailiang
    Tuo, Xianguo
    Li, Linjia
    Wen, Junnan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [4] Unsupervised Clustering of Microseismic Signals Using a Contrastive Learning Model
    Yang, Zhen
    Li, Huailiang
    Tuo, Xianguo
    Li, Linjia
    Wen, Junnan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [5] Improving Unsupervised Image Clustering With Robust Learning
    Park, Sungwon
    Han, Sungwon
    Kim, Sundong
    Kim, Danu
    Park, Sungkyu
    Hong, Seunghoon
    Cha, Meeyoung
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 12273 - 12282
  • [6] Supervised contrastive learning for robust text adversarial training
    Weidong Li
    Bo Zhao
    Yang An
    Chenhan Shangguan
    Minzi Ji
    Anqi Yuan
    Neural Computing and Applications, 2023, 35 : 7357 - 7368
  • [7] A Robust and Effective Text Detector Supervised by Contrastive Learning
    Wei, Ran
    Li, Yaoyi
    Li, Haiyan
    Tang, Ze
    Lu, Hongtao
    Cai, Nengbin
    Zhao, Xuejun
    IEEE ACCESS, 2021, 9 : 26431 - 26441
  • [8] Supervised contrastive learning for robust text adversarial training
    Li, Weidong
    Zhao, Bo
    An, Yang
    Shangguan, Chenhan
    Ji, Minzi
    Yuan, Anqi
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (10): : 7357 - 7368
  • [9] Graph Debiased Contrastive Learning with Joint Representation Clustering
    Zhao, Han
    Yang, Xu
    Wang, Zhenru
    Yang, Erkun
    Deng, Cheng
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3434 - 3440
  • [10] Contrastive Learning with Transformer Initialization and Clustering Prior for Text Representation
    Liu, Chenjing
    Chen, Xiangru
    Hu, Peng
    Lin, Jie
    Wang, Junfeng
    Geng, Xue
    APPLIED SOFT COMPUTING, 2024, 166