Deep Multirepresentation Learning for Data Clustering

被引:2
|
作者
Sadeghi, Mohammadreza [1 ,2 ]
Armanfard, Narges [1 ,2 ]
机构
[1] McGill Univ, Dept Elect & Comp Engn, Montreal, PQ H3A 0E9, Canada
[2] Mila Quebec AI Inst, Montreal, PQ H2S 3H1, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Clustering algorithms; Training; Task analysis; Decoding; Optimization; Learning systems; Image reconstruction; Autoencoder (AE); cluster-specific AEs; data clustering; multiple representation learning; FUZZY C-MEANS; RECOGNITION;
D O I
10.1109/TNNLS.2023.3289158
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep clustering incorporates embedding into clustering in order to find a lower-dimensional space suitable for clustering tasks. Conventional deep clustering methods aim to obtain a single global embedding subspace (aka latent space) for all the data clusters. In contrast, in this article, we propose a deep multirepresentation learning (DML) framework for data clustering whereby each difficult-to-cluster data group is associated with its own distinct optimized latent space and all the easy-to-cluster data groups are associated with a general common latent space. Autoencoders (AEs) are employed for generating cluster-specific and general latent spaces. To specialize each AE in its associated data cluster(s), we propose a novel and effective loss function which consists of weighted reconstruction and clustering losses of the data points, where higher weights are assigned to the samples more probable to belong to the corresponding cluster(s). Experimental results on benchmark datasets demonstrate that the proposed DML framework and loss function outperform state-of-the-art clustering approaches. In addition, the results show that the DML method significantly outperforms the SOTA on imbalanced datasets as a result of assigning an individual latent space to the difficult clusters.
引用
收藏
页码:15675 / 15686
页数:12
相关论文
共 50 条
  • [1] Deep Successive Subspace Learning for Data Clustering
    Sadeghi, Mohammadreza
    Armanfard, Narges
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [2] Clustering Mixed Data: Bridging the Gap with Deep Learning
    Yerra, Harini
    Kommu, Siddartha
    Kumar, B. Vijay
    Sudam, Rachana
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON DATA SCIENCE, MACHINE LEARNING AND APPLICATIONS, VOL 1, ICDSMLA 2023, 2025, 1273 : 695 - 702
  • [3] A Deep Unsupervised Learning Algorithm for Dynamic Data Clustering
    Pantula, Priyanka D.
    Miriyala, Srinivas S.
    Mitra, Kishalay
    2021 SEVENTH INDIAN CONTROL CONFERENCE (ICC), 2021, : 147 - 152
  • [4] Grid Clustering Analysis the Big Data of Spectrum by Deep Learning
    Chen Shuxin
    Sun Weimin
    2017 2ND INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING (ICIVC 2017), 2017, : 1002 - 1005
  • [5] Deep clustering of tabular data by weighted Gaussian distribution learning
    Rabbani, Shourav B.
    Medri, Ivan, V
    Samad, Manar D.
    NEUROCOMPUTING, 2025, 623
  • [6] Classification and unsupervised clustering of LIGO data with Deep Transfer Learning
    George, Daniel
    Shen, Hongyu
    Huerta, E. A.
    PHYSICAL REVIEW D, 2018, 97 (10)
  • [7] Metaheuristic Based Clustering with Deep Learning Model for Big Data Classification
    Krishnaswamy, R.
    Subramaniam, Kamalraj
    Nandini, V
    Vijayalakshmi, K.
    Kadry, Seifedine
    Nam, Yunyoung
    COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2023, 44 (01): : 391 - 406
  • [8] Deep probability multi-view feature learning for data clustering
    Zhao, Liang
    Wang, Xiao
    Liu, Zhenjiao
    Yuan, Hong
    Zhao, Jingyuan
    Zhou, Shuang
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 217
  • [9] Application of Clustering Analysis in Brain Gene Data Based on Deep Learning
    Suo, Yina
    Liu, Tingwei
    Jia, Xueyong
    Yu, Fuxing
    IEEE ACCESS, 2019, 7 : 2947 - 2956
  • [10] Traffic prediction via clustering and deep transfer learning with limited data
    Zou, Xiexin
    Chung, Edward
    COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2024, 39 (17) : 2683 - 2700