Combining Semi-supervised Clustering and Classification Under a Generalized Framework

被引:0
|
作者
Jiang, Zhen [1 ,2 ]
Zhao, Lingyun [1 ]
Lu, Yu [1 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Commun Engn, Zhenjiang, Peoples R China
[2] Jiangsu Prov Big Data Ubiquitous Percept & Intelli, Zhenjiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Co-training; Classification; Semi-supervised clustering; Cluster-splitting;
D O I
10.1007/s00357-024-09489-9
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Most machine learning algorithms rely on having a sufficient amount of labeled data to train a reliable classifier. However, labeling data is often costly and time-consuming, while unlabeled data can be readily accessible. Therefore, learning from both labeled and unlabeled data has become a hot topic of interest. Inspired by the co-training algorithm, we present a learning framework called CSCC, which combines semi-supervised clustering and classification to learn from both labeled and unlabeled data. Unlike existing co-training style methods that construct diverse classifiers to learn from each other, CSCC leverages the diversity between semi-supervised clustering and classification models to achieve mutual enhancement. Existing classification algorithms can be easily adapted to CSCC, allowing them to generalize from a few labeled data. Especially, in order to bridge the gap between class information and clustering, we propose a semi-supervised hierarchical clustering algorithm that utilizes labeled data to guide the process of cluster-splitting. Within the CSCC framework, we introduce two loss functions to supervise the iterative updating of the semi-supervised clustering and classification models, respectively. Extensive experiments conducted on a variety of benchmark datasets validate the superiority of CSCC over other state-of-the-art methods.
引用
收藏
页码:181 / 204
页数:24
相关论文
共 50 条
  • [41] Semi-Supervised Clustering Under a "Compact-Cluster" Assumption
    Jiang, Zhen
    Zhan, Yongzhao
    Mao, Qirong
    Du, Yang
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (05) : 5244 - 5256
  • [42] An Effective Semi-Supervised Learning Framework for Temporal Student Classification
    Vo Thi Ngoc Chau
    Nguyen Hua Phung
    PROCEEDINGS OF 2019 6TH NATIONAL FOUNDATION FOR SCIENCE AND TECHNOLOGY DEVELOPMENT (NAFOSTED) CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS), 2019, : 363 - 369
  • [43] Robust adaptive learning framework for semi-supervised pattern classification
    Ma, Jun
    Yu, Guolin
    SIGNAL PROCESSING, 2024, 224
  • [44] Semi-supervised graph learning framework for apicomplexan parasite classification
    Ha, Yan
    Meng, Xiangjie
    Du, Zeyu
    Tian, Junfeng
    Yuan, Yu
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 81
  • [45] A Semi-Supervised and Incremental Modeling Framework for Wafer Map Classification
    Kong, Yuting
    Ni, Dong
    IEEE TRANSACTIONS ON SEMICONDUCTOR MANUFACTURING, 2020, 33 (01) : 62 - 71
  • [46] A semi-supervised autoencoder framework for joint generation and classification of breathing
    Pastor-Serrano, Oscar
    Lathouwers, Danny
    Perko, Zoltan
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2021, 209
  • [47] A novel semi-supervised learning framework for hyperspectral image classification
    Ye, Zhijing
    Li, Hong
    Song, Yalong
    Wang, Jianzhong
    Benediktsson, Jon Atli
    INTERNATIONAL JOURNAL OF WAVELETS MULTIRESOLUTION AND INFORMATION PROCESSING, 2016, 14 (02)
  • [48] SIAVC: Semi-Supervised Framework for Industrial Accident Video Classification
    Li, Zuoyong
    Lin, Qinghua
    Fan, Haoyi
    Zhao, Tiesong
    Zhang, David
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2603 - 2615
  • [49] Semi-supervised classification trees
    Levatic, Jurica
    Ceci, Michelangelo
    Kocev, Dragi
    Dzeroski, Saso
    JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2017, 49 (03) : 461 - 486
  • [50] Watersheds for Semi-Supervised Classification
    Challa, Aditya
    Danda, Sravan
    Sagar, B. S. Daya
    Najman, Laurent
    IEEE SIGNAL PROCESSING LETTERS, 2019, 26 (05) : 720 - 724