Spectral clustering based on iterative optimization for large-scale and high-dimensional data

被引:29
|
作者
Zhao, Yang [1 ,2 ]
Yuan, Yuan [1 ]
Nie, Feiping [3 ,4 ]
Wang, Qi [3 ,4 ,5 ]
机构
[1] Chinese Acad Sci, Ctr OPT IMagery Anal & Learning OPTIMAL, Xian Inst Opt & Precis Mech, Xian 710119, Shaanxi, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 049, Peoples R China
[3] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Shaanxi, Peoples R China
[4] Northwestern Polytech Univ, Ctr OPT IMagery Anal & Learning OPTIMAL, Xian 710072, Shaanxi, Peoples R China
[5] Northwestern Polytech Univ, Unmanned Syst Res Inst, Xian 710072, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Manifold learning; Spectral clustering; Multi-task learning; CUTS;
D O I
10.1016/j.neucom.2018.08.059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spectral graph theoretic methods have been a fundamental and important topic in the field of manifold learning and it has become a vital tool in data clustering. However, spectral clustering approaches are limited by their computational demands. It would be too expensive to provide an optimal approximation for spectral decomposition in dealing with large-scale and high-dimensional data sets. On the other hand, the rapid development of data on the Web has posed many rising challenges to the traditional single-task clustering, while the multi-task clustering provides many new thoughts for real-world applications such as video segmentation. In this paper, we will study a Spectral Clustering based on Iterative Optimization (SCIO), which solves the spectral decomposition problem of large-scale and high-dimensional data sets and it well performs on multi-task clustering. Extensive experiments on various synthetic data sets and real-world data sets demonstrate that the proposed method provides an efficient solution for spectral clustering. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:227 / 235
页数:9
相关论文
共 50 条
  • [21] Large-scale spectral clustering based on pairwise constraints
    Semertzidis, T.
    Rafailidis, D.
    Strintzis, M. G.
    Daras, P.
    INFORMATION PROCESSING & MANAGEMENT, 2015, 51 (05) : 616 - 624
  • [22] Efficient distributed optimization for large-scale high-dimensional sparse penalized Huber regression
    Pan, Yingli
    Xu, Kaidong
    Wei, Sha
    Wang, Xiaojuan
    Liu, Zhan
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2024, 53 (07) : 3106 - 3125
  • [23] High-dimensional data clustering
    Bouveyron, C.
    Girard, S.
    Schmid, C.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2007, 52 (01) : 502 - 519
  • [24] Clustering High-Dimensional Data
    Masulli, Francesco
    Rovetta, Stefano
    CLUSTERING HIGH-DIMENSIONAL DATA, CHDD 2012, 2015, 7627 : 1 - 13
  • [25] Spectral clustering with linear embedding: A discrete clustering method for large-scale data
    Gao, Chenhui
    Chen, Wenzhi
    Nie, Feiping
    Yu, Weizhong
    Wang, Zonghui
    PATTERN RECOGNITION, 2024, 151
  • [26] A study of large-scale data clustering based on fuzzy clustering
    Li, Yangyang
    Yang, Guoli
    He, Haiyang
    Jiao, Licheng
    Shang, Ronghua
    SOFT COMPUTING, 2016, 20 (08) : 3231 - 3242
  • [27] A study of large-scale data clustering based on fuzzy clustering
    Yangyang Li
    Guoli Yang
    Haiyang He
    Licheng Jiao
    Ronghua Shang
    Soft Computing, 2016, 20 : 3231 - 3242
  • [28] Visualizing large-scale high-dimensional data via hierarchical embedding of KNN graphs
    Zhu, Haiyang
    Zhu, Minfeng
    Feng, Yingchaojie
    Cai, Deng
    Hu, Yuanzhe
    Wu, Shilong
    Wu, Xiangyang
    Chen, Wei
    VISUAL INFORMATICS, 2021, 5 (02) : 51 - 59
  • [29] Monitoring high-dimensional data for failure detection and localization in large-scale computing systems
    Chen, Haifeng
    Jiang, Guofei
    Yoshihira, Kenji
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2008, 20 (01) : 13 - 25
  • [30] Asynchronous Distributed ADMM for Learning with Large-Scale and High-Dimensional Sparse Data Set
    Wang, Dongxia
    Lei, Yongmei
    ADVANCED HYBRID INFORMATION PROCESSING, ADHIP 2019, PT II, 2019, 302 : 259 - 274