DIMENSIONALITY REDUCTION OF HIGH-DIMENSIONAL DATA WITH A NONLINEAR PRINCIPAL COMPONENT ALIGNED GENERATIVE TOPOGRAPHIC MAPPING

被引:2
|
作者
Griebel, M. [1 ]
Hullmann, A. [1 ]
机构
[1] Univ Bonn, Inst Numer Simulat, D-53115 Bonn, Germany
来源
SIAM JOURNAL ON SCIENTIFIC COMPUTING | 2014年 / 36卷 / 03期
关键词
dimensionality reduction; generative topographic mapping; principal component analysis; density estimation; additive model; classification; EM ALGORITHM;
D O I
10.1137/130931382
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Most high-dimensional real-life data exhibit some dependencies such that data points do not populate the whole data space but lie approximately on a lower-dimensional manifold. A major problem in many data mining applications is the detection of such a manifold and the expression of the given data in terms of a moderate number of latent variables. We present a method which is derived from the generative topographic mapping (GTM) and can be seen as a nonlinear generalization of the principal component analysis (PCA). It can detect certain nonlinearities in the data but does not suffer from the curse of dimensionality with respect to the latent space dimension as the original GTM and thus allows for higher embedding dimensions. We provide experiments that show that our approach leads to an improved data reconstruction compared to the purely linear PCA and that it can furthermore be used for classification.
引用
收藏
页码:A1027 / A1047
页数:21
相关论文
共 50 条
  • [31] Adaptive local Principal Component Analysis improves the clustering of high-dimensional data
    Migenda, Nico
    Moeller, Ralf
    Schenck, Wolfram
    PATTERN RECOGNITION, 2024, 146
  • [32] Cauchy robust principal component analysis with applications to high-dimensional data sets
    Aisha Fayomi
    Yannis Pantazis
    Michail Tsagris
    Andrew T. A. Wood
    Statistics and Computing, 2024, 34
  • [33] Efficient Dimensionality Reduction for High-Dimensional Network Estimation
    Celik, Safiye
    Logsdon, Benjamin A.
    Lee, Su-In
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1953 - 1961
  • [34] High-dimensional principal component analysis with heterogeneous missingness
    Zhu, Ziwei
    Wang, Tengyao
    Samworth, Richard J.
    JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2022, 84 (05) : 2000 - 2031
  • [35] PRINCIPAL COMPONENT ANALYSIS IN VERY HIGH-DIMENSIONAL SPACES
    Lee, Young Kyung
    Lee, Eun Ryung
    Park, Byeong U.
    STATISTICA SINICA, 2012, 22 (03) : 933 - 956
  • [36] Test for high-dimensional outliers with principal component analysis
    Nakayama, Yugo
    Yata, Kazuyoshi
    Aoshima, Makoto
    JAPANESE JOURNAL OF STATISTICS AND DATA SCIENCE, 2024, 7 (02) : 739 - 766
  • [37] Efficient high-dimensional indexing by sorting principal component
    Cui, Jiangtao
    Zhou, Shuisheng
    Sun, Junding
    PATTERN RECOGNITION LETTERS, 2007, 28 (16) : 2412 - 2418
  • [38] DYNAMICAL COMPONENT ANALYSIS (DYCA): DIMENSIONALITY REDUCTION FOR HIGH-DIMENSIONAL DETERMINISTIC TIME-SERIES
    Seifert, Bastian
    Korn, Katharina
    Hartmann, Steffen
    Uhl, Christian
    2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [39] High-dimensional covariance forecasting based on principal component analysis of high-frequency data
    Jian, Zhihong
    Deng, Pingjun
    Zhu, Zhican
    ECONOMIC MODELLING, 2018, 75 : 422 - 431
  • [40] Data-Efficient Dimensionality Reduction and Surrogate Modeling of High-Dimensional Stress Fields
    Samaddar, Anirban
    Ravi, Sandipp Krishnan
    Ramachandra, Nesar
    Luan, Lele
    Madireddy, Sandeep
    Bhaduri, Anindya
    Pandita, Piyush
    Sun, Changjie
    Wang, Liping
    JOURNAL OF MECHANICAL DESIGN, 2025, 147 (03)