Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data

被引:998
|
作者
Donoho, DL [1 ]
Grimes, C [1 ]
机构
[1] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
关键词
manifold learning; ISOMAP; tangent coordinates; isometry; Laplacian eigenmaps;
D O I
10.1073/pnas.1031596100
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We describe a method for recovering the underlying parametrization of scattered data (m(i)) lying on a manifold M embedded in high-dimensional Euclidean space. The method, Hessian-based locally linear embedding, derives from a conceptual framework of local isometry in which the manifold M, viewed as a Riemannian submanifold of the ambient Euclidean space R-n, is locally isometric to an open, connected subset Theta of Euclidean space R-d. Because Theta does not have to be convex, this framework is able to handle a significantly wider class of situations than the original ISOMAP algorithm. The theoretical framework revolves around a quadratic form H(f) = integral(M)parallel toH(f)(m)parallel to(F)(2)dm defined on functions f : M --> R. Here Hf denotes the Hessian of f, and H(f) averages the Frobenius norm of the Hessian over M. To define the Hessian, we use orthogonal coordinates on the tangent planes of M. The key observation is that, if M truly is locally isometric to an open, connected subset of R-d, then H(f) has a (d + 1)-dimensional null space consisting of the constant functions and a d-dimensional space of functions spanned by the original isometric coordinates. Hence, the isometric coordinates can be recovered up to a linear isometry. Our method may be viewed as a modification of locally linear embedding and our theoretical framework as a modification of the Laplacian eigen-maps framework, where we substitute a quadratic form based on the Hessian in place of one based on the Laplacian.
引用
收藏
页码:5591 / 5596
页数:6
相关论文
共 50 条
  • [31] Quantization techniques for similarity search in high-dimensional data spaces
    Garcia-Arellano, C
    Sevcik, K
    NEW HORIZONS IN INFORMATION MANAGEMENT, 2003, 2712 : 75 - 94
  • [32] High-dimensional data
    Geubbelmans, Melvin
    Rousseau, Axel-Jan
    Valkenborg, Dirk
    Burzykowski, Tomasz
    AMERICAN JOURNAL OF ORTHODONTICS AND DENTOFACIAL ORTHOPEDICS, 2023, 164 (03) : 453 - 456
  • [33] High-dimensional data
    Amaratunga, Dhammika
    Cabrera, Javier
    JOURNAL OF THE NATIONAL SCIENCE FOUNDATION OF SRI LANKA, 2016, 44 (01): : 3 - 9
  • [34] Generative adversarial networks and hessian locally linear embedding for geometric variations management in manufacturing
    Qie, Yifan
    Schleich, Benjamin
    Anwer, Nabil
    JOURNAL OF INTELLIGENT MANUFACTURING, 2025, 36 (02) : 1033 - 1062
  • [35] Optimal Linear Discriminant Analysis for High-Dimensional Functional Data
    Xue, Kaijie
    Yang, Jin
    Yao, Fang
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024, 119 (546) : 1055 - 1064
  • [36] Generalized autoregressive linear models for discrete high-dimensional data
    Pandit P.
    Sahraee-Ardakan M.
    Amini A.A.
    Rangan S.
    Fletcher A.K.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (03): : 884 - 896
  • [37] Test on the linear combinations of covariance matrices in high-dimensional data
    Zhidong Bai
    Jiang Hu
    Chen Wang
    Chao Zhang
    Statistical Papers, 2021, 62 : 701 - 719
  • [38] Test on the linear combinations of mean vectors in high-dimensional data
    Huiqin Li
    Jiang Hu
    Zhidong Bai
    Yanqing Yin
    Kexin Zou
    TEST, 2017, 26 : 188 - 208
  • [39] Scaling log-linear analysis to high-dimensional data
    Petitjean, Francois
    Webb, Geoffrey I.
    Nicholson, Ann E.
    2013 IEEE 13TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2013, : 597 - 606
  • [40] Generalized Linear Discriminant Analysis for High-Dimensional Genomic Data
    Li, Sisi
    Lewinger, Juan Pablo
    GENETIC EPIDEMIOLOGY, 2017, 41 (07) : 704 - 704