Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data

被引:998
|
作者
Donoho, DL [1 ]
Grimes, C [1 ]
机构
[1] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
关键词
manifold learning; ISOMAP; tangent coordinates; isometry; Laplacian eigenmaps;
D O I
10.1073/pnas.1031596100
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We describe a method for recovering the underlying parametrization of scattered data (m(i)) lying on a manifold M embedded in high-dimensional Euclidean space. The method, Hessian-based locally linear embedding, derives from a conceptual framework of local isometry in which the manifold M, viewed as a Riemannian submanifold of the ambient Euclidean space R-n, is locally isometric to an open, connected subset Theta of Euclidean space R-d. Because Theta does not have to be convex, this framework is able to handle a significantly wider class of situations than the original ISOMAP algorithm. The theoretical framework revolves around a quadratic form H(f) = integral(M)parallel toH(f)(m)parallel to(F)(2)dm defined on functions f : M --> R. Here Hf denotes the Hessian of f, and H(f) averages the Frobenius norm of the Hessian over M. To define the Hessian, we use orthogonal coordinates on the tangent planes of M. The key observation is that, if M truly is locally isometric to an open, connected subset of R-d, then H(f) has a (d + 1)-dimensional null space consisting of the constant functions and a d-dimensional space of functions spanned by the original isometric coordinates. Hence, the isometric coordinates can be recovered up to a linear isometry. Our method may be viewed as a modification of locally linear embedding and our theoretical framework as a modification of the Laplacian eigen-maps framework, where we substitute a quadratic form based on the Hessian in place of one based on the Laplacian.
引用
收藏
页码:5591 / 5596
页数:6
相关论文
共 50 条
  • [41] HIGH-DIMENSIONAL LINEAR REGRESSION FOR DEPENDENT DATA WITH APPLICATIONS TO NOWCASTING
    Han, Yuefeng
    Tsay, Ruey S.
    STATISTICA SINICA, 2020, 30 (04) : 1797 - 1827
  • [42] Test on the linear combinations of covariance matrices in high-dimensional data
    Bai, Zhidong
    Hu, Jiang
    Wang, Chen
    Zhang, Chao
    STATISTICAL PAPERS, 2021, 62 (02) : 701 - 719
  • [43] Test on the linear combinations of mean vectors in high-dimensional data
    Li, Huiqin
    Hu, Jiang
    Bai, Zhidong
    Yin, Yanqing
    Zou, Kexin
    TEST, 2017, 26 (01) : 188 - 208
  • [44] Generalized linear discriminant analysis for high-dimensional genomic data
    Li, Sisi
    Lewinger, Juan Pablo
    GENETIC EPIDEMIOLOGY, 2018, 42 (07) : 713 - 713
  • [45] Incremental Isometric Embedding of High-Dimensional Data Using Connected Neighborhood Graphs
    Zhao, Dongfang
    Yang, Li
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (01) : 86 - 98
  • [46] ASPECTS OF HIGH-DIMENSIONAL THEORIES IN EMBEDDING SPACES
    MAIA, MD
    MECKLENBURG, W
    JOURNAL OF MATHEMATICAL PHYSICS, 1984, 25 (10) : 3047 - 3050
  • [47] Tensorized Projection for High-Dimensional Binary Embedding
    Hong, Weixiang
    Meng, Jingjing
    Yuan, Junsong
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 69 - 76
  • [48] High-Dimensional Dueling Optimization with Preference Embedding
    Zhang, Yangwenhui
    Qian, Hong
    Shu, Xiang
    Zhou, Aimin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 11280 - 11288
  • [49] A Generalized-Momentum-Accelerated Hessian-Vector Algorithm for High-Dimensional and Sparse Data
    Li, Weiling
    Luo, Xin
    20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 1112 - 1117
  • [50] Hessian spectrum at the global minimum of high-dimensional random landscapes
    Fyodorov, Yan, V
    Le Doussal, Pierre
    JOURNAL OF PHYSICS A-MATHEMATICAL AND THEORETICAL, 2018, 51 (47)