Adaptive Riemannian stochastic gradient descent and reparameterization for Gaussian mixture model fitting

被引:0
|
作者
Ji, Chunlin [1 ]
Fu, Yuhao [1 ,2 ]
He, Ping [3 ]
机构
[1] Kuang Chi Inst Adv Technol, Shenzhen, Peoples R China
[2] Origin Artificial Intelligence Technol Co, Shenzhen, Peoples R China
[3] HeGuangLiangZi Tech, Shenzhen, Peoples R China
基金
国家重点研发计划;
关键词
Gaussian mixture model; Reparameterization; Symmetric positive definite matrix manifold; Riemannian stochastic gradient descent; Riemannian Adam algorithm; EM ALGORITHM;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in manifold optimization for the Gaussian mixture model (GMM) have gained increasing interest. In this work, instead of directly addressing the manifold optimization on covariance matrices of GMM, we consider the GMM fitting as an optimization of the density function over a statistical manifold and seek the natural gradient to speed up the optimization process. We present an upper bound for the Kullback-Leibler (KL) divergence between two GMMs and obtain simple closed-form expressions for the natural gradients. With the natural gradients, we then apply the Riemannian stochastic gradient descent (RSGD) algorithm to optimize covariance matrices on a symmetric and positive definite (SPD) matrix manifold. We further propose a Riemannian Adam (RAdam) algorithm that extends the momentum method and adaptive learning in the Euclidean space to the SPD manifold space. Extensive simulations show that the proposed algorithms scale well to high-dimensional large-scale datasets and outperform expectation maximization (EM) algorithms in fitted log-likelihood.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Stochastic Gradient Descent on Riemannian Manifolds
    Bonnabel, Silvere
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2013, 58 (09) : 2217 - 2229
  • [2] STOCHASTIC MODIFIED FLOWS FOR RIEMANNIAN STOCHASTIC GRADIENT DESCENT
    Gess, Benjamin
    Kassing, Sebastian
    Rana, Nimit
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2024, 62 (06) : 3288 - 3314
  • [3] A Gradient-Descent Method for Curve Fitting on Riemannian Manifolds
    Samir, Chafik
    Absil, P. -A.
    Srivastava, Anuj
    Klassen, Eric
    FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2012, 12 (01) : 49 - 73
  • [4] A Gradient-Descent Method for Curve Fitting on Riemannian Manifolds
    Chafik Samir
    P.-A. Absil
    Anuj Srivastava
    Eric Klassen
    Foundations of Computational Mathematics, 2012, 12 : 49 - 73
  • [5] CONVERGENCE OF RIEMANNIAN STOCHASTIC GRADIENT DESCENT ON HADAMARD MANIFOLD
    Sakai, Hiroyuki
    Iiduka, Hideaki
    PACIFIC JOURNAL OF OPTIMIZATION, 2024, 20 (04): : 743 - 767
  • [6] Dynamical mean-field theory for stochastic gradient descent in Gaussian mixture classification*
    Mignacco, Francesca
    Krzakala, Florent
    Urbani, Pierfrancesco
    Zdeborova, And Lenka
    JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2021, 2021 (12):
  • [7] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    arXiv, 2019,
  • [8] On the Convergence of Stochastic Gradient Descent with Adaptive Stepsizes
    Li, Xiaoyu
    Orabona, Francesco
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [9] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [10] Riemannian proximal stochastic gradient descent for sparse 2DPCA
    Zhang, Zhuan
    Zhou, Shuisheng
    Li, Dong
    Yang, Ting
    DIGITAL SIGNAL PROCESSING, 2022, 122