Adaptive Riemannian stochastic gradient descent and reparameterization for Gaussian mixture model fitting

被引:0
|
作者
Ji, Chunlin [1 ]
Fu, Yuhao [1 ,2 ]
He, Ping [3 ]
机构
[1] Kuang Chi Inst Adv Technol, Shenzhen, Peoples R China
[2] Origin Artificial Intelligence Technol Co, Shenzhen, Peoples R China
[3] HeGuangLiangZi Tech, Shenzhen, Peoples R China
基金
国家重点研发计划;
关键词
Gaussian mixture model; Reparameterization; Symmetric positive definite matrix manifold; Riemannian stochastic gradient descent; Riemannian Adam algorithm; EM ALGORITHM;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in manifold optimization for the Gaussian mixture model (GMM) have gained increasing interest. In this work, instead of directly addressing the manifold optimization on covariance matrices of GMM, we consider the GMM fitting as an optimization of the density function over a statistical manifold and seek the natural gradient to speed up the optimization process. We present an upper bound for the Kullback-Leibler (KL) divergence between two GMMs and obtain simple closed-form expressions for the natural gradients. With the natural gradients, we then apply the Riemannian stochastic gradient descent (RSGD) algorithm to optimize covariance matrices on a symmetric and positive definite (SPD) matrix manifold. We further propose a Riemannian Adam (RAdam) algorithm that extends the momentum method and adaptive learning in the Euclidean space to the SPD manifold space. Extensive simulations show that the proposed algorithms scale well to high-dimensional large-scale datasets and outperform expectation maximization (EM) algorithms in fitted log-likelihood.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] An alternative to EM for Gaussian mixture models: batch and stochastic Riemannian optimization
    Hosseini, Reshad
    Sra, Suvrit
    MATHEMATICAL PROGRAMMING, 2020, 181 (01) : 187 - 223
  • [22] An alternative to EM for Gaussian mixture models: batch and stochastic Riemannian optimization
    Reshad Hosseini
    Suvrit Sra
    Mathematical Programming, 2020, 181 : 187 - 223
  • [23] Revisiting the Noise Model of Stochastic Gradient Descent
    Battash, Barak
    Wolf, Lior
    Lindenbaum, Ofir
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [24] Bolstering stochastic gradient descent with model building
    Birbil, S. Ilker
    Martin, Ozgur
    Onay, Gonenc
    Oztoprak, Figen
    TOP, 2024, 32 (03) : 517 - 536
  • [25] Sampling from Gaussian Process Posteriors using Stochastic Gradient Descent
    Lin, Jihao Andreas
    Antoran, Javier
    Padhy, Shreyas
    Janz, David
    Hernandez-Lobato, Jose Miguel
    Terenin, Alexander
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Stochastic Gradient Descent on a Tree: an Adaptive and Robust Approach to Stochastic Convex Optimization
    Vakili, Sattar
    Salgia, Sudeep
    Zhao, Qing
    2019 57TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2019, : 432 - 438
  • [27] FTSGD: An Adaptive Stochastic Gradient Descent Algorithm for Spark MLlib
    Zhang, Hong
    Liu, Zixia
    Huang, Hai
    Wang, Liqiang
    2018 16TH IEEE INT CONF ON DEPENDABLE, AUTONOM AND SECURE COMP, 16TH IEEE INT CONF ON PERVAS INTELLIGENCE AND COMP, 4TH IEEE INT CONF ON BIG DATA INTELLIGENCE AND COMP, 3RD IEEE CYBER SCI AND TECHNOL CONGRESS (DASC/PICOM/DATACOM/CYBERSCITECH), 2018, : 828 - 835
  • [28] Adaptive Sampling for Incremental Optimization Using Stochastic Gradient Descent
    Papa, Guillaume
    Bianchi, Pascal
    Clemencon, Stephan
    ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 317 - 331
  • [29] FITTING A GAUSSIAN MIXTURE MODEL THROUGH THE GINI INDEX
    Lopez-Lobato, Adriana Laura
    Avendano-Garrido, Martha Lorena
    INTERNATIONAL JOURNAL OF APPLIED MATHEMATICS AND COMPUTER SCIENCE, 2021, 31 (03) : 487 - 500
  • [30] A Stochastic Gradient Descent Algorithm Based on Adaptive Differential Privacy
    Deng, Yupeng
    Li, Xiong
    He, Jiabei
    Liu, Yuzhen
    Liang, Wei
    COLLABORATIVE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, COLLABORATECOM 2022, PT II, 2022, 461 : 133 - 152