Learning Mixtures of Gaussians Using the DDPM Objective

被引:0
|
作者
Shah, Kulin [1 ]
Chen, Sitan [2 ]
Klivans, Adam [1 ]
机构
[1] UT Austin, Austin, TX 78712 USA
[2] Harvard Univ, Boston, MA USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
关键词
ALGORITHM; EM;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent works have shown that diffusion models can learn essentially any distribution provided one can perform score estimation. Yet it remains poorly understood under what settings score estimation is possible, let alone when practical gradient-based algorithms for this task can provably succeed. In this work, we give the first provably efficient results along these lines for one of the most fundamental distribution families, Gaussian mixture models. We prove that gradient descent on the denoising diffusion probabilistic model (DDPM) objective can efficiently recover the ground truth parameters of the mixture model in the following two settings: 1. We show gradient descent with random initialization learns mixtures of two spherical Gaussians in d dimensions with 1/poly(d)-separated centers. 2. We show gradient descent with a warm start learns mixtures of K spherical Gaussians with Omega(root log(min(K, d)))-separated centers. A key ingredient in our proofs is a new connection between score-based methods and two other approaches to distribution learning, the expectation-maximization (EM) algorithm and spectral methods.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Efficiently Learning Mixtures of Two Gaussians
    Kalai, Adam Tauman
    Moitra, Ankur
    Valiant, Gregory
    STOC 2010: PROCEEDINGS OF THE 2010 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2010, : 553 - 562
  • [2] Robustly Learning General Mixtures of Gaussians
    Liu, Allen
    Moitra, Ankur
    JOURNAL OF THE ACM, 2023, 70 (03)
  • [3] Learning Mixtures of Gaussians in High Dimensions
    Ge, Rong
    Huang, Qingqing
    Kakade, Sham M.
    STOC'15: PROCEEDINGS OF THE 2015 ACM SYMPOSIUM ON THEORY OF COMPUTING, 2015, : 761 - 770
  • [4] Learning mixtures of separated nonspherical gaussians
    Arora, S
    Kannan, R
    ANNALS OF APPLIED PROBABILITY, 2005, 15 (1A): : 69 - 92
  • [5] Density estimation using mixtures of mixtures of Gaussians
    Abd-Almageed, Wael
    Davis, Larry S.
    COMPUTER VISION - ECCV 2006, PT 4, PROCEEDINGS, 2006, 3954 : 410 - 422
  • [6] Robustly Learning Mixtures of k Arbitrary Gaussians
    Bakshi, Ainesh
    Diakonikolas, Ilias
    Jia, He
    Kane, Daniel M.
    Kothari, Pravesh K.
    Vempala, Santosh S.
    PROCEEDINGS OF THE 54TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING (STOC '22), 2022, : 1234 - 1247
  • [7] Formant analysis using mixtures of Gaussians
    Zolfaghari, P
    Robinson, T
    ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 1229 - 1232
  • [8] Robot docking using Mixtures of Gaussians
    Williamson, M
    Murray-Smith, R
    Hansen, V
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 11, 1999, 11 : 945 - 951
  • [9] On Learning Mixtures of Well-Separated Gaussians
    Regev, Oded
    Vijayaraghavan, Aravindan
    2017 IEEE 58TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2017, : 85 - 96
  • [10] Differentially Private Algorithms for Learning Mixtures of Separated Gaussians
    Kamath, Gautam
    Sheffet, Or
    Singhal, Vikrant
    Ullman, Jonathan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32