On convergence and parameter selection of the EM and DA-EM algorithms for Gaussian mixtures

被引:29
|
作者
Yu, Jian [1 ]
Chaomurilige, Chaomu [1 ]
Yang, Miin-Shen [2 ]
机构
[1] Beijing Jiaotong Univ, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] Chung Yuan Christian Univ, Dept Appl Math, Chungli 32023, Taiwan
关键词
Expectation & maximization (EM) algorithm; Deterministic annealing EM (DA-EM); GAUSSIAN mixtures; Self-annealing; Convergence; Parameter selection; MAXIMUM-LIKELIHOOD; MODELS;
D O I
10.1016/j.patcog.2017.12.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The expectation & maximization (EM) for Gaussian mixtures is popular as a clustering algorithm. However, the EM algorithm is sensitive to initial values, and so Ueda and Nakano [4] proposed the deterministic annealing EM (DA-EM) algorithm to improve it. In this paper, we investigate theoretical behaviors of the EM and DA-EM algorithms. We first derive a general Jacobian matrix of the DA-EM algorithm with respect to posterior probabilities. We then propose a theoretical lower bound for initialization of the annealing parameter in the DA-EM algorithm. On the other hand, some researches mentioned that the EM algorithm exhibits a self-annealing behavior, that is, the equal posterior probability with small random perturbations can avoid the EM algorithm to output the mass center for Gaussian mixtures. However, there is no theoretical analysis on this self-annealing property. Since the DA-EM will become the EM when the annealing parameter is 1, according to the Jacobian matrix of the DA-EM, we can prove the self-annealing property of the EM algorithm for Gaussian mixtures. Based on these results, we give not only convergence behaviors of the equal posterior probabilities and initialization lower bound of the temperature parameter of the DA-EM, but also a theoretical explanation why the EM algorithm for Gaussian mixtures exhibits a self-annealing behavior. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:188 / 203
页数:16
相关论文
共 50 条
  • [21] An MCMC Based EM Algorithm for Mixtures of Gaussian Processes
    Wu, Di
    Chen, Ziyi
    Ma, Jinwen
    ADVANCES IN NEURAL NETWORKS - ISNN 2015, 2015, 9377 : 327 - 334
  • [22] Estimation of structured Gaussian mixtures: The inverse EM algorithm
    Snoussi, Hichem
    Mohammad-Djafari, Ali
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2007, 55 (07) : 3185 - 3191
  • [23] Convergence results for the EM approach to mixtures of experts architectures
    Jordan, MI
    Xu, L
    NEURAL NETWORKS, 1995, 8 (09) : 1409 - 1431
  • [24] Convergence properties of the EM algorithm in constrained parameter spaces
    Nettleton, D
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 1999, 27 (03): : 639 - 648
  • [25] GMM Parameter Estimation by Means of EM and Genetic Algorithms
    Zablotskiy, Sergey
    Pitakrat, Teerat
    Zablotskaya, Kseniya
    Minker, Wolfgang
    HUMAN-COMPUTER INTERACTION: DESIGN AND DEVELOPMENT APPROACHES, PT I, 2011, 6761 : 527 - 536
  • [26] Initializing EM using the properties of its trajectories in Gaussian mixtures
    Biernacki, C
    STATISTICS AND COMPUTING, 2004, 14 (03) : 267 - 279
  • [27] The competitive EM algorithm for Gaussian mixtures with BYY harmony criterion
    Wang, Hengyu
    Li, Lei
    Ma, Jinwen
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS, PROCEEDINGS: WITH ASPECTS OF THEORETICAL AND METHODOLOGICAL ISSUES, 2008, 5226 : 552 - +
  • [28] Initializing EM using the properties of its trajectories in Gaussian mixtures
    Christophe Biernacki
    Statistics and Computing, 2004, 14 : 267 - 279
  • [29] Degeneracy in the maximum likelihood estimation of univariate Gaussian mixtures with EM
    Biernacki, C
    Chrétien, S
    STATISTICS & PROBABILITY LETTERS, 2003, 61 (04) : 373 - 382
  • [30] An Efficient EM Approach to Parameter Learning of the Mixture of Gaussian Processes
    Yang, Yan
    Ma, Jinwen
    ADVANCES IN NEURAL NETWORKS - ISNN 2011, PT II, 2011, 6676 : 165 - +