Information geometry of the EM and em algorithms for neural networks

被引:177
|
作者
Amari, SI
机构
关键词
EM algorithm; information geometry; stochastic model of neural networks; learning; identification of neural network; e-projection; m-projection; hidden variable;
D O I
10.1016/0893-6080(95)00003-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To realize an input-output relation given by noise-contaminated examples, it is effective to use a stochastic model of neural networks. When the model network includes hidden units whose activation values are not specified nor observed, it is useful to estimate the hidden variables from the observed or specified input-output data based on the stochastic model. Two algorithms, the EM and em algorithms, have so far been proposed for this purpose. The EM algorithm is an iterative statistical technique of using the conditional expectation, and the em algorithm is a geometrical one given by information geometry. The em algorithm minimizes iteratively the Kullback-Leibler divergence in the manifold of neural networks. These two algorithms are equivalent in most cases. The present paper gives a unified information geometrical framework for studying stochastic models of neural networks, by focusing on the EM and em algorithms, and proves a condition that guarantees their equivalence. Examples include: (1) stochastic multilayer perceptron, (2) mixtures of experts, and (3) normal mixture model.
引用
收藏
页码:1379 / 1408
页数:30
相关论文
共 50 条
  • [1] THE EM ALGORITHM AND INFORMATION GEOMETRY IN NEURAL-NETWORK LEARNING
    AMARI, S
    NEURAL COMPUTATION, 1995, 7 (01) : 13 - 18
  • [2] Geometry of EM and related iterative algorithms
    Hino, Hideitsu
    Akaho, Shotaro
    Murata, Noboru
    INFORMATION GEOMETRY, 2024, 7 (SUPPL1) : 39 - 77
  • [3] Dynamic learning with the EM algorithm for neural networks
    de Freitas, JFG
    Niranjan, M
    Gee, AH
    JOURNAL OF VLSI SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2000, 26 (1-2): : 119 - 131
  • [4] Modeling Dst with Recurrent EM Neural Networks
    Mirikitani, Derrick Takeshi
    Ouarbya, Lahcen
    ARTIFICIAL NEURAL NETWORKS - ICANN 2009, PT I, 2009, 5768 : 975 - 984
  • [5] Dynamic Learning with the EM Algorithm for Neural Networks
    J.F.G. de Freitas
    M. Niranjan
    A.H. Gee
    Journal of VLSI signal processing systems for signal, image and video technology, 2000, 26 : 119 - 131
  • [6] Capsule Graph Neural Networks with EM Routing
    Lei, Yu
    Zhang, Jing
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3191 - 3195
  • [7] Distributed EM algorithms for density estimation in sensor networks
    Nowak, R
    2003 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL IV, PROCEEDINGS: SIGNAL PROCESSING FOR COMMUNICATIONS SPECIAL SESSIONS, 2003, : 836 - 839
  • [8] A Legacy of EM Algorithms
    Lange, Kenneth
    Zhou, Hua
    INTERNATIONAL STATISTICAL REVIEW, 2022, 90 : S52 - S66
  • [9] Combinatorial EM algorithms
    Ian C. Marschner
    Statistics and Computing, 2014, 24 : 921 - 940
  • [10] Combinatorial EM algorithms
    Marschner, Ian C.
    STATISTICS AND COMPUTING, 2014, 24 (06) : 921 - 940