Discriminative training of feed-forward and recurrent sum-product networks by extended Baum-Welch

被引:2
|
作者
Duan, Haonan [1 ,2 ]
Rashwan, Abdullah [1 ,2 ]
Poupart, Pascal [1 ,2 ]
Chen, Zhitang [3 ]
机构
[1] Univ Waterloo, Waterloo AI Inst, Waterloo, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
[3] Huawei Technol, Hong Kong, Peoples R China
关键词
Sum-product network; Extended Baum-Welch; Discriminative learning; PROBABILISTIC FUNCTIONS; STATISTICAL ESTIMATION; MAXIMIZATION; INEQUALITY;
D O I
10.1016/j.ijar.2020.02.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a discriminative learning algorithm for feed-forward Sum-Product Networks (SPNs) [42] and recurrent SPNs [31] based on the Extended Baum-Welch (EBW) algorithm [4]. We formulate the conditional data likelihood in the SPN framework as a rational function, and we use EBW to monotonically maximize it. We derive the algorithm for SPNs and RSPNs with both discrete and continuous variables. The experiments show that this algorithm performs better than both generative Expectation-Maximization, and discriminative gradient descent on a wide variety of applications. We also demonstrate the robustness of the algorithm in the case of missing features by comparing its performance to Support Vector Machines and Neural Networks. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页码:66 / 81
页数:16
相关论文
共 50 条
  • [21] Subspace Based Sequence Discriminative Training of LSTM Acoustic Models with Feed-Forward Layers
    Samarakoon, Lahiru
    Mak, Brian
    Lam, Albert Y. S.
    2018 11TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2018, : 136 - 140
  • [22] Limits to the development of feed-forward structures in large recurrent neuronal networks
    Kunkel, Susanne
    Diesmann, Markus
    Morrison, Abigail
    Frontiers in Computational Neuroscience, 2010, 4
  • [23] Limits to the development of feed-forward structures in large recurrent neuronal networks
    Kunkel, Susanne
    Diesmann, Markus
    Morrison, Abigail
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2011, 4
  • [24] Feed-forward chains of recurrent attractor neural networks near saturation
    Coolen, ACC
    Viana, L
    JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1996, 29 (24): : 7855 - 7866
  • [25] A Comparison of Feed-forward and Recurrent Neural Networks in Time Series Forecasting
    Brezak, Danko
    Bacek, Tomislav
    Majetic, Dubravko
    Kasac, Josip
    Novakovic, Branko
    2012 IEEE CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR FINANCIAL ENGINEERING & ECONOMICS (CIFER), 2012, : 206 - 211
  • [26] An ensemble of differential evolution and Adam for training feed-forward neural networks
    Xue, Yu
    Tong, Yiling
    Neri, Ferrante
    INFORMATION SCIENCES, 2022, 608 : 453 - 471
  • [27] Unsupervised, smooth training of feed-forward neural networks for mismatch compensation
    Surendran, AC
    Lee, CH
    Rahim, M
    1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, PROCEEDINGS, 1997, : 482 - 489
  • [28] Salp Swarm Algorithm (SSA) for Training Feed-Forward Neural Networks
    Bairathi, Divya
    Gopalani, Dinesh
    SOFT COMPUTING FOR PROBLEM SOLVING, SOCPROS 2017, VOL 1, 2019, 816 : 521 - 534
  • [29] Hybrid learning schemes for fast training of feed-forward neural networks
    Karayiannis, NB
    MATHEMATICS AND COMPUTERS IN SIMULATION, 1996, 41 (1-2) : 13 - 28
  • [30] Hybrid training of feed-forward neural networks with particle swarm optimization
    Carvalho, M.
    Ludermir, T. B.
    NEURAL INFORMATION PROCESSING, PT 2, PROCEEDINGS, 2006, 4233 : 1061 - 1070