Mixture of Experts with Entropic Regularization for Data Classification

被引:5
|
作者
Peralta, Billy [1 ]
Saavedra, Ariel [2 ]
Caro, Luis [2 ]
Soto, Alvaro [3 ]
机构
[1] Andres Bello Univ, Dept Engn Sci, Santiago 7500971, Chile
[2] Catholic Univ Temuco, Dept Informat Engn, Temuco 4781312, Chile
[3] Pontificia Univ Catolica Chile, Dept Comp Sci, Santiago 7820436, Chile
关键词
mixture-of-experts; regularization; entropy; classification;
D O I
10.3390/e21020190
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. Mixture-of-experts is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a winner-takes-all output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3-6% in some datasets. In future work, we plan to embed feature selection into this model.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] A Proposal for Mixture of Experts with Entropic Regularization
    Peralta, Billy
    Saavedra, Ariel
    Caro, Luis
    2017 XLIII LATIN AMERICAN COMPUTER CONFERENCE (CLEI), 2017,
  • [2] Mixture of experts for stellar data classification
    Jiang, YG
    Guo, P
    ADVANCES IN NEURAL NETWORKS - ISNN 2005, PT 2, PROCEEDINGS, 2005, 3497 : 310 - 315
  • [3] Dropout regularization in hierarchical mixture of experts
    Irsoy, Ozan
    Alpaydin, Ethem
    NEUROCOMPUTING, 2021, 419 : 148 - 156
  • [4] Regularization and error bars for the mixture of experts network
    Ramamurti, V
    Ghosh, J
    1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, 1997, : 221 - 225
  • [5] Extension of mixture-of-experts networks for binary classification of hierarchical data
    Ng, Shu-Kay
    McLachlan, Geoffrey J.
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2007, 41 (01) : 57 - 67
  • [6] Mixture of experts classification using a hierarchical mixture model
    Titsias, MK
    Likas, A
    NEURAL COMPUTATION, 2002, 14 (09) : 2221 - 2244
  • [7] Incorporation of a Regularization Term to Control Negative Correlation in Mixture of Experts
    Saeed Masoudnia
    Reza Ebrahimpour
    Seyed Ali Asghar Abbaszadeh Arani
    Neural Processing Letters, 2012, 36 : 31 - 47
  • [8] Incorporation of a Regularization Term to Control Negative Correlation in Mixture of Experts
    Masoudnia, Saeed
    Ebrahimpour, Reza
    Arani, Seyed Ali Asghar Abbaszadeh
    NEURAL PROCESSING LETTERS, 2012, 36 (01) : 31 - 47
  • [9] Unseen Family Member Classification Using Mixture of Experts
    Ghahramani, M.
    Wang, H. L.
    Yau, W. Y.
    Teoh, E. K.
    ICIEA 2010: PROCEEDINGS OF THE 5TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS, VOL 1, 2010, : 359 - +
  • [10] Improved learning algorithms for mixture of experts in multiclass classification
    Chen, K
    Xu, L
    Chi, H
    NEURAL NETWORKS, 1999, 12 (09) : 1229 - 1252