Hierarchical Routing Mixture of Experts

被引:2
|
作者
Zhao, Wenbo [1 ]
Gao, Yang [1 ]
Memon, Shahan Ali [1 ]
Raj, Bhiksha [1 ]
Singh, Rita [1 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
SUPPORT VECTOR MACHINES; APPROXIMATION; PREDICTION;
D O I
10.1109/ICPR48806.2021.9412813
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In regression tasks, the data distribution is often too complex to be fitted by a single model. In contrast, partition-based models are developed where data is divided and fitted by local models. These models partition the input space and do not leverage the input-output dependency of multimodal-distributed data, and strong local models are needed to make good predictions. Addressing these problems, we propose a binary tree-structured hierarchical routing mixture of experts (HRME) model that has classifiers as non-leaf node experts and simple regression models as leaf node experts. The classifier nodes jointly soft-partition the input-output space based on the natural separateness of multimodal data. This enables simple leaf experts to be effective for prediction. Further, we develop a probabilistic framework for the HRME model and propose a recursive Expectation-Maximization (EM) based algorithm to learn both the tree structure and the expert models. Experiments on a collection of regression tasks validate our method's effectiveness compared to various other regression models.
引用
收藏
页码:7900 / 7906
页数:7
相关论文
共 50 条
  • [21] Extension of mixture-of-experts networks for binary classification of hierarchical data
    Ng, Shu-Kay
    McLachlan, Geoffrey J.
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2007, 41 (01) : 57 - 67
  • [22] Model Selection of Bayesian Hierarchical Mixture of Experts based on Variational Inference
    Iikubo, Yuji
    Horii, Shunsuke
    Matsushima, Toshiyasu
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 3474 - 3479
  • [23] A Hierarchical Mixture-Of-Experts Framework for Few Labeled Node Classification
    Wang, Yimeng
    Yang, Zhiyao
    Che, Xiangjiu
    NEURAL NETWORKS, 2025, 188
  • [24] Hierarchical Mixture-of-Experts approach for neural compact modeling of MOSFETs
    Park, Chanwoo
    Vincent, Premkumar
    Chong, Soogine
    Park, Junghwan
    Cha, Ye Sle
    Cho, Hyunbo
    SOLID-STATE ELECTRONICS, 2023, 199
  • [25] Merging experts' opinions: A Bayesian hierarchical model with mixture of prior distributions
    Rufo, M. J.
    Perez, C. J.
    Martin, J.
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2010, 207 (01) : 284 - 289
  • [26] IMPROVEMENTS TO A HIERARCHICAL MIXTURE OF EXPERTS SYSTEM USED FOR CHARACTERIZATION OF RESIDENT SPACE OBJECTS
    Pinon, Elfego, III
    Anderson, Jessica
    Ceniceros, Angelica
    Jones, Brandon
    Russell, Ryan
    Hatten, Noble
    Ravago, Nicholas
    ASTRODYNAMICS 2017, PTS I-IV, 2018, 162 : 3819 - 3838
  • [27] HIERARCHICAL LEARNING OF SPARSE IMAGE REPRESENTATIONS USING STEERED MIXTURE-OF-EXPERTS
    Jongebloed, Rolf
    Verhack, Ruben
    Lange, Lieven
    Sikora, Thomas
    2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW 2018), 2018,
  • [28] Specifying a hierarchical mixture of experts for hydrologic modeling: Gating function variable selection
    Jeremiah, Erwin
    Marshall, Lucy
    Sisson, Scott A.
    Sharma, Ashish
    WATER RESOURCES RESEARCH, 2013, 49 (05) : 2926 - 2939
  • [29] Language-Routing Mixture of Experts for Multilingual and Code-Switching Speech Recognition
    Wang, Wenxuan
    Ma, Guodong
    Li, Yuke
    Du, Binbin
    INTERSPEECH 2023, 2023, : 1389 - 1393
  • [30] Hierarchical mixture of experts for autonomous unmanned aerial vehicles utilizing thrust models and acoustics
    Kawamura, Evan
    Azimov, Dilmurat
    Allen, John S.
    Ippolito, Corey
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 162