Fast deep mixtures of Gaussian process experts

被引:0
|
作者
Clement Etienam
Kody J. H. Law
Sara Wade
Vitaly Zankin
机构
[1] University of Manchester,Department of Mathematics
[2] University of Edinburgh,School of Mathematics
[3] The Alan Turing Institute,undefined
[4] NVIDIA,undefined
来源
Machine Learning | 2024年 / 113卷
关键词
Bayesian inference; DNN; Gaussian process; Mixture of experts;
D O I
暂无
中图分类号
学科分类号
摘要
Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, allowing not only the mean function but the entire density of the output to change with the inputs. Sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models, and in this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster–Classify–Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is significantly lower for higher-dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
引用
收藏
页码:1483 / 1508
页数:25
相关论文
共 50 条
  • [21] A fast and efficient Modal EM algorithm for Gaussian mixtures
    Scrucca, Luca
    STATISTICAL ANALYSIS AND DATA MINING, 2021, 14 (04) : 305 - 314
  • [22] Batch process modelling with mixtures of Gaussian processes
    Xiaoling Ou
    Elaine Martin
    Neural Computing and Applications, 2008, 17 : 471 - 479
  • [23] Batch process modelling with mixtures of Gaussian processes
    Ou, Xiaoling
    Martin, Elaine
    NEURAL COMPUTING & APPLICATIONS, 2008, 17 (5-6): : 471 - 479
  • [24] Regularized Estimation and Feature Selection in Mixtures of Gaussian-Gated Experts Models
    Chamroukhi, Faicel
    Lecocq, Florian
    Nguyen, Hien D.
    STATISTICS AND DATA SCIENCE, RSSDS 2019, 2019, 1150 : 42 - 56
  • [25] Sequential Inference for Deep Gaussian Process
    Wang, Yali
    Brubaker, Marcus
    Chaib-draa, Brahim
    Urtasun, Raquel
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 694 - 703
  • [26] Image Matting With Deep Gaussian Process
    Zheng, Yuanjie
    Yang, Yunshuai
    Che, Tongtong
    Hou, Sujuan
    Huang, Wenhui
    Gao, Yue
    Tan, Ping
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8879 - 8893
  • [27] Sparse Mixtures of Shallow Linear Experts for Interpretable and Fast Outcome Prediction
    Folino, Francesco
    Pontieri, Luigi
    Sabatino, Pietro
    PROCESS MINING WORKSHOPS, ICPM 2023, 2024, 503 : 141 - 152
  • [28] Fast Gaussian Process Regression for Big Data
    Das, Sourish
    Roy, Sasanka
    Sambasivan, Rajiv
    BIG DATA RESEARCH, 2018, 14 : 12 - 26
  • [29] Fast identification algorithms for Gaussian process model
    Hong, Xia
    Gao, Junbin
    Jiang, Xinwei
    Harris, Chris J.
    NEUROCOMPUTING, 2014, 133 : 25 - 31
  • [30] Fast Gaussian Process Posteriors with Product Trees
    Moore, David A.
    Russell, Stuart
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2014, : 613 - 622