Fast deep mixtures of Gaussian process experts

被引:0
|
作者
Clement Etienam
Kody J. H. Law
Sara Wade
Vitaly Zankin
机构
[1] University of Manchester,Department of Mathematics
[2] University of Edinburgh,School of Mathematics
[3] The Alan Turing Institute,undefined
[4] NVIDIA,undefined
来源
Machine Learning | 2024年 / 113卷
关键词
Bayesian inference; DNN; Gaussian process; Mixture of experts;
D O I
暂无
中图分类号
学科分类号
摘要
Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, allowing not only the mean function but the entire density of the output to change with the inputs. Sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models, and in this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster–Classify–Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is significantly lower for higher-dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
引用
收藏
页码:1483 / 1508
页数:25
相关论文
共 50 条
  • [31] Gaussian Process Domain Experts for Model Adaptation in Facial Behavior Analysis
    Eleftheriadis, Stefanos
    Rudovic, Ognjen
    Deisenroth, Marc P.
    Pantic, Maja
    PROCEEDINGS OF 29TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, (CVPRW 2016), 2016, : 1469 - 1477
  • [32] STATISTICAL CHARACTERISTICS OF THE GAUSSIAN PROCESS AND NONGAUSSIAN DISTRIBUTION PROCESS MIXTURES RATIO
    MELITITSKY, VA
    IZVESTIYA VYSSHIKH UCHEBNYKH ZAVEDENII RADIOELEKTRONIKA, 1979, 22 (07): : 52 - 56
  • [33] Hierarchical Gaussian Processes and Mixtures of Experts to Model COVID-19 Patient Trajectories
    Cui, Sunny
    Yoo, Elizabeth C.
    Li, Didong
    Laudanski, Krzysztof
    Engelhardt, Barbara E.
    BIOCOMPUTING 2022, PSB 2022, 2022, : 266 - 277
  • [34] A DAEM Algorithm for Mixtures of Gaussian Process Functional Regressions
    Wu, Di
    Ma, Jinwen
    INTELLIGENT COMPUTING METHODOLOGIES, ICIC 2016, PT III, 2016, 9773 : 294 - 303
  • [35] Mixtures of Gaussian process models for human pose estimation
    Fergie, Martin
    Galata, Aphrodite
    IMAGE AND VISION COMPUTING, 2013, 31 (12) : 949 - 957
  • [36] Fast and reliable prediction of scramjet flowfields via Gaussian process latent variable model and deep learning
    Fujio, Chihiro
    Akiyama, Kento
    Ogawa, Hideaki
    PHYSICS OF FLUIDS, 2023, 35 (04)
  • [37] Manifold learning by a deep Gaussian process autoencoder
    Camastra, Francesco
    Casolaro, Angelo
    Iannuzzo, Gennaro
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (21): : 15573 - 15582
  • [38] Active Learning for Deep Gaussian Process Surrogates
    Sauer, Annie
    Gramacy, Robert B.
    Higdon, David
    TECHNOMETRICS, 2023, 65 (01) : 4 - 18
  • [39] Manifold learning by a deep Gaussian process autoencoder
    Francesco Camastra
    Angelo Casolaro
    Gennaro Iannuzzo
    Neural Computing and Applications, 2023, 35 : 15573 - 15582
  • [40] Deep learning with differential Gaussian process flows
    Hegde, Pashupati
    Heinonen, Markus
    Lahdesmaki, Harri
    Kaski, Samuel
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89