Fast deep mixtures of Gaussian process experts

被引:0
|
作者
Clement Etienam
Kody J. H. Law
Sara Wade
Vitaly Zankin
机构
[1] University of Manchester,Department of Mathematics
[2] University of Edinburgh,School of Mathematics
[3] The Alan Turing Institute,undefined
[4] NVIDIA,undefined
来源
Machine Learning | 2024年 / 113卷
关键词
Bayesian inference; DNN; Gaussian process; Mixture of experts;
D O I
暂无
中图分类号
学科分类号
摘要
Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, allowing not only the mean function but the entire density of the output to change with the inputs. Sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models, and in this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster–Classify–Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is significantly lower for higher-dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
引用
收藏
页码:1483 / 1508
页数:25
相关论文
共 50 条
  • [41] Deep Gaussian Process autoencoders for novelty detection
    Rémi Domingues
    Pietro Michiardi
    Jihane Zouaoui
    Maurizio Filippone
    Machine Learning, 2018, 107 : 1363 - 1383
  • [42] Deep Gaussian Process autoencoders for novelty detection
    Domingues, Remi
    Michiardi, Pietro
    Zouaoui, Jihane
    Filippone, Maurizio
    MACHINE LEARNING, 2018, 107 (8-10) : 1363 - 1383
  • [43] Longitudinal Deep Kernel Gaussian Process Regression
    Liang, Junjie
    Wu, Yanting
    Xu, Dongkuan
    Honavar, Vasant G.
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8556 - 8564
  • [44] Posterior Contraction for Deep Gaussian Process Priors
    Finocchio, Gianluca
    Schmidt-Hieber, Johannes
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [45] A Deep Gaussian Process Approach for Predictive Maintenance
    Zeng, Junqi
    Liang, Zhenglin
    IEEE TRANSACTIONS ON RELIABILITY, 2023, 72 (03) : 916 - 933
  • [46] Gaussian Process and Deep Learning Atmospheric Correction
    Basener, Bill
    Basener, Abigail
    REMOTE SENSING, 2023, 15 (03)
  • [47] Fast Gaussian process regression using representative data
    Yoshioka, T
    Ishii, S
    IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 132 - 137
  • [48] Nested aggregation of experts using inducing points for approximated Gaussian process regression
    Nakai-Kasai, Ayano
    Tanaka, Toshiyuki
    MACHINE LEARNING, 2022, 111 (05) : 1671 - 1694
  • [49] Nested aggregation of experts using inducing points for approximated Gaussian process regression
    Ayano Nakai-Kasai
    Toshiyuki Tanaka
    Machine Learning, 2022, 111 : 1671 - 1694
  • [50] Sparse Information Filter for Fast Gaussian Process Regression
    Kania, Lucas
    Schuerch, Manuel
    Azzimonti, Dario
    Benavoli, Alessio
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 : 527 - 542