Fast deep mixtures of Gaussian process experts

被引:0
|
作者
Clement Etienam
Kody J. H. Law
Sara Wade
Vitaly Zankin
机构
[1] University of Manchester,Department of Mathematics
[2] University of Edinburgh,School of Mathematics
[3] The Alan Turing Institute,undefined
[4] NVIDIA,undefined
来源
Machine Learning | 2024年 / 113卷
关键词
Bayesian inference; DNN; Gaussian process; Mixture of experts;
D O I
暂无
中图分类号
学科分类号
摘要
Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, allowing not only the mean function but the entire density of the output to change with the inputs. Sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models, and in this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster–Classify–Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is significantly lower for higher-dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
引用
收藏
页码:1483 / 1508
页数:25
相关论文
共 50 条
  • [1] Fast deep mixtures of Gaussian process experts
    Etienam, Clement
    Law, Kody J. H.
    Wade, Sara
    Zankin, Vitaly
    MACHINE LEARNING, 2024, 113 (03) : 1483 - 1508
  • [2] Infinite mixtures of Gaussian process experts
    Rasmussen, CE
    Ghahramani, Z
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 14, VOLS 1 AND 2, 2002, 14 : 881 - 888
  • [3] Fast Allocation of Gaussian Process Experts
    Nguyen, Trung, V
    Bonilla, Edwin, V
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 1), 2014, 32
  • [4] Enriched mixtures of generalised Gaussian process experts
    Gadd, Charles W. L.
    Wade, Sara
    Boukouvalas, Alexis
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 3144 - 3153
  • [5] Gaussian Process-Gated Hierarchical Mixtures of Experts
    Liu, Yuhao
    Ajirak, Marzieh
    Djuric, Petar M.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6443 - 6453
  • [6] Convergence Rates for Gaussian Mixtures of Experts
    Ho, Nhat
    Yang, Chiao-Yu
    Jordan, Michael I.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [7] Convergence Rates for Gaussian Mixtures of Experts
    Ho, Nhat
    Yang, Chiao-Yu
    Jordan, Michael I.
    Journal of Machine Learning Research, 2022, 23
  • [8] Healing Products of Gaussian Process Experts
    Cohen, Samuel
    Mbuvha, Rendani
    Marwala, Tshilidzi
    Deisenroth, Marc Peter
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [9] Gaussian Process Experts for Voice Conversion
    Pilkington, Nicholas C. V.
    Zen, Heiga
    Gales, Mark J. F.
    12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 2772 - +
  • [10] Healing Products of Gaussian Process Experts
    Cohen, Samuel
    Mbuvha, Rendani
    Marwala, Tshilidzi
    Deisenroth, Marc Peter
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,