Learning to Adapt Dynamic Clinical Event Sequences with Residual Mixture of Experts

被引:1
|
作者
Lee, Jeong Min [1 ]
Hauskrecht, Milos [1 ]
机构
[1] Univ Pittsburgh, Dept Comp Sci, Pittsburgh, PA 15260 USA
关键词
D O I
10.1007/978-3-031-09342-5_15
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Clinical event sequences in Electronic Health Records (EHRs) record detailed information about the patient condition and patient care as they occur in time. Recent years have witnessed increased interest of machine learning community in developing machine learning models solving different types of problems defined upon information in EHRs. More recently, neural sequential models, such as RNN and LSTM, became popular and widely applied models for representing patient sequence data and for predicting future events or outcomes based on such data. However, a single neural sequential model may not properly represent complex dynamics of all patients and the differences in their behaviors. In this work, we aim to alleviate this limitation by refining a one-fits-all model using aMixture-of-Experts (MoE) architecture. The architecture consists of multiple (expert) RNN models covering patient sub-populations and refining the predictions of the base model. That is, instead of training expert RNN models from scratch we define them on the residual signal that attempts to model the differences from the population-wide model. The heterogeneity of various patient sequences is modeled through multiple experts that consist of RNN. Particularly, instead of directly training MoE from scratch, we augment MoE based on the prediction signal from pretrained base GRU model. With this way, the mixture of experts can provide flexible adaptation to the (limited) predictive power of the single base RNN model. We experiment with the newly proposed model on real-world EHRs data and the multivariate clinical event prediction task. We implement RNN using Gated Recurrent Units (GRU). We show 4.1% gain on AUPRC statistics compared to a single RNN prediction.
引用
收藏
页码:155 / 166
页数:12
相关论文
共 50 条
  • [11] Boosted mixture of experts: An ensemble learning scheme
    Avnimelech, R
    Intrator, N
    NEURAL COMPUTATION, 1999, 11 (02) : 483 - 497
  • [12] Learning ambiguities using Bayesian mixture of experts
    Kanaujia, Atul
    Metaxas, Dimitris
    ICTAI-2006: EIGHTEENTH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2006, : 436 - +
  • [13] IMPLICIT LEARNING OF EVENT SEQUENCES
    BUCHNER, A
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 1992, 27 (3-4) : 88 - 88
  • [14] LEARNING THE STRUCTURE OF EVENT SEQUENCES
    CLEEREMANS, A
    MCCLELLAND, JL
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-GENERAL, 1991, 120 (03) : 235 - 253
  • [15] Learning Event Representations for Temporal Segmentation of Image Sequences by Dynamic Graph Embedding
    Dimiccoli, Mariella
    Wendt, Herwig
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1476 - 1486
  • [16] On the Benefits of Learning to Route in Mixture-of-Experts Models
    Dikkala, Nishanth
    Ghosh, Nikhil
    Meka, Raghu
    Panigrahy, Rina
    Vyas, Nikhil
    Wang, Xin
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9376 - 9396
  • [17] Improved learning algorithms for mixture of experts in multiclass classification
    Chen, K
    Xu, L
    Chi, H
    NEURAL NETWORKS, 1999, 12 (09) : 1229 - 1252
  • [18] Robust expectation maximization learning algorithm for mixture of experts
    Torres, R
    Salas, R
    Allende, H
    Moraga, C
    COMPUTATIONAL METHODS IN NEURAL MODELING, PT 1, 2003, 2686 : 238 - 245
  • [19] Dynamic factor mixture of experts for functional time series modeling
    Same, Allou
    2016 15TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2016), 2016, : 19 - 25
  • [20] Discontinuity-Sensitive Optimal Control Learning by Mixture of Experts
    Tang, Gao
    Hauser, Kris
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 7892 - 7898