Spatial Mixture-of-Experts

被引:0
|
作者
Dryden, Nikoli [1 ]
Hoefler, Torsten [1 ]
机构
[1] Swiss Fed Inst Technol, Zurich, Switzerland
基金
欧盟地平线“2020”;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many data have an underlying dependence on spatial location; it may be weather on the Earth, a simulation on a mesh, or a registered image. Yet this feature is rarely taken advantage of, and violates common assumptions made by many neural network layers, such as translation equivariance. Further, many works that do incorporate locality fail to capture fine-grained structure. To address this, we introduce the Spatial Mixture-of-Experts (SMOE) layer, a sparsely-gated layer that learns spatial structure in the input domain and routes experts at a fine-grained level to utilize it. We also develop new techniques to train SMOEs, including a self-supervised routing loss and damping expert errors. Finally, we show strong results for SMOEs on numerous tasks, and set new state-of-the-art results for medium-range weather prediction and post-processing ensemble weather forecasts.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] New estimation and feature selection methods in mixture-of-experts models
    Khalili, Abbas
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 2010, 38 (04): : 519 - 539
  • [32] Hierarchical Mixture-of-Experts approach for neural compact modeling of MOSFETs
    Park, Chanwoo
    Vincent, Premkumar
    Chong, Soogine
    Park, Junghwan
    Cha, Ye Sle
    Cho, Hyunbo
    SOLID-STATE ELECTRONICS, 2023, 199
  • [33] Multiscale Segmentation of Elevation Images Using a Mixture-of-Experts Framework
    Nagarajan, K.
    Slatton, K. C.
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2009, 6 (04) : 865 - 869
  • [34] A Hierarchical Mixture-Of-Experts Framework for Few Labeled Node Classification
    Wang, Yimeng
    Yang, Zhiyao
    Che, Xiangjiu
    NEURAL NETWORKS, 2025, 188
  • [35] COLOR PREDICTION IN IMAGE CODING USING STEERED MIXTURE-OF-EXPERTS
    Verhack, Ruben
    Van de Keer, Simon
    Van Wallendael, Glenn
    Sikora, Thomas
    Lambert, Peter
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 1288 - 1292
  • [36] GeRM: A Generalist Robotic Model with Mixture-of-experts for Quadruped Robot
    Song, Wenxuan
    Zhao, Han
    Ding, Pengxiang
    Cui, Can Can
    Liu, Shangke
    Fan, Yaning
    Wang, Donglin
    2024 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2024), 2024, : 11879 - 11886
  • [37] Phase-type mixture-of-experts regression for loss severities
    Bladt, Martin
    Yslas, Jorge
    SCANDINAVIAN ACTUARIAL JOURNAL, 2023, 2023 (04) : 303 - 329
  • [38] ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
    Zhao, Yue
    Zheng, Guoqing
    Mukherjee, Subhabrata
    McCann, Robert
    Awadallah, Ahmed
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 4937 - 4945
  • [39] A Mixture-of-Experts Prediction Framework for Evolutionary Dynamic Multiobjective Optimization
    Rambabu, Rethnaraj
    Vadakkepat, Prahlad
    Tan, Kay Chen
    Jiang, Min
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (12) : 5099 - 5112
  • [40] A MULTIMODAL MIXTURE-OF-EXPERTS MODEL FOR DYNAMIC EMOTION PREDICTION IN MOVIES
    Goyal, Ankit
    Kumar, Naveen
    Guha, Tanaya
    Narayanan, Shrikanth S.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2822 - 2826