Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach

被引:0
|
作者
Du, Hongyang [1 ]
Liu, Guangyuan [1 ]
Lin, Yijing [2 ]
Niyato, Dusit [1 ]
Kang, Jiawen [3 ,4 ,5 ]
Xiong, Zehui [6 ]
Kim, Dong In [7 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[4] Minist Educ, Key Lab Intelligent Informat Proc & Syst Integrat, Guangzhou 510006, Peoples R China
[5] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Peoples R China
[6] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore 487372, Singapore
[7] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon 16419, South Korea
来源
20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024 | 2024年
基金
新加坡国家研究基金会; 中国国家自然科学基金;
关键词
Generative AI (GAI); large language model; mixture of experts; network optimization;
D O I
10.1109/IWCMC61514.2024.10592370
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optimizing various wireless user tasks poses a significant challenge for networking systems because of the expanding range of user requirements. Despite advancements in Deep Reinforcement Learning (DRL), the need for customized optimization tasks for individual users complicates developing and applying numerous DRL models, leading to substantial computation resource and energy consumption and can lead to inconsistent outcomes. To address this issue, we propose a novel approach utilizing a Mixture of Experts (MoE) framework, augmented with Large Language Models (LLMs), to analyze user objectives and constraints effectively, select specialized DRL experts, and weigh each decision from the participating experts. Specifically, we develop a gate network to oversee the expert models, allowing a collective of experts to tackle a wide array of new tasks. Furthermore, we innovatively substitute the traditional gate network with an LLM, leveraging its advanced reasoning capabilities to manage expert model selection for joint decisions. Our proposed method reduces the need to train new DRL models for each unique optimization problem, decreasing energy consumption and AI model implementation costs. The LLM-enabled MoE approach is validated through a general maze navigation task and a specific network service provider utility maximization task, demonstrating its effectiveness and practical applicability in optimizing complex networking systems.
引用
收藏
页码:531 / 536
页数:6
相关论文
共 50 条
  • [21] Analysis on Energy Efficiency of Large Scale Intelligent Reflecting Surface-Enabled Networks
    Chen, Youjia
    Zhang, Baoxian
    Hu, Jinsong
    Lopez-Perez, David
    Ding, Ming
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (10) : 2802 - 2806
  • [22] A novel individual blood glucose control model based on Mixture of Experts neural networks
    Wang, W
    Bian, ZZ
    Yan, LF
    Su, J
    ADVANCES IN NEURAL NETWORKS - ISNN 2004, PT 2, 2004, 3174 : 453 - 458
  • [23] Large language models enabled intelligent microstructure optimization and defects classification of welded titanium alloys
    Zhang, Suyang
    Wang, William Yi
    Wang, Xinzhao
    Li, Gaonan
    Ren, Yong
    Gao, Xingyu
    Sun, Feng
    Tang, Bin
    Song, Haifeng
    Li, Jinshan
    JOURNAL OF MATERIALS INFORMATICS, 2024, 4 (04):
  • [24] Hierarchical mixture of experts and diagnostic modeling approach to reduce hydrologic model structural uncertainty
    Moges, Edom
    Demissie, Yonas
    Li, Hong-Yi
    WATER RESOURCES RESEARCH, 2016, 52 (04) : 2551 - 2570
  • [25] Efficient scaling of large language models with mixture of experts and 3D analog in-memory computing
    Buchel, Julian
    Vasilopoulos, Athanasios
    Simon, William Andrew
    Boybat, Irem
    Tsai, Hsinyu
    Burr, Geoffrey W.
    Castro, Hernan
    Filipiak, Bill
    Le Gallo, Manuel
    Rahimi, Abbas
    Narayanan, Vijay
    Sebastian, Abu
    NATURE COMPUTATIONAL SCIENCE, 2025, 5 (01): : 13 - 26
  • [26] Artificially Intelligent Billing in Spine Surgery: An Analysis of a Large Language Model
    Kong, Xiuhua
    Wang, Lingling
    Liu, Changhua
    GLOBAL SPINE JOURNAL, 2024, 14 (05) : 1684 - 1684
  • [27] Large Language Model Firewall for AIGC Protection with Intelligent Detection Policy
    Huang, Tianrui
    You, Lina
    Cai, Nishui
    Huang, Ting
    2024 2ND INTERNATIONAL CONFERENCE ON MOBILE INTERNET, CLOUD COMPUTING AND INFORMATION SECURITY, MICCIS 2024, 2024, : 247 - 252
  • [28] Ontology-integrated tuning of large language model for intelligent maintenance
    Wang, Peng
    Karigiannis, John
    Gao, Robert X.
    CIRP ANNALS-MANUFACTURING TECHNOLOGY, 2024, 73 (01) : 361 - 364
  • [29] GeoLLM: A specialized large language model framework for intelligent geotechnical design
    Xu, Hao-Ruo
    Zhang, Ning
    Yin, Zhen-Yu
    Njock, Pierre Guy Atangana
    COMPUTERS AND GEOTECHNICS, 2025, 177
  • [30] A Framework for Agricultural Intelligent Analysis Based on a Visual Language Large Model
    Yu, Piaofang
    Lin, Bo
    APPLIED SCIENCES-BASEL, 2024, 14 (18):