Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach

被引:0
|
作者
Du, Hongyang [1 ]
Liu, Guangyuan [1 ]
Lin, Yijing [2 ]
Niyato, Dusit [1 ]
Kang, Jiawen [3 ,4 ,5 ]
Xiong, Zehui [6 ]
Kim, Dong In [7 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[4] Minist Educ, Key Lab Intelligent Informat Proc & Syst Integrat, Guangzhou 510006, Peoples R China
[5] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Peoples R China
[6] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore 487372, Singapore
[7] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon 16419, South Korea
基金
新加坡国家研究基金会; 中国国家自然科学基金;
关键词
Generative AI (GAI); large language model; mixture of experts; network optimization;
D O I
10.1109/IWCMC61514.2024.10592370
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optimizing various wireless user tasks poses a significant challenge for networking systems because of the expanding range of user requirements. Despite advancements in Deep Reinforcement Learning (DRL), the need for customized optimization tasks for individual users complicates developing and applying numerous DRL models, leading to substantial computation resource and energy consumption and can lead to inconsistent outcomes. To address this issue, we propose a novel approach utilizing a Mixture of Experts (MoE) framework, augmented with Large Language Models (LLMs), to analyze user objectives and constraints effectively, select specialized DRL experts, and weigh each decision from the participating experts. Specifically, we develop a gate network to oversee the expert models, allowing a collective of experts to tackle a wide array of new tasks. Furthermore, we innovatively substitute the traditional gate network with an LLM, leveraging its advanced reasoning capabilities to manage expert model selection for joint decisions. Our proposed method reduces the need to train new DRL models for each unique optimization problem, decreasing energy consumption and AI model implementation costs. The LLM-enabled MoE approach is validated through a general maze navigation task and a specific network service provider utility maximization task, demonstrating its effectiveness and practical applicability in optimizing complex networking systems.
引用
收藏
页码:531 / 536
页数:6
相关论文
共 50 条
  • [1] A large language model-enabled machining process knowledge graph construction method for intelligent process planning
    Xu, Qingfeng
    Qiu, Fei
    Zhou, Guanghui
    Zhang, Chao
    Ding, Kai
    Chang, Fengtian
    Lu, Fengyi
    Yu, Yongrui
    Ma, Dongxu
    Liu, Jiancong
    ADVANCED ENGINEERING INFORMATICS, 2025, 65
  • [2] Generative AI Agents With Large Language Model for Satellite Networks via a Mixture of Experts Transmission
    Zhang, Ruichen
    Du, Hongyang
    Liu, Yinqiu
    Niyato, Dusit
    Kang, Jiawen
    Xiong, Zehui
    Jamalipour, Abbas
    Kim, Dong In
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2024, 42 (12) : 3581 - 3596
  • [3] Human-Autonomy Teaming on Autonomous Vehicles with Large Language Model-Enabled Human Digital Twins
    Cui, Can
    Ma, Yunsheng
    Cao, Xu
    Ye, Wenqian
    Wang, Ziran
    2023 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING, SEC 2023, 2023, : 319 - 324
  • [4] Denoising diffusion probabilistic model-enabled data augmentation method for intelligent machine fault diagnosis
    Zhao, Pengcheng
    Zhang, Wei
    Cao, Xiaoshan
    Li, Xiang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [5] Virtual Co-Pilot: Multimodal Large Language Model-enabled Quick-access Procedures for Single Pilot Operations
    Li, Fan
    Feng, Shanshan
    Yan, Yuqi
    Lee, Ching-Hung
    Ong, Yew Soon
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 1501 - 1506
  • [6] Dialogue Summarization with Mixture of Experts based on Large Language Models
    Tian, Yuanhe
    Xia, Fei
    Song, Yan
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 7143 - 7155
  • [7] Few-Shot Learning in Wireless Networks: A Meta-Learning Model-Enabled Scheme
    Xiong, Kexin
    Zhao, Zhongyuan
    Hong, Wei
    Peng, Mugen
    Quek, Tony Q. S.
    2022 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2022, : 415 - 420
  • [8] BIM and individual physical needs-trained model-enabled approach to spatial redesign and visualization
    Lee, Jin-Kook
    Chae, Sumin
    Yang, Eunhwa
    Hong, Seung Wan
    JOURNAL OF ASIAN ARCHITECTURE AND BUILDING ENGINEERING, 2024,
  • [9] Getting MoRE out of Mixture of Language Model Reasoning Experts
    Si, Chenglei
    Shit, Weijia
    Zhao, Chen
    Zettlemoyer, Luke
    Boyd-Graber, Jordan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8234 - 8249
  • [10] CECA: An intelligent large-language-model-enabled method for accounting embodied carbon in buildings
    Gu, Xierong
    Chen, Cheng
    Fang, Yuan
    Mahabir, Ron
    Fan, Lei
    BUILDING AND ENVIRONMENT, 2025, 272