Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach

被引:0
|
作者
Du, Hongyang [1 ]
Liu, Guangyuan [1 ]
Lin, Yijing [2 ]
Niyato, Dusit [1 ]
Kang, Jiawen [3 ,4 ,5 ]
Xiong, Zehui [6 ]
Kim, Dong In [7 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[4] Minist Educ, Key Lab Intelligent Informat Proc & Syst Integrat, Guangzhou 510006, Peoples R China
[5] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Peoples R China
[6] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore 487372, Singapore
[7] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon 16419, South Korea
来源
20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024 | 2024年
基金
新加坡国家研究基金会; 中国国家自然科学基金;
关键词
Generative AI (GAI); large language model; mixture of experts; network optimization;
D O I
10.1109/IWCMC61514.2024.10592370
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optimizing various wireless user tasks poses a significant challenge for networking systems because of the expanding range of user requirements. Despite advancements in Deep Reinforcement Learning (DRL), the need for customized optimization tasks for individual users complicates developing and applying numerous DRL models, leading to substantial computation resource and energy consumption and can lead to inconsistent outcomes. To address this issue, we propose a novel approach utilizing a Mixture of Experts (MoE) framework, augmented with Large Language Models (LLMs), to analyze user objectives and constraints effectively, select specialized DRL experts, and weigh each decision from the participating experts. Specifically, we develop a gate network to oversee the expert models, allowing a collective of experts to tackle a wide array of new tasks. Furthermore, we innovatively substitute the traditional gate network with an LLM, leveraging its advanced reasoning capabilities to manage expert model selection for joint decisions. Our proposed method reduces the need to train new DRL models for each unique optimization problem, decreasing energy consumption and AI model implementation costs. The LLM-enabled MoE approach is validated through a general maze navigation task and a specific network service provider utility maximization task, demonstrating its effectiveness and practical applicability in optimizing complex networking systems.
引用
收藏
页码:531 / 536
页数:6
相关论文
共 50 条
  • [41] Adversarial Text Purification: A Large Language Model Approach for Defense
    Moraffah, Raha
    Khandelwal, Shubh
    Bhattacharjee, Amrita
    Liu, Huan
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PT V, PAKDD 2024, 2024, 14649 : 65 - 77
  • [42] A Large Language Model Approach to Educational Survey Feedback Analysis
    Parker, Michael J.
    Anderson, Caitlin
    Stone, Claire
    Oh, Yearim
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2024,
  • [43] LLM-CDM: A Large Language Model Enhanced Cognitive Diagnosis for Intelligent Education
    Chen, Xin
    Zhang, Jin
    Zhou, Tong
    Zhang, Feng
    IEEE ACCESS, 2025, 13 : 47165 - 47180
  • [44] Large Language Model-Driven Curriculum Design for Mobile Networks
    Erak, Omar
    Alhussein, Omar
    Naser, Shimaa
    Alabbasi, Nouf
    Mi, De
    Muhaidat, Sami
    2024 IEEE/CIC INTERNATIONAL CONFERENCE ON COMMUNICATIONS IN CHINA, ICCC, 2024,
  • [45] Intelligent extraction of reservoir dispatching information integrating large language model and structured prompts
    Yangrui Yang
    Sisi Chen
    Yaping Zhu
    Xuemei Liu
    Wei Ma
    Ling Feng
    Scientific Reports, 14 (1)
  • [46] Accelerating Mixture-of-Experts language model inference via plug-and-play lookahead gate on a single GPU
    Ou, Jie
    Chen, Yueming
    Xiong, Buyao
    Wang, Zhaokun
    Tian, Wenhong
    COMPUTER STANDARDS & INTERFACES, 2025, 94
  • [47] Intelligent extraction of reservoir dispatching information integrating large language model and structured prompts
    Yang, Yangrui
    Chen, Sisi
    Zhu, Yaping
    Liu, Xuemei
    Ma, Wei
    Feng, Ling
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [48] Special issue on "Neuro-Symbolic Intelligence: large Language Model enabled Knowledge Engineering"
    Wang, Haofen
    Khan, Arijit
    Liu, Jun
    Witbrock, Michael
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2025, 28 (01):
  • [49] IPDCA: Intelligent Proficient Data Collection Approach for IoT-Enabled Wireless Sensor Networks in Smart Environments
    Osamy, Walid
    Khedr, Ahmed M.
    El-Sawy, Ahmed A.
    Salim, Ahmed
    Vijayan, Dilna
    ELECTRONICS, 2021, 10 (09)
  • [50] Robust Event Boundary Detection in Sensor Networks - A Mixture Model Based Approach
    Ding, Min
    Cheng, Xiuzhen
    IEEE INFOCOM 2009 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, VOLS 1-5, 2009, : 2991 - 2995