Dialogue Summarization with Mixture of Experts based on Large Language Models

被引:0
|
作者
Tian, Yuanhe [1 ,2 ]
Xia, Fei [2 ]
Song, Yan [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Univ Washington, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dialogue summarization is an important task that requires to generate highlights for a conversation from different aspects (e.g., content of various speakers). While several studies successfully employ large language models (LLMs) and achieve satisfying results, they are limited by using one model at a time or treat it as a black box, which makes it hard to discriminatively learn essential content in a dialogue from different aspects, therefore may lead to anticipation bias and potential loss of information in the produced summaries. In this paper, we propose an LLM-based approach with role-oriented routing and fusion generation to utilize mixture of experts (MoE) for dialogue summarization. Specifically, the role-oriented routing is an LLM-based module that selects appropriate experts to process different information; fusion generation is another LLM-based module to locate salient information and produce finalized dialogue summaries. The proposed approach offers an alternative solution to employing multiple LLMs for dialogue summarization by leveraging their capabilities of in-context processing and generation in an effective manner. We run experiments on widely used benchmark datasets for this task, where the results demonstrate the superiority of our approach in producing informative and accurate dialogue summarization.(1)
引用
收藏
页码:7143 / 7155
页数:13
相关论文
共 50 条
  • [31] Evaluating the Factual Consistency of Large Language Models Through News Summarization
    Tam, Derek
    Mascarenhas, Anisha
    Zhang, Shiyue
    Kwan, Sarah
    Bansal, Mohit
    Raffel, Colin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5220 - 5255
  • [32] Leveraging large language models for abstractive summarization of Italian legal news
    Benedetto, Irene
    Cagliero, Luca
    Ferro, Michele
    Tarasconi, Francesco
    Bernini, Claudia
    Giacalone, Giuseppe
    ARTIFICIAL INTELLIGENCE AND LAW, 2025,
  • [33] Assessing the Impact of Prompt Strategies on Text Summarization with Large Language Models
    Onan, Aytug
    Alhumyani, Hesham
    COMPUTER APPLICATIONS IN INDUSTRY AND ENGINEERING, CAINE 2024, 2025, 2242 : 41 - 55
  • [34] Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach
    Du, Hongyang
    Liu, Guangyuan
    Lin, Yijing
    Niyato, Dusit
    Kang, Jiawen
    Xiong, Zehui
    Kim, Dong In
    20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 531 - 536
  • [35] Empirical Analysis of Dialogue Relation Extraction with Large Language Models
    Li, Guozheng
    Xu, Zijie
    Shang, Ziyu
    Liu, Jiajun
    Ji, Ke
    Guo, Yikai
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 6359 - 6367
  • [36] Fine-Tuning Pretrained Language Models to Enhance Dialogue Summarization in Customer Service Centers
    Yun, Jiseon
    Sohn, Jae Eui
    Kyeong, Sunghyon
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 365 - 373
  • [37] Language Model as an Annotator: Exploring DialoGPT for Dialogue Summarization
    Feng, Xiachong
    Feng, Xiaocheng
    Qin, Libo
    Qin, Bing
    Liu, Ting
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1479 - 1491
  • [38] Kernel-Based Mixture of Experts Models For Linear Regression
    Santarcangelo, Joseph
    Zhang, Xiao-Ping
    2015 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2015, : 1526 - 1529
  • [39] Large language models surpass human experts in predicting neuroscience results
    Luo, Xiaoliang
    Rechardt, Akilles
    Sun, Guangzhi
    Nejad, Kevin K.
    Yanez, Felipe
    Yilmaz, Bati
    Lee, Kangjoo
    Cohen, Alexandra O.
    Borghesani, Valentina
    Pashkov, Anton
    Marinazzo, Daniele
    Nicholas, Jonathan
    Salatiello, Alessandro
    Sucholutsky, Ilia
    Minervini, Pasquale
    Razavi, Sepehr
    Rocca, Roberta
    Yusifov, Elkhan
    Okalova, Tereza
    Gu, Nianlong
    Ferianc, Martin
    Khona, Mikail
    Patil, Kaustubh R.
    Lee, Pui-Shee
    Mata, Rui
    Myers, Nicholas E.
    Bizley, Jennifer K.
    Musslick, Sebastian
    Bilgin, Isil Poyraz
    Niso, Guiomar
    Ales, Justin M.
    Gaebler, Michael
    Ratan Murty, N. Apurva
    Loued-Khenissi, Leyla
    Behler, Anna
    Hall, Chloe M.
    Dafflon, Jessica
    Bao, Sherry Dongqi
    Love, Bradley C.
    NATURE HUMAN BEHAVIOUR, 2025, 9 (02): : 305 - 315
  • [40] Large language models as a substitute for human experts in annotating political text
    Heseltine, Michael
    von Hohenberg, Bernhard Clemm
    RESEARCH & POLITICS, 2024, 11 (01)