Disfluent Cues for Enhanced Speech Understanding in Large Language Models

被引:0
|
作者
Rohanian, Morteza [1 ]
Nooralahzadeh, Farhad [1 ]
Rohanian, Omid [2 ]
Clifton, David [2 ]
Krauthammer, Michael [1 ]
机构
[1] Univ Zurich, Dept Quantit Biomed, Zurich, Switzerland
[2] Univ Oxford, Dept Engn Sci, Oxford, England
关键词
REPAIR;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In computational linguistics, the common practice is to "clean" disfluent content from spontaneous speech. However, we hypothesize that these disfluencies might serve as more than mere noise, potentially acting as informative cues. We use a range of pre-trained models for a reading comprehension task involving disfluent queries, specifically featuring different types of speech repairs. The findings indicate that certain disfluencies can indeed improve model performance, particularly those stemming from context-based adjustments. However, large-scale language models struggle to handle repairs involving decision-making or the correction of lexical or syntactic errors, suggesting a crucial area for potential improvement. This paper thus highlights the importance of a nuanced approach to disfluencies, advocating for their potential utility in enhancing model performance rather than their removal.
引用
收藏
页码:3676 / 3684
页数:9
相关论文
共 50 条
  • [31] Believing Anthropomorphism: Examining the Role of Anthropomorphic Cues on Trust in Large Language Models
    Cohn, Michelle
    Pushkarna, Mahima
    Olanubi, Gbolahan O.
    Moran, Joseph M.
    Padgett, Daniel
    Mengesha, Zion
    Heldreth, Courtney
    EXTENDED ABSTRACTS OF THE 2024 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2024, 2024,
  • [32] SPEECH PRODUCTION AND LANGUAGE UNDERSTANDING
    CONWAY, DF
    VOLTA REVIEW, 1992, 94 (03) : 239 - 240
  • [33] Towards an understanding of large language models in software engineering tasks
    Zheng, Zibin
    Ning, Kaiwen
    Zhong, Qingyuan
    Chen, Jiachi
    Chen, Wenqing
    Guo, Lianghong
    Wang, Weicheng
    Wang, Yanlin
    EMPIRICAL SOFTWARE ENGINEERING, 2025, 30 (02)
  • [34] LLMEffiChecker: : Understanding and Testing Efficiency Degradation of Large Language Models
    Feng, Xiaoning
    Han, Xiaohong
    Chen, Simin
    Yang, Wei
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (07)
  • [35] Semantic Understanding of Traffic Scenes with Large Vision Language Models
    Jain, Sandesh
    Thapa, Surendrabikram
    Chen, Kuan-Ting
    Abbott, A. Lynn
    Sarkar, Abhijit
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 1580 - 1587
  • [36] The debate over understanding in AI?s large language models
    Mitchell, Melanie
    Krakauer, David C.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2023, 120 (13)
  • [37] Semantic Understanding of Traffic Scenes with Large Vision Language Models
    Jain, Sandesh
    Thapa, Surendrabikram
    Chen, Kuan-Ting
    Abbott, A. Lynn
    Sarkar, Abhijit
    IEEE Intelligent Vehicles Symposium, Proceedings, 2024, : 1580 - 1587
  • [38] Assess and Summarize: Improve Outage Understanding with Large Language Models
    Jin, Pengxiang
    Zhang, Shenglin
    Ma, Minghua
    Li, Haozhe
    Kang, Yu
    Li, Liqun
    Liu, Yudong
    Qiao, Bo
    Zhang, Chaoyun
    Zhao, Pu
    He, Shilin
    Sarro, Federica
    Dang, Yingnong
    Rajmohan, Saravana
    Lin, Qingwei
    Zhang, Dongmei
    PROCEEDINGS OF THE 31ST ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2023, 2023, : 1657 - 1668
  • [39] Understanding models understanding language
    Sogaard, Anders
    SYNTHESE, 2022, 200 (06)
  • [40] Understanding models understanding language
    Anders Søgaard
    Synthese, 200