Disfluent Cues for Enhanced Speech Understanding in Large Language Models

被引:0
|
作者
Rohanian, Morteza [1 ]
Nooralahzadeh, Farhad [1 ]
Rohanian, Omid [2 ]
Clifton, David [2 ]
Krauthammer, Michael [1 ]
机构
[1] Univ Zurich, Dept Quantit Biomed, Zurich, Switzerland
[2] Univ Oxford, Dept Engn Sci, Oxford, England
关键词
REPAIR;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In computational linguistics, the common practice is to "clean" disfluent content from spontaneous speech. However, we hypothesize that these disfluencies might serve as more than mere noise, potentially acting as informative cues. We use a range of pre-trained models for a reading comprehension task involving disfluent queries, specifically featuring different types of speech repairs. The findings indicate that certain disfluencies can indeed improve model performance, particularly those stemming from context-based adjustments. However, large-scale language models struggle to handle repairs involving decision-making or the correction of lexical or syntactic errors, suggesting a crucial area for potential improvement. This paper thus highlights the importance of a nuanced approach to disfluencies, advocating for their potential utility in enhancing model performance rather than their removal.
引用
收藏
页码:3676 / 3684
页数:9
相关论文
共 50 条
  • [1] Adding Filled Pauses and Disfluent Events into Language Models for Speech Recognition
    Stas, Jan
    Hladek, Daniel
    Juhar, Jozef
    2016 7TH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFOCOMMUNICATIONS (COGINFOCOM), 2016, : 133 - 137
  • [2] Exploring Synergies between Causal Models and Large Language Models for Enhanced Understanding and Inference
    Sun, Yaru
    Yang, Ying
    Fu, Wenhao
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [3] The Importance of Understanding Language in Large Language Models
    Youssef, Alaa
    Stein, Samantha
    Clapp, Justin
    Magnus, David
    AMERICAN JOURNAL OF BIOETHICS, 2023, 23 (10): : 6 - 7
  • [4] Combining large language models with enterprise knowledge graphs: a perspective on enhanced natural language understanding
    Mariotti, Luca
    Guidetti, Veronica
    Mandreoli, Federica
    Belli, Andrea
    Lombardi, Paolo
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 7
  • [5] Meaning and understanding in large language models
    Havlik, Vladimir
    SYNTHESE, 2024, 205 (01)
  • [6] Understanding Telecom Language Through Large Language Models
    Bariah, Lina
    Zou, Hang
    Zhao, Qiyang
    Mouhouche, Belkacem
    Bader, Faouzi
    Debbah, Merouane
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 6542 - 6547
  • [7] On the Evaluation of Speech Foundation Models for Spoken Language Understanding
    Arora, Siddhant
    Pasad, Ankita
    Chien, Chung-Ming
    Han, Jionghao
    Sharma, Roshan
    Jung, Jee-weon
    Dhamyal, Hira
    Chen, William
    Shona, Suwon
    Lee, Hung-yi
    Livescu, Karen
    Watanabe, Shinji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11923 - 11938
  • [8] JOINT LANGUAGE MODELS FOR AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING
    Bayer, Ali Orkan
    Riccardi, Giuseppe
    2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 199 - 203
  • [9] Understanding HTML']HTML with Large Language Models
    Gur, Izzeddin
    Nachum, Ofir
    Miao, Yingjie
    Safdari, Mustafa
    Huang, Austin
    Chowdhery, Aakanksha
    Narang, Sharan
    Fiedel, Noah
    Faust, Aleksandra
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2803 - 2821
  • [10] Shortcut Learning of Large Language Models in Natural Language Understanding
    Du, Mengnan
    He, Fengxiang
    Zou, Na
    Tao, Dacheng
    Hu, Xia
    COMMUNICATIONS OF THE ACM, 2024, 67 (01) : 110 - 120