CARTGPT: Improving CART Captioning using Large Language Models

被引:0
|
作者
Wu, Liang-Yuan [1 ]
Kleiver, Andrea
Jain, Dhruv [1 ]
机构
[1] Univ Michigan, Comp Sci & Engn, Ann Arbor, MI 48109 USA
关键词
Accessibility; Deaf and hard of hearing; real-time captioning;
D O I
10.1145/3663548.3688494
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Communication Access Realtime Translation (CART) is a commonly used real-time captioning technology used by deaf and hard of hearing (DHH) people, due to its accuracy, reliability, and ability to provide a holistic view of the conversational environment (e.g., by displaying speaker names). However, in many real-world situations (e.g., noisy environments, long meetings), the CART captioning accuracy can considerably decline, thereby affecting the comprehension of DHH people. In this work-in-progress paper, we introduce CARTGPT, a system to assist CART captioners in improving their transcription accuracy. CARTGPT takes in errored CART captions and inaccurate automatic speech recognition (ASR) captions as input and uses a large language model to generate corrected captions in real-time. We quantified performance on a noisy speech dataset, showing that our system outperforms both CART (+5.6% accuracy) and a state-of-the-art ASR model (+17.3%). A preliminary evaluation with three DHH users further demonstrates the promise of our approach.
引用
收藏
页数:5
相关论文
共 50 条
  • [41] Using Large Language Models to Improve Sentiment Analysis in Latvian Language
    Purvins, Pauls
    Urtans, Evalds
    Caune, Vairis
    BALTIC JOURNAL OF MODERN COMPUTING, 2024, 12 (02): : 165 - 175
  • [42] Driving and suppressing the human language network using large language models
    Tuckute, Greta
    Sathe, Aalok
    Srikant, Shashank
    Taliaferro, Maya
    Wang, Mingye
    Schrimpf, Martin
    Kay, Kendrick
    Fedorenko, Evelina
    NATURE HUMAN BEHAVIOUR, 2024, 8 (03) : 544 - 561
  • [43] Driving and suppressing the human language network using large language models
    Greta Tuckute
    Aalok Sathe
    Shashank Srikant
    Maya Taliaferro
    Mingye Wang
    Martin Schrimpf
    Kendrick Kay
    Evelina Fedorenko
    Nature Human Behaviour, 2024, 8 : 544 - 561
  • [44] Improving Web Element Localization by Using a Large Language Model
    Nass, Michel
    Alegroth, Emil
    Feldt, Robert
    SOFTWARE TESTING VERIFICATION & RELIABILITY, 2024, 34 (07):
  • [45] Classification of Chronic Dizziness Using Large Language Models
    Xu, Xiaowei
    Jiang, Ruixuan
    Zheng, Si
    Wang, Min
    Ju, Yi
    Li, Jiao
    JOURNAL OF HEALTHCARE INFORMATICS RESEARCH, 2025, 9 (01) : 88 - 102
  • [46] Smart process automation using Large Language Models
    Matuschek, Mario
    ATP MAGAZINE, 2025, (03):
  • [47] Verbal lie detection using Large Language Models
    Loconte, Riccardo
    Russo, Roberto
    Capuozzo, Pasquale
    Pietrini, Pietro
    Sartori, Giuseppe
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [48] The dangers of using large language models for peer review
    Carr, Edward J.
    Wu, Mary Y.
    Gahir, Joshua
    Harvey, Ruth
    Townsley, Hermaleigh
    Bailey, Chris
    Fowler, Ashley S.
    Dowgier, Giulia
    Hobbs, Agnieszka
    Herman, Lou
    Ragno, Martina
    Miah, Murad
    Bawumia, Phillip
    Smith, Callie
    Miranda, Mauro
    Mears, Harriet, V
    Adams, Lorin
    Haptipoglu, Emine
    O'Reilly, Nicola
    Warchal, Scott
    Sawyer, Chelsea
    Ambrose, Karen
    Kelly, Gavin
    Beale, Rupert
    Papineni, Padmasayee
    Corrah, Tumena
    Gilson, Richard
    Gamblin, Steve
    Kassiotis, George
    Libri, Vincenzo
    Williams, Bryan
    Swanton, Charles
    Gandhi, Sonia
    Bauer, David L., V
    Wall, Emma C.
    LANCET INFECTIOUS DISEASES, 2023, 23 (07): : 781 - 781
  • [49] Using large language models to create narrative events
    Bartalesi, Valentina
    Lenzi, Emanuele
    De Martino, Claudio
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [50] An overview of diagnostics and therapeutics using large language models
    Malgaroli, Matteo
    Mcduff, Daniel
    JOURNAL OF TRAUMATIC STRESS, 2024, 37 (05) : 754 - 760