Improving Speech Understanding Accuracy with Limited Training Data Using Multiple Language Models and Multiple Understanding Models

被引:0
|
作者
Katsumaru, Masaki [1 ]
Nakano, Mikio [2 ]
Komatani, Kazunori [1 ]
Funakoshi, Kotaro [2 ]
Ogata, Tetsuya [1 ]
Okuno, Hiroshi G. [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto, Japan
[2] Honda Res Inst Japan Co Ltd, Kisarazu, Chiba, Japan
关键词
speech understanding; multiple language models and language understanding models; limited training data;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.
引用
收藏
页码:2699 / +
页数:2
相关论文
共 50 条
  • [1] Improving Conversation-Context Language Models with Multiple Spoken Language Understanding Models
    Masumura, Ryo
    Tanaka, Tomohiro
    Ando, Atsushi
    Kamiyama, Hosana
    Oba, Takanobu
    Kobashikawa, Satoshi
    Aono, Yushi
    INTERSPEECH 2019, 2019, : 834 - 838
  • [2] Language understanding using hidden understanding models
    Schwartz, R
    Miller, S
    Stallard, D
    Makhoul, J
    ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 997 - 1000
  • [3] Understanding models understanding language
    Sogaard, Anders
    SYNTHESE, 2022, 200 (06)
  • [4] Understanding models understanding language
    Anders Søgaard
    Synthese, 200
  • [5] On the Evaluation of Speech Foundation Models for Spoken Language Understanding
    Arora, Siddhant
    Pasad, Ankita
    Chien, Chung-Ming
    Han, Jionghao
    Sharma, Roshan
    Jung, Jee-weon
    Dhamyal, Hira
    Chen, William
    Shona, Suwon
    Lee, Hung-yi
    Livescu, Karen
    Watanabe, Shinji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11923 - 11938
  • [6] JOINT LANGUAGE MODELS FOR AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING
    Bayer, Ali Orkan
    Riccardi, Giuseppe
    2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 199 - 203
  • [7] Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
    Meng, Yu
    Huang, Jiaxin
    Zhang, Yu
    Han, Jiawei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [8] Automatic Allocation of Training Data for Speech Understanding Based on Multiple Model Combinations
    Komatani, Kazunori
    Nakano, Mikio
    Katsumaru, Masaki
    Funakoshi, Kotaro
    Ogata, Tetsuya
    Okuno, Hiroshi G.
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2012, E95D (09): : 2298 - 2307
  • [9] Deep Study of CRF Models for Speech understanding in Limited Task
    Graja, Marwa
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (02) : 220 - 226
  • [10] Improving Medical Speech-to-Text Accuracy using Vision-Language Pre-training Models
    Huh, Jaeyoung
    Park, Sangjoon
    Lee, Jeong Eun
    Ye, Jong Chul
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (03) : 1692 - 1703