Improving Speech Understanding Accuracy with Limited Training Data Using Multiple Language Models and Multiple Understanding Models

被引:0
|
作者
Katsumaru, Masaki [1 ]
Nakano, Mikio [2 ]
Komatani, Kazunori [1 ]
Funakoshi, Kotaro [2 ]
Ogata, Tetsuya [1 ]
Okuno, Hiroshi G. [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto, Japan
[2] Honda Res Inst Japan Co Ltd, Kisarazu, Chiba, Japan
关键词
speech understanding; multiple language models and language understanding models; limited training data;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.
引用
收藏
页码:2699 / +
页数:2
相关论文
共 50 条
  • [21] Towards situated speech understanding: visual context priming of language models
    Roy, D
    Mukherjee, N
    COMPUTER SPEECH AND LANGUAGE, 2005, 19 (02): : 227 - 248
  • [22] Markovian combination of language and prosodic models for better speech understanding and recognition
    Stolcke, A
    Shriberg, E
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 177 - 177
  • [23] Phrase language models for detection and verification-based speech understanding
    Kawahara, T
    Doshita, S
    Lee, CH
    1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, PROCEEDINGS, 1997, : 49 - 56
  • [24] Utilizing Multiple Speech Recognizers to Improve Spoken Language Understanding Performance
    Choi, Stanley Jungkyu
    Lee, Kyehwan
    Hahn, Minsoo
    18TH IEEE INTERNATIONAL SYMPOSIUM ON CONSUMER ELECTRONICS (ISCE 2014), 2014,
  • [25] Data Augmentation for Spoken Language Understanding via Pretrained Language Models
    Peng, Baolin
    Zhu, Chenguang
    Zeng, Michael
    Gao, Jianfeng
    INTERSPEECH 2021, 2021, : 1219 - 1223
  • [27] Understanding language understanding: Computational models of reading.
    Capo, T
    APPLIED COGNITIVE PSYCHOLOGY, 2002, 16 (07) : 864 - 865
  • [28] Understanding disability trajectories in a contemporary multiple sclerosis cohort using multistate models
    Harding, K.
    Joseph, F.
    Pickersgill, T.
    Tomassini, V.
    Tallantyre, E.
    Robertson, N.
    MULTIPLE SCLEROSIS JOURNAL, 2019, 25 : 604 - 605
  • [29] Using Natural Sentences for Understanding Biases in Language Models
    Alnegheimish, Sarah
    Guo, Alicia
    Sun, Yi
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 2824 - 2830
  • [30] Language understanding using n-multigram models
    Hurtado, L
    Segarra, E
    García, F
    Sanchis, E
    ADVANCES IN NATURAL LANGUAGE PROCESSING, 2004, 3230 : 207 - 219