Improving Speech Understanding Accuracy with Limited Training Data Using Multiple Language Models and Multiple Understanding Models

被引:0
|
作者
Katsumaru, Masaki [1 ]
Nakano, Mikio [2 ]
Komatani, Kazunori [1 ]
Funakoshi, Kotaro [2 ]
Ogata, Tetsuya [1 ]
Okuno, Hiroshi G. [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto, Japan
[2] Honda Res Inst Japan Co Ltd, Kisarazu, Chiba, Japan
关键词
speech understanding; multiple language models and language understanding models; limited training data;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.
引用
收藏
页码:2699 / +
页数:2
相关论文
共 50 条
  • [31] Improving the robustness and accuracy of biomedical language models through adversarial training
    Moradi, Milad
    Samwald, Matthias
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 132
  • [32] Verbs in Action: Improving verb understanding in video-language models
    Momeni, Liliane
    Caron, Mathilde
    Nagrani, Arsha
    Zisserman, Andrew
    Schmid, Cordelia
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 15533 - 15545
  • [33] Data fusion using multiple models
    Sworder, DD
    Boyd, JE
    Elliott, RJ
    Hutchins, RG
    CONFERENCE RECORD OF THE THIRTY-FOURTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2000, : 1749 - 1753
  • [34] The Importance of Understanding Language in Large Language Models
    Youssef, Alaa
    Stein, Samantha
    Clapp, Justin
    Magnus, David
    AMERICAN JOURNAL OF BIOETHICS, 2023, 23 (10): : 6 - 7
  • [35] The Journey of Language Models in Understanding Natural Language
    Liu, Yuanrui
    Zhou, Jingping
    Sang, Guobiao
    Huang, Ruilong
    Zhao, Xinzhe
    Fang, Jintao
    Wang, Tiexin
    Li, Bohan
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 331 - 363
  • [36] Improve accuracy of empirical models with multiple models
    Zhang, YP
    Fan, MH
    PHOTOMASK AND NEXT-GENERATION LITHOGRAPHY MASK TECHNOLOGY X, 2003, 5130 : 698 - 709
  • [37] CLONAL COMPETITION MODELS FOR UNDERSTANDING PROGRESSION AND ENDURANCE IN MULTIPLE MYELOMA
    Haertle, L.
    Martin, L.
    Munawar, U.
    Cuenca, I
    Vogt, C.
    Da-Via, M.
    Garitano Trojaola, A.
    Rasche, L.
    Gallardo, M.
    Stuehmer, T.
    Martinez Lopez, J.
    Kortum, M.
    Barrio, S.
    HAEMATOLOGICA, 2019, 104 : 7 - 8
  • [38] Multiple models of attachment: A new way of understanding perinatal depression
    Vanwalleghem, S.
    Sirparanta, A.
    Leclercq, S.
    Deborde, A. -S
    Miljkovitch, R.
    ENCEPHALE-REVUE DE PSYCHIATRIE CLINIQUE BIOLOGIQUE ET THERAPEUTIQUE, 2022, 48 (05): : 593 - 594
  • [39] MODELS OF NATURAL-LANGUAGE UNDERSTANDING
    BATES, M
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 1995, 92 (22) : 9977 - 9982
  • [40] Meaning and understanding in large language models
    Havlik, Vladimir
    SYNTHESE, 2024, 205 (01)