Improving transformer-based acoustic model performance using sequence discriminative training

被引:0
|
作者
Lee, Chae-Won [1 ]
Chang, Joon-Hyuk [1 ]
机构
[1] Hanyang Univ, Dept Elect Engn, 222,Wangsimni Ro, Seoul 04763, South Korea
来源
关键词
Speech recognition; Transformer; Sequence discriminative training; Weighted finite state transducer;
D O I
10.7776/ASK.2022.41.3.335
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we adopt a transformer that shows remarkable performance in natural language processing as an acoustic model of hybrid speech recognition. The transformer acoustic model uses attention structures to process sequential data and shows high performance with low computational cost. This paper proposes a method to improve the performance of transformer AM by applying each of the four algorithms of sequence discriminative training, a weighted finite-state transducer (wFST)-based learning used in the existing DNN-HMM model. In addition, compared to the Cross Entropy (CE) learning method, sequence discriminative method shows 5 % of the relative Word Error Rate (WER).
引用
收藏
页码:335 / 341
页数:7
相关论文
共 50 条
  • [1] Improving Short Answer Grading Using Transformer-Based Pre-training
    Sung, Chul
    Dhamecha, Tejas Indulal
    Mukhi, Nirmal
    ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2019), PT I, 2019, 11625 : 469 - 481
  • [2] Training and analyzing a Transformer-based machine translation model
    Pimentel, Clovis Henrique Martins
    Pires, Thiago Blanch
    TEXTO LIVRE-LINGUAGEM E TECNOLOGIA, 2024, 17
  • [3] Improving soil surface evaporation estimates with transformer-based model
    Zou, Mijun
    Zhong, Lei
    Jia, Weijia
    Ge, Yangfei
    Mamtimin, Ali
    ATMOSPHERIC RESEARCH, 2025, 316
  • [4] Sequence discriminative training for deep learning based acoustic keyword spotting
    Chen, Zhehuai
    Qian, Yanmin
    Yu, Kai
    SPEECH COMMUNICATION, 2018, 102 : 100 - 111
  • [5] Pre-training a Transformer-Based Generative Model Using a Small Sepedi Dataset
    Ramalepe, Simon Phetole
    Modipa, Thipe I.
    Davel, Marelie H.
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 319 - 333
  • [6] Improving the quality of dental crown using a Transformer-based method
    Hosseinimanesh, Golriz
    Ghadiri, Farnoosh
    Alsheghri, Ammar
    Zhang, Ying
    Keren, Julia
    Cheriet, Farida
    Guibault, Francois
    arXiv, 2023,
  • [7] Training data selection for improving discriminative training of acoustic models
    Liu, Shih-Hung
    Chu, Fang-Hui
    Lin, Shih-Hsiang
    Lee, Hung-Shin
    Chen, Berlin
    2007 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, VOLS 1 AND 2, 2007, : 284 - 289
  • [8] Training data selection for improving discriminative training of acoustic models
    Chen, Berlin
    Liu, Shih-Hung
    Chu, Fang-Hui
    PATTERN RECOGNITION LETTERS, 2009, 30 (13) : 1228 - 1235
  • [9] Predicting the formation of NADES using a transformer-based model
    Ayres, Lucas B.
    Gomez, Federico J. V.
    Silva, Maria Fernanda
    Linton, Jeb R.
    Garcia, Carlos D.
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [10] Generating Music Transition by Using a Transformer-Based Model
    Hsu, Jia-Lien
    Chang, Shuh-Jiun
    ELECTRONICS, 2021, 10 (18)