Mandarin Recognition Based on Self-Attention Mechanism with Deep Convolutional Neural Network (DCNN)-Gated Recurrent Unit (GRU)

被引:0
|
作者
Chen, Xun [1 ]
Wang, Chengqi [1 ]
Hu, Chao [1 ]
Wang, Qin [1 ]
机构
[1] Hainan Univ, Sch Informat & Commun Engn, Haikou 570228, Peoples R China
基金
中国国家自然科学基金;
关键词
self-attention mechanism; CTC; gated circulation units;
D O I
10.3390/bdcc8120195
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Speech recognition technology is an important branch in the field of artificial intelligence, aiming to transform human speech into computer-readable text information. However, speech recognition technology still faces many challenges, such as noise interference, and accent and speech rate differences. An aim of this paper is to explore a deep learning-based speech recognition method to improve the accuracy and robustness of speech recognition. Firstly, this paper introduces the basic principles of speech recognition and existing mainstream technologies, and then focuses on the deep learning-based speech recognition method. Through comparative experiments, it is found that the self-attention mechanism performs best in speech recognition tasks. In order to further improve speech recognition performance, this paper proposes a deep learning model based on the self-attention mechanism with DCNN-GRU. The model realizes the dynamic attention to an input speech by introducing the self-attention mechanism in a neural network model instead of an RNN and with a deep convolutional neural network, which improves the robustness and recognition accuracy of this model. This experiment uses 170 h of Chinese dataset AISHELL-1. Compared with the deep convolutional neural network, the deep learning model based on the self-attention mechanism with DCNN-GRU accomplishes a reduction of at least 6% in CER. Compared with a bidirectional gated recurrent neural network, the deep learning model based on the self-attention mechanism with DCNN-GRU accomplishes a reduction of 0.7% in CER. And finally, this experiment is performed on a test set analyzed the influencing factors affecting the CER. The experimental results show that this model exhibits good performance in various noise environments and accent conditions.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Recurrent Neural Network Model with Self-Attention Mechanism for Fault Detection and Diagnosis
    Zhang, Rui
    Xiong, Zhihua
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 4706 - 4711
  • [32] Multiple Protein Subcellular Locations Prediction Based on Deep Convolutional Neural Networks with Self-Attention Mechanism
    Cong, Hanhan
    Liu, Hong
    Cao, Yi
    Chen, Yuehui
    Liang, Cheng
    INTERDISCIPLINARY SCIENCES-COMPUTATIONAL LIFE SCIENCES, 2022, 14 (02) : 421 - 438
  • [33] Multiple Protein Subcellular Locations Prediction Based on Deep Convolutional Neural Networks with Self-Attention Mechanism
    Hanhan Cong
    Hong Liu
    Yi Cao
    Yuehui Chen
    Cheng Liang
    Interdisciplinary Sciences: Computational Life Sciences, 2022, 14 : 421 - 438
  • [34] Learning model combining convolutional deep neural network with a self-attention mechanism for AC optimal power flow
    Tran, Quan
    Mitra, Joydeep
    Nguyen, Nga
    ELECTRIC POWER SYSTEMS RESEARCH, 2024, 231
  • [35] Neural network based on convolution and self-attention fusion mechanism for plant leaves disease recognition
    Zhao, Yun
    Li, Yang
    Wu, Na
    Xu, Xing
    CROP PROTECTION, 2024, 180
  • [36] Vehicle speed prediction using a convolutional neural network combined with a gated recurrent unit with attention
    Zhang, Dongxue
    Wang, Zhennan
    Jiao, Xiaohong
    Zhang, Zhao
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2024,
  • [37] Masked face recognition based on knowledge distillation and convolutional self-attention network
    Wan, Weiguo
    Wen, Runlin
    Yao, Li
    Yang, Yong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, : 2269 - 2284
  • [38] 2D Self-attention Convolutional Recurrent Network for Offline Handwritten Text Recognition
    Ly, Nam Tuan
    Nguyen, Hung Tuan
    Nakagawa, Masaki
    DOCUMENT ANALYSIS AND RECOGNITION - ICDAR 2021, PT I, 2021, 12821 : 191 - 204
  • [39] Masked face recognition with convolutional visual self-attention network
    Ge, Yiming
    Liu, Hui
    Du, Junzhao
    Li, Zehua
    Wei, Yuheng
    NEUROCOMPUTING, 2023, 518 : 496 - 506
  • [40] Gesture Recognition and Master–Slave Control of a Manipulator Based on sEMG and Convolutional Neural Network–Gated Recurrent Unit
    Ge, Zhaojie
    Wu, Zhile
    Han, Xu
    Zhao, Ping
    Journal of Engineering and Science in Medical Diagnostics and Therapy, 2023, 6 (02)