Compressing Speech Recognition Networks with MLP via Tensor-Train Decomposition

被引:0
|
作者
He, Dan [1 ,2 ]
Zhong, Yubin [1 ]
机构
[1] Guangzhou Univ, Guangzhou, Peoples R China
[2] Tsinghua Ununiv, CSLT, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep neural networks (DNNs) have produced state-of-the-art performance in automatic speech recognition (ASR). This success is often associated with a large DNN structure with millions or even billions of parameters. Such large-scale networks take large disk space and require huge computational resources at run-time, therefore not suitable for applications in mobile or wearable devices. In this paper, we investigate a compression approach for DNNs based on Tensor-Train (TT) decomposition and apply it to the ASR task. Our results on the TIMIT database reveals that the compressed networks can maintain the performance of the original full-connected network, while greatly reducing the number of parameters. In particular, we found that the rate of model size decreasing is much larger than the rate of WER (word error rate) increasing, which means that the performance loss caused by the TT-based compression can be well compensated by the model size reduction. Moreover, how many layers and which layer can be substituted by TT is application dependent and should be carefully designed according to the application scenario.
引用
收藏
页码:1215 / 1219
页数:5
相关论文
共 50 条
  • [1] Compressing End-to-end ASR Networks by Tensor-Train Decomposition
    Mori, Takuma
    Tjandra, Andros
    Sakti, Sakriani
    Nakamura, Satoshi
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 806 - 810
  • [2] Tensor-Train decomposition for image recognition
    D. Brandoni
    V. Simoncini
    Calcolo, 2020, 57
  • [3] Tensor-Train decomposition for image recognition
    Brandoni, D.
    Simoncini, V
    CALCOLO, 2020, 57 (01)
  • [4] TENSOR-TRAIN DECOMPOSITION
    Oseledets, I. V.
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2011, 33 (05): : 2295 - 2317
  • [5] SPECTRAL TENSOR-TRAIN DECOMPOSITION
    Bigoni, Daniele
    Engsig-Karup, Allan P.
    Marzouk, Youssef M.
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2016, 38 (04): : A2405 - A2439
  • [6] A continuous analogue of the tensor-train decomposition
    Gorodetsky, Alex
    Karaman, Sertac
    Marzouk, Youssef
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2019, 347 : 59 - 84
  • [7] Accelerating Tensor Contraction Products via Tensor-Train Decomposition [Tips & Tricks]
    Kisil, Ilya
    Calvi, Giuseppe G.
    Konstantinidis, Kriton
    Xu, Yao Lei
    Mandic, Danilo P.
    IEEE SIGNAL PROCESSING MAGAZINE, 2022, 39 (05) : 63 - 70
  • [8] Compact lossy compression of tensors via neural tensor-train decomposition
    Kwon, Taehyung
    Ko, Jihoon
    Jung, Jinhong
    Jang, Jun-Gi
    Shin, Kijung
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (02) : 1169 - 1211
  • [9] Completion of High Order Tensor Data with Missing Entries via Tensor-Train Decomposition
    Yuan, Longhao
    Zhao, Qibin
    Cao, Jianting
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT I, 2017, 10634 : 222 - 229
  • [10] EXPLOITING HYBRID MODELS OF TENSOR-TRAIN NETWORKS FOR SPOKEN COMMAND RECOGNITION
    Qi, Jun
    Tejedor, Javier
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3114 - 3118