LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS

被引:0
|
作者
Sainath, Tara N. [1 ]
Kingsbury, Brian [1 ]
Sindhwani, Vikas [1 ]
Arisoy, Ebru [1 ]
Ramabhadran, Bhuvana [1 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Deep Neural Networks; Speech Recognition;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
While Deep Neural Networks (DNNs) have achieved tremendous success for large vocabulary continuous speech recognition (LVCSR) tasks, training of these networks is slow. One reason is that DNNs are trained with a large number of training parameters (i.e., 10-50 million) Because networks are trained with a large number of output targets to achieve good performance, the majority of these parameters are in the final weight layer. In this paper, we propose a low-rank matrix factorization of the final weight layer. We apply this low-rank technique to DNNs for both acoustic modeling and language modeling. We show on three different LVCSR tasks ranging between 50-400 hrs, that a low-rank factorization reduces the number of parameters of the network by 30-50%. This results in roughly an equivalent reduction in training time, without a significant loss in final recognition accuracy, compared to a full-rank representation.
引用
收藏
页码:6655 / 6659
页数:5
相关论文
共 50 条
  • [1] Learning and Compressing: Low-Rank Matrix Factorization for Deep Neural Network Compression
    Cai, Gaoyuan
    Li, Juhu
    Liu, Xuanxin
    Chen, Zhibo
    Zhang, Haiyan
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [2] Reconstruction of a high-dimensional low-rank matrix
    Yata, Kazuyoshi
    Aoshima, Makoto
    ELECTRONIC JOURNAL OF STATISTICS, 2016, 10 (01): : 895 - 917
  • [3] EXTRACTING DEEP NEURAL NETWORK BOTTLENECK FEATURES USING LOW-RANK MATRIX FACTORIZATION
    Zhang, Yu
    Chuangsuwanich, Ekapol
    Glass, James
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [4] Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks
    Povey, Daniel
    Cheng, Gaofeng
    Wang, Yiming
    Li, Ke
    Xu, Hainan
    Yarmohamadi, Mahsa
    Khudanpur, Sanjeev
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3743 - 3747
  • [5] ESTIMATION OF HIGH-DIMENSIONAL LOW-RANK MATRICES
    Rohde, Angelika
    Tsybakov, Alexandre B.
    ANNALS OF STATISTICS, 2011, 39 (02): : 887 - 930
  • [6] High-dimensional VAR with low-rank transition
    Alquier, Pierre
    Bertin, Karine
    Doukhan, Paul
    Garnier, Remy
    STATISTICS AND COMPUTING, 2020, 30 (04) : 1139 - 1153
  • [7] High-dimensional VAR with low-rank transition
    Pierre Alquier
    Karine Bertin
    Paul Doukhan
    Rémy Garnier
    Statistics and Computing, 2020, 30 : 1139 - 1153
  • [8] High-dimensional covariance matrix estimation using a low-rank and diagonal decomposition
    Wu, Yilei
    Qin, Yingli
    Zhu, Mu
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 2020, 48 (02): : 308 - 337
  • [9] Low-rank Riemannian eigensolver for high-dimensional Hamiltonians
    Rakhuba, Maxim
    Novikov, Alexander
    Oseledets, Ivan
    JOURNAL OF COMPUTATIONAL PHYSICS, 2019, 396 : 718 - 737
  • [10] On the Minimal Problems of Low-Rank Matrix Factorization
    Jiang, Fangyuan
    Oskarsson, Magnus
    Astrom, Kalle
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 2549 - 2557