LOW-RANK MATRIX FACTORIZATION FOR DEEP NEURAL NETWORK TRAINING WITH HIGH-DIMENSIONAL OUTPUT TARGETS

被引:0
|
作者
Sainath, Tara N. [1 ]
Kingsbury, Brian [1 ]
Sindhwani, Vikas [1 ]
Arisoy, Ebru [1 ]
Ramabhadran, Bhuvana [1 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
来源
2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2013年
关键词
Deep Neural Networks; Speech Recognition;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
While Deep Neural Networks (DNNs) have achieved tremendous success for large vocabulary continuous speech recognition (LVCSR) tasks, training of these networks is slow. One reason is that DNNs are trained with a large number of training parameters (i.e., 10-50 million) Because networks are trained with a large number of output targets to achieve good performance, the majority of these parameters are in the final weight layer. In this paper, we propose a low-rank matrix factorization of the final weight layer. We apply this low-rank technique to DNNs for both acoustic modeling and language modeling. We show on three different LVCSR tasks ranging between 50-400 hrs, that a low-rank factorization reduces the number of parameters of the network by 30-50%. This results in roughly an equivalent reduction in training time, without a significant loss in final recognition accuracy, compared to a full-rank representation.
引用
收藏
页码:6655 / 6659
页数:5
相关论文
共 50 条
  • [31] Low-rank nonnegative matrix factorization on Stiefel manifold
    He, Ping
    Xu, Xiaohua
    Ding, Jie
    Fan, Baichuan
    INFORMATION SCIENCES, 2020, 514 : 131 - 148
  • [32] PARALLEL MATRIX FACTORIZATION FOR LOW-RANK TENSOR COMPLETION
    Xu, Yangyang
    Hao, Ruru
    Yin, Wotao
    Su, Zhixun
    INVERSE PROBLEMS AND IMAGING, 2015, 9 (02) : 601 - 624
  • [33] Speeding up Distributed Low-rank Matrix Factorization
    Qin, Chengjie
    Rusu, Florin
    2013 INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND BIG DATA (CLOUDCOM-ASIA), 2013, : 521 - 528
  • [34] Accurate and fast matrix factorization for low-rank learning
    Godaz, Reza
    Monsefi, Reza
    Toutounian, Faezeh
    Hosseini, Reshad
    JOURNAL OF MATHEMATICAL MODELING, 2022, 10 (02): : 263 - 278
  • [35] SNIG PROPERTY OF MATRIX LOW-RANK FACTORIZATION MODEL
    Wang, Hong
    Liu, Xin
    Chen, Xiaojun
    Yuan, Ya-xiang
    JOURNAL OF COMPUTATIONAL MATHEMATICS, 2018, 36 (03) : 374 - 390
  • [36] Structured Low-Rank Matrix Factorization for Haplotype Assembly
    Cai, Changxiao
    Sanghavi, Sujay
    Vikalo, Haris
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2016, 10 (04) : 647 - 657
  • [37] Cluster structure augmented deep nonnegative matrix factorization with low-rank tensor learning
    Zhong, Bo
    Wu, Jian-Sheng
    Huang, Wei
    Zheng, Wei-Shi
    INFORMATION SCIENCES, 2024, 670
  • [38] Projection Methods for Dynamical Low-Rank Approximation of High-Dimensional Problems
    Kieri, Emil
    Vandereycken, Bart
    COMPUTATIONAL METHODS IN APPLIED MATHEMATICS, 2019, 19 (01) : 73 - 92
  • [39] High-Dimensional Multivariate Forecasting with Low-Rank Gaussian Copula Processes
    Salinas, David
    Bohlke-Schneider, Michael
    Callot, Laurent
    Medico, Roberto
    Gasthaus, Jan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [40] ESTIMATION OF (NEAR) LOW-RANK MATRICES WITH NOISE AND HIGH-DIMENSIONAL SCALING
    Negahban, Sahand
    Wainwright, Martin J.
    ANNALS OF STATISTICS, 2011, 39 (02): : 1069 - 1097