DEEP RECURRENT MIXTURE OF EXPERTS FOR SPEECH ENHANCEMENT

被引:0
|
作者
Chazan, Shlomo E. [1 ]
Goldberger, Jacob [1 ]
Gannot, Sharon [1 ]
机构
[1] Bar Ilan Univ, Fac Engn, Ramat Gan, Israel
关键词
recurrent neural network; long short-term memory; speech presence probability; RECOGNITION; NOISE;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
deep neural networks (DNNs) have recently became a viable methodology for single microphone speech enhancement. The most common approach, is to feed the noisy speech features into a fully-connected DNN to either directly enhance the speech signal or to infer a mask which can be used for the speech enhancement. In this case, one network has to deal with the large variability of the speech signal. Most approaches also discard the speech continuity. In this paper, we propose a deep recurrent mixture of experts (DRMoE) architecture that addresses these two issues. In order to reduce the large speech variability, we split the network into a mixture of networks (denoted experts), each of which specializes in a specific and simpler task and a gating network. The time-continuity of the speech signal is taken into account by implementing the experts and the gating network as a recurrent neural network (RNN). Experimental study shows that the proposed algorithm produces higher objective measurements scores compared to both a single RNN and a deep mixture of experts (DMoE) architectures.
引用
收藏
页码:359 / 363
页数:5
相关论文
共 50 条
  • [1] Speech Enhancement Based on Deep Mixture of Distinguishing Experts
    Jia, Xupeng
    Li, Dongmei
    2019 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2019, : 684 - 689
  • [2] Speech enhancement using deep mixture of experts based on hard expectation maximization
    Karjol, Pavan
    Ghosh, Prasanta Kumar
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3254 - 3258
  • [3] SPEECH ENHANCEMENT WITH MIXTURE OF DEEP EXPERTS WITH CLEAN CLUSTERING PRE-TRAINING
    Chazan, Shlomo E.
    Goldberger, Jacob
    Gannot, Sharon
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 716 - 720
  • [4] Sparse Mixture of Local Experts for Efficient Speech Enhancement
    Sivaraman, Aswin
    Kim, Minje
    INTERSPEECH 2020, 2020, : 4526 - 4530
  • [5] Aphasic Speech Recognition using a Mixture of Speech Intelligibility Experts
    Perez, Matthew
    Aldeneh, Zakaria
    Provost, Emily Mower
    INTERSPEECH 2020, 2020, : 4986 - 4990
  • [6] MIXTURE OF INFORMED EXPERTS FOR MULTILINGUAL SPEECH RECOGNITION
    Gaur, Neeraj
    Farris, Brian
    Haghani, Parisa
    Leal, Isabel
    Moreno, Pedro J.
    Prasad, Manasa
    Ramabhadran, Bhuvana
    Zhu, Yun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6234 - 6238
  • [7] Speech Enhancement for Speaker Recognition Using Deep Recurrent Neural Networks
    Tkachenko, Maxim
    Yamshinin, Alexander
    Lyubimov, Nikolay
    Kotov, Mikhail
    Nastasenko, Marina
    SPEECH AND COMPUTER, SPECOM 2017, 2017, 10458 : 690 - 699
  • [8] Deep Mixture of Experts via Shallow Embedding
    Wang, Xin
    Yu, Fisher
    Dunlap, Lisa
    Ma, Yi-An
    Wang, Ruth
    Mirhoseini, Azalia
    Darrell, Trevor
    Gonzalez, Joseph E.
    35TH UNCERTAINTY IN ARTIFICIAL INTELLIGENCE CONFERENCE (UAI 2019), 2020, 115 : 552 - 562
  • [9] Deep Mixture of Experts with Diverse Task Spaces
    Fan, Jianping
    Zhao, Tianyi
    Kuang, Zhenzhong
    Yu, Zhou
    Yu, Jun
    2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 721 - 725
  • [10] A mixture of experts image enhancement scheme for CCTV images
    Singh, M
    Singh, S
    Porter, M
    INTELLIGENT DAA ENGINEERING AND AUTOMATED LEARNING IDEAL 2004, PROCEEDINGS, 2004, 3177 : 185 - 190