Monaural Source Separation in Complex Domain With Long Short-Term Memory Neural Network

被引:29
|
作者
Sun, Yang [1 ]
Xian, Yang [1 ]
Wang, Wenwu [2 ]
Naqvi, Syed Mohsen [1 ]
机构
[1] Newcastle Univ, Sch Engn, Intelligent Sensing & Commun Res Grp, Newcastle Upon Tyne NE1 7RU, Tyne & Wear, England
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Dept Elect & Elect Engn, Surrey GU2 7XH, England
关键词
Deep neural networks; monaural speech separation; long short-term memory; complex signal approximation; SPEECH DEREVERBERATION; MASKING; RECOGNITION; FEATURES; NOISE;
D O I
10.1109/JSTSP.2019.2908760
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In recent research, deep neural network (DNN) has been used to solve the monaural source separation problem. According to the training objectives, DNN-based monaural speech separation is categorized into three aspects, namely masking, mapping, and signal approximation based techniques. However, the performance of the traditional methods is not robust due to variations in real-world environments. Besides, in the vanilla DNN-based methods, the temporal information cannot be fully utilized. Therefore, in this paper, the long short-term memory (LSTM) neural network is applied to exploit the long-term speech contexts. Then, we propose the complex signal approximation (cSA), which is operated in the complex domain to utilize the phase information of the desired speech signal to improve the separation performance. The IEEE and the TIMIT corpora are used to generate mixtures with noise and speech interferences to evaluate the efficacy of the proposed method. The experimental results demonstrate the advantages of the proposed cSA-based LSTM recurrent neural network method in terms of different objective performance measures.
引用
收藏
页码:359 / 369
页数:11
相关论文
共 50 条
  • [21] Complex network structure influences processing in long-term and short-term memory
    Vitevitch, Michael S.
    Chan, Kit Ying
    Roodenrys, Steven
    JOURNAL OF MEMORY AND LANGUAGE, 2012, 67 (01) : 30 - 44
  • [22] A hybrid convolutional neural network with long short-term memory for statistical arbitrage
    Eggebrecht, P.
    Luetkebohmert, E.
    QUANTITATIVE FINANCE, 2023, 23 (04) : 595 - 613
  • [23] Predicting Playa Inundation Using a Long Short-Term Memory Neural Network
    Solvik, Kylen
    Bartuszevige, Anne M.
    Bogaerts, Meghan
    Joseph, Maxwell B.
    WATER RESOURCES RESEARCH, 2021, 57 (12)
  • [24] Application of Long Short-Term Memory Neural Network to Crack Propagation Prognostics
    Abbasi, Amirhassan
    Nazari, Foad
    Nataraj, C.
    2020 IEEE INTERNATIONAL CONFERENCE ON PROGNOSTICS AND HEALTH MANAGEMENT (ICPHM), 2020,
  • [25] Microgrid Equivalent Modeling Based on Long Short-Term Memory Neural Network
    Cai, Changchun
    Liu, Haolin
    Tao, Yuan
    Deng, Zhixiang
    Dai, Weil
    Chen, Jie
    IEEE ACCESS, 2020, 8 : 23120 - 23133
  • [26] Long Short-Term Memory Bayesian Neural Network for Air Pollution Forecast
    Nguyen, Huynh A. D.
    Ha, Quang P.
    Duc, Hiep
    Azzi, Merched
    Jiang, Ningbo
    Barthelemy, Xavier
    Riley, Matthew
    IEEE ACCESS, 2023, 11 : 35710 - 35725
  • [27] A COMPACT AND CONFIGURABLE LONG SHORT-TERM MEMORY NEURAL NETWORK HARDWARE ARCHITECTURE
    Chen, Kewei
    Huang, Leilei
    Li, Minjiang
    Zeng, Xiaoyang
    Fan, Yibo
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 4168 - 4172
  • [28] Chinese Lyrics Generation Using Long Short-Term Memory Neural Network
    Wu, Xing
    Du, Zhikang
    Zhong, Mingyu
    Dai, Shuji
    Liu, Yazhou
    ADVANCES IN ARTIFICIAL INTELLIGENCE: FROM THEORY TO PRACTICE (IEA/AIE 2017), PT II, 2017, 10351 : 419 - 427
  • [29] On extended long short-term memory and dependent bidirectional recurrent neural network
    Su, Yuanhang
    Kuo, C-C Jay
    NEUROCOMPUTING, 2019, 356 : 151 - 161
  • [30] Applying Long Short-Term Memory Recurrent Neural Network for Intrusion Detection
    Althubiti, Sara
    Nick, William
    Mason, Janelle
    Yuan, Xiaohong
    Esterline, Albert
    IEEE SOUTHEASTCON 2018, 2018,