Self-Supervised Learning and Multi-Task Pre-Training Based Single-Channel Acoustic Denoising

被引:1
|
作者
Li, Yi [1 ]
Sun, Yang [2 ]
Naqvi, Syed Mohsen [1 ]
机构
[1] Newcastle Univ, Sch Engn, Intelligent Sensing & Commun Grp, Newcastle Upon Tyne NE1 7RU, England
[2] Univ Oxford, Big Data Inst, Oxford OX3 7LF, England
关键词
MONAURAL SOURCE SEPARATION; SPEECH; ENVIRONMENTS;
D O I
10.1109/MFI55806.2022.9913855
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In self-supervised learning-based single-channel speech denoising problem, it is challenging to reduce the gap between the denoising performance on the estimated and target speech signals with existed pre-tasks. In this paper, we propose a multi-task pre-training method to improve the speech denoising performance within self-supervised learning. In the proposed pre-training autoencoder (PAE), only a very limited set of unpaired and unseen clean speech signals are required to learn speech latent representations. Meanwhile, to solve the limitation of existing single pre-task, the proposed masking module exploits the dereverberated mask and estimated ratio mask to denoise the mixture as the new pre-task. The downstream task autoencoder (DAE) utilizes unlabeled and unseen reverberant mixtures to generate the estimated mixtures. The DAE is trained to share a latent representation with the clean examples from the learned representation in the PAE. Experimental results on a benchmark dataset demonstrate that the proposed method outperforms the state-of-the-art approaches.
引用
收藏
页数:5
相关论文
共 50 条
  • [41] SELF-SUPERVISED MULTI-TASK LEARNING FOR SEMANTIC SEGMENTATION OF URBAN SCENES
    Santiago, Jonathan Gonzalez
    Schenkel, Fabian
    Middelmann, Wolfgang
    IMAGE AND SIGNAL PROCESSING FOR REMOTE SENSING XXVII, 2021, 11862
  • [42] Self-supervised Heterogeneous Graph Pre-training Based on Structural Clustering
    Yang, Yaming
    Guan, Ziyu
    Wang, Zhe
    Zhao, Wei
    Xu, Cai
    Lu, Weigang
    Huang, Jianbin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [43] Multi-task self-supervised time-series representation learning
    Choi, Heejeong
    Kang, Pilsung
    INFORMATION SCIENCES, 2024, 671
  • [44] The Effectiveness of Self-supervised Pre-training for Multi-modal Endometriosis Classification
    Butler, David
    Wang, Hu
    Zhang, Yuan
    To, Minh-Son
    Condous, George
    Leonardi, Mathew
    Knox, Steven
    Avery, Jodie
    Hull, M. Louise
    Carneiro, Gustavo
    2023 45TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY, EMBC, 2023,
  • [45] Reducing Barriers to Self-Supervised Learning: HuBERT Pre-training with Academic Compute
    Chen, William
    Chang, Xuankai
    Peng, Yifan
    Ni, Zhaoheng
    Maiti, Soumi
    Watanabe, Shinji
    INTERSPEECH 2023, 2023, : 4404 - 4408
  • [46] A Stacked Denoising Autoencoder Based on Supervised Pre-training
    Wang, Xiumei
    Mu, Shaomin
    Shi, Aiju
    Lin, Zhongqi
    SMART INNOVATIONS IN COMMUNICATION AND COMPUTATIONAL SCIENCES, VOL 2, 2019, 670 : 139 - 146
  • [47] Pre-training Multi-task Contrastive Learning Models for Scientific Literature Understanding
    Zhang, Yu
    Cheng, Hao
    Shen, Zhihong
    Liu, Xiaodong
    Wang, Ye-Yi
    Gao, Jianfeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12259 - 12275
  • [48] ConvMTL: Multi-task Learning via Self-supervised Learning for Simultaneous Dense Predictions
    Iyer, Vijayasri
    Thangavel, Senthil Kumar
    Nalluri, Madhusudana Rao
    Chang, Maiga
    COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT I, 2024, 2009 : 455 - 466
  • [49] A Multi-Task Dense Network with Self-Supervised Learning for Retinal Vessel Segmentation
    Tu, Zhonghao
    Zhou, Qian
    Zou, Hua
    Zhang, Xuedong
    ELECTRONICS, 2022, 11 (21)
  • [50] GMSS: Graph-Based Multi-Task Self-Supervised Learning for EEG Emotion Recognition
    Li, Yang
    Chen, Ji
    Li, Fu
    Fu, Boxun
    Wu, Hao
    Ji, Youshuo
    Zhou, Yijin
    Niu, Yi
    Shi, Guangming
    Zheng, Wenming
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2512 - 2525