Environment Sound Event Classification With a Two-Stream Convolutional Neural Network

被引:26
|
作者
Dong, Xifeng [1 ]
Yin, Bo [1 ,2 ]
Cong, Yanping [1 ]
Du, Zehua [1 ]
Huang, Xianqing [1 ]
机构
[1] Ocean Univ China, Sch Informat Sci & Engn, Qingdao 266100, Peoples R China
[2] Pilot Natl Lab Marine Sci & Technol, Qingdao 266237, Peoples R China
基金
中国国家自然科学基金;
关键词
Environmental sound classification; sound recognition; convolutional neural networks; data processing; pre-emphasis; two stream model; RECOGNITION; REPRESENTATIONS;
D O I
10.1109/ACCESS.2020.3007906
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, with the construction of intelligent cities, the importance of environmental sound classification (ESC) research has become increasingly prominent. However, due to the non-stationary nature of environment sound and the strong interference of ambient noise, the recognition accuracy of ESC is not high enough. Even with deep learning methods, it is difficult to fully extract features from models with a single input. Aiming to improve the recognition accuracy of ESC, this paper proposes a two-stream convolutional neural network (CNN) based on raw audio CNN (RACNN) and logmel CNN (LMCNN). In this method, a pre-emphasis module is first constructed to deal with raw audio signal. The processed audio data and logmel data are imported into RACNN and LMCNN, respectively to obtain both of time and frequency features of audio. In addition, a random-padding method is proposed to patch shorter data sequences. In such a way, the available data for experiment are greatly increased. Finally, the effectiveness of the methods has been verified based on UrbanSound8K dataset in experimental part.
引用
收藏
页码:125714 / 125721
页数:8
相关论文
共 50 条
  • [1] Two-Stream Convolutional Neural Network for Multimodal Matching
    Zhang, Youcai
    Gu, Yiwei
    Gu, Xiaodong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 14 - 21
  • [2] Deep Convolutional Neural Network Based on Two-Stream Convolutional Unit
    Hou Congcong
    He Yuqing
    Jiang Xiaoheng
    Pan Jing
    LASER & OPTOELECTRONICS PROGRESS, 2018, 55 (02)
  • [3] Two-Stream Convolutional Neural Network for Video Action Recognition
    Qiao, Han
    Liu, Shuang
    Xu, Qingzhen
    Liu, Shouqiang
    Yang, Wanggan
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2021, 15 (10): : 3668 - 3684
  • [4] Stage classification using two-stream deep convolutional neural networks
    Alexander Chefranov
    Altaf Khan
    Hasan Demirel
    Signal, Image and Video Processing, 2022, 16 : 311 - 319
  • [5] Research on remote sensing image classification method using two-stream convolutional neural network
    Peng K.
    Hu J.
    Liu S.
    Qu F.
    Yang H.
    Chen J.
    International Journal of Wireless and Mobile Computing, 2022, 23 (3-4) : 250 - 255
  • [6] Transformer-enhanced two-stream complementary convolutional neural network for hyperspectral image classification
    Pan, Haizhu
    Yan, Hui
    Ge, Haimiao
    Liu, Moqi
    Shi, Cuiping
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (12):
  • [7] A Refined Non-Driving Activity Classification Using a Two-Stream Convolutional Neural Network
    Yang, Lichao
    Yang, Ting-Yu
    Liu, Haochen
    Shan, Xiaocai
    Brighton, James
    Skrypchuk, Lee
    Mouzakitis, Alexandros
    Zhao, Yifan
    IEEE SENSORS JOURNAL, 2021, 21 (14) : 15574 - 15583
  • [8] Deep Feature Fusion via Two-Stream Convolutional Neural Network for Hyperspectral Image Classification
    Li, Xian
    Ding, Mingli
    Pizurica, Aleksandra
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2020, 58 (04): : 2615 - 2629
  • [9] Stage classification using two-stream deep convolutional neural networks
    Chefranov, Alexander
    Khan, Altaf
    Demirel, Hasan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2022, 16 (02) : 311 - 319
  • [10] Target Classification in Unattended Ground Sensors With a Two-Stream Convolutional Network
    Qian, Yanling
    Tang, Hongying
    Ran, Yue
    Li, Baoqing
    IEEE SENSORS JOURNAL, 2023, 23 (04) : 3747 - 3755