Deep Learning-Based Portable Device for Audio Distress Signal Recognition in Urban Areas

被引:6
|
作者
Felipe Gaviria, Jorge [1 ]
Escalante-Perez, Alejandra [1 ]
Camilo Castiblanco, Juan [1 ]
Vergara, Nicolas [1 ]
Parra-Garces, Valentina [1 ]
David Serrano, Juan [1 ]
Felipe Zambrano, Andres [1 ]
Felipe Giraldo, Luis [1 ]
机构
[1] Univ Los Andes, Dept Elect & Elect Engn, Bogota 111711, DC, Colombia
来源
APPLIED SCIENCES-BASEL | 2020年 / 10卷 / 21期
关键词
acoustic signal processing; smart cities; convolutional neural network; raspberry Pi; deep learning;
D O I
10.3390/app10217448
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Real-time automatic identification of audio distress signals in urban areas is a task that in a smart city can improve response times in emergency alert systems. The main challenge in this problem lies in finding a model that is able to accurately recognize these type of signals in the presence of background noise and allows for real-time processing. In this paper, we present the design of a portable and low-cost device for accurate audio distress signal recognition in real urban scenarios based on deep learning models. As real audio distress recordings in urban areas have not been collected and made publicly available so far, we first constructed a database where audios were recorded in urban areas using a low-cost microphone. Using this database, we trained a deep multi-headed 2D convolutional neural network that processed temporal and frequency features to accurately recognize audio distress signals in noisy environments with a significant performance improvement to other methods from the literature. Then, we deployed and assessed the trained convolutional neural network model on a Raspberry Pi that, along with the low-cost microphone, constituted a device for accurate real-time audio recognition. Source code and database are publicly available. Dataset: https://github.com/jfgf11/Problema-Especial.git
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [21] Deep Learning for Audio Signal Processing
    Purwins, Hendrik
    Li, Bo
    Virtanen, Tuomas
    Schlueter, Jan
    Chang, Shuo-Yiin
    Sainath, Tara
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2019, 13 (02) : 206 - 219
  • [22] True Detect: Deep Learning-based Device-Free Activity Recognition using WiFi
    Sulaiman, Muhammad
    Hassan, Syed Ali
    Jung, Haejoon
    2020 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE WORKSHOPS (WCNCW), 2020,
  • [23] Deep learning-based signal detection in OFDM systems
    Chang D.
    Zhou J.
    Dongnan Daxue Xuebao (Ziran Kexue Ban)/Journal of Southeast University (Natural Science Edition), 2020, 50 (05): : 912 - 917
  • [24] Deep Reinforcement Learning-based Traffic Signal Control
    Ruan, Junyun
    Tang, Jinzhuo
    Gao, Ge
    Shi, Tianyu
    Khamis, Alaa
    2023 IEEE INTERNATIONAL CONFERENCE ON SMART MOBILITY, SM, 2023, : 21 - 26
  • [25] A Survey on Deep Learning-Based Traffic Signal Control
    Si, Qinbatu
    Yang, Lirun
    Bao, Jingjing
    Lin, Yangfei
    Bao, Wugedele
    Wu, Celimuge
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2025, 34 (02)
  • [26] Deep Learning-Based Resource Allocation for Device-to-Device Communication
    Lee, Woongsup
    Schober, Robert
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (07) : 5235 - 5250
  • [27] Metric Learning-Based Multimodal Audio-Visual Emotion Recognition
    Ghaleb, Esam
    Popa, Mirela
    Asteriadis, Stylianos
    IEEE MULTIMEDIA, 2020, 27 (01) : 37 - 48
  • [28] Audio signal based danger detection using signal processing and deep learning
    Fime, Awal Ahmed
    Ashikuzzaman, Md.
    Aziz, Abdul
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 237
  • [29] Sample Balancing for Deep Learning-Based Visual Recognition
    Chen, Xin
    Weng, Jian
    Luo, Weiqi
    Lu, Wei
    Wu, Huimin
    Xu, Jiaming
    Tian, Qi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 3962 - 3976
  • [30] Deep learning-based image recognition for autonomous driving
    Fujiyoshi, Hironobu
    Hirakawa, Tsubasa
    Yamashita, Takayoshi
    IATSS RESEARCH, 2019, 43 (04) : 244 - 252