Direction Estimation of Instrumental Sound Sources Using Regression Analysis by Convolutional Neural Network

被引:0
|
作者
Yamamoto, Kaho [1 ]
Ogihara, Akio [2 ]
Murata, Harumi [3 ]
机构
[1] Shonan Inst Technol, Fac Informat, 1-1-25 Tsujido Nishikaigan, Fujisawa, Kanagawa 2518511, Japan
[2] Kindai Univ, Fac Engn, 1 Takaya Umenobe, Higashihiroshima, Hiroshima 7392116, Japan
[3] Chukyo Univ, Sch Engn, 101 Tokodachi,Kaizu Cho, Toyota 4700393, Japan
关键词
Direction estimation of sound source; Instrument sound; Overtone structure; MUSIC spectrum; CNN; Regression analysis;
D O I
10.1007/s00034-023-02433-z
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
There has been much research on estimating noise and speech source direction, but there have not been many studies on estimating the source direction of instrumental sound sources. In this study, we considered the source direction estimation of a single instrumental sound. Direction estimation of sound sources by the multiple signal classification (MUSIC) method often causes large estimation errors. Then, we propose a technique for estimating the direction of musical instrument sound sources by applying regression analysis using a convolutional neural network (CNN), a type of neural network. We calculated the MUSIC spectrum obtained using MUSIC that uses the fundamental and harmonic components, which have relatively large amplitudes, and we estimated the direction of the sound source using the CNN with these components as input. We achieved this by focusing on the overtone structure of the instrumental sound source. This study demonstrated the effectiveness of this method using simulations in a monaural environment.
引用
收藏
页码:7004 / 7021
页数:18
相关论文
共 50 条
  • [1] Direction Estimation of Instrumental Sound Sources Using Regression Analysis by Convolutional Neural Network
    Kaho Yamamoto
    Akio Ogihara
    Harumi Murata
    Circuits, Systems, and Signal Processing, 2023, 42 : 7004 - 7021
  • [2] Direction Estimation of Instrumental Sound Sources Using Convolutional Neural Network Classification
    Yamamoto, Kaho
    Murata, Harumi
    Ogihara, Akio
    2024 INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS, AND COMMUNICATIONS, ITC-CSCC 2024, 2024,
  • [3] Direction of Arrival Estimation for Multiple Sound Sources Using Convolutional Recurrent Neural Network
    Adavanne, Sharath
    Politis, Archontis
    Virtanen, Tuomas
    2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 1462 - 1466
  • [4] Sound Source Localization Using a Convolutional Neural Network and Regression Model
    Tan, Tan-Hsu
    Lin, Yu-Tang
    Chang, Yang-Lang
    Alkhaleefah, Mohammad
    SENSORS, 2021, 21 (23)
  • [5] Separation of Multiple Stationary Sound Sources using convolutional neural network
    Mali, Swapnil G.
    Dhale, Mohit, V
    Mahajan, Shriniwas P.
    2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,
  • [6] Estimation of Pipe Wall Thinning Using a Convolutional Neural Network for Regression
    Kim, Jonghwan
    Jung, Byunyoung
    Park, Junhong
    Choi, Youngchul
    NUCLEAR TECHNOLOGY, 2022, 208 (07) : 1184 - 1191
  • [7] Object Direction Estimation by Constrained Convolutional Neural Network
    Liu, Jin
    Luo, Chuxuan
    Gao, Fang
    MIPPR 2019: PATTERN RECOGNITION AND COMPUTER VISION, 2020, 11430
  • [8] Estimation of Sound Source Direction of Arrival Map Using Convolutional Neural Network and Cross-Correlation in Frequency Bands
    Sakavicius, Saulius
    Serackis, Arturas
    2019 OPEN CONFERENCE OF ELECTRICAL, ELECTRONIC AND INFORMATION SCIENCES (ESTREAM), 2019,
  • [9] Direction Estimation of Sound Source Using Neural Network in Consideration of the Overtone Structure of Instrument Sound
    Yamamoto, Kaho
    Ogihara, Akio
    Murata, Harumi
    2021 36TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC), 2021,
  • [10] Broadband Direction of Arrival Estimation Based on Convolutional Neural Network
    Zhu, Wenli
    Zhang, Min
    Wu, Chenxi
    Zeng, Lingqing
    IEICE TRANSACTIONS ON COMMUNICATIONS, 2020, E103B (03) : 148 - 154