Vocal control of sound synthesis personalized by unsupervised machine listening and learning

被引:0
|
作者
Fasciani S. [1 ]
Wyse L. [2 ]
机构
[1] Faculty of Engineering and Information Sciences, University of Wollongong in Dubai, P.O. Box 20183, Dubai
[2] Communication and New Media Department, National University of Singapore, 11 Computing Drive, Singapore
关键词
All Open Access; Green;
D O I
10.1162/comj_a_00450
中图分类号
学科分类号
摘要
In this article we describe a user-driven adaptive method to control the sonic response of digital musical instruments using information extracted from the timbre of the human voice. The mapping between heterogeneous attributes of the input and output timbres is determined from data collected through machine-listening techniques and then processed by unsupervised machine-learning algorithms. This approach is based on a minimum-loss mapping that hides any synthesizer-specific parameters and that maps the vocal interaction directly to perceptual characteristics of the generated sound. The mapping adapts to the dynamics detected in the voice and maximizes the timbral space covered by the sound synthesizer. The strategies for mapping vocal control to perceptual timbral features and for automating the customization of vocal interfaces for different users and synthesizers, in general, are evaluated through a variety of qualitative and quantitative methods. © 2018 Massachusetts Institute of Technology.
引用
收藏
页码:37 / 59
页数:22
相关论文
共 50 条
  • [1] Vocal Control of Sound Synthesis Personalized by Unsupervised Machine Listening and Learning
    Fasciani, Stefano
    Wyse, Lonce
    COMPUTER MUSIC JOURNAL, 2018, 42 (01) : 37 - 59
  • [2] Machine Learning and Sound Processing in Vocal Disease Detection
    Costandache, Mihai-Andrei
    COMPUTER SCIENCE JOURNAL OF MOLDOVA, 2024, 32 (02) : 223 - 241
  • [3] Supervised and Unsupervised Sound Retrieval by Vocal Imitation
    Zhang, Yichi
    Duan, Zhiyao
    JOURNAL OF THE AUDIO ENGINEERING SOCIETY, 2016, 64 (7-8): : 533 - 543
  • [4] Unsupervised Anomalous Sound Detection Using Hybrid Machine Learning Techniques
    Yun, Eunsun
    Jeong, Minjoong
    2024 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING, IEEE BIGCOMP 2024, 2024, : 347 - 348
  • [5] IMISOUND: AN UNSUPERVISED SYSTEM FOR SOUND QUERY BY VOCAL IMITATION
    Zhang, Yichi
    Duan, Zhiyao
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2269 - 2273
  • [6] Development of manufacturing control strategies using unsupervised machine learning
    Bowden, R
    Bullington, SF
    IIE TRANSACTIONS, 1996, 28 (04) : 319 - 331
  • [7] Unsupervised Filterbank Learning Using Convolutional Restricted Boltzmann Machine for Environmental Sound Classification
    Sailor, Hardik B.
    Agrawal, Dharmesh M.
    Patil, Hemant A.
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 3107 - 3111
  • [8] Personalized Automated Machine Learning
    Kulbach, Cedric
    Philipp, Patrick
    Thoma, Steffen
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1246 - 1253
  • [9] Sound from silence: On listening in organizational learning
    Jacobs, C
    Coghlan, D
    HUMAN RELATIONS, 2005, 58 (01) : 115 - 138
  • [10] A machine learning approach for personalized autonomous lane change initiation and control
    Vallon, Charlott
    Ercan, Ziya
    Carvalho, Ashwin
    Borrelli, Francesco
    2017 28TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV 2017), 2017, : 1590 - 1595