Speech-based Human-Robot Interaction Robust to Acoustic Reflections in Real Environment

被引:0
|
作者
Gomez, Randy [1 ]
Inoue, Koji
Nakamura, Keisuke [1 ]
Mizumoto, Takeshi [1 ]
Nakadai, Kazuhiro [1 ]
机构
[1] Honda Res Inst Japan Ltd Co, Wako, Saitama, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Acoustic reflection inside an enclosed environment is detrimental to human-robot interaction. Reflection may manifest as phantom sources emanating from unknown directions. In effect, a single speaker may falsely manifest as multiple speakers to the robot audition system, impeding the robot's ability to correctly associate the speech command to the actual speaker. Moreover, speech reflection smears the original speech signal due to reverberation. This degrades speech recognition and understanding performance. Conventional robot audition schemes that rely purely on acoustics and spatial information are very sensitive to acoustic reflection which ultimately leads to the failure in human-robot interaction. We propose a method for human-robot interaction robust to the effect of acoustic reflection. First, visual information is utilized and head tracking scheme is employed to reinforce the acoustic information with the visual presence of a prospect user. Second, we employ a model-based sound event identification scheme and scrutinize whether the acoustic information is likely to be speech or non-speech. Using all the information we have gathered, we create a simple rule construct to effectively discriminate the original source (actual speaker) from phantom sources (reflection). Consequently, the corresponding source identified as phantom (reflection) is used to estimate the unwanted smearing for effective suppression via speech enhancement. Experiments are conducted in human-robot interaction setting in which the proposed method outperforms the conventional method.
引用
收藏
页码:1367 / 1373
页数:7
相关论文
共 50 条
  • [21] Intelligent Speech Control System for Human-Robot Interaction
    Liu, Xiaomei
    Ge, Shuzhi Sam
    Jiang, Rui
    Goh, Cher-Hiang
    PROCEEDINGS OF THE 35TH CHINESE CONTROL CONFERENCE 2016, 2016, : 6154 - 6159
  • [22] Object recognition through human-robot interaction by speech
    Kurnia, R
    Hossain, A
    Nakamura, A
    Kuno, Y
    RO-MAN 2004: 13TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, PROCEEDINGS, 2004, : 619 - 624
  • [23] Unified Understanding of Environment, Task, and Human for Human-Robot Interaction in Real-World Environments
    Yano, Yuga
    Mizutani, Akinobu
    Fukuda, Yukiya
    Kanaoka, Daiju
    Ono, Tomohiro
    Tamukoh, Hakaru
    2024 33RD IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, ROMAN 2024, 2024, : 224 - 230
  • [24] Paralinguistic Cues in Speech to Adapt Robot Behavior in Human-Robot Interaction
    Ashok, Ashita
    Pawlak, Jakub
    Paplu, Sarwar
    Zafar, Zuhair
    Berns, Karsten
    2022 9TH IEEE RAS/EMBS INTERNATIONAL CONFERENCE ON BIOMEDICAL ROBOTICS AND BIOMECHATRONICS (BIOROB 2022), 2022,
  • [25] Robust Motion Detection and Tracking for Human-Robot Interaction
    Martinez-Martin, Ester
    del Pobil, Angel P.
    COMPANION OF THE 2017 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI'17), 2017, : 401 - 402
  • [26] Human-Robot Interaction Based on Biosignals
    Liu, Yang
    Yang, Chenguang
    Wang, Min
    2020 INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS), 2020, : 58 - 63
  • [27] Speech-Based Human and Service Robot Interaction: An Application for Mexican Dysarthric People Regular Paper
    Caballero Morales, Santiago Omar
    Bonilla Enriquez, Gladys
    Trujillo Romero, Felipe
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2013, 10
  • [28] Robust real-time hand detection and localization for space human-robot interaction based on deep learning
    Gao, Qing
    Liu, Jinguo
    Ju, Zhaojie
    NEUROCOMPUTING, 2020, 390 : 198 - 206
  • [29] Development of a Robot Quizmaster with Auditory Functions for Speech-based Multiparty Interaction
    Nishimuta, Izaya
    Yoshii, Kazuyoshi
    Itoyama, Katsutoshi
    Okuno, Hiroshi G.
    2014 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2014, : 328 - 333
  • [30] Immersion type virtual environment for human-robot interaction
    Odashima, T
    Onishi, M
    Luo, ZW
    Hosoe, S
    2003 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-5, CONFERENCE PROCEEDINGS, 2003, : 3651 - 3656