Speech-based Human-Robot Interaction Robust to Acoustic Reflections in Real Environment

被引:0
|
作者
Gomez, Randy [1 ]
Inoue, Koji
Nakamura, Keisuke [1 ]
Mizumoto, Takeshi [1 ]
Nakadai, Kazuhiro [1 ]
机构
[1] Honda Res Inst Japan Ltd Co, Wako, Saitama, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Acoustic reflection inside an enclosed environment is detrimental to human-robot interaction. Reflection may manifest as phantom sources emanating from unknown directions. In effect, a single speaker may falsely manifest as multiple speakers to the robot audition system, impeding the robot's ability to correctly associate the speech command to the actual speaker. Moreover, speech reflection smears the original speech signal due to reverberation. This degrades speech recognition and understanding performance. Conventional robot audition schemes that rely purely on acoustics and spatial information are very sensitive to acoustic reflection which ultimately leads to the failure in human-robot interaction. We propose a method for human-robot interaction robust to the effect of acoustic reflection. First, visual information is utilized and head tracking scheme is employed to reinforce the acoustic information with the visual presence of a prospect user. Second, we employ a model-based sound event identification scheme and scrutinize whether the acoustic information is likely to be speech or non-speech. Using all the information we have gathered, we create a simple rule construct to effectively discriminate the original source (actual speaker) from phantom sources (reflection). Consequently, the corresponding source identified as phantom (reflection) is used to estimate the unwanted smearing for effective suppression via speech enhancement. Experiments are conducted in human-robot interaction setting in which the proposed method outperforms the conventional method.
引用
收藏
页码:1367 / 1373
页数:7
相关论文
共 50 条
  • [31] Human-robot interaction in the home ubiquitous network environment
    Ueda, Hirotada
    Minoh, Michihiko
    Chikama, Masaki
    Satake, Junji
    Kobayashi, Akihiro
    Miyawaki, Kenzabro
    Kidode, Masatsugu
    HUMAN-COMPUTER INTERACTION, PT 2, PROCEEDINGS, 2007, 4551 : 990 - +
  • [32] Safe human-robot interaction in a life science environment
    Fritzsche, Markus
    Schulenburg, Erik
    Elkmann, Norbert
    Girstl, Angelika
    Stiene, Stefan
    Teutsch, Christian
    2007 IEEE INTERNATIONAL WORKSHOP ON SAFETY, SECURITY AND RESCUE ROBOTICS, 2007, : 156 - +
  • [33] Safety strategies for human-robot interaction in surgical environment
    Park, Shinsuk
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 4681 - 4685
  • [34] Real-time safety for human-robot interaction
    Kulic, D
    Croft, EA
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2006, 54 (01) : 1 - 12
  • [35] Real-time safety for human-robot interaction
    Kulic, D
    Croft, EA
    2005 12TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS, 2005, : 719 - 724
  • [36] Real-time acoustic source localization in noisy environments for human-robot multimodal interaction
    Trifa, Vlad M.
    Koene, Ansgar
    Moren, Jan
    Cheng, Gordon
    2007 RO-MAN: 16TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1-3, 2007, : 392 - +
  • [37] Multimodal Human-Robot Interaction from the Perspective of a Speech Scientist
    Rigoll, Gerhard
    SPEECH AND COMPUTER (SPECOM 2015), 2015, 9319 : 3 - 10
  • [38] Robust Speech-Based Happiness Recognition
    Lin, Chang-Hong
    Siahaan, Ernestasia
    Chin, Yu-Hau
    Chen, Bo-Wei
    Wang, Jia-Ching
    Wang, Jhing-Fa
    1ST INTERNATIONAL CONFERENCE ON ORANGE TECHNOLOGIES (ICOT 2013), 2013, : 227 - 230
  • [39] Child Speech Recognition in Human-Robot Interaction: Evaluations and Recommendations
    Kennedy, James
    Lemaignan, Severin
    Montassier, Caroline
    Lavalade, Pauline
    Irfan, Bahar
    Papadopoulos, Fotios
    Senft, Emmanuel
    Belpaeme, Tony
    PROCEEDINGS OF THE 2017 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI'17), 2017, : 82 - 90
  • [40] Toward More Expressive Speech Communication in Human-Robot Interaction
    Delic, Vlado
    Borovac, Branislav
    Gnjatovic, Milan
    Tasevski, Jovica
    Miskovic, Dragisa
    Pekar, Darko
    Secujski, Milan
    INTERACTIVE COLLABORATIVE ROBOTICS, ICR 2018, 2018, 11097 : 44 - 51