A Dynamic Head Gesture Recognition Method for Real-time Intention Inference and Its Application to Visual Human-robot Interaction

被引:1
|
作者
Xie, Jialong [1 ]
Zhang, Botao [1 ]
Lu, Qiang [1 ]
Borisov, Oleg [2 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou, Peoples R China
[2] ITMO Univ, Fac Control Syst & Robot, St Petersburg, Russia
基金
中国国家自然科学基金;
关键词
Computer vision; deep learning; head gesture; human-robot interaction; MOTION;
D O I
10.1007/s12555-022-0051-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Head gesture is a natural and non-verbal communication method for human-computer and human-robot interaction, conveying attitudes and intentions. However, the existing vision-based recognition methods cannot meet the precision and robustness of interaction requirements. Due to the limited computational resources, applying most high-accuracy methods to mobile and onboard devices is challenging. Moreover, the wearable device-based approach is inconvenient and expensive. To deal with these problems, an end-to-end two-stream fusion network named TSIR3D is proposed to identify head gestures from videos for analyzing human attitudes and intentions. Inspired by Inception and ResNet architecture, the width and depth of the network are increased to capture motion features sufficiently. Meanwhile, convolutional kernels are expanded from the spatial domain to the spatiotemporal domain for temporal feature extraction. The fusion position of the two-stream channel is explored under an accuracy/complexity trade-off to a certain extent. Furthermore, a dynamic head gesture dataset named DHG and a behavior tree are designed for human-robot interaction. Experimental results show that the proposed method has advantages in real-time performance on the remote server or the onboard computer. Furthermore, its accuracy on the DHG can surpass most state-of-the-art vision-based methods and is even better than most previous approaches based on head-mounted sensors. Finally, TSIR3D is applied on Pepper Robot equipped with Jetson TX2.
引用
收藏
页码:252 / 264
页数:13
相关论文
共 50 条
  • [31] A descriptive behavior intention inference framework using spatio-temporal semantic features for human-robot real-time interaction
    Wang, Liangliang
    Huo, Guanglei
    Li, Ruifeng
    Liang, Peidong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 128
  • [32] Real-time Framework for Multimodal Human-Robot Interaction
    Gast, Juergen
    Bannat, Alexander
    Rehrl, Tobias
    Wallhoff, Frank
    Rigoll, Gerhard
    Wendt, Cornelia
    Schmidt, Sabrina
    Popp, Michael
    Faerber, Berthold
    HSI: 2009 2ND CONFERENCE ON HUMAN SYSTEM INTERACTIONS, 2009, : 273 - 280
  • [33] Human-robot interaction based on gesture and movement recognition
    Li, Xing
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 81
  • [34] Continuous Gesture Recognition for Flexible Human-Robot Interaction
    Iengo, Salvatore
    Rossi, Silvia
    Staffa, Mariacarla
    Finzi, Alberto
    2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2014, : 4863 - 4868
  • [35] Efficient Hand Gesture Recognition for Human-Robot Interaction
    Peral, Marc
    Sanfeliu, Alberto
    Garrell, Anais
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 10272 - 10279
  • [36] Upper Body Gesture Recognition for Human-Robot Interaction
    Oh, Chi-Min
    Islam, Md Zahidul
    Lee, Jun-Sung
    Lee, Chil-Woo
    Kweon, In-So
    HUMAN-COMPUTER INTERACTION: INTERACTION TECHNIQUES AND ENVIRONMENTS, PT II, 2011, 6762 : 294 - 303
  • [37] Real-time human motion analysis for human-robot interaction
    Molina-Tanco, L
    Bandera, JP
    Marfil, R
    Sandoval, F
    2005 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4, 2005, : 1808 - 1813
  • [38] Automatic gesture recognition for intelligent human-robot interaction
    Lee, Seong-Whan
    PROCEEDINGS OF THE SEVENTH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION - PROCEEDINGS OF THE SEVENTH INTERNATIONAL CONFERENCE, 2006, : 645 - 650
  • [39] Human-robot interaction using facial gesture recognition
    Zelinsky, A
    Heinzmann, J
    RO-MAN '96 - 5TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS, 1996, : 256 - 261
  • [40] Multiple people gesture recognition for human-robot interaction
    Hong, Seok-ju
    Setiawan, Nurul Arif
    Lee, Chil-woo
    HUMAN-COMPUTER INTERACTION, PT 3, PROCEEDINGS, 2007, 4552 : 625 - +