Human-robot interaction-oriented video understanding of human actions

被引:1
|
作者
Wang, Bin [1 ]
Chang, Faliang [1 ]
Liu, Chunsheng [1 ]
Wang, Wenqian [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, 17923 Jingshi Rd, Jinan 250061, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Action recognition; Human-robot interaction; Temporal modeling; Contextual scene reasoning; NETWORK; FEATURES;
D O I
10.1016/j.engappai.2024.108247
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper focuses on action recognition tasks oriented to the field of human-robot interaction, which is one of the major challenges in the robotic video understanding field. Previous approaches focus on designing temporal models, lack the ability to capture motion information and build contextual correlation models. This may result in robots being unable to effectively understand long-term video actions. To solve these two problems, this paper propose a novel video understanding framework including: an Adaptive Temporal Sensitivity and Motion Capture Network (ATSMC-Net) and a contextual scene reasoning module called Knowledge Function Graph Module (KFG-Module). The proposed ATSMC-Net can adaptively adjust the frame-level and pixel-level sensitive regions of temporal features to effectively capture motion information. To fuse contextual scene information for cross-temporal inference, the KFG-Module is introduced to achieve fine-grained video understanding based on the relationship between objects and actions. We evaluate the method using three public video understanding benchmarks, including Something-Something-V1&V2 and HMDB51. In addition, we present a dataset with real-world application scenarios of human-robot interactions to verify the effectiveness of our approach on mobile robots. The experimental results show that the proposed method can significantly improve the video understanding of the robots.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Social Human-Robot Interaction: A New Cognitive and Affective Interaction-Oriented Architecture
    Adam, Carole
    Johal, Wafa
    Pellier, Damien
    Fiorino, Humbert
    Pesty, Sylvie
    SOCIAL ROBOTICS, (ICSR 2016), 2016, 9979 : 253 - 263
  • [2] Expectations, Intentions, and Actions in Human-Robot Interaction
    Hanheide, Marc
    Lohse, Manja
    Zender, Hendrik
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (02) : 107 - 108
  • [3] Expectations, Intentions, and Actions in Human-Robot Interaction
    Marc Hanheide
    Manja Lohse
    Hendrik Zender
    International Journal of Social Robotics, 2012, 4 : 107 - 108
  • [4] Understanding Human-Robot Interaction in Virtual Reality
    Liu, Oliver
    Rakita, Daniel
    Mutlu, Bilge
    Gleicher, Michael
    2017 26TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2017, : 751 - 757
  • [5] Semantic Scene Understanding for Human-Robot Interaction
    Patel, Maithili
    Dogan, Fethiye Irmak
    Zeng, Zhen
    Baraka, Kim
    Chernova, Sonia
    COMPANION OF THE ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2023, 2023, : 941 - 943
  • [6] Impedance Control with Bounded Actions for Human-Robot Interaction
    Ramirez-Vera, Victor, I
    Mendoza-Gutierrez, Marco O.
    Bonilla-Gutierrez, Isela
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2022, 47 (11) : 14989 - 15000
  • [7] Planning Ergonomic Sequences of Actions in Human-Robot Interaction
    Busch, Baptiste
    Toussaint, Marc
    Lopes, Manuel
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 1916 - 1923
  • [8] Understanding the Perception of Incremental Robot Response in Human-Robot Interaction
    Jensen, Lars Christian
    Langedijk, Rosalyn Melissa
    Fischer, Kerstin
    2020 29TH IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2020, : 41 - 47
  • [9] Human-Robot Interaction
    Jia, Yunyi
    Zhang, Biao
    Li, Miao
    King, Brady
    Meghdari, Ali
    JOURNAL OF ROBOTICS, 2018, 2018
  • [10] Advbot: Towards Understanding Human Preference in a Human-Robot Interaction Scenario
    Wong, Clarice J.
    Tay, Yong Ling
    Lew, Lincoln W. C.
    Koh, Hui Fang
    Xiong, Yijing
    Wu, Yan
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 1305 - 1309