Human-robot interaction-oriented video understanding of human actions

被引:1
|
作者
Wang, Bin [1 ]
Chang, Faliang [1 ]
Liu, Chunsheng [1 ]
Wang, Wenqian [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, 17923 Jingshi Rd, Jinan 250061, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Action recognition; Human-robot interaction; Temporal modeling; Contextual scene reasoning; NETWORK; FEATURES;
D O I
10.1016/j.engappai.2024.108247
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper focuses on action recognition tasks oriented to the field of human-robot interaction, which is one of the major challenges in the robotic video understanding field. Previous approaches focus on designing temporal models, lack the ability to capture motion information and build contextual correlation models. This may result in robots being unable to effectively understand long-term video actions. To solve these two problems, this paper propose a novel video understanding framework including: an Adaptive Temporal Sensitivity and Motion Capture Network (ATSMC-Net) and a contextual scene reasoning module called Knowledge Function Graph Module (KFG-Module). The proposed ATSMC-Net can adaptively adjust the frame-level and pixel-level sensitive regions of temporal features to effectively capture motion information. To fuse contextual scene information for cross-temporal inference, the KFG-Module is introduced to achieve fine-grained video understanding based on the relationship between objects and actions. We evaluate the method using three public video understanding benchmarks, including Something-Something-V1&V2 and HMDB51. In addition, we present a dataset with real-world application scenarios of human-robot interactions to verify the effectiveness of our approach on mobile robots. The experimental results show that the proposed method can significantly improve the video understanding of the robots.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] The Science of Human-Robot Interaction
    Kiesler, Sara
    Goodrich, Michael A.
    ACM TRANSACTIONS ON HUMAN-ROBOT INTERACTION, 2018, 7 (01)
  • [42] Human-robot interaction and psychoanalysis
    Scalzone, Franco
    Tamburrini, Guglielmo
    AI & SOCIETY, 2013, 28 (03) : 297 - 307
  • [43] Anthropomorphism and Human-Robot Interaction
    Kim, Rae Yule
    COMMUNICATIONS OF THE ACM, 2024, 67 (02) : 80 - 85
  • [44] Human-Robot Medical Interaction
    Scimeca, Luca
    Iida, Fumiya
    Maiolino, Perla
    Nanayakkara, Thrishantha
    HRI'20: COMPANION OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2020, : 660 - 661
  • [45] Sound in Human-Robot Interaction
    Pelikan, Hannah
    Robinson, Frederic Anthony
    Keevallik, Leelo
    Velonaki, Mari
    Broth, Mathias
    Bown, Oliver
    HRI '21: COMPANION OF THE 2021 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2021, : 706 - 708
  • [46] Benefits of Human-Robot Interaction
    Salvini, Pericle
    Nicolescu, Monica
    Ishiguro, Hiroshi
    IEEE ROBOTICS & AUTOMATION MAGAZINE, 2011, 18 (04) : 98 - 99
  • [47] Semiotics and human-robot interaction
    Sequeira, Joao Silva
    Ribeiro, Maria Isabel
    ICINCO 2006: PROCEEDINGS OF THE THIRD INTERNATIONAL CONFERENCE ON INFORMATICS IN CONTROL, AUTOMATION AND ROBOTICS: ROBOTICS AND AUTOMATION, 2006, : 58 - 65
  • [48] ROS for Human-Robot Interaction
    Mohamed, Youssef
    Lemaignan, Severin
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 3020 - 3027
  • [49] Immersive Human-Robot Interaction
    Sandygulova, Anara
    Campbell, Abraham G.
    Dragone, Mauro
    O'Hare, G. M. P.
    HRI'12: PROCEEDINGS OF THE SEVENTH ANNUAL ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2012, : 227 - 228
  • [50] Haptic Human-Robot Interaction
    Karniel, Amir
    Peer, Angelika
    Donchin, Opher
    Mussa-Ivaldi, Ferdinando A.
    Loeb, Gerald E.
    IEEE TRANSACTIONS ON HAPTICS, 2012, 5 (03) : 193 - 195