Attention-Based Variational Autoencoder Models for Human-Human Interaction Recognition via Generation

被引:1
|
作者
Banerjee, Bonny [1 ,2 ]
Baruah, Murchana [1 ,2 ]
机构
[1] Univ Memphis, Inst Intelligent Syst, Memphis, TN 38152 USA
[2] Univ Memphis, Dept Elect & Comp Engn, Memphis, TN 38152 USA
关键词
embodied AI agent; intent prediction; human-human interaction recognition; human-human interaction generation; attention; perception; proprioception; multimodal; variational autoencoder; recurrent neural network (RNN); long-short term memory (LSTM); FRAMEWORK; EMERGENCE; FUSION;
D O I
10.3390/s24123922
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The remarkable human ability to predict others' intent during physical interactions develops at a very early age and is crucial for development. Intent prediction, defined as the simultaneous recognition and generation of human-human interactions, has many applications such as in assistive robotics, human-robot interaction, video and robotic surveillance, and autonomous driving. However, models for solving the problem are scarce. This paper proposes two attention-based agent models to predict the intent of interacting 3D skeletons by sampling them via a sequence of glimpses. The novelty of these agent models is that they are inherently multimodal, consisting of perceptual and proprioceptive pathways. The action (attention) is driven by the agent's generation error, and not by reinforcement. At each sampling instant, the agent completes the partially observed skeletal motion and infers the interaction class. It learns where and what to sample by minimizing the generation and classification errors. Extensive evaluation of our models is carried out on benchmark datasets and in comparison to a state-of-the-art model for intent prediction, which reveals that classification and generation accuracies of one of the proposed models are comparable to those of the state of the art even though our model contains fewer trainable parameters. The insights gained from our model designs can inform the development of efficient agents, the future of artificial intelligence (AI).
引用
收藏
页数:35
相关论文
共 50 条
  • [21] A Novel Attention-Based Convolution Neural Network for Human Activity Recognition
    Zheng, Ge
    IEEE SENSORS JOURNAL, 2021, 21 (23) : 27015 - 27025
  • [22] Radar Human Activity Recognition with an Attention-Based Deep Learning Network
    Huan, Sha
    Wu, Limei
    Zhang, Man
    Wang, Zhaoyue
    Yang, Chao
    SENSORS, 2023, 23 (06)
  • [23] HAMLET: A Hierarchical Multimodal Attention-based Human Activity Recognition Algorithm
    Islam, Md Mofijul
    Iqbal, Tariq
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10285 - 10292
  • [24] Learn From Others and Be Yourself in Federated Human Activity Recognition via Attention-Based Pairwise Collaborations
    Bu, Can
    Zhang, Lei
    Cui, Hengtao
    Cheng, Dongzhou
    Wu, Hao
    Song, Aiguo
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [25] Attention-based Fusion for Multi-source Human Image Generation
    Lathuiliere, Stephane
    Sangineto, Enver
    Siarohin, Aliaksandr
    Sebe, Nicu
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 428 - 437
  • [26] HHI-AttentionNet: An Enhanced Human-Human Interaction Recognition Method Based on a Lightweight Deep Learning Model with Attention Network from CSI
    Shafiqul, Islam Md
    Jannat, Mir Kanon Ara
    Kim, Jin-Woo
    Lee, Soo-Wook
    Yang, Sung-Hyun
    SENSORS, 2022, 22 (16)
  • [27] Towards Understanding Attention-Based Speech Recognition Models
    Qin, Chu-Xiong
    Qu, Dan
    IEEE ACCESS, 2020, 8 : 24358 - 24369
  • [28] Deep learning and RGB-D based human action, human-human and human-object interaction recognition: A survey?
    Khaire, Pushpajit
    Kumar, Praveen
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 86
  • [29] Modeling human–human interaction with attention-based high-order GCN for trajectory prediction
    Yanyan Fang
    Zhiyu Jin
    Zhenhua Cui
    Qiaowen Yang
    Tianyi Xie
    Bo Hu
    The Visual Computer, 2022, 38 : 2257 - 2269
  • [30] Predicting response to joint attention performance in human-human interaction based on human-robot interaction for young children with autism spectrum disorder
    Nie, Guangtao
    Zheng, Zhi
    Johnson, Jazette
    Swanson, Amy R.
    Weitlauf, Amy S.
    Warren, Zachary E.
    Sarkar, Nilanjan
    2018 27TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (IEEE RO-MAN 2018), 2018, : 1069 - 1074