Learning Multimodal Representations for Sample-efficient Recognition of Human Actions

被引:0
|
作者
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
de Matos, David Martins [1 ,2 ]
Paiva, Ana [1 ,2 ]
Inamura, Tetsunari [3 ,4 ]
机构
[1] Univ Lisbon, INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
[3] SOKENDAI Grad Univ Adv Studies, Natl Inst Informat, Chiyoda Ku, 2-1-2 Hitotsubashi, Tokyo, Japan
[4] SOKENDAI Grad Univ Adv Studies, Dept Informat, Chiyoda Ku, 2-1-2 Hitotsubashi, Tokyo, Japan
关键词
D O I
10.1109/iros40897.2019.8967635
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans interact in rich and diverse ways with the environment. However, the representation of such behavior by artificial agents is often limited. In this work we present motion concepts, a novel multimodal representation of human actions in a household environment. A motion concept encompasses a probabilistic description of the kinematics of the action along with its contextual background, namely the location and the objects held during the performance. We introduce a novel algorithm which learns and recognizes motion concepts from action demonstrations, named Online Motion Concept Learning (OMCL). The algorithm is evaluated on a virtual-reality household environment with the presence of a human avatar. OMCL outperforms standard motion recognition algorithms on an one-shot recognition task, attesting to its potential for sample-efficient recognition of human actions.
引用
收藏
页码:4288 / 4293
页数:6
相关论文
共 50 条
  • [21] Sample-efficient reinforcement learning for CERN accelerator control
    Kain, Verena
    Hirlander, Simon
    Goddard, Brennan
    Velotti, Francesco Maria
    Porta, Giovanni Zevi Della
    Bruchon, Niky
    Valentino, Gianluca
    PHYSICAL REVIEW ACCELERATORS AND BEAMS, 2020, 23 (12)
  • [22] Conditional Abstraction Trees for Sample-Efficient Reinforcement Learning
    Dadvar, Mehdi
    Nayyar, Rashmeet Kaur
    Srivastava, Siddharth
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 485 - 495
  • [23] A New Sample-Efficient PAC Reinforcement Learning Algorithm
    Zehfroosh, Ashkan
    Tanner, Herbert G.
    2020 28TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2020, : 788 - 793
  • [24] Learning Sample-Efficient Target Reaching for Mobile Robots
    Khan, Arbaaz
    Kumar, Vijay
    Ribeiro, Alejandro
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 3080 - 3087
  • [25] Sample-efficient Deep Reinforcement Learning with Imaginary Rollouts for Human-Robot Interaction
    Thabet, Mohammad
    Patacchiola, Massimiliano
    Cangelosi, Angelo
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 5079 - 5085
  • [26] Sample-Efficient Reinforcement Learning for Pose Regulation of a Mobile Robot
    Brescia, Walter
    De Cicco, Luca
    Mascolo, Saverio
    2022 11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS), 2022, : 42 - 47
  • [27] Safe and Sample-Efficient Reinforcement Learning Algorithms for Factored Environments
    Simao, Thiago D.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 6460 - 6461
  • [28] Sample-Efficient Deep Reinforcement Learning with Directed Associative Graph
    Yang, Dujia
    Qin, Xiaowei
    Xu, Xiaodong
    Li, Chensheng
    Wei, Guo
    CHINA COMMUNICATIONS, 2021, 18 (06) : 100 - 113
  • [29] Sample-Efficient Policy Learning based on Completely Behavior Cloning
    Zou, Qiming
    Wang, Ling
    Li, Yu
    Liu, Jie
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 2543 - 2548
  • [30] Sample-Efficient Reinforcement Learning with Stochastic Ensemble Value Expansion
    Buckman, Jacob
    Hafner, Danijar
    Tucker, George
    Brevdo, Eugene
    Lee, Honglak
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31