Multi-modal Open World User Identification

被引:8
|
作者
Irfan, Bahar [1 ]
Ortiz, Michael Garcia [2 ,3 ]
Lyubova, Natalia [4 ]
Belpaeme, Tony [1 ,5 ]
机构
[1] Univ Plymouth, Ctr Robot & Neural Syst, Plymouth PL4 8AA, Devon, England
[2] SoftBank Robot Europe, AI Lab, 43 Rue Colonel Pierre Avia, Paris, France
[3] City Univ London, Northampton Sq, London EC1V 0HB, England
[4] Prophesee, 74 Rue Faubourg St Antoine, F-75012 Paris, France
[5] Univ Ghent, IDLab Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
基金
欧盟地平线“2020”;
关键词
Open world recognition; Bayesian network; soft biometrics; incremental learning; online learning; multi-modal dataset; long-term user recognition; Human-Robot Interaction; WEIGHTED BAYESIAN NETWORK; FACE; MODELS; RECOGNITION; IMITATION;
D O I
10.1145/3477963
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
User identification is an essential step in creating a personalised long-term interaction with robots. This requires learning the users continuously and incrementally, possibly starting from a state without any known user. In this article, we describe a multi-modal incremental Bayesian network with online learning, which is the first method that can be applied in such scenarios. Face recognition is used as the primary biometric, and it is combined with ancillary information, such as gender, age, height, and time of interaction to improve the recognition. The Multi-modal Long-term User Recognition Dataset is generated to simulate various human-robot interaction (HRI) scenarios and evaluate our approach in comparison to face recognition, soft biometrics, and a state-of-the-art open world recognition method (Extreme Value Machine). The results show that the proposed methods significantly outperform the baselines, with an increase in the identification rate up to 47.9% in open-set and closed-set scenarios, and a significant decrease in long-term recognition performance loss. The proposed models generalise well to new users. provide stability, improve over time, and decrease the bias of face recognition. The models were applied in HRI studies for user recognition, personalised rehabilitation, and customer-oriented service, which showed that they are suitable for long-term HRI in the real world.
引用
收藏
页数:50
相关论文
共 50 条
  • [31] Multi-modal anchor adaptation learning for multi-modal summarization
    Chen, Zhongfeng
    Lu, Zhenyu
    Rong, Huan
    Zhao, Chuanjun
    Xu, Fan
    NEUROCOMPUTING, 2024, 570
  • [32] Multi-modal User Interface for Multi-robot Control in Underground Environments*
    Chen, Shengkang
    O'Brien, Matthew J.
    Talbot, Fletcher
    Williams, Jason
    Tidd, Brendan
    Pitt, Alex
    Arkin, Ronald C.
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 9995 - 10002
  • [33] Flexible Dual Multi-Modal Hashing for Incomplete Multi-Modal Retrieval
    Wei, Yuhong
    An, Junfeng
    INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2024,
  • [34] A multi-modal open dataset for mental-disorder analysis
    Hanshu Cai
    Zhenqin Yuan
    Yiwen Gao
    Shuting Sun
    Na Li
    Fuze Tian
    Han Xiao
    Jianxiu Li
    Zhengwu Yang
    Xiaowei Li
    Qinglin Zhao
    Zhenyu Liu
    Zhijun Yao
    Minqiang Yang
    Hong Peng
    Jing Zhu
    Xiaowei Zhang
    Guoping Gao
    Fang Zheng
    Rui Li
    Zhihua Guo
    Rong Ma
    Jing Yang
    Lan Zhang
    Xiping Hu
    Yumin Li
    Bin Hu
    Scientific Data, 9
  • [35] A multi-modal open dataset for mental-disorder analysis
    Cai, Hanshu
    Yuan, Zhenqin
    Gao, Yiwen
    Sun, Shuting
    Li, Na
    Tian, Fuze
    Xiao, Han
    Li, Jianxiu
    Yang, Zhengwu
    Li, Xiaowei
    Zhao, Qinglin
    Liu, Zhenyu
    Yao, Zhijun
    Yang, Minqiang
    Peng, Hong
    Zhu, Jing
    Zhang, Xiaowei
    Gao, Guoping
    Zheng, Fang
    Li, Rui
    Guo, Zhihua
    Ma, Rong
    Yang, Jing
    Zhang, Lan
    Hu, Xiping
    Li, Yumin
    Hu, Bin
    SCIENTIFIC DATA, 2022, 9 (01)
  • [36] Multi-Modal 2020: Multi-Modal Argumentation 30 Years Later
    Gilbert, Michael A.
    INFORMAL LOGIC, 2022, 42 (03): : 487 - 506
  • [37] User behavior fusion in dialog management with multi-modal history cues
    Minghao Yang
    Jianhua Tao
    Linlin Chao
    Hao Li
    Dawei Zhang
    Hao Che
    Tingli Gao
    Bin Liu
    Multimedia Tools and Applications, 2015, 74 : 10025 - 10051
  • [38] Natural User Interfaces for Human-Drone Multi-Modal Interaction
    Suarez Fernandez, Ramon A.
    Luis Sanchez-Lopez, Jose
    Sampedro, Carlos
    Bavle, Hriday
    Molina, Martin
    Campoy, Pascual
    2016 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2016, : 1013 - 1022
  • [39] Emoji Helps! A Multi-modal Siamese Architecture for Tweet User Verification
    Suman, Chanchal
    Saha, Sriparna
    Bhattacharyya, Pushpak
    Chaudhari, Rohit Shyamkant
    COGNITIVE COMPUTATION, 2021, 13 (02) : 261 - 276
  • [40] Dataset and Models for Item Recommendation Using Multi-Modal User Interactions
    Bruun, Simone Borg
    Balog, Krisztian
    Maistro, Maria
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 709 - 718