Multi-modal Open World User Identification

被引:8
|
作者
Irfan, Bahar [1 ]
Ortiz, Michael Garcia [2 ,3 ]
Lyubova, Natalia [4 ]
Belpaeme, Tony [1 ,5 ]
机构
[1] Univ Plymouth, Ctr Robot & Neural Syst, Plymouth PL4 8AA, Devon, England
[2] SoftBank Robot Europe, AI Lab, 43 Rue Colonel Pierre Avia, Paris, France
[3] City Univ London, Northampton Sq, London EC1V 0HB, England
[4] Prophesee, 74 Rue Faubourg St Antoine, F-75012 Paris, France
[5] Univ Ghent, IDLab Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
基金
欧盟地平线“2020”;
关键词
Open world recognition; Bayesian network; soft biometrics; incremental learning; online learning; multi-modal dataset; long-term user recognition; Human-Robot Interaction; WEIGHTED BAYESIAN NETWORK; FACE; MODELS; RECOGNITION; IMITATION;
D O I
10.1145/3477963
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
User identification is an essential step in creating a personalised long-term interaction with robots. This requires learning the users continuously and incrementally, possibly starting from a state without any known user. In this article, we describe a multi-modal incremental Bayesian network with online learning, which is the first method that can be applied in such scenarios. Face recognition is used as the primary biometric, and it is combined with ancillary information, such as gender, age, height, and time of interaction to improve the recognition. The Multi-modal Long-term User Recognition Dataset is generated to simulate various human-robot interaction (HRI) scenarios and evaluate our approach in comparison to face recognition, soft biometrics, and a state-of-the-art open world recognition method (Extreme Value Machine). The results show that the proposed methods significantly outperform the baselines, with an increase in the identification rate up to 47.9% in open-set and closed-set scenarios, and a significant decrease in long-term recognition performance loss. The proposed models generalise well to new users. provide stability, improve over time, and decrease the bias of face recognition. The models were applied in HRI studies for user recognition, personalised rehabilitation, and customer-oriented service, which showed that they are suitable for long-term HRI in the real world.
引用
收藏
页数:50
相关论文
共 50 条
  • [1] User identification and authentication using multi-modal behavioral biometrics
    Bailey, Kyle O.
    Okolica, James S.
    Peterson, Gilbert L.
    COMPUTERS & SECURITY, 2014, 43 : 77 - 89
  • [2] Multi-modal user identification and object recognition surveillance system
    Clapes, Albert
    Reyes, Miguel
    Escalera, Sergio
    PATTERN RECOGNITION LETTERS, 2013, 34 (07) : 799 - 808
  • [3] Multi-modal Prototypes for Open-World Semantic Segmentation
    Yang, Yuhuan
    Ma, Chaofan
    Ju, Chen
    Zhang, Fei
    Yao, Jiangchao
    Zhang, Ya
    Wang, Yanfeng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 6004 - 6020
  • [4] Fusion of Multi-modal Information of User Profile Across Social Networks for User Identification
    Ye, Cuicui
    Yang, Jing
    Mao, Yan
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT V, ICIC 2024, 2024, 14866 : 416 - 429
  • [5] Multi-modal human identification system
    Ivanov, Y
    WACV 2005: SEVENTH IEEE WORKSHOP ON APPLICATIONS OF COMPUTER VISION, PROCEEDINGS, 2005, : 164 - 170
  • [6] Design and implementation of a multi-modal user interface of the virtual world database system (VWDB)
    Masunaga, Y
    Watanabe, C
    SEVENTH INTERNATIONAL CONFERENCE ON DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, PROCEEDINGS, 2001, : 294 - 301
  • [7] Mudra: A Multi-Modal Smartwatch Interactive System with Hand Gesture Recognition and User Identification
    Guo, Kaiwen
    Zhou, Hao
    Tian, Ye
    Zhou, Wangqiu
    Ji, Yusheng
    Li, Xiang-Yang
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022), 2022, : 100 - 109
  • [8] User Identification from Gait Analysis Using Multi-Modal Sensors in Smart Insole
    Choi, Sang-Il
    Moon, Jucheol
    Park, Hee-Chan
    Choi, Sang Tae
    SENSORS, 2019, 19 (17)
  • [9] Multi-Modal Open-Domain Dialogue
    Shuster, Kurt
    Smith, Eric Michael
    Ju, Da
    Weston, Jason
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 4863 - 4883
  • [10] Multi-modal Interaction System for Enhanced User Experience
    Jeong, Yong Mu
    Min, Soo Young
    Lee, Seung Eun
    COMPUTER APPLICATIONS FOR WEB, HUMAN COMPUTER INTERACTION, SIGNAL AND IMAGE PROCESSING AND PATTERN RECOGNITION, 2012, 342 : 287 - +