Multi-modal Open World User Identification

被引:8
|
作者
Irfan, Bahar [1 ]
Ortiz, Michael Garcia [2 ,3 ]
Lyubova, Natalia [4 ]
Belpaeme, Tony [1 ,5 ]
机构
[1] Univ Plymouth, Ctr Robot & Neural Syst, Plymouth PL4 8AA, Devon, England
[2] SoftBank Robot Europe, AI Lab, 43 Rue Colonel Pierre Avia, Paris, France
[3] City Univ London, Northampton Sq, London EC1V 0HB, England
[4] Prophesee, 74 Rue Faubourg St Antoine, F-75012 Paris, France
[5] Univ Ghent, IDLab Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
基金
欧盟地平线“2020”;
关键词
Open world recognition; Bayesian network; soft biometrics; incremental learning; online learning; multi-modal dataset; long-term user recognition; Human-Robot Interaction; WEIGHTED BAYESIAN NETWORK; FACE; MODELS; RECOGNITION; IMITATION;
D O I
10.1145/3477963
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
User identification is an essential step in creating a personalised long-term interaction with robots. This requires learning the users continuously and incrementally, possibly starting from a state without any known user. In this article, we describe a multi-modal incremental Bayesian network with online learning, which is the first method that can be applied in such scenarios. Face recognition is used as the primary biometric, and it is combined with ancillary information, such as gender, age, height, and time of interaction to improve the recognition. The Multi-modal Long-term User Recognition Dataset is generated to simulate various human-robot interaction (HRI) scenarios and evaluate our approach in comparison to face recognition, soft biometrics, and a state-of-the-art open world recognition method (Extreme Value Machine). The results show that the proposed methods significantly outperform the baselines, with an increase in the identification rate up to 47.9% in open-set and closed-set scenarios, and a significant decrease in long-term recognition performance loss. The proposed models generalise well to new users. provide stability, improve over time, and decrease the bias of face recognition. The models were applied in HRI studies for user recognition, personalised rehabilitation, and customer-oriented service, which showed that they are suitable for long-term HRI in the real world.
引用
收藏
页数:50
相关论文
共 50 条
  • [41] Inferring Relevant People from Multi-Modal User Communication Activity
    Dhara, Krishna Kishore
    Krishnaswamy, Venkatesh
    Katre, Prasad
    2012 IEEE CONSUMER COMMUNICATIONS AND NETWORKING CONFERENCE (CCNC), 2012, : 538 - 543
  • [42] Multi-modal User Intent Classification Under the Scenario of Smart Factory
    Chiu, Yu-Ching
    Chang, Bo-Hao
    Chen, Tzu-Yu
    Yang, Cheng-Fu
    Bi, Nanyi
    Tsai, Richard Tzong-Han
    Lee, Hung-yi
    Hsu, Jane Yung-jen
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 15771 - 15772
  • [43] Learning the User's Deeper Preferences for Multi-modal Recommendation Systems
    Lei, Fei
    Cao, Zhongqi
    Yang, Yuning
    Ding, Yibo
    Zhang, Cong
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (03)
  • [44] Rataplan: Resilient Automation of User Interface Actions with Multi-modal Proxies
    Veuskens, Tom
    Luyten, Kris
    Ramakers, Raf
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2020, 4 (02):
  • [45] A Cognitive User Interface for a Multi-modal Human-Machine Interaction
    Tschoepe, Constanze
    Duckhorn, Frank
    Huber, Markus
    Meyer, Werner
    Wolff, Matthias
    SPEECH AND COMPUTER (SPECOM 2018), 2018, 11096 : 707 - 717
  • [46] User behavior fusion in dialog management with multi-modal history cues
    Yang, Minghao
    Tao, Jianhua
    Chao, Linlin
    Li, Hao
    Zhang, Dawei
    Che, Hao
    Gao, Tingli
    Liu, Bin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2015, 74 (22) : 10025 - 10051
  • [47] Holographic Raman tweezers controlled by multi-modal natural user interface
    Tomori, Zoltan
    Kesa, Peter
    Nikorovic, Matej
    Kanka, Jan
    Jakl, Petr
    Sery, Mojmir
    Bernatova, Silvie
    Valusova, Eva
    Antalik, Marian
    Zemanek, Pavel
    JOURNAL OF OPTICS, 2016, 18 (01)
  • [48] Refining Parent SMART: User feedback to optimize a multi-modal intervention
    Becker, Sara J.
    Shiller, Hannah
    Fan, Yiqing
    Dibartolo, Emily
    Olson, Miranda B.
    Casline, Elizabeth
    Wijaya, Clarisa
    Helseth, Sarah A.
    Kelly, Lourah M.
    JOURNAL OF SUBSTANCE USE & ADDICTION TREATMENT, 2024, 166
  • [49] Multi-modal Scene-compliant User Intention Estimation in Navigation
    Katuwandeniya, Kavindie
    Kiss, Stefan H.
    Shi, Lei
    Miro, Jaime Valls
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 1001 - 1006
  • [50] Emoji Helps! A Multi-modal Siamese Architecture for Tweet User Verification
    Chanchal Suman
    Sriparna Saha
    Pushpak Bhattacharyya
    Rohit Shyamkant Chaudhari
    Cognitive Computation, 2021, 13 : 261 - 276