Automating Gaze Target Annotation in Human-Robot Interaction

被引:0
|
作者
Cheng, Linlin [1 ]
Hindriks, Koen V. [1 ]
Belopolsky, Artem V. [2 ]
机构
[1] Vrije Univ Amsterdam, Fac Sci, Comp Sci, Amsterdam, Netherlands
[2] Vrije Univ Amsterdam, Dept Human Movement Sci, Amsterdam, Netherlands
关键词
D O I
10.1109/RO-MAN60168.2024.10731455
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Identifying gaze targets in videos of human-robot interaction is useful for measuring engagement. In practice, this requires manually annotating for a fixed set of objects that a participant is looking at in a video, which is very time-consuming. To address this issue, we propose an annotation pipeline for automating this effort. In this work, we focus on videos in which the objects looked at do not move. As input for the proposed pipeline, we therefore only need to annotate object bounding boxes for the first frame of each video. The benefit, moreover, of manually annotating these frames is that we can also draw bounding boxes for objects outside of it, which enables estimating gaze targets in videos where not all objects are visible. A second issue that we address is that the models used for automating the pipeline annotate individual video frames. In practice, however, manual annotation is done at the event level for video segments instead of single frames. Therefore, we also introduce and investigate several variants of algorithms for aggregating frame-level to event-level annotations, which are used in the last step in our annotation pipeline. We compare two versions of our pipeline: one that uses a state-of-the-art gaze estimation model (GEM) and a second one using a state-of-the-art target detection model (TDM). Our results show that both versions successfully automate the annotation, but the GEM pipeline performs slightly (approximate to 10%) better for videos where not all objects are visible. Analysis of our aggregation algorithm, moreover, shows that there is no need for manual video segmentation because a fixed time interval for segmentation yields very similar results. We conclude that the proposed pipeline can be used to automate almost all of the annotation effort.
引用
收藏
页码:991 / 998
页数:8
相关论文
共 50 条
  • [1] Human Gaze Following for Human-Robot Interaction
    Saran, Akanksha
    Majumdar, Srinjoy
    Short, Elaine Schaertl
    Thomaz, Andrea
    Niekum, Scott
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 8615 - 8621
  • [2] Neural Network Implementation of Gaze-Target Prediction for Human-Robot Interaction
    Somashekarappa, Vidya
    Sayeed, Asad
    Howes, Christine
    2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 2238 - 2244
  • [3] The Importance of Mutual Gaze in Human-Robot Interaction
    Kompatsiari, Kyveli
    Tikhanoff, Vadim
    Ciardo, Francesca
    Metta, Giorgio
    Wykowska, Agnieszka
    SOCIAL ROBOTICS, ICSR 2017, 2017, 10652 : 443 - 452
  • [4] Active gaze tracking for human-robot interaction
    Atienza, R
    Zelinsky, A
    FOURTH IEEE INTERNATIONAL CONFERENCE ON MULTIMODAL INTERFACES, PROCEEDINGS, 2002, : 261 - 266
  • [5] Robot Gaze Behavior Affects Honesty in Human-Robot Interaction
    Schellen, Elef
    Bossi, Francesco
    Wykowska, Agnieszka
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
  • [6] Towards Automating the Generation of Human-Robot Interaction Scenarios
    Fontaine, Matthew C.
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12876 - 12877
  • [7] Implementation of Gaze Estimation in Dialogue to Human-Robot Interaction
    Somashekarappa, Vidya
    2022 10TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS, ACIIW, 2022,
  • [8] Gaze Cueing and the Role of Presence in Human-Robot Interaction
    Friebe, Kassandra
    Samporova, Sabina
    Malinovska, Kristina
    Hoffimann, Matej
    SOCIAL ROBOTICS, ICSR 2022, PT I, 2022, 13817 : 402 - 414
  • [9] Pseudo-Randomization in Automating Robot Behaviour during Human-Robot Interaction
    Paplu, Sarwar Hussain
    Mishra, Chinmaya
    Berns, Karsten
    10TH IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL-EPIROB 2020), 2020,
  • [10] Using Gaze Patterns to Infer Human Intention for Human-Robot Interaction
    Li, Kang
    Wu, Jinting
    Zhao, Xiaoguang
    Tan, Min
    2018 13TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2018, : 933 - 938