Iris Geometric Transformation Guided Deep Appearance-Based Gaze Estimation

被引:0
|
作者
Nie, Wei [1 ]
Wang, Zhiyong [1 ]
Ren, Weihong [1 ]
Zhang, Hanlin [1 ]
Liu, Honghai [1 ,2 ]
机构
[1] Harbin Inst Technol Shenzhen, State Key Lab Robot & Syst, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518000, Peoples R China
基金
中国国家自然科学基金;
关键词
Estimation; Iris; Artificial neural networks; Three-dimensional displays; Feature extraction; Faces; Multitasking; Training; Heating systems; Vectors; Appearance-based gaze estimation; deep learning; geometric priors; eye landmarks; EYE; VISION;
D O I
10.1109/TIP.2025.3546465
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The geometric alterations in the iris's appearance are intricately linked to the gaze direction. However, current deep appearance-based gaze estimation methods mainly rely on latent feature sharing to leverage iris features for improving deep representation learning, often neglecting the explicit modeling of their geometric relationships. To address this issue, this paper revisits the physiological structure of the eyeball and introduces a set of geometric assumptions, such as "the normal vector of the iris center approximates the gaze direction". Building on these assumptions, we propose an Iris Geometric Transformation Guided Gaze estimation (IGTG-Gaze) module, which establishes an explicit geometric parameter sharing mechanism to link gaze direction and sparse iris landmark coordinates directly. Extensive experimental results demonstrate that IGTG-Gaze seamlessly integrates into various deep neural networks, flexibly extends from sparse iris landmarks to dense eye mesh, and consistently achieves leading performance in both within- and cross-dataset evaluations, all while maintaining end-to-end optimization. These advantages highlight IGTG-Gaze as a practical and effective approach for enhancing deep gaze representation from appearance.
引用
收藏
页码:1616 / 1631
页数:16
相关论文
共 50 条
  • [21] Gaze-Net: Appearance-Based Gaze Estimation using Capsule Networks
    Mahanama, Bhanuka
    Jayawardana, Yasith
    Jayarathna, Sampath
    AUGMENTED HUMAN 2020: PROCEEDINGS OF THE 11TH AUGMENTED HUMAN INTERNATIONAL CONFERENCE, 2020,
  • [22] Appearance-Based Gaze Block Estimation via CNN Classification
    Wu, Xuemei
    Li, Jing
    Wu, Qiang
    Sun, Jiande
    2017 IEEE 19TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2017,
  • [23] Appearance-based gaze estimation under slight head motion
    Zhizhi Guo
    Qianxiang Zhou
    Zhongqi Liu
    Multimedia Tools and Applications, 2017, 76 : 2203 - 2222
  • [24] Manifold Alignment for Person Independent Appearance-based Gaze Estimation
    Schneider, Timo
    Schauerte, Boris
    Stiefelhagen, Rainer
    2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 1167 - 1172
  • [25] Appearance-based gaze estimation under slight head motion
    Guo, Zhizhi
    Zhou, Qianxiang
    Liu, Zhongqi
    MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (02) : 2203 - 2222
  • [26] Appearance-Based Gaze Estimation Using Dilated-Convolutions
    Chen, Zhaokang
    Shi, Bertram E.
    COMPUTER VISION - ACCV 2018, PT VI, 2019, 11366 : 309 - 324
  • [27] Appearance-based Gaze Estimation using Attention and Difference Mechanism
    Murthy, L. R. D.
    Biswas, Pradipta
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3137 - 3146
  • [28] A Head Pose-free Approach for Appearance-based Gaze Estimation
    Lu, Feng
    Okabe, Takahiro
    Sugano, Yusuke
    Sato, Yoichi
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2011, 2011,
  • [29] Improving Domain Generalization in Appearance-Based Gaze Estimation With Consistency Regularization
    Back, Moon-Ki
    Yoo, Cheol-Hwan
    Yoo, Jang-Hee
    IEEE ACCESS, 2023, 11 : 137948 - 137956
  • [30] Evaluating the Robustness of an Appearance-based Gaze Estimation Method for Multimodal Interfaces
    Li, Nanxiang
    Busso, Carlos
    ICMI'13: PROCEEDINGS OF THE 2013 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2013, : 91 - 98