Head Pose-Aware Regression for Pupil Localization From a-Pillar Cameras

被引:0
|
作者
Kang, Donghwa [1 ]
Kang, Dongwoo [1 ]
机构
[1] Hongik Univ, Sch Elect & Elect Engn, Seoul 04066, South Korea
基金
新加坡国家研究基金会;
关键词
Pupil center localization; remote eye tracking; head pose-aware pupil regression; eye-nose points regression; head pose estimation; A-pillar camera; augmented reality (AR) 3D head-up displays (HUDs); driver monitoring system (DMS); SUPPORT VECTOR MACHINES; TRACKING; FACES;
D O I
10.1109/ACCESS.2024.3354373
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In vehicular applications, remote eye pupil tracking is essential, particularly for advanced augmented reality (AR) 3D head-up displays (HUDs), and driver monitoring systems (DMS). However, achieving accurate pupil center localization under varied head poses presents significant challenges, especially when cameras are positioned on a vehicle's A-pillar. This placement introduces substantial head pose variations, complicating traditional tracking methods. In response, this study presents a remote pupil localization method designed to address the unique challenges posed by a camera situated on a vehicle's A-pillar, a spot causing significant head pose variations. The proposed technique relies on a head pose-aware pupil localization strategy utilizing A-pillar cameras. Our pupil localization algorithm adopts a Transformer regression approach, into which we integrate head pose estimation data, enhancing its capability across diverse head poses. To further enhance our approach, we used an optimized nine-point eye-nose landmark set, to minimize the pupil center localization loss. To demonstrate the robustness of our method, we conducted evaluations using both the public WIDER Facial Landmarks in-the-wild (WFLW) dataset and a custom in-house dataset focused on A-pillar camera captures. Results indicate a Normalized Mean Error (NME) of 2.79% and a failure rate (FR) of 1.28% on the WFLW dataset. On our in-house dataset, the method achieved an NME of 2.96% and a FR of 0.72%. These impressive results demonstrate the robustness and efficacy of our method, suggesting its potential for implementation in commercial eye tracking systems using A-pillar mounted cameras, especially for AR 3D HUD and DMS applications.
引用
收藏
页码:11083 / 11094
页数:12
相关论文
共 17 条
  • [1] Pose-aware instance segmentation framework from cone beam CT images for tooth segmentation
    Chung, Minyoung
    Lee, Minkyung
    Hong, Jioh
    Park, Sanguk
    Lee, Jusang
    Lee, Jingyu
    Yang, Il-Hyung
    Lee, Jeongjin
    Shin, Yeong-Gil
    COMPUTERS IN BIOLOGY AND MEDICINE, 2020, 120 (120)
  • [2] Rethinking Reprojection: Closing the Loop for Pose-aware Shape Reconstruction from a Single Image
    Zhu, Rui
    Galoogahi, Hamed Kiani
    Wang, Chaoyang
    Lucey, Simon
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 57 - 65
  • [3] From one to many: Pose-Aware Metric Learning for single-sample face recognition
    Deng, Weihong
    Hu, Jiani
    Wu, Zhongjun
    Guo, Jun
    PATTERN RECOGNITION, 2018, 77 : 426 - 437
  • [4] A Pose-Aware Auto-Augmentation Framework for 3D Human Pose and Shape Estimation from Partial Point Clouds
    Wang, Kangkan
    Yin, Sliihao
    Fang, Chenghao
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VI, 2025, 15036 : 64 - 79
  • [5] Absolute head pose estimation from overhead wide-angle cameras
    Tian, YL
    Brown, L
    Connell, J
    Pankanti, S
    Hampapur, A
    Senior, A
    Bolle, R
    IEEE INTERNATIONAL WORKSHOP ON ANALYSIS AND MODELING OF FACE AND GESTURES, 2003, : 92 - 99
  • [6] Head Pose Estimation Based on 3-D Facial Landmarks Localization and Regression
    Derkach, Dmytro
    Ruiz, Adria
    Sukno, Federico M.
    2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, : 820 - 827
  • [7] Joint head pose and facial landmark regression from depth images
    Wang J.
    Zhang J.
    Luo C.
    Chen F.
    Computational Visual Media, 2017, 3 (3) : 229 - 241
  • [8] Joint head pose and facial landmark regression from depth images
    Jie Wang
    Juyong Zhang
    Changwei Luo
    Falai Chen
    Computational Visual Media, 2017, 3 (03) : 229 - 241
  • [9] Hand pose estimation based on regression method from monocular RGB cameras for handling occlusion
    Roumaissa, Bekiri
    Chaouki, Babahenini Mohamed
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (07) : 21497 - 21523
  • [10] Hand pose estimation based on regression method from monocular RGB cameras for handling occlusion
    Bekiri Roumaissa
    Babahenini Mohamed Chaouki
    Multimedia Tools and Applications, 2024, 83 : 21497 - 21523