Gaze Estimation Using Residual Neural Network

被引:0
|
作者
Wong, En Teng [1 ]
Yean, Seanglidet [1 ]
Hu, Qingyao [1 ]
Lee, Bu Sung [1 ]
Liu, Jigang [1 ]
Deepu, Rajan [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
关键词
eye track; mobile; ResNet; deep learning;
D O I
10.1109/percomw.2019.8730846
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Eye gaze tracking has become an prominent research topic in human-computer interaction and computer vision. It is due to its application in numerous fields, such as the market research, medical, neuroscience and psychology. Eye gaze tracking is implemented by estimating gaze (gaze estimation) for each individual frame in offline or real-time video captured. Therefore, in order to produce the secure the accurate tracking, especially in the emerging use in medical and community, innovation on the gaze estimation posts a challenge in research field. In this paper, we explored the use of the deep learning model, Residual Neural Network (ResNet-18), to predict the eye gaze on mobile device. The model is trained using the large-scale eye tracking public dataset called GazeCapture. We aim to innovate by incorporating methods/techniques of removing the blinking data, applying image histogram normalisation, head pose, and face grid features. As a result, we achieved 3.05cm average error, which is better performance than iTracker (4.11cm average error), the recent gaze tracking deep-learning model using AlexNet architecture. Upon observation, adaptive normalisation of the images was found to produce better results compared to histogram normalisation. Additionally, we found that head pose information was useful contribution to the proposed deep-learning network, while face grid information does not help to reduce test error.
引用
收藏
页码:411 / 414
页数:4
相关论文
共 50 条
  • [1] Gaze Estimation Approach Using Deep Differential Residual Network
    Huang, Longzhao
    Li, Yujie
    Wang, Xu
    Wang, Haoyu
    Bouridane, Ahmed
    Chaddad, Ahmad
    [J]. SENSORS, 2022, 22 (14)
  • [2] Gaze Estimation Using Neural Network And Logistic Regression
    Xia, Yifan
    Liang, Baosheng
    Li, Zhaotong
    Gao, Song
    [J]. COMPUTER JOURNAL, 2022, 65 (08): : 2034 - 2043
  • [3] Gaze Estimation Based on Neural Network
    Luo, Mingyuan
    Liu, Xi
    Huang, Wei
    [J]. PROCEEDINGS OF 2019 IEEE 2ND INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION AND COMMUNICATION TECHNOLOGY (ICEICT 2019), 2019, : 590 - 594
  • [4] Style transformed synthetic images for real world gaze estimation by using residual neural network with embedded personal identities
    Quan Wang
    Hui Wang
    Ruo-Chen Dang
    Guang-Pu Zhu
    Hai-Feng Pi
    Frederick Shic
    Bing-liang Hu
    [J]. Applied Intelligence, 2023, 53 : 2026 - 2041
  • [5] Style transformed synthetic images for real world gaze estimation by using residual neural network with embedded personal identities
    Wang, Quan
    Wang, Hui
    Dang, Ruo-Chen
    Zhu, Guang-Pu
    Pi, Hai-Feng
    Shic, Frederick
    Hu, Bing-liang
    [J]. APPLIED INTELLIGENCE, 2023, 53 (02) : 2026 - 2041
  • [6] Hybrid LASSO and Neural Network Estimator for Gaze Estimation
    Iyer, S. Deepthi
    Ratnasangu, Hariharan
    [J]. PROCEEDINGS OF THE 2016 IEEE REGION 10 CONFERENCE (TENCON), 2016, : 2579 - 2582
  • [7] GFNet: Gaze Focus Network using Attention for Gaze Estimation
    Hu, Daosong
    Huang, Kai
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2399 - 2404
  • [8] Estimation of Residual Traveling Distance for Power Wheelchair Using Neural Network
    Chen, Pei-Chung
    Li, Xiao-Qin
    Koh, Yong-Fa
    [J]. PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON INTELLIGENT TECHNOLOGIES AND ENGINEERING SYSTEMS (ICITES2014), 2016, 345 : 43 - 49
  • [9] Gaze estimation using convolutional neural networks
    Rawdha Karmi
    Ines Rahmany
    Nawres Khlifa
    [J]. Signal, Image and Video Processing, 2024, 18 : 389 - 398
  • [10] Gaze estimation using convolutional neural networks
    Karmi, Rawdha
    Rahmany, Ines
    Khlifa, Nawres
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (01) : 389 - 398