Gaze-Net: Appearance-Based Gaze Estimation using Capsule Networks

04/16/2020
by   Bhanuka Mahanama, et al.
0

Recent studies on appearance based gaze estimation indicate the ability of Neural Networks to decode gaze information from facial images encompassing pose information. In this paper, we propose Gaze-Net: A capsule network capable of decoding, representing, and estimating gaze information from ocular region images. We evaluate our proposed system using two publicly available datasets, MPIIGaze (200,000+ images in the wild) and Columbia Gaze (5000+ images of users with 21 gaze directions observed at 5 camera angles/positions). Our model achieves a Mean Absolute Error (MAE) of 2.84^∘ for Combined angle error estimate within dataset for MPI-IGaze dataset. Further, model achieves a MAE of 10.04^∘ for across dataset gaze estimation error for Columbia gaze dataset. Through transfer learning, the error is reduced to 5.9^∘. The results show this approach is promising with implications towards using commodity webcams to develop low-cost multi-user gaze tracking systems.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset