September 2019
Volume 19, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2019
Towards End to End head-free gaze classification
Author Affiliations & Notes
  • Rakshit S Kothari
    Chester F. Carlson Institute of Imaging Science, Rochester Institute of Technology
  • Zhizhuo Yang
    B. Thomas Golisano College of Computing & Information Sciences, Rochester Institute of Technology
  • Chris Kanan
    Chester F. Carlson Institute of Imaging Science, Rochester Institute of Technology
  • Jeff Pelz
    Chester F. Carlson Institute of Imaging Science, Rochester Institute of Technology
  • Reynold Bailey
    B. Thomas Golisano College of Computing & Information Sciences, Rochester Institute of Technology
  • Gabriel J Diaz
    Chester F. Carlson Institute of Imaging Science, Rochester Institute of Technology
Journal of Vision September 2019, Vol.19, 147. doi:https://doi.org/10.1167/19.10.147
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Rakshit S Kothari, Zhizhuo Yang, Chris Kanan, Jeff Pelz, Reynold Bailey, Gabriel J Diaz; Towards End to End head-free gaze classification. Journal of Vision 2019;19(10):147. https://doi.org/10.1167/19.10.147.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

The study of gaze behavior benefits from the classification of the time series into distinct movement types, or events, such as saccade, pursuit, and fixation. Because the manual identification of events is time consuming and subjective, there is a need for automated classifiers. Although there are several solutions for the classification of the eye-in-head signal, there are no established solutions for classification of the coordinated movements of the eyes and head that will occur in less constrained contexts, for example, when wearing a virtual or augmented reality display. Our approach involves training various temporal classifiers on our new Gaze-in-Wild dataset, recorded from over 20 unrestrained participants and hand-coded by 5 practiced labellers. Subjects were instrumented with a 6-axis 100 Hz inertial measurement unit (mean drift: 0.03 deg/sec), a 30 Hz ZED stereo camera, and a 120Hz Pupil labs eye tracker (mean calibration AngError < 1 deg within 10 degrees from calibration pattern center) to record eye and head orientation. The effort culminated in over 2 hours and 20 minutes of hand labelled head-free gaze behavior data, with approximately 20000 detected fixational movements, 18000 saccades, 1400 pursuit events, and 4000 blinks. We use this hand labelled data to benchmark our dataset using standard machine learning classifiers and train our recurrent network model which leverages multiple Neural Arithmetic Logic Units to classify gaze behavior directly from raw, unfiltered eye-in-head and head vectors. Activation map of various hidden units provides insight on learned representations of eye and head coordination and velocity based feature representations which are directly comparable with hand-crafted features. The performance of our classifier is evaluated using various event based metrics and shows that it can attain near-human level classification (>kappa 0.70, event F1> 0.85).

Acknowledgement: Google Daydream grant 
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×