June 2006
Volume 6, Issue 6
Free
Vision Sciences Society Annual Meeting Abstract  |   June 2006
From eye-tracking data to information: Lessons from dynamic scenes
Author Affiliations
  • Ran Carmi
    Neuroscience Program, University of Southern California, Los Angeles
  • Laurent Itti
    Neuroscience Program, University of Southern California, Los Angeles
Journal of Vision June 2006, Vol.6, 493. doi:https://doi.org/10.1167/6.6.493
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ran Carmi, Laurent Itti; From eye-tracking data to information: Lessons from dynamic scenes. Journal of Vision 2006;6(6):493. https://doi.org/10.1167/6.6.493.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

A common simplifying assumption for dealing with vast amounts of raw eye-tracking data is to focus on spatial rather than temporal analyses. This assumption is supported by studies with still images, which showed that spatial rather than temporal correlations provide the only source of information in eye-tracking data. Here we establish the extent to which this assumption is violated during inspection of dynamic scenes.

We collected 50 video clips depicting a heterogeneous collection of natural scenes. These clips were cut into clip segments, which were re-assembled into 50 scene-shuffled clips (MTV-style). Human observers inspected either continuous or scene-shuffled clips, and inter-observer agreement in gaze position was quantified across conditions and over time.

On average, the instantaneous eye-positions of 4 human observers were clustered within a rectangle covering 8.51% and 6.04% of the display area in the continuous and scene-shuffled conditions, respectively. These values increased to 11.48% (p<0.01) and 9.36% (p<0.01) when eye-positions were sampled from the same eye traces in random order. The average cluster area increased further to 35.88% (p<0.01) when 4 eye-positions were chosen at random from a uniform distribution of spatial locations. Moreover, preserving time information led to previously unreported patterns of inter-observer agreement.

These results demonstrate that increasing stimulus dynamics triggers eye-movement patterns that diverge increasingly from previous accounts based on still images. The limited scalability of conclusions based on still images is likely to be further accentuated by future enhancements in the realism of laboratory stimuli, such as larger field of view and reduced central bias.

Carmi, R. Itti, L. (2006). From eye-tracking data to information: Lessons from dynamic scenes [Abstract]. Journal of Vision, 6(6):493, 493a, http://journalofvision.org/6/6/493/, doi:10.1167/6.6.493. [CrossRef]
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×