August 2014
Volume 14, Issue 10
Free
Vision Sciences Society Annual Meeting Abstract  |   August 2014
A data-driven approach to learning strategies for the visual control of navigation.
Author Affiliations
  • Youssef Barhomi
    Department of Cognitive, Linguistics, and Psychological Sciences, Brown University
  • Abigail Yanke
    Department of Cognitive, Linguistics, and Psychological Sciences, Brown University
  • Stephane Bonneaud
    Department of Cognitive, Linguistics, and Psychological Sciences, Brown University
  • William Warren
    Department of Cognitive, Linguistics, and Psychological Sciences, Brown University
  • Thomas Serre
    Department of Cognitive, Linguistics, and Psychological Sciences, Brown University
Journal of Vision August 2014, Vol.14, 1347. doi:https://doi.org/10.1167/14.10.1347
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Youssef Barhomi, Abigail Yanke, Stephane Bonneaud, William Warren, Thomas Serre; A data-driven approach to learning strategies for the visual control of navigation.. Journal of Vision 2014;14(10):1347. https://doi.org/10.1167/14.10.1347.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Rapid visual presentation paradigms, characterized by short presentation times and speeded behavioral responses, have played a central role in the study of our core visual recognition ability. Here, we adapted the rapid presentation paradigm for a visual navigation task to study how pre-attentive vision controls behavior. Method. We used CryEngine, a state-of-the-art gaming engine, to synthesize videos simulating self-motion through naturalistic scenes from a first-person view at a human walking speed (~1.4m/sec). Modern gaming engines offer the possibility to create relatively controlled, yet realistic, visual environments. Participants were presented with these 300ms long masked video sequences and used the mouse to report their preferred best direction to steer in order to navigate through the scene while avoiding obstacles. The best steering direction was selected from a continuum indicated by arrows along a circle projected on the ground. Participants were instructed to answer as accurately and as fast as possible to limit cortical feedback. Results. We computed a navigability index for each video by computing the proportion of responses from all participants in each steering direction (18 bins of 10 deg azimuth). Overall we found a high degree of consistency across participants, suggesting that they understood the task and relied on similar scene information. Behavioral data were consistent with steering directions predicted by a model for the behavioral dynamics of steering (Fajen & Warren, 2003) based on ground-truth depth data. We further used the behavioral responses to evaluate a variety of visual cues, from motion and shape to saliency, and to learn to predict participants' steering directions. Overall, the relative success of the proposed approach suggests that it may be possible to learn visual strategies (e.g. visual equalization, saccade and clutter response, etc.) directly from data without explicitly implementing them.

Meeting abstract presented at VSS 2014

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×