September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Unveiling task-dependent action affordance representations: Insights from scene-selective cortex and deep neural networks
Author Affiliations
  • Clemens G. Bartnik
    Video & Image Sense Lab, Informatics Institute, University of Amsterdam, The Netherlands
  • Nikolina Vukšić
    Psychology Research Institute, University of Amsterdam Amsterdam, The Netherlands
  • Steven Bommer
    Psychology Research Institute, University of Amsterdam Amsterdam, The Netherlands
  • Iris I.A. Groen
    Video & Image Sense Lab, Informatics Institute, University of Amsterdam, The Netherlands
    Psychology Research Institute, University of Amsterdam Amsterdam, The Netherlands
Journal of Vision September 2024, Vol.24, 897. doi:https://doi.org/10.1167/jov.24.10.897
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Clemens G. Bartnik, Nikolina Vukšić, Steven Bommer, Iris I.A. Groen; Unveiling task-dependent action affordance representations: Insights from scene-selective cortex and deep neural networks. Journal of Vision 2024;24(10):897. https://doi.org/10.1167/jov.24.10.897.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Humans effortlessly know how and where to move in the immediate environment using a wide range of navigational actions, from walking and driving to climbing. Yet little is known about where and how action affordances are computed in the brain. Some work implicates scene-selective cortex in navigational affordance representation, reflecting visual features computed in mid-level DNN layers (Bonner et al., 2017, 2018), while others report a lack of affordance representation therein (Groen et al., 2018). Here, we curated a novel set of real-world scenes that afford distinct navigational actions in both indoor and outdoor environments, for which we collected rich behavioral annotations (N=152) for seven commonly used visual properties. The behavioral annotations indicate that navigational actions form a distinct space separate from representations of objects or materials; even in combination, visual properties explain only around 20% of the variance in navigational action annotations. We collected human fMRI measurements (N=20) to a subset of 90 images while subjects performed three distinct tasks (action affordance recognition, object recognition, and fixation). Using representational similarity analysis, we confirm that scene-selective brain regions, especially the Parahippocampal Place Area and Occipital Place Area, represent navigational action affordances. Furthermore, elevated behavioral correlations in scene-selective regions during action affordance and object recognition tasks relative to fixation suggests these representations are task-dependent. Unlike prior findings, however, we find that DNNs trained for scene and object classification poorly represent these action affordances. Interestingly, language-supervised models like Contrastive Language-Image Pre-training (CLIP) show enhanced predictions for behavior and brain activity, suggesting they capture affordance representation better. These findings strengthen evidence for action affordances in the scene-selective cortex and reveal their task dependency. However, the underlying computations remain elusive, but our work suggests that integrating semantic information in computational models of affordance perception is a promising direction.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×