September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Interpreting distributed population codes with feature-accentuated visual encoding models
Author Affiliations & Notes
  • Jacob S. Prince
    Harvard University
  • Christopher Hamblin
    Harvard University
  • George A. Alvarez
    Harvard University
  • Talia Konkle
    Harvard University
    Kempner Institute for the Study of Natural and Artificial Intelligence
  • Footnotes
    Acknowledgements  This research was supported by NSF CAREER BCS-1942438 (TK), and an NDSEG research fellowship (JSP).
Journal of Vision September 2024, Vol.24, 1223. doi:https://doi.org/10.1167/jov.24.10.1223
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Jacob S. Prince, Christopher Hamblin, George A. Alvarez, Talia Konkle; Interpreting distributed population codes with feature-accentuated visual encoding models. Journal of Vision 2024;24(10):1223. https://doi.org/10.1167/jov.24.10.1223.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

A typical view of a natural scene will often contain many different people and objects in a broader surrounding environment, generating an expansive pattern of activity along the ventral visual stream. Longstanding highly productive paradigms in visual neuroscience have focused on understanding different regions of the ventral stream in isolation, by identifying the kinds of stimuli that activate each highly (e.g. in an fMRI localizer study). As such, these approaches do not directly assess how different parts of the broader population code operate in parallel to encode a single complex natural image. Here we introduce a new analytical paradigm aimed at this goal. First, we fit voxel-wise encoding models using the Natural Scenes Dataset and focus our analysis on voxels whose responses are accurately predicted for new images. Then, we apply a new interpretability method called “feature accentuation”, which identifies the features of an image that are critical for driving a voxel’s response, by synthesizing a new version of the image with the relevant features emphasized. As a proof of concept, we show that in everyday images of people in different scene contexts––where both face- and scene-selective voxels are moderately active––we can attribute the activation of face-selective voxels to the people within the scene, and the scene-selective voxels to the surrounding scene context, all within the same image. These initial demonstrations offer a roadmap for subsequent analyses along high-level visual cortex, especially targeting voxels with less-well-understood tuning properties. Critically, this method is general, effective for any voxel or neuron over any image, without presupposing specific content distinctions or tuning ahead of time. As such, this analytical approach enables the dissection of the joint operation of a distributed activation profile, which may provide new insight into how the ventral stream encodes a glance of the rich, complex visual world.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×