September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Spatial filters in neural network models of visual cortex do not need to be learned
Author Affiliations
  • Ananya Passi
    Johns Hopkins University
  • Atlas Kazemian
    Johns Hopkins University
  • Michael Bonner
    Johns Hopkins University
Journal of Vision September 2024, Vol.24, 895. doi:https://doi.org/10.1167/jov.24.10.895
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ananya Passi, Atlas Kazemian, Michael Bonner; Spatial filters in neural network models of visual cortex do not need to be learned. Journal of Vision 2024;24(10):895. https://doi.org/10.1167/jov.24.10.895.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

The performance of convolutional neural networks as models of visual cortex relies on pre-training millions of parameters, optimizing them for a specific classification task. This process not only requires massive computational resources, but also results in learned features whose effectiveness is limited by the dataset richness. Furthermore, the time and resource intensive nature of this training process discourages iterative parameter studies, further reducing the interpretability of high-performing models of visual cortex. Here we propose a theoretically grounded convolutional architecture in which the training process is limited to learning linear combinations of pre-defined wavelet filters. This simplified model is based on an iterative process of expanding and subsequently reducing dimensionality in a deep hierarchy of modules, where each module consists of a filtering operation, followed by a non-linearly and channel mixing. We show that this model rivals a traditional pre-trained CNN in explaining stimuli-evoked neural responses to natural scenes in the human visual cortex Our model generates a useful set of features that can be combined to extract information from a wide range of stimuli, and it reduces the number of learned parameters by orders of magnitude. This model can enable neuroscientists to more efficiently perform in-silico analyses and controlled rearing experiments on deep learning models. Moreover, it can also give insights about how visual computation occurs in the brain, owing to its simple organization and reduced dependence on training.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×