September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Comparing artificial neural network models with varied objectives to probe the role of sensory representation in primate visual memorability
Author Affiliations & Notes
  • Ram Ahuja
    Department of Biology, York University
  • Kohitij Kar
    Department of Biology, York University
  • Footnotes
    Acknowledgements  Google Research, CFREF, Brain Canada, SFARI
Journal of Vision September 2024, Vol.24, 1239. doi:https://doi.org/10.1167/jov.24.10.1239
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ram Ahuja, Kohitij Kar; Comparing artificial neural network models with varied objectives to probe the role of sensory representation in primate visual memorability. Journal of Vision 2024;24(10):1239. https://doi.org/10.1167/jov.24.10.1239.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Imagine a typical experience of scrolling through numerous photos on social media. While most images blend into obscurity, certain ones, like a playful kitten tangled in yarn, linger in our memories. This selective retention, known as image memorability, raises intriguing questions about its neural basis, particularly how the sensory representations in the cortex facilitate this behavior. In this study, we investigate this process, leveraging recent strides in computer vision through artificial neural network (ANN) models. These models, designed to mimic the ventral visual pathway in primates, offer a vast hypotheses space for the brain function that are critical for determining the memorability of an image. We compare two distinct types of ANNs: models geared toward basic object categorization (e.g., ResNet-50, AlexNet, GoogLeNet) and those tailored to predict image memorability (e.g., MemNet, ViTMem) on a set of 200 images (20 images from 10 distinct object categories) from the MS-COCO dataset. Consistent with previous results, we observed that both these model classes can predict which images humans find most memorable. However, our results show that they produce significantly different internal representations as assessed by representational similarity analysis (comparing representations across model classes in architecture-specific and non-specific manner). In addition, here, we used neural data recorded across the macaque inferior temporal (IT) in 6 monkeys to address two primary questions. 1) Can both model classes accurately predict variance in the neural data? 2) Do they predict distinct variances in the responses of the individual neurons? Surprisingly, our results show that memorability models (MemNet) predict significantly more variance in the neural data than object categorization models with matched architecture (AlexNet) – emphasizing a) the need to further probe these model classes as encoding models of the ventral stream, b) provide a new normative framework to think about the evolution of sensory representations.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×