September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Does Leveraging the Human Ventral Visual Stream Improve Neural Network Robustness?
Author Affiliations & Notes
  • Zhenan Shao
    Department of Psychology, University of Illinois Urbana-Champaign
    Beckman Institute, University of Illinois Urbana-Champaign
  • Linjian Ma
    Department of Computer Science, University of Illinois Urbana-Champaign
  • Bo Li
    Department of Computer Science, University of Illinois Urbana-Champaign
    Department of Computer Science, University of Chicago
  • Diane M. Beck
    Department of Psychology, University of Illinois Urbana-Champaign
    Beckman Institute, University of Illinois Urbana-Champaign
  • Footnotes
    Acknowledgements  This work used NCSA Delta GPU through allocation SOC230011 from the Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program, which is supported by National Science Foundation grants #2138259, #2138286, #2138307, #2137603, and #2138296.
Journal of Vision September 2024, Vol.24, 277. doi:https://doi.org/10.1167/jov.24.10.277
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Zhenan Shao, Linjian Ma, Bo Li, Diane M. Beck; Does Leveraging the Human Ventral Visual Stream Improve Neural Network Robustness?. Journal of Vision 2024;24(10):277. https://doi.org/10.1167/jov.24.10.277.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Human object recognition is robust to a variety of object transformations, including changes in lighting, rotations, and translations, as well as other image manipulations, including the addition of various forms of noise. Invariance has been shown to emerge gradually along the ventral visual stream with later regions showing higher tolerance to object transformations. In contrast, despite their unprecedented performance on numerous visual tasks, Deep Neural Networks (DNNs) fall short in achieving human-level robustness to image perturbations (adversarial attacks), even those that are visually imperceptible to humans. One potential explanation for this difference is that brains, but not DNNs, build increasingly disentangled and therefore robust object representations with each successive stage of the ventral visual stream. Here, we asked whether training DNNs to emulate human representation can enhance their robustness and, more importantly, whether different stages of the ventral visual stream enable progressively increased robustness, reflecting the potentially evolving representation crucial for human perceptual invariance. We extracted neural activity patterns in five hierarchical regions of interest (ROIs) in the ventral visual stream: V1, V2, V4, LO, and TO from a 7T fMRI dataset (Allen et al., 2022) obtained when human participants viewed natural images. DNN models were trained to perform image classification tasks while aligning their penultimate layer representations with neural activity from each ROI. Our findings reveal not only a significant improvement in DNN robustness but also a hierarchical effect: greater robustness gains were observed when trained with neural representations from later stages of the visual hierarchy. Our results not only show that ventral visual cortex representations improve DNN robustness but also support the gradual emergence of robustness along the ventral visual stream.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×