September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Investigating the neural computations underlying visual social inference with graph neural network and inverse planning models
Author Affiliations & Notes
  • Manasi Malik
    Johns Hopkins University
  • Tianmin Shu
    Johns Hopkins University
  • Josh Tenenbaum
    Massachusetts Institute of Technology
  • Leyla Isik
    Johns Hopkins University
  • Footnotes
    Acknowledgements  This work was funded by NIMH R01MH132826 awarded to L.I.
Journal of Vision September 2024, Vol.24, 352. doi:https://doi.org/10.1167/jov.24.10.352
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Manasi Malik, Tianmin Shu, Josh Tenenbaum, Leyla Isik; Investigating the neural computations underlying visual social inference with graph neural network and inverse planning models. Journal of Vision 2024;24(10):352. https://doi.org/10.1167/jov.24.10.352.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Recognizing social interactions in visual scenes is a crucial human ability, however, the neural computations that enable this remain undetermined. Prior work has shown that two distinct computational models, a bottom-up graph neural network (SocialGNN) based solely on visual information and a generative inverse planning model (SIMPLE) based on mental state inference, both uniquely explain significant variance in human judgements. Here, we compare both models to neural data to understand how the brain combines these two types of computations for social scene understanding. We collected fMRI data from adults while they watched videos of two animated agents interacting. We compared neural representations with human behavior judgements and each computational model of social interaction recognition. Preliminary analysis using whole-brain searchlight RSA showed a significant correlation between neural RDMs and behavioral RDM in the visual cortex, lateral occipital temporal cortex and the superior temporal sulcus (STS). With the computational models, we find that SocialGNN exhibited a significantly higher correlation than SIMPLE in more posterior and dorsal regions including the lateral occipital cortex (LOTC) and posterior STS - regions previously implicated in social perception. On the other hand, SIMPLE demonstrated significantly higher correlation than SocialGNN in more anterior regions, including anterior STS and medial prefrontal cortex (mPFC). Further, both SocialGNN and SIMPLE explain significant variance in posterior and mid regions of the STS suggesting these regions as a potential site of integration of social perception and mental state inference. This work provides a novel framework for testing computational theories of social perception and cognition, as well as preliminary evidence for how the brain combines bottom-up vision and mental state inference during social scene understanding.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×