August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
Statistical inference on representational geometries
Author Affiliations & Notes
  • Heiko Schütt
    New York University
    Zuckerman Institute, Columbia University
  • Alexander D. Kipnis
    Max Planck Institute for Biological Cybernetics
  • Jörn Diedrichsen
    Western University
  • Nikolaus Kriegeskorte
    Zuckerman Institute, Columbia University
  • Footnotes
    Acknowledgements  This study was funded in part by the German Research Foundation (DFG) through grant SCHU 3351/1-1 to HHS.
Journal of Vision August 2023, Vol.23, 5749. doi:https://doi.org/10.1167/jov.23.9.5749
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Heiko Schütt, Alexander D. Kipnis, Jörn Diedrichsen, Nikolaus Kriegeskorte; Statistical inference on representational geometries. Journal of Vision 2023;23(9):5749. https://doi.org/10.1167/jov.23.9.5749.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Models and brain measurements of visual processing have substantially increased in complexity in recent years. Summarizing and comparing their high-dimensional representations to each other requires specialized statistical methods. Here, we introduce new inference methods to evaluate models based on their predicted representational geometries, i.e. based on how well they match the distances or dissimilarities between the representations. Our inference methods are based on cross-validation and bootstrapping. We introduce a novel 2-factor bootstrap technique wrapped around a cross-validation procedure with analytically derived adjustments for the biases induced by 2-factor inflation of measurement noise and the choice of cross-validation folds. We validate our new inference methods using extensive simulations. We first simulate fMRI-like data based on local averages of deep neural network activations for images sampled from ecoset. In these simulations, we have full access to the true data generating process and can thus test a wide range of experiments. Additionally, we performed simulations based on subsampling data from large scale calcium imaging and fMRI experiments. These simulations are less flexible, but we are more confident that the patterns and their variability are representative of true experimental data. In all simulations, our new methods yield good estimates of the variance of model evaluations and thus valid statistical tests. In contrast, uncorrected bootstrap methods substantially overestimate variance and thus yield overly conservative tests. Also, ignoring the desired generalization to new stimuli leads to underestimated variance and thus to overly liberal tests. Similar statistical problems occur whenever other bootstrap methods aim to generalize to new stimuli and new subjects and/or are combined with cross-validation. Our new methods are available as part of the open source rsatoolbox in python at https://github.com/rsagroup/rsatoolbox.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×