September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Evaluating the Alignment of Machine and Human Explanations in Visual Object Recognition through a Novel Behavioral Approach
Author Affiliations & Notes
  • Yousif Kashef Alghetaa
    Department of Biology, York University
  • Simon Kornblith
    Anthropic PBC
  • Kohitij Kar
    Department of Biology, York University
  • Footnotes
    Acknowledgements  Google Research, CFREF, Brain Canada, SFARI
Journal of Vision September 2024, Vol.24, 1259. doi:https://doi.org/10.1167/jov.24.10.1259
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Yousif Kashef Alghetaa, Simon Kornblith, Kohitij Kar; Evaluating the Alignment of Machine and Human Explanations in Visual Object Recognition through a Novel Behavioral Approach. Journal of Vision 2024;24(10):1259. https://doi.org/10.1167/jov.24.10.1259.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Understanding how computer vision models make decisions is paramount, particularly with increasing scrutiny from various institutions. The field of Explainable Artificial Intelligence (XAI) provides tools to interpret these model decisions, but the explanations are often at odds. Kar et al. (2022) suggested evaluating the goodness of machine explanations based on their alignment with human cognitive processes. This study builds on that concept, addressing the challenge of reliably approximating human explanations, a task complicated by the limitations of existing psychophysical tools like 'bubbles' and classification-images. Our study introduces a novel method to assess the alignment between human and machine explanations in object discrimination tasks. We establish a two-model framework: a target (ResNet-50, whose explanations are under scrutiny) and a reference model (a fully differentiable model, AlexNet, as a stand-in for humans). The objective is to eventually compare the target model's explanations with human explanations. We begin by analyzing feature attribution maps (heat maps showing how image features influence model outputs) from both models. We compare these maps using various metrics to create a baseline ranking of explanation similarity between ResNet-50 and AlexNet. Following this, we create explanation-masked images (EMIs) by retaining only the most informative pixels based on ResNet-50's (Target) feature attributions. We hypothesize that the impact of these EMIs on both model behaviors could reflect the similarity of their underlying explanations. We then estimate the object discrimination accuracy of both ResNet-50 and AlexNet on these EMIs. The correlation between their performances provides a ranking of explanation similarity. Our results showed a significant correlation (Spearman R=0.65, p=0.003), indicating a strong alignment between the two models' explanations. This finding sets the stage for extending our method to human subjects, using their behavioral responses to EMIs to evaluate the accuracy of ResNet-50's explanations, offering a new direction for comparing machine and human explanations.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×