September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Comparing Convolutional Neural Networks to Traditional Models of Covert Attention During Visual Search
Author Affiliations
  • Ansh K. Soni
    University of California - Santa Barbara
  • Sudhanshu Srivastava
    University of California - Santa Barbara
  • Miguel P. Eckstein
    University of California - Santa Barbara
Journal of Vision September 2024, Vol.24, 1497. doi:https://doi.org/10.1167/jov.24.10.1497
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ansh K. Soni, Sudhanshu Srivastava, Miguel P. Eckstein; Comparing Convolutional Neural Networks to Traditional Models of Covert Attention During Visual Search. Journal of Vision 2024;24(10):1497. https://doi.org/10.1167/jov.24.10.1497.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Introduction: Performance degradation in Visual Search with an increased number of distractors (the set-size effect) is often used to make inferences about the properties of covert attention. Recently, studies have utilized Convolutional Neural Network (CNN) models (Srivastava, 2023; Nicholson, 2022; Poder, 2022), although their relationship to traditional models is not well understood. Our goal is to compare CNN models to traditional models of visual search. Methods: We implement six models of covert attention during visual search to predict target detection accuracy (yes/no) in feature and conjunction tasks. For each model, we adjust the two feature searches (line angle or luminance) to match model performance. These feature values were used to build the conjunction task (angle and luminance). Three of the models are image-computable, acting directly on the pixels of the images: Image computable Bayesian ideal observer (IC-BIO), a fully-trained 5-layer CNN, and a large network trained on image classification (VGG-16 network) with transfer learning. The remaining three models operated on assumed extracted feature values (normally distributed activations from the target and each distractor): a Signal Detection Theory (SDT; Green and Swets, 1966) model without or with capacity limits (SDTc; Poder, 2019) and a Guided Search Accuracy model (GSA; Wolfe 1989; Eckstein, 2000). Results: For feature and conjunction search, the CNN and VGG-16 models showed similar set-size effects to unlimited capacity models (IC-BIO/SBIO and SDT) and smaller than the models with capacity limitations (SDTc and GSA with serial attention). The accuracy degradation from feature to conjunction search was lowest to highest for: IC-BIO, CNN, VGG-16, SDT (assuming independent processing of features; Ecktein, 1998), and SDTc. Conclusion: Our findings benchmark newer CNN models against traditional search models, showing a correspondence between CNN set-size effects and Signal Detection/Ideal Observer models but distinct feature/conjunction search accuracy relationships.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×