August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
Reward Maps Predict Target-present and Target-absent Visual Search
Author Affiliations & Notes
  • Gregory J. Zelinsky
    Stony Brook University
  • Seoyoung Ahn
    Stony Brook University
  • Zhibo Yang
    Stony Brook University
  • Yupei Chen
    Stony Brook University
  • Sounak Mondal
    Stony Brook University
  • Minh Hoai
    Stony Brook University
  • Dimitrios Samaras
    Stony Brook University
  • Footnotes
    Acknowledgements  We would like to thank the National Science Foundation for supporting this work through award 1763981.
Journal of Vision August 2023, Vol.23, 5161. doi:https://doi.org/10.1167/jov.23.9.5161
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Gregory J. Zelinsky, Seoyoung Ahn, Zhibo Yang, Yupei Chen, Sounak Mondal, Minh Hoai, Dimitrios Samaras; Reward Maps Predict Target-present and Target-absent Visual Search. Journal of Vision 2023;23(9):5161. https://doi.org/10.1167/jov.23.9.5161.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Predictions of attention control are typically formalized as priority maps, and many types have been proposed (e.g., saliency maps). We add to this list by introducing a reward map, a map of the reward expected by fixating different locations in an image. In the context of goal-directed attention, our premise is that achieving a goal is rewarding and that goal-directed fixations are controlled by computing expected reward and attempting to maximize its receipt. We obtained reward maps for the COCO-Search18 target-object categories using Inverse-Reinforcement-Learning (IRL), which learns from observations of search fixations (in the training data) a reward function for predicting the scanpaths of fixations made by people searching scenes for categories of target objects. Reward maps are therefore priority maps that are reverse-engineered from the training fixations using IRL. From this purely data-driven approach, we found that reward maps explained the combined variability of saliency maps, target maps, object maps, and meaning maps in predicting target-present and target-absent search behavior, thus supporting our hypothesis that the pursuit of expected reward is the common thread stitching together these other attention biases. Moreover, scanpath predictions from reward maps approached the noise ceiling imposed by agreement in participant behavior (meaning there is little room for improvement) and came close to achieving state-of-the-art (SOTA) against scanpath-prediction models in computer vision. This SOTA, however, is a black box, whereas predictions from reward maps are highly interpretable as reward. We conclude that goal-directed attention control can be understood as seeking out expected goal-related reward, and that a reward map may be THE priority map—the common priority representation into which other bottom-up and top-down biases collectively exert their control over behavior. Our work also enables scanpath prediction researchers to weigh interpretability benefits against the (often negligible) performance costs incurred by making biologically plausible modeling decisions.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×