August 2009
Volume 9, Issue 8
Free
Vision Sciences Society Annual Meeting Abstract  |   August 2009
Task-driven saliency using natural statistics (SUN)
Author Affiliations
  • Matthew Tong
    Department of Computer Science and Engineering, University of California at San Diego
  • Christopher Kanan
    Department of Computer Science and Engineering, University of California at San Diego
  • Lingyun Zhang
    Department of Computer Science and Engineering, University of California at San Diego
  • Garrison Cottrell
    Department of Computer Science and Engineering, University of California at San Diego
Journal of Vision August 2009, Vol.9, 392. doi:https://doi.org/10.1167/9.8.392
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Matthew Tong, Christopher Kanan, Lingyun Zhang, Garrison Cottrell; Task-driven saliency using natural statistics (SUN). Journal of Vision 2009;9(8):392. https://doi.org/10.1167/9.8.392.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Based on an assumption that one main goal of the visual attention system is to direct attention towards objects of interest, we have derived a probabilistic model of salience. The resulting model, Saliency Using Natural Statistics (SUN), is grounded in probabilities learned through the experience of the natural world. These probabilities decompose into three parts: knowledge about what features are rare or novel, the visual appearance of particular objects of interest, and where those objects are likely to occur in a scene. Bottom-up saliency is defined by SUN as rare combinations of features, and an implementation of this component of the SUN model has been shown to achieve state-of-the-art performance predicting human eye-movement fixations when free viewing static images (Zhang et al., in press) and video. SUN's bottom-up saliency model also predicts visual search asymmetries that other models of bottom-up salience based only on the current image fail to capture. However, when interacting with the world, we do so with a focus. Models of visual attention likewise need to be driven by the task at hand. Here we implement the remaining portions of SUN, a location prior which guides attention to likely locations of the target and a probabilistic appearance model in the spirit of the Guided Search (Wolfe, 1994) and Iconic Search (Rao et al., 1996) models. We evaluate our model on the publicly available dataset of Torralba et al. (2006) that contains eye-tracking data collected from subjects asked to count people, cups, or paintings in indoor or outdoor scenes. We show that the full SUN model achieves superior performance in predicting human fixations, suggesting that learned knowledge of targets' appearance, location, and the rarity of features play a role in determining where to fixate.

Tong, M. Kanan, C. Zhang, L. Cottrell, G. (2009). Task-driven saliency using natural statistics (SUN) [Abstract]. Journal of Vision, 9(8):392, 392a, http://journalofvision.org/9/8/392/, doi:10.1167/9.8.392. [CrossRef]
Footnotes
 NIH, NSF, and the James S. McDonnell Foundation.
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×