September 2019
Volume 19, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2019
Gaze attraction toward higher-order image features generated by deep convolutional neural network
Author Affiliations & Notes
  • Rina Watanabe
    Faculty of Informatics and Engineering, The University of Electro-Communications
  • Tomohiro Nishino
    Faculty of Informatics and Engineering, The University of Electro-Communications
  • Kazuaki Akamatsu
    Graduate School of Informatics and Engineering, The University of Electro-Communications
  • Yoichi Miyawaki
    Graduate School of Informatics and Engineering, The University of Electro-Communications
    JST, PRESTO
Journal of Vision September 2019, Vol.19, 47c. doi:https://doi.org/10.1167/19.10.47c
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Rina Watanabe, Tomohiro Nishino, Kazuaki Akamatsu, Yoichi Miyawaki; Gaze attraction toward higher-order image features generated by deep convolutional neural network. Journal of Vision 2019;19(10):47c. https://doi.org/10.1167/19.10.47c.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Humans acquire visual information of the external environment while moving their eyes sequentially from one location to another. Previous studies showed that salient locations attract human gaze frequently (Itti, Koch, 1998), but more recent evidence suggests that higher-order image features might have higher predictability of gaze frequency (Kümmerer et al., 2016) and temporal characteristics (Kümmerer et al., 2017; Akamatsu, Miyawaki, 2018) than the classical saliency theory. However, it remains unclear whether higher-order image features per se serve as strong gaze attractors because previous experiments used natural scene images and the results could be influenced by semantic information from object categories and scene contexts. To resolve this issue, we designed a new experiment using “feature images” that can contain a pre-specified order of image features while suppressing object-categorical and scene-contextual information. The feature images were artificially generated so that they selectively maximized the response of a specific layer of a pre-trained deep convolutional neural network (DCNN), using gradient ascent optimization of image pixel values. Subject’s eye movement was recorded while they were observing a pair of feature images, each of which corresponded to a different layer of DCNN. Results showed that feature images corresponding to a higher layer of DCNN (higher-layer feature images) attracted the gaze more frequently than the simultaneously-presented lower-layer feature images, and the gaze frequency progressively increased as the DCNN layer tied to feature images. Control analyses confirmed that higher-layer feature images did not possess higher saliency and thus classical saliency theory is unlikely to explain the observed gaze frequency bias. These results suggest that higher-order image features serve as a significant gaze attractor independently of semantic information embedded in natural scenes.

Acknowledgement: JST PRESTO (JPMJPR1778), JSPS KAKENHI (17H01755) 
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×