August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
A deep convolutional neural network trained to infer surface reflectance is deceived by mid-level lightness illusions
Author Affiliations & Notes
  • Jaykishan Patel
    York University
  • Alban Flachot
    York University
  • Javier Vazquez-Corral
    Universitat Autònoma de Barcelona
  • David H. Brainard
    University of Pennsylvania
  • Thomas S. A. Wallis
    Technische Universität Darmstadt
  • Marcus A. Brubaker
    York University
  • Richard F. Murray
    York University
  • Footnotes
    Acknowledgements  Funded in part by grants from NSERC and VISTA to R.F.M.
Journal of Vision August 2023, Vol.23, 4817. doi:https://doi.org/10.1167/jov.23.9.4817
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Jaykishan Patel, Alban Flachot, Javier Vazquez-Corral, David H. Brainard, Thomas S. A. Wallis, Marcus A. Brubaker, Richard F. Murray; A deep convolutional neural network trained to infer surface reflectance is deceived by mid-level lightness illusions. Journal of Vision 2023;23(9):4817. https://doi.org/10.1167/jov.23.9.4817.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

A long-standing view is that lightness illusions are by-products of strategies employed by the visual system to stabilize its perceptual representation of surface reflectance against changes in illumination. Computationally, one such strategy is to infer reflectance from the retinal image, and to base the lightness percept on this inference. CNNs trained to infer reflectance from images have proven successful at solving this problem under limited conditions. To evaluate whether these CNNs provide suitable starting points for computational models of human lightness perception, we tested a state-of-the-art CNN on several lightness illusions, and compared its behaviour to prior measurements of human performance. We trained a CNN (Yu & Smith, 2019) to infer reflectance from luminance images. The network had a 30-layer hourglass architecture with skip connections. We trained the network via supervised learning on 100K images, rendered in Blender, each showing randomly placed geometric objects (surfaces, cubes, tori, etc.), with random Lambertian reflectance patterns (solid, Voronoi, or low-pass noise), under randomized point+ambient lighting. The renderer also provided the ground-truth reflectance images required for training. After training, we applied the network to several visual illusions. These included the argyle, Koffka-Adelson, snake, White’s, checkerboard assimilation, and simultaneous contrast illusions, along with their controls where appropriate. The CNN correctly predicted larger illusions in the argyle, Koffka-Adelson, and snake images than in their controls. It also correctly predicted an assimilation effect in White's illusion. It did not, however, account for the checkerboard assimilation or simultaneous contrast effects. These results are consistent with the view that at least some lightness phenomena are by-products of a rational approach to inferring stable representations of physical properties from intrinsically ambiguous retinal images. Furthermore, they suggest that CNN models may be a promising starting point for new models of human lightness perception.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×