October 2020
Volume 20, Issue 11
Open Access
Vision Sciences Society Annual Meeting Abstract  |   October 2020
Configural processing of 2D shape
Author Affiliations & Notes
  • Shaiyan Keshvari
    York University
  • Ingo Fruend
    York University
  • James H. Elder
    York University
  • Footnotes
    Acknowledgements  VISTA Postdoctoral Fellowship
Journal of Vision October 2020, Vol.20, 1647. doi:https://doi.org/10.1167/jov.20.11.1647
  • Views
  • Share
  • Tools
    • Alerts
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Shaiyan Keshvari, Ingo Fruend, James H. Elder; Configural processing of 2D shape. Journal of Vision 2020;20(11):1647. https://doi.org/10.1167/jov.20.11.1647.

      Download citation file:

      © ARVO (1962-2015); The Authors (2016-present)

  • Supplements

Background: Deep network models are relatively successful at predicting both performance and neural activations on object recognition tasks. However, recent work suggests that these models rely largely on local features rather than global shape, whereas humans, while sensitive to local 2D shape (curvature), are easily able to discriminate natural shapes from synthetic shapes with matched curvature statistics. Here we assess two alternative shape models that could account for this human sensitivity to 2D shape beyond local curvature: 1) Pooling – shape information is pooled over a collection of independently-coded fragments or parts; 2) Configural – the representation depends on the arrangement of these parts over the entire shape. Method: We employed a dataset of 2D animal shapes approximated as 120-segment outline polygons and local ‘metamers’ – closed contours that match the local curvature statistics of the animal shapes. In a two-interval task, five observers discriminated between a stimulus containing only animal contour fragments and a second stimulus containing only metamer fragments, while the length of the fragments was varied from 2 segments (local) to 120 segments (global). There were two conditions: 1. A single fragment displayed centrally. 2. Multiple fragments displayed within a 7.5 deg circular window. The number of fragments was selected to yield a total of 120 turning angles, matching the full-shape condition. Results: For both single- and multi-fragment conditions, performance rises from chance to near 100% as fragment length increases from 2 to 120, reflecting human sensitivity to 2D shape beyond local curvature. Interestingly, there is little difference in the psychometric functions for the single- and multi-fragment conditions (75%-correct thresholds of 24 +/- 7 vs 18 +/- 6 segments), indicating very little pooling across fragments. This suggests that human shape perception is highly configural, posing a challenge to recent deep learning accounts of object coding.


This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.