Open Access
Article  |   November 2018
The role of global cues in the perceptual grouping of natural shapes
Author Affiliations
Journal of Vision November 2018, Vol.18, 14. doi:https://doi.org/10.1167/18.12.14
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      James H. Elder, Timothy D. Oleskiw, Ingo Fruend; The role of global cues in the perceptual grouping of natural shapes. Journal of Vision 2018;18(12):14. https://doi.org/10.1167/18.12.14.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Perceptual grouping of the bounding contours of objects is a crucial step in visual scene understanding and object recognition. The standard perceptual model for this task, supported by a convergence of physiological and psychophysical evidence, is based upon an association field that governs local grouping, and a Markov or transitivity assumption that allows global contours to be inferred solely from these local cues. However, computational studies suggest that these local cues may not be sufficient for reliable identification of object boundaries in natural scenes. Here we employ a novel psychophysical method to assess the potential role of more global factors in the perceptual grouping of natural object contours. Observers were asked to detect briefly presented fragmented target contours in oriented element noise. We employed natural animal shape stimuli, which in addition to local grouping cues possess global regularities that could potentially be exploited to guide grouping and thereby improve target detection performance. To isolate the role of these global regularities we contrasted performance with open and closed control target stimuli we call local metamers, as they afford the same local grouping cues as animal shapes. We found that performance for closed metamers exceeded performance for open metamers, while performance for animal targets exceeded both, indicating that global grouping cues represented in higher visual areas codetermine the association between orientation signals coded in early visual cortex. These results demand a revision to the standard model for perceptual grouping of contours to accommodate feedback from higher visual areas coding global shape properties.

Introduction
Real-world scenes can be highly cluttered, often causing an object to fragment into many pieces before it hits our retinae. Perceptual grouping is the process of reassociating these fragments to generate the percept of a whole object. Here we consider in particular the problem of grouping the fragments of the bounding contour based on their geometric relationships. 
While the shape of these contour fragments can be arbitrary, approximating them as piecewise linear allows the problem to be modeled as the grouping of local orientation signals, coded by neurons in primary visual cortex (Hubel & Wiesel, 1968; Heitger, Rosenthaler, von der Heydt, Peterhans, & Kulber, 1992), to form a sequence or chain that describes the entire contour (Figure 1). Note the combinatorial nature of the problem: Given n oriented elements in the image, there are n!/(2(nk)!) different sequences of length k to consider. This is far too many to evaluate for most contours in complex natural scenes—the brain must employ some prior knowledge or principles to limit the search. 
Figure 1
 
The problem of contour grouping. Neurons in primary visual cortex (V1) code the local orientation of contours in the image, but without explicit representation of how they are linked together. Our percept, on the other hand, is of coherent and complete objects, segmented from their background.
Figure 1
 
The problem of contour grouping. Neurons in primary visual cortex (V1) code the local orientation of contours in the image, but without explicit representation of how they are linked together. Our percept, on the other hand, is of coherent and complete objects, segmented from their background.
The study of these perceptual organization principles has a long history stretching back to the early days of Gestalt psychology (Koffka, 1935; Wertheimer, 1938). Two key Gestalt principles relevant to contour grouping are proximity and good continuation. We will refer to these grouping cues as local cues because they can be defined for pairs of neighboring contour elements. In particular, proximity can be encoded in terms of the distance dij between elements i and j, and good continuation can be coded in terms of the angles αij, βij induced by a linear interpolation between the elements (Figure 2a). 
Figure 2
 
(a) Local geometric cues to contour grouping. Smoothness properties of objects give rise to the classical Gestalt cues of proximity and good continuation (Wertheimer, 1938). For a pair of local oriented elements these can be expressed through a variable d encoding the separation of the elements and two angles (α,β) induced by a linear interpolation between the elements. (b) Posterior probability, derived from ecological statistics, that a local oriented element (shown in black) should be grouped directly with a second observed element, as a function of its proximity, direction, and relative orientation (Elder & Goldberg 2002). (c) Local Gestalt cues, together with a Markov independence assumption, lead to an efficient framework for extracting extended contours.
Figure 2
 
(a) Local geometric cues to contour grouping. Smoothness properties of objects give rise to the classical Gestalt cues of proximity and good continuation (Wertheimer, 1938). For a pair of local oriented elements these can be expressed through a variable d encoding the separation of the elements and two angles (α,β) induced by a linear interpolation between the elements. (b) Posterior probability, derived from ecological statistics, that a local oriented element (shown in black) should be grouped directly with a second observed element, as a function of its proximity, direction, and relative orientation (Elder & Goldberg 2002). (c) Local Gestalt cues, together with a Markov independence assumption, lead to an efficient framework for extracting extended contours.
What are the brain mechanisms underlying these Gestalt principles? Physiological experiments first in cat (Gilbert & Wiesel, 1989) and later in nonhuman primate (Li & Gilbert, 2002; Stettler, Das, Bennett, & Gilbert, 2002; Li, Piech, & Gilbert, 2006) have revealed a striking bias in the long-range horizontal connections between orientation columns in primary visual cortex that serves to connect neurons coding local orientation signals that can be smoothly interpolated, consistent with the classical Gestalt principles of proximity and good continuation. Psychophysical experiments suggest that these connections serve to facilitate geometrically consistent orientation responses in the human striate cortex (Cass & Spehar, 2005). 
These physiological results have motivated more detailed computational models of the neural circuits in primary visual cortex responsible for contour grouping (e.g., Li, 1998; Yen & Finkel, 1998). In parallel, computational research has shown that these “cocircularity” cues can be helpful in guiding the extraction of contours from natural imagery (Parent & Zucker, 1989), and psychophysical studies revealed that the human visual system is highly sensitive to these local geometric relationships (Field, Hayes, & Hess, 1993). Finally, statistical models of these relationships rooted in empirical measurements of natural imagery have yielded accurate quantitative normative models for these local Gestalt cues that validate their inferential power (Figure 2b; Kruger, 1998; Geisler, Perry, Super, B. J., & Gallogly, 2001; Elder & Goldberg, 2002; Sigman, Cecchi, Gilbert, & Magnasco, 2007; Geisler & Perry, 2009). Together, these converging results have established a local “Association Field” model rooted in the circuitry of early visual cortex as the standard perceptual model for contour grouping. 
Although these local Gestalt cues directly relate only pairs of contour elements, given certain assumptions they can be used to infer global contours spanning many elements (Elder & Zucker, 1996; Elder, Krupnik, & Johnston, 2003; Elder, 2013, 2015). To see this inference, it is helpful to introduce some notation. Let yij = {dij, αij, βij} represent the local Gestalt cues relating elements i and j. Now consider the hypothesis Display Formula\(\def\upalpha{\unicode[Times]{x3B1}}\)\(\def\upbeta{\unicode[Times]{x3B2}}\)\(\def\upgamma{\unicode[Times]{x3B3}}\)\(\def\updelta{\unicode[Times]{x3B4}}\)\(\def\upvarepsilon{\unicode[Times]{x3B5}}\)\(\def\upzeta{\unicode[Times]{x3B6}}\)\(\def\upeta{\unicode[Times]{x3B7}}\)\(\def\uptheta{\unicode[Times]{x3B8}}\)\(\def\upiota{\unicode[Times]{x3B9}}\)\(\def\upkappa{\unicode[Times]{x3BA}}\)\(\def\uplambda{\unicode[Times]{x3BB}}\)\(\def\upmu{\unicode[Times]{x3BC}}\)\(\def\upnu{\unicode[Times]{x3BD}}\)\(\def\upxi{\unicode[Times]{x3BE}}\)\(\def\upomicron{\unicode[Times]{x3BF}}\)\(\def\uppi{\unicode[Times]{x3C0}}\)\(\def\uprho{\unicode[Times]{x3C1}}\)\(\def\upsigma{\unicode[Times]{x3C3}}\)\(\def\uptau{\unicode[Times]{x3C4}}\)\(\def\upupsilon{\unicode[Times]{x3C5}}\)\(\def\upphi{\unicode[Times]{x3C6}}\)\(\def\upchi{\unicode[Times]{x3C7}}\)\(\def\uppsy{\unicode[Times]{x3C8}}\)\(\def\upomega{\unicode[Times]{x3C9}}\)\(\def\bialpha{\boldsymbol{\alpha}}\)\(\def\bibeta{\boldsymbol{\beta}}\)\(\def\bigamma{\boldsymbol{\gamma}}\)\(\def\bidelta{\boldsymbol{\delta}}\)\(\def\bivarepsilon{\boldsymbol{\varepsilon}}\)\(\def\bizeta{\boldsymbol{\zeta}}\)\(\def\bieta{\boldsymbol{\eta}}\)\(\def\bitheta{\boldsymbol{\theta}}\)\(\def\biiota{\boldsymbol{\iota}}\)\(\def\bikappa{\boldsymbol{\kappa}}\)\(\def\bilambda{\boldsymbol{\lambda}}\)\(\def\bimu{\boldsymbol{\mu}}\)\(\def\binu{\boldsymbol{\nu}}\)\(\def\bixi{\boldsymbol{\xi}}\)\(\def\biomicron{\boldsymbol{\micron}}\)\(\def\bipi{\boldsymbol{\pi}}\)\(\def\birho{\boldsymbol{\rho}}\)\(\def\bisigma{\boldsymbol{\sigma}}\)\(\def\bitau{\boldsymbol{\tau}}\)\(\def\biupsilon{\boldsymbol{\upsilon}}\)\(\def\biphi{\boldsymbol{\phi}}\)\(\def\bichi{\boldsymbol{\chi}}\)\(\def\bipsy{\boldsymbol{\psy}}\)\(\def\biomega{\boldsymbol{\omega}}\)\(\def\bupalpha{\unicode[Times]{x1D6C2}}\)\(\def\bupbeta{\unicode[Times]{x1D6C3}}\)\(\def\bupgamma{\unicode[Times]{x1D6C4}}\)\(\def\bupdelta{\unicode[Times]{x1D6C5}}\)\(\def\bupepsilon{\unicode[Times]{x1D6C6}}\)\(\def\bupvarepsilon{\unicode[Times]{x1D6DC}}\)\(\def\bupzeta{\unicode[Times]{x1D6C7}}\)\(\def\bupeta{\unicode[Times]{x1D6C8}}\)\(\def\buptheta{\unicode[Times]{x1D6C9}}\)\(\def\bupiota{\unicode[Times]{x1D6CA}}\)\(\def\bupkappa{\unicode[Times]{x1D6CB}}\)\(\def\buplambda{\unicode[Times]{x1D6CC}}\)\(\def\bupmu{\unicode[Times]{x1D6CD}}\)\(\def\bupnu{\unicode[Times]{x1D6CE}}\)\(\def\bupxi{\unicode[Times]{x1D6CF}}\)\(\def\bupomicron{\unicode[Times]{x1D6D0}}\)\(\def\buppi{\unicode[Times]{x1D6D1}}\)\(\def\buprho{\unicode[Times]{x1D6D2}}\)\(\def\bupsigma{\unicode[Times]{x1D6D4}}\)\(\def\buptau{\unicode[Times]{x1D6D5}}\)\(\def\bupupsilon{\unicode[Times]{x1D6D6}}\)\(\def\bupphi{\unicode[Times]{x1D6D7}}\)\(\def\bupchi{\unicode[Times]{x1D6D8}}\)\(\def\buppsy{\unicode[Times]{x1D6D9}}\)\(\def\bupomega{\unicode[Times]{x1D6DA}}\)\(\def\bupvartheta{\unicode[Times]{x1D6DD}}\)\(\def\bGamma{\bf{\Gamma}}\)\(\def\bDelta{\bf{\Delta}}\)\(\def\bTheta{\bf{\Theta}}\)\(\def\bLambda{\bf{\Lambda}}\)\(\def\bXi{\bf{\Xi}}\)\(\def\bPi{\bf{\Pi}}\)\(\def\bSigma{\bf{\Sigma}}\)\(\def\bUpsilon{\bf{\Upsilon}}\)\(\def\bPhi{\bf{\Phi}}\)\(\def\bPsi{\bf{\Psi}}\)\(\def\bOmega{\bf{\Omega}}\)\(\def\iGamma{\unicode[Times]{x1D6E4}}\)\(\def\iDelta{\unicode[Times]{x1D6E5}}\)\(\def\iTheta{\unicode[Times]{x1D6E9}}\)\(\def\iLambda{\unicode[Times]{x1D6EC}}\)\(\def\iXi{\unicode[Times]{x1D6EF}}\)\(\def\iPi{\unicode[Times]{x1D6F1}}\)\(\def\iSigma{\unicode[Times]{x1D6F4}}\)\(\def\iUpsilon{\unicode[Times]{x1D6F6}}\)\(\def\iPhi{\unicode[Times]{x1D6F7}}\)\(\def\iPsi{\unicode[Times]{x1D6F9}}\)\(\def\iOmega{\unicode[Times]{x1D6FA}}\)\(\def\biGamma{\unicode[Times]{x1D71E}}\)\(\def\biDelta{\unicode[Times]{x1D71F}}\)\(\def\biTheta{\unicode[Times]{x1D723}}\)\(\def\biLambda{\unicode[Times]{x1D726}}\)\(\def\biXi{\unicode[Times]{x1D729}}\)\(\def\biPi{\unicode[Times]{x1D72B}}\)\(\def\biSigma{\unicode[Times]{x1D72E}}\)\(\def\biUpsilon{\unicode[Times]{x1D730}}\)\(\def\biPhi{\unicode[Times]{x1D731}}\)\(\def\biPsi{\unicode[Times]{x1D733}}\)\(\def\biOmega{\unicode[Times]{x1D734}}\)\({\cal H}\) that a particular sequence of elements i1,…, in forms a global contour, and let Display Formula\(\bar {\cal H}\) represent the negation of this hypothesis. The likelihood ratio L for the proposed grouping of elements can then be written as  
\begin{equation}\tag{1}L = {{p\left( {{y_{{i_1}{i_2}}}, \ldots ,{y_{{i_{n - 1}}{i_n}}}|{\cal H}} \right)} \over {p\left( {{y_{{i_1}{i_2}}}, \ldots ,{y_{{i_{n - 1}}{i_n}}}|\bar {\cal H}} \right)}}\end{equation}
If we assume conditional independence between the cues relating pairs of elements, the likelihoods can be factored, so that  
\begin{equation}\tag{2}\log L = \log \prod\limits_{k = 1}^{n - 1} {{{p\left( {{y_{{i_k}{i_{k + 1}}}}|{\cal H}} \right)} \over {p\left( {{y_{{i_k}{i_{k + 1}}}}|\bar {\cal H}} \right)}}} = \sum\limits_{k = 1}^{n - 1} {\log } {{p\left( {{y_{{i_k}{i_{k + 1}}}}|{\cal H}} \right)} \over {p\left( {{y_{{i_k}{i_{k + 1}}}}|\bar {\cal H}} \right)}}\end{equation}
 
This conditional independence assumption implies that the elements form a Markov chain in which, given a sequence of elements, the location and orientation of the next element depends only upon the location and orientation of the element immediately preceding it in the chain. This Markov property gives rise to an optimal substructure property that allows maximum probability contours to be identified in polynomial time despite the exponential size of the search space (Figure 2c). 
Although this Markov assumption has been used by computer vision algorithms to extract global contours from complex natural images (Elder & Zucker, 1996), there are a number of reasons to be skeptical that it represents a complete model of human perceptual grouping. First, it has been noted that these algorithms can fail in some cases where the human visual system seems to have little trouble (Movahedi & Elder, 2013). Second, the distribution of intrinsic distances between high-curvature events on natural contours has been shown to be incompatible with the Markov assumption (Ren, Fowlkes, & Malik, 2008). Third, the Markov assumption does not accommodate important global topological constraints such as closure and avoidance of self-intersections (Elder & Zucker, 1996; Elder et al., 2003). Finally, we note that Equation 2 implies that the whole is exactly equal to the sum of its parts: not a particularly Gestalt assertion! 
For all of these reasons, it seems likely that the standard model is incomplete, and that the human visual system is somehow capable of combining these important local cues with additional global cues for the perceptual grouping of contours. The goal of this paper is to psychophysically test this hypothesis. 
Has it not already been tested? One would think so. Particularly in the last 25 years there have been a number of excellent quantitative psychophysical studies on the perceptual determinants of contour grouping. Most of these employ a method due to Field, Hayes and Hess (1993), in which a sequence of oriented elements following a curvilinear path is embedded in uniformly distributed and randomly oriented distractor elements. The observer's task is to detect the path, and properties such as the spacing, orientation jitter, and photometric appearance of the elements can be varied to identify their influence on the underlying grouping mechanisms (e.g., Hess & Dakin, 1999; Geisler et al., 2001; Ledgeway, Hess, & Geisler, 2005; Watt, Ledgeway, & Dakin, 2008). However, in these studies the target contours are typically constructed to afford only local cues to grouping—the placement and orientation of each element on the path depends only upon the previous element. It is thus impossible to infer anything about global factors in the perceptual grouping of contours from these experiments. 
In order to test whether global cues play a role in contour grouping, we need to incorporate global cues into the target contour. But exactly which cues should be incorporated? There are many possibilities, including closure, convexity, parallelism and symmetry (Zhu, 1999; Elder, 2013, 2015, 2018). One potential approach is to take each global cue in turn and test for its role in isolation. Although this reductionist approach has advantages, these stimuli would not be typical of the complex natural objects we see around us, which afford a number of these cues at the same time. We therefore elect to use natural shape stimuli, in particular animal shapes. We know from prior work that animals in natural scenes are detected efficiently (Thorpe, Fize, & Marlot, 1996), and that rapid animal detection is driven in large part by shape cues (Elder & Velisavljević, 2009). These animal shapes exhibit a diversity of global regularities to which the human visual system may have adapted to support grouping and recognition in highly cluttered (e.g., forest) environments. The key to isolating the effect of these global regularities will be the careful design of control stimuli that afford the same local cues as our animal stimuli but none of the global cues. 
Methods
Stimuli
The stimuli were based upon 264 blue-screened images of animal objects taken from the Hemera Photo-Objects dataset. We tried to select shapes that were recognizable as animals, as opposed to more generic shapes (e.g., snakes). The original images had maximum dimension (either height or width) of 1,024 pixels; the other dimension varied depending upon the shape of the animal. The alpha channel for each image signaled figure/ground for the blue-screened animal object. Contours were extracted by tracing points with nonzero gradient magnitude in the alpha channel using Moore neighborhood tracing (Pradhan, Kumar, Agarwal, Pradhan, & Ghose, 2010), and each contour was then represented as a high resolution polygon with this sequence of points as vertices. The number of vertices N ranged from 1,506–4,668 over shapes. See Figure 3 for examples. 
Figure 3
 
Example animal shapes extracted from the Hemera photo-object database.
Figure 3
 
Example animal shapes extracted from the Hemera photo-object database.
From these high-resolution polygons, lower resolution polygons of length 2n were derived by selecting a random point on the original polygon as a first sample and then selecting 2n −1 additional samples at roughly equal arc-length intervals from this point. Alternate segments of these polygons were then labelled as “visible” and “gap” (Figure 4); rendering just the visible segments yields a roughly uniformly fragmented animal target shape of length n visible segments (Figure 5, top). In what follows we refer to the number of visible segments n as the target length; this was varied in order to assess how the influence of nonlocal factors in perceptual grouping and target detection depends upon level of detail. 
Figure 4
 
Alternating visible and gap segments on polygonal target shapes.
Figure 4
 
Alternating visible and gap segments on polygonal target shapes.
Figure 5
 
Example stimuli.
Figure 5
 
Example stimuli.
Each of the gap segments in a polygon forms two angles with respect to the adjacent visible segments. Let the gap length and two angles identified with gap segment i be denoted di, αi, and βi, respectively (Figure 4). Note that the visible and gap segments varied slightly in length. 
Our experiments assessed human ability to detect these animal shape targets in noise, compared with various forms of local metamer targets that have the same first-order statistical geometric properties between neighboring elements but lack the global regularities of natural shapes (Figure 5). We detail the method for constructing these metamer stimuli within the appropriate experiment-specific methodology sections below. 
In these experiments, the shape targets subtended roughly 2.5° visual arc and were randomly oriented and positioned within a 20° diameter circular stimulus window. The task was made challenging by embedding the targets within a field of randomly positioned and oriented distractor segments. The lengths of the distractor segments were drawn randomly and uniformly from the lengths of the visible target segments when the target was present, and from a randomly selected target when the target was absent. Antialiased segments were drawn white on a midgray background. Figure 6 shows a sample target-present stimulus display. 
Figure 6
 
Stimulus sequence.
Figure 6
 
Stimulus sequence.
Procedure
A yes/no design was employed. Each trial sequence consisted of a fixation display (0.5 s), stimulus display (160 ms in Experiments 1 and 2, 200 ms in Experiment 3), and then the fixation display again until response (Figure 6). A target shape appeared within the stimulus window with 0.5 probability. The observer pressed the left arrow key on a keyboard to indicate target absent, and the right arrow key to indicate target present. Feedback was provided in the form of a tone for incorrect responses. The number of distractor elements at criterion (75% correct) performance was estimated using an adaptive psychometric procedure (QUEST; Watson & Pelli, 1983). Each experiment was broken into three 100-trial blocks per condition; blocks were counterbalanced across conditions. QUEST was reinitialized for each block, yielding three independent threshold estimates for each condition. 
Experiment 1. Detection of animal and open metamer targets
Methods
Stimuli
For our first experiment, we used a simple procedure to construct metamer stimuli that have exactly the same first-order (local) geometric properties as our animal shapes, but none of the global regularities. The process is illustrated in Figure 7a. To construct a metamer, a fragmented animal shape was first selected, and lists of the lengths li of the visible segments and of the gap geometry (di, αi, βi) were extracted, starting from a randomly selected visible segment on the animal shape. These lists were then randomly and independently permuted. Finally, the fragmented metamer shape was constructed by sequentially rendering the sequence of visible segment lengths and gap geometries from the permuted lists. Since the individual gap geometries were copied faithfully, this transcription preserves the local pairwise grouping relationships between visible elements. However, since the order of these gap geometries is randomized, higher order nonlocal structure has been shuffled away. Figure 7b shows the result for an example animal shape. 
Figure 7
 
(a) Open metamer construction. The lengths li of elements and the geometric relationships (di, αi, βi) between successive elements of animal contours (left) are transcribed in a random order to form first-order metamers (right). Only the transcription of the geometric relationships is shown, for clarity. (b) Example of animal and (open) metamer stimuli.
Figure 7
 
(a) Open metamer construction. The lengths li of elements and the geometric relationships (di, αi, βi) between successive elements of animal contours (left) are transcribed in a random order to form first-order metamers (right). Only the transcription of the geometric relationships is shown, for clarity. (b) Example of animal and (open) metamer stimuli.
Since each animal polygon is simple and closed, the sum of its turning angles Display Formula\(\sum\nolimits_i {\left( {{\alpha _i} + {\beta _i}} \right)} \) is ±2π. This property is commutative and so preserved in the metamer despite the shuffling. This regularity can be considered a “convexity bias” as it favors one sign of curvature over the other. As this convexity bias is a nonlocal property that could feasibly be a cue to perceptual grouping, we generated a second form of metamer without this bias by randomizing the sign of all of the turning angles. We refer to these two forms of metamer stimuli as M1 and M2, respectively. 
Line segments were drawn in white on a midgray disk 20° in diameter, within a black rectangular screen. 
We emphasize that although the metamer stimuli created for this experiment are specific to animal shapes, the same technique could be applied to any class of shape. 
Observers
There were 10 observers in our first experiment. Each performed 3 blocks of 100 trials each for 3 target lengths (5, 10, 20) and 3 target conditions (animal, M1, M2), for a total of 2,700 trials per observer. One of the observers was an author; the others were naïve to the goals of the experiment. All observers had normal or corrected-to-normal vision. 
Results
Figure 8 shows the results. Mean noise thresholds over observers for the three conditions are shown on the left, and pairwise contrasts between conditions are shown on the right. We see that results for the two metamer conditions are nearly identical: The convexity bias induced by retaining the signs of the turning angles does not seem to impact target detection. However, there is a clear difference between these metamer conditions and the animal shape condition: Thresholds are significantly higher for the animal condition than for the metamer conditions at higher target lengths (Table 1). Since the information present in the pairwise relationships between successive elements is identical for animal and metamer stimuli, this result indicates that observers are making use of information in the animal targets beyond these local cues. 
Figure 8
 
Experiment 1 results: animal contours versus first-order metamers M1 (with convexity bias) and M2 (without convexity bias). Left: mean noise thresholds as a function of the number of oriented elements sampling the target shape (target length). Right: pairwise differences between conditions. Error bars indicate standard error of the mean.
Figure 8
 
Experiment 1 results: animal contours versus first-order metamers M1 (with convexity bias) and M2 (without convexity bias). Left: mean noise thresholds as a function of the number of oriented elements sampling the target shape (target length). Right: pairwise differences between conditions. Error bars indicate standard error of the mean.
Table 1
 
Experiment 1: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (with convexity bias) and M2 (without convexity bias).
Table 1
 
Experiment 1: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (with convexity bias) and M2 (without convexity bias).
What is the nature of the nonlocal differences between animal and metamer stimuli responsible for the observed differences in detectability? There are many possibilities; we first seek to determine whether simple factors such as target size, elongation, eccentricity, self-intersections, or number of inflections could explain our results. 
Target size
Whereas the animal and metamer stimuli consist of the same number of oriented elements, the metamer stimuli tend to be less compact, and thus subtend a larger portion of the visual field. One way to quantify the size of the target stimulus is by the area of its convex hull. Figure 9a shows the results of this analysis. Observe that the metamer stimuli are considerably larger by this measure. To understand whether this difference could account for the observed differences in noise thresholds, we analyzed a subset of our data for animal and metamer (M1) stimuli at a target length of 20 elements, equalized according to the area of their convex hulls. Specifically, we sorted animal and M1 stimuli separately according to the area of their convex hull. We then sequentially selected animal stimuli in decreasing order of area and metamers in increasing order of area until the mean area of the two subsets matched (Figure 9b); this procedure resulted in area-matched subsets containing roughly 87% of the original dataset. Finally, we used QUEST to re-estimate noise thresholds from the original human data using only these area-matched stimulus subsets. The results for the animal and M1 targets at a target length of 20 elements are shown in Figure 9c. We find that results are not substantially affected by this equalization; comparing results for all target stimuli shown in Figure 8, thresholds are still significantly higher for animal targets relative to metamer targets: two-tailed matched sample t(9) = 4.4, p = 0.002. Thus it seems that the difference in the average area of the animal and metamer stimuli does not account for the observed difference in their detectability. 
Figure 9
 
Experiment 1 analysis: target area. (a) Mean area of the convex hull of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Histogram of areas of equalized subsets of animal and first-order metamer M1 stimuli (with convexity bias). (c) Mean noise thresholds for area-equalized animal and M1 targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 9
 
Experiment 1 analysis: target area. (a) Mean area of the convex hull of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Histogram of areas of equalized subsets of animal and first-order metamer M1 stimuli (with convexity bias). (c) Mean noise thresholds for area-equalized animal and M1 targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Target elongation
Since the metamers are not constrained to be closed, they tend to be more elongated than the animal stimuli. We measured the elongation of the stimuli by the length of the maximal inscribing chord of the convex hull. Figure 10a verifies that by this measure the metamer stimuli are more elongated than the animal stimuli. Applying the same equalizing technique we used for area now to elongation for 20-element stimuli yielded a matched subset comprising 63% of the original data. Again we find that results are essentially unchanged by this equalization (Figure 10b); thresholds are still significantly higher for animal targets relative to metamer targets: two-tailed matched sample t(9) = 4.0, p = 0.003. Thus it seems that the difference in the average elongation of the animal and metamer stimuli also does not account for the observed difference in their detectability. 
Figure 10
 
Experiment 1 analysis: target elongation. (a) Mean elongation of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Mean noise thresholds for elongation-equalized animal and M1 metamer targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 10
 
Experiment 1 analysis: target elongation. (a) Mean elongation of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Mean noise thresholds for elongation-equalized animal and M1 metamer targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Target eccentricity
We also assessed whether there were systematic differences in the eccentricity of the stimuli. Figure 11a shows the mean eccentricity of the target shape centroids as a function of target length. Note that on average, the animal stimuli tend to be more eccentric than the metamer stimuli. This is because targets were positioned randomly to lie entirely within the central 20° of the display. Since the animal stimuli tended to be more compact, they tended to assume more eccentric locations. Since more eccentric presentation would be expected to reduce performance, these differences are not likely to account for the superior detectability of animal targets. 
Figure 11
 
Experiment 1 analysis: eccentricity and self-intersections. (a) Mean eccentricity of the centroid of the target animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. Error bars indicate standard error of the mean. (b) Mean noise thresholds for animal targets and the subset of open metamer targets that do not self-intersect. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 11
 
Experiment 1 analysis: eccentricity and self-intersections. (a) Mean eccentricity of the centroid of the target animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. Error bars indicate standard error of the mean. (b) Mean noise thresholds for animal targets and the subset of open metamer targets that do not self-intersect. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Self-intersections
Although the animal targets are simple (nonintersecting) closed contours, our method for generating metamers does not prevent them from turning back on themselves, potentially generating multiple self-intersections (Figure 5). Is it possible that these self-intersections somehow obscure the target, lowering detectability? To assess this, we analyzed the subset of 20-element metamer stimuli that do not self-intersect. (This consists of 20.6% of M1 stimuli and 30.5% of M2 stimuli.) Results are substantially unchanged, and animal targets are still significantly more detectable than the metamers: Figure 11b, two-tailed matched sample t(9) = 4.2, p = 0.002 for Animal versus M1 and t(9) = 6.7, p = 0.00008 for Animal versus M2. Thus the absence of self-intersections does not appear to be responsible for the superior detectability of the animal targets. 
Number of inflections
Pettet, Mckee, and Grzywacz (1998; see also Pettet, 1999) found that introducing larger turning angles or inflections in a contour lowers detectability. This raises the question: Could the enhanced detectability of closed animal shapes relative to open metamers be due to some difference in the average magnitude or sign of the turning angles? By construction, the magnitude of the turning angles for animals and metamers are identical, so certainly a difference in the average magnitude of the turning angles cannot explain our results. What about sign? While the signs of the turning angles for M1 metamers match those for the animal contours, the signs of the turning angles were selected randomly for M2 metamers, yet no effect on performance was observed. Thus it seems that differences in the signs of the turning angles also cannot explain our results. 
It is important to note, however, that the number of inflections in the contour depends not only on the signs of the turning angles but also upon their sequencing. For the animal contours the signs of the turning angles at adjacent elements are correlated, and this could be expected to produce fewer inflections on average than for metamers, for which turning angles at adjacent elements are statistically independent. 
Figure 12a shows the mean probability of an inflection occurring at any given element in our animal and metamer stimuli. We see that indeed for longer target lengths, the metamers contain slightly more inflections than the animal stimuli, as predicted. Could this explain why animals are more easily detected? 
Figure 12
 
Experiment 1 analysis: mean number of inflections for animal and metamer stimuli. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 12
 
Experiment 1 analysis: mean number of inflections for animal and metamer stimuli. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
To answer this question, we analyzed the number of inflections in target contours that were successfully detected (hits) versus the number in target contours that were not detected (misses). Figure 12b shows the results. We see that there is no significant difference between the number of inflections in hit trials versus miss trials, for any of the stimulus conditions (p > 0.05 for matched-sample t tests over our 10 observers for all three stimulus conditions at all three target lengths). We thus conclude that a difference in the number of inflections does not explain the observed difference in detectability of animal and metamer stimuli. 
Summary
These control analyses appear to rule out explanations for the superior detectability of animal shapes over metamers based upon size, length, eccentricity, self-intersection, and number of inflections. However, one important difference between animal and metamer stimuli that we have not analyzed is closure. Closure has been suggested as potentially important both as a cue for contour grouping (Kovacs & Julesz, 1993) and as a bridge from one-dimensional to two-dimensional shape (Koffka, 1935; Elder & Zucker, 1993; Elder & Zucker, 1994; Garrigan, 2012), and thus could plausibly contribute to the detectability of our animal stimuli. Our next two experiments will explicitly assess the potential role of closure in the perceptual organization of fragmented shapes. 
Experiment 2. Detection as a function of closure
Methods
Methods were identical to those of Experiment 1, except as noted below. 
Stimuli
The goal of this experiment is to compare detection for partial fragmented animal shapes as a function of their closure, compared with metamer shapes matched in target length (number of elements). Because the results of Experiment 1 suggest that the convexity bias that distinguishes our two types of metamer stimuli has very little impact on detectability, we elected to employ only the M1 metamer stimulus for this comparison. 
We selected a target length of n = 40 for the complete animal shapes, and measured detection performance for 50% closure (n = 20), 75% closure (n = 30), and 100% closure (n = 40), and for length-matched M1 metamers. (For animal shapes with 100% closure [n = 40] we compare with metamer stimuli of length n = 41 to equate the number of pairwise connections between neighboring elements.) 
Line segments were drawn in white on a midgray disk 20° in diameter, within a black rectangular screen. 
Observers
Ten observers participated in Experiment 2. Five of these observers had already participated in Experiment 1 (including one author); the other five were new. All had normal or corrected-to-normal vision, and all but the author were naïve with respect to the goals of the experiment. 
Results
Figure 13 shows the results. Noise thresholds are consistently higher for the partial animal targets relative to the partial metamer shapes, and this detectability advantage becomes statistically significant for higher target lengths (Table 2). These results suggest that the detectability advantage for animal targets is not dependent upon perfect closure: Seventy-five percent closure is sufficient to show a statistically significant advantage. This is consistent with prior work suggesting that perceptual closure, unlike mathematical closure, is a continuous property of contours, rather than a binary, all-or-nothing feature (Elder & Zucker, 1993, 1994). This experiment therefore leaves us with uncertainty about why the animal targets are easier to detect: Is it perceptual closure, or some other nonlocal property of these animal shapes? Our final experiment will address this question. 
Figure 13
 
Experiment 2 results: mean noise thresholds and contrasts for animal and first-order metamer M1 (with convexity bias) target stimuli, as a function of target length and degree of closure for animal targets. Error bars indicate standard error of the mean.
Figure 13
 
Experiment 2 results: mean noise thresholds and contrasts for animal and first-order metamer M1 (with convexity bias) target stimuli, as a function of target length and degree of closure for animal targets. Error bars indicate standard error of the mean.
Table 2
 
Experiment 2: Statistical analysis of noise threshold contrasts for animal targets with progressive degrees of closure vs. M1 metamer controls, based on two-tailed matched-sample t tests over observers.
Table 2
 
Experiment 2: Statistical analysis of noise threshold contrasts for animal targets with progressive degrees of closure vs. M1 metamer controls, based on two-tailed matched-sample t tests over observers.
Experiment 3. Detection of animal, open metamer and closed metamer targets
Methods
Methods were identical to those of Experiment 1, except as noted below. 
Stimuli
Construction of open metamers that match the local pairwise geometry of animal shapes was a simple matter, requiring only a shuffle of pairwise geometric properties. Creating simple (nonintersecting) and closed versions of these metamers is more complicated, as the constraints of nonintersection and closure introduce nonlocal dependencies between the placement of the oriented elements comprising the stimulus. 
To solve this problem, we employed a very different approach. The process is somewhat involved: We summarize it here and provide additional detail in the appendix. We first used a novel method to construct from our high-resolution animal polygons, low-resolution 120-segment equilateral approximations. (The equilateral constraint simplifies the process, as the polygons can now be described entirely by the sequence of turning angles, without reference to the segment lengths.) 
From the resulting 120-segment equilateral polygonal approximations of our animal shapes, we learned a nonparametric kernel density model of the turning angles. We then developed a method for creating novel equilateral polygons whose turning angles match this density model but which are otherwise maximum entropy. The method is seeded with a regular polygon (all angles equal). Proposals for new nonintersecting polygons are then formed using an isoperimetric constraint on inscribed quadrilaterals (Goldin, Delosme, & Bruckstein, 2010). Metropolis-Hastings sampling (Metropolis, Rosenbluth, Rosenbluth, Teller, & Teller, 1953; Hastings, 1970) is then used to fairly sample from these proposals, using the kernel density model for turning angles of animal shapes as the target density. The probability density of turning angles for the resulting closed metamer samples matches the density for the original animal shapes (Figure 14). We will use the label M3 to refer to metamers generated from these shapes, to distinguish them from the open metamers used in Experiments 1 and 2
Figure 14
 
Experiment 3. The probability density of turning angles for our closed metamer shapes is a close match to the turning angle density for the original animal shapes.
Figure 14
 
Experiment 3. The probability density of turning angles for our closed metamer shapes is a close match to the turning angle density for the original animal shapes.
We emphasize that although the closed metamer stimuli created for this experiment are specific to animal shapes, the same technique could be applied to any class of shape, given a sufficient number of exemplars to learn the statistical model for local turning angles. 
Line segments were drawn in white on a black rectangular screen. 
Observers
There were 13 observers. One of the observers was an author; the others were naïve to the goals of the experiment. All observers had normal or corrected-to-normal vision. 
Results
Figure 15 shows the results of this experiment. We see that as the target length is increased to 20 elements and beyond, a dissociation between the three conditions emerges: Animal target are significantly more easily detected than either form of metamer, but at the same time the closed metamers are significantly more easily detected than open metamers (Table 3). These results indicate that closure is certainly one nonlocal cue that aids detection, but it is not the only nonlocal cue: There must be additional nonlocal regularities of the animal targets that play a role in the perceptual organization of shape.1 
Figure 15
 
Experiment 3 results. A: Animal targets; M1: first-order metamers (open, with convexity bias); M3: first-order metamers (closed). Left: mean noise thresholds. Right: pairwise contrasts. Error bars indicate standard error of the mean.
Figure 15
 
Experiment 3 results. A: Animal targets; M1: first-order metamers (open, with convexity bias); M3: first-order metamers (closed). Left: mean noise thresholds. Right: pairwise contrasts. Error bars indicate standard error of the mean.
Table 3
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Table 3
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
As for open metamers, closed metamers have slightly more inflections than animal shapes, but the number of inflections did not seem to influence thresholds. This was verified by a four-way ANOVA analysis of the number of inflections in the target for target-present trials with observer, target shape, target length, and response (hit or miss) as factors (Table 4). Whereas both target shape and target length are significant predictors of the number of inflections, observer and response are not. 
Table 4
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Table 4
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Discussion
With three experiments we have tested the hypothesis that nonlocal cues play a role in the perceptual grouping of contours. Rather than narrowing focus on a specific cue, we employed natural animal shapes that afford many nonlocal cues, on the assumption that the brain is adapted to take advantage of cues available in ecologically important stimuli. Critical to our method is the design of control stimuli that match the local cues afforded by our animal stimuli, but none of the global cues. 
In our first experiment we used a simple shuffling trick to achieve this, which produced first-order metamer contours with exactly the same local geometry as animal contours. The results (Figure 8) clearly indicate that the animal shapes are easier to detect in oriented noise than the first-order controls. A series of control experiments ruled out trivial explanations based on differences in size, elongation, eccentricity, or self-intersections, pointing to the importance of more meaningful nonlocal regularities. A specific and somewhat controversial candidate is closure. 
Closure
Closure has long been identified as an important perceptual bridge from one-dimensional to two-dimensional shape (Koffka, 1935). The empirical evidence for this claim derives in part from visual search experiments using outline figures in which observers must detect a unique 2D target shape embedded in a field of distractor shapes (Figure 16; Elder & Zucker, 1993, 1994). Search becomes progressively faster as the closure of the shapes is increased. There is also evidence that our ability to encode shape information for later recall is a strong function of closure (Garrigan, 2012). 
Figure 16
 
Closure as a bridge from 1D to 2D shape (Elder & Zucker, 1993, 1994). (a) Shape discrimination is easy when good continuation and closure are strong. (b) Discrimination becomes hard when good continuation and closure are weak.
Figure 16
 
Closure as a bridge from 1D to 2D shape (Elder & Zucker, 1993, 1994). (a) Shape discrimination is easy when good continuation and closure are strong. (b) Discrimination becomes hard when good continuation and closure are weak.
However, the present experiments do not involve the discrimination of 2D shapes but rather the detection of a fragmented contour in random clutter. Using the Field, Hayes, and Hess methodology, Kovacs and Julesz (1993) found that a roughly circular closed contour was more detectable than an open contour of the same length, and argued from this that closure enhances the detectability of contours in clutter. However, the magnitude and sign of the turning angles used to form the open and closed contour stimuli employed in this experiment were not equated, making it unclear whether it was closure per se that was enhancing detection. 
Pettet et al. (1998) addressed this question by fixing each turning angle to be exactly 30° in magnitude. For the closed contour, the sign of the turning angle was constant, whereas for the open contour it was randomly and uniformly selected to be either positive or negative. As a result, the open control contours typically had multiple inflections not present in the closed contours. 
They found that the closed contours were more easily detected than the open contours, replicating the Kovacs & Julesz result. In order to isolate the underlying cause, they (see also Pettet, 1999) also compared the detectability of a circular closed contour with constant turning angles and no inflections to the detectability of a noncircular closed contour with no inflections but nonconstant turning angles (a “bean”), and to a noncircular closed contour with two inflections and nonconstant turning angles (a “moon”). They found that variation in turning angle (curvature) and the presence of inflections reduced detection performance. In a critical control, they showed that when closed and open contours were better matched for the presence of inflections, the advantage of closure disappeared. Tversky, Geisler, and Perry (2004) have also presented evidence that small enhancements in detectability found for closed contours may derive from probability summation over local configurations of elements rather than a global closure mechanism. 
In summary, prior work suggests that, despite the clear role of closure in the formation of 2D shape percepts (Koffka, 1935; Elder & Zucker, 1993, 1994; Garrigan, 2012), its role as a global cue for detection of contours in clutter may be minimal, with evidence to the contrary attributed to local geometric covariates, in particular the magnitude and sign of turning angles. 
Our novel metamer stimuli provide an opportunity to more carefully test for a role of closure in contour detection while controlling carefully for local geometry. In Experiment 2, we compared the detectability of fragments of our animal stimuli ranging from 50% closure to 100% closure against length-matched local metamer controls. The results show that the superior detectability of the animal shapes does not depend upon perfect closure: Seventy-five percent closure was sufficient to generate a statistically significant advantage. These results are consistent with prior work (Elder & Zucker, 1994) showing that perceptual closure is a graded phenomenon, increasing smoothly in influence as the degree of closure increases. However, they are not consistent with a more simplistic explanation of the effects of closure in terms of proximity and good continuation cues (Tversky et al., 2004), as these are the same for the partial animal shapes and the metamer stimuli. 
Nevertheless, Experiment 2 leaves the factor of closure confounded with other potential nonlocal factors determining perceptual grouping. The relative improvement in detectability of the animal stimuli with increased target length could be due to closure but could also be due to other concomitant global shape cues that are progressively revealed. Experiment 3 serves to more clearly separate these two possibilities by introducing a new class of closed metamer stimuli. Local geometry (turning angles) of both open and closed metamers is by construction matched to each other and to the animal shapes. The closed metamers are also constrained to be closed but otherwise are maximum entropy. Thus the sole difference between closed and open metamers is closure. The finding that closed metamers are more detectable than open metamers is thus clear evidence that closure does play a role in the detection of contours in clutter. 
Other global cues
In addition to closure, there are a number of other global shape attributes known to be perceptually important that might play a role in perceptual grouping. These include convexity (Jacobs, 1996; Liu, Jacobs, & Basri, 1999; Corcoran, Mooney, & Tilton, 2011) and symmetry and parallelism (Lowe, 1985; Mohan & Nevatia, 1992; Wagemans, 1995; Zisserman et al., 1995; Jepson, Richards, & Knill, 1996; Behrmann, Zemel, & Mozer, 1998; Jacobs, 2003; Feldman, 2007; Sasaki, 2007; Stahl & Wang, 2008; Machilsen, Pauwels, Wagemans, 2009). Wilder, Feldman, and Singh (2016) have also advanced a more general theory of shape detectability based on efficient coding. They suggest that shapes that are more likely to occur in our visual environment will be coded more concisely by the brain, and that these more compact encodings for common stimuli will lead to more efficient detection. Our results are at least partially consistent with this theory, as the animal shapes we use here are surely more probable ecological stimuli than the metamers. It also seems plausible that our closed metamers are more ecological than our open metamers. 
Neural basis
What are the brain mechanisms underlying the integration of local and global cues for contour grouping? Neurons in early visual cortex have highly localized, oriented receptive fields well suited to coding the local edges of object boundaries (Hubel & Wiesel, 1968; Dobbins, Zucker, & Cynader, 1987; Heitger et al., 1992). The standard view is that perceptual grouping of these edges into contours relies upon long-range horizontal connections between orientation columns in primary visual cortex (Gilbert & Wiesel, 1989; Li & Gilbert, 2002; Stettler et al., 2002; Li et al., 2006). 
The significance of the results reported here is that global cues beyond the local association field are seen to codetermine contour grouping. This evidence demands a revision to the standard model, a revision that allows for global shape properties to influence local grouping. These global shape and object properties are coded deeper in the ventral stream, in Area V4 and TE/TEO, for example (Pasupathy & Connor, 1999; Kourtzi & Kanwisher, 2000, 2001; Op de Beeck, Wagemans, & Vogels, 2001; Connor, Brincat, & Pasupathy, 2007; Op de Beeck, Torfs, & Wagemans, 2008) and the responses of neurons in these areas depend upon input from the early visual areas coding local grouping relationships. We consider two alternative neural architectures that might account for the influence of global shape properties on perceptual grouping. 
Feedforward architecture
One possible account is that early visual cortex passes multiple, competing local grouping hypotheses forward to intermediate and higher visual areas, which through a combination of feedforward computation and lateral interactions bring to bear more global factors to select a unique solution that is globally consistent with the visual input. Under this model, no trace of this global computation would necessarily be evident in early visual cortex. 
This kind of computational progression is consistent with some studies that examine the dynamics of neural tuning in intermediate and higher areas of the object pathway. Brincat and Connor (2004) studied the dynamics of contour shape selectivity in posterior inferotemporal cortex (PIT). They found that early PIT responses were selective for individual straight and curved contour fragments, but that tuning evolved toward nonlinear selectivity to multifragment configurations of contours over a time course of about 60 ms. 
Yau, Pasupathy, Brincat, and Connor (2013) extended this analysis to neurons in Area V4. Analogously to PIT, they found that V4 neural response evolved over a time course of about 50 ms from selectivity to contour orientation to selectivity for curvature. They posited an intracortical facilitatory circuit between similarly tuned V4 neurons that would generate a sustained selectivity for curvature. 
The temporal dynamics in V4 and PIT are consistent with a feed-forward progression of progressively more complex shape tuning. V4 neural selectivity for orientation peaks at roughly 100 ms poststimulus, and selectivity for curvature peaks at roughly 150 ms. In PIT, neural selectivity for individual straight and curved contour fragments peaks at roughly 150 ms poststimulus, while selectivity for multifragment configurations peaks at roughly 200 ms. 
Orientation and turning angles (local curvature) are equated for our animal and metamer stimuli, suggesting that the greatest differentiation in neural response to our target shapes would be seen in the later phase of PIT response, when selectivity for multifragment configurations emerges. However, it is unclear whether the simple local facilitatory circuit proposed here would be sufficient to resolve local grouping ambiguities in cluttered displays like the ones we use here. 
Feedback architecture
A second possibility is that higher visual areas receive grouping hypotheses from early visual areas and then provide feedback to these early visual areas to support hypotheses that are consistent with global evidence while suppressing hypotheses that are inconsistent with global evidence. This kind of model extends the function of V1 beyond a transient initial stage of local processing to include a more sustained role in representing high-resolution geometric detail (Lee, Mumford, Romero, & Lamme, 1998; Lee & Mumford, 2003; Roelfsema & de Lange, 2016). Anatomically, substantial feedback connections from higher to lower visual areas in ventral stream are known to exist (Ungerleider, 1995), and the feedback of global shape information to inform local grouping in early visual cortex could be one of their functions. Under this second model, one would expect to find a trace of this global influence in the later phase of response of the early visual neurons coding local contour geometry. 
There is growing evidence for this feedback hypothesis. Perceptual learning and attention have been found to dynamically influence contour grouping mechanisms in early visual cortex (Gilbert & Sigman, 2007; Li, Piech, & Gilbert, 2008; Kravitz, Saleem, Baker, Ungerleider, & Mishkin, 2013), and V1 neurons can adapt dynamically to different nonlocal shape properties depending on the immediate task (McManus, Li, & Gilbert, 2011; Gilbert & Li, 2013; Piech, Li, Reeke, & Gilbert, 2013; Ramalingam, McManus, Li, & Gilbert, 2013). Both of these effects are thought to be the result of feedback from higher shape-selective visual areas. In more recent physiological experiments involving simultaneous recording from neurons in macaque areas V1 and V4, selectivity for global contour information is first seen in V4, emerging in V1 roughly 40 ms later (Chen, Yan, Gong, Gilbert, Liang, & Li, 2014). These results point to a V1–V4 recurrent network underlying the global perceptual integration of contours. 
Human brain studies of illusory contour perception using event-related potential (ERP; Halgren, Mendola, Chong, & Dale, 2003), magnetoencephalography (MEG; Yoshino, Kawamoto, Yoshida, Kobayashi, & Shigemura, 2006) and trans-cranial magnetic stimulation (TMS; Wokke, Vandenbroucke, Scholte, & Lamme, 2013) methodologies point to an analogous cooperative recurrent computation involving early and late visual areas in ventral stream. Similar feedback mechanisms have been proposed to account for the perceptual assignment of figure/ground relationships (Lamme, 1995; Zipser, Lamme, & Schiller, 1996; Lamme, Super, & Spekreijse, 1998; Peterson & Cacciamani, 2013; Cacciamani, Scalf, & Peterson, 2015; Poort, Self, van Vugt, Malkki, & Roelfsema, 2016) and border ownership (Craft, Schutze, Niebur, & von der Heydt, 2007). 
Human behavioral studies also point to a recurrent circuit for perceptual grouping and shape perception. Most of these studies have been based on backward masking, where performance on a shape detection or discrimination task is disrupted by the subsequent presentation of a masking stimulus that is thought to interfere with feedback from higher shape areas to early visual cortex (Enns & Di Lollo, 1997, 2000; Habak, Wilkinson, & Wilson, 2006). Whereas these studies rely on a disruption of visual processing to infer feedback, Drewes, Goren, Zhu, and Elder (2016) used the animal and open metamer stimuli employed here to demonstrate that a behavioral facilitation effect can also be observed. Specifically, target shapes were presented twice at the same location but with variable delay, in dynamic oriented element noise. Performance was found to peak when the two target presentations were separated by a delay of roughly 60 ms, suggesting a fast recurrent circuit underlying the perceptual organization of shape. 
These primate and human studies have all focused on the perceptual organization of shape. However, recent human fMRI evidence suggests that feedback modulation may also extend to more semantic object category information (Uyar, Shomstein, Greenberg, & Behrmann, 2016). This is consistent with psychophysical studies that show an influence of object category on figure/ground organization (Peterson, 1994). 
In summary, recent anatomical, physiological and behavioral evidence suggests that cortical feedback of global shape information from higher visual areas to lower visual areas may play an important role in the perceptual grouping of contours, and our psychophysical results are consistent with this hypothesis. That said, we suspect that perceptual object formation is the outcome of a cooperative computation involving diverse feedforward, lateral, and feedback connections between a number of early and late visual areas. A computation that is progressive, generating quick approximate solutions that are refined over time, would have adaptive advantage in supporting rapid real-time vision, and such mechanisms could be probed through future behavioral and physiological experiments that jointly manipulate local and global grouping cues as well as stimulus dynamics. 
A role for parietal cortex?
Our discussion so far has focused on the so-called object pathway in ventral stream, but posterior areas of parietal cortex are known to be involved in the processing of object configural information (Kravitz et al., 2013; Freud, Plaut & Behrmann, 2016), and could also be involved in the detection of contours in clutter. Kuai, Li, Yu, and Kourtzi (2017) have assessed to what degree the presence of a contour in a cluttered display can be decoded from visual areas in the object pathway but also from areas in parietal cortex. They employed a standard static form of the Field, Hayes, and Hess stimulus, but also an interesting dynamic form, in which the stimulus is revealed over time through a narrow slit. For smaller slit widths, the elements constituting the collinear contour appeared sequentially, moving either up or down the slit, depending on the orientation of the contour and the direction of motion (left or right). Under these conditions, not only do the elements not appear simultaneously, their relative geometry when they do appear is not colinear. As a consequence the stimulus is unlikely to strongly activate long-range horizontal connections in V1, which are thought to be primarily selective for simultaneously presented, colinear stimuli. 
Kuai et al. found that even for narrow slit widths for which only one element of the contour was visible at a time, observers were still well above chance on the contour detection task. fMRI analysis revealed that contour presence could be decoded from higher dorsal visual areas, including V3B, and intraparietal sulcus (IPS) as well as LOC, but not from early visual areas (V1, V2) or from ventral visual areas (V3v, hV4). Most accurate decoding was from posterior IPS. In contrast, when the stimuli were presented simultaneously, without the vertical slit, decoding was possible from early as well as late visual areas. Interestingly, decoding of contour presence (but not contour orientation) for simultaneous stimuli was still possible from higher dorsal as well as ventral areas. These results suggest that higher areas of dorsal stream are also involved with contour integration over both time and space. Exactly how these visual areas collaborate with higher areas in ventral stream and early visual areas to support detection of contours in cluttered displays is not yet understood. 
Acknowledgments
We thank the NSERC Discovery, York Vision: Science to Applications, York Research Chair, and York Bridging Fund programs for supporting this research. 
Commercial relationships: none. 
Corresponding author: James H. Elder. 
Email: jelder@yorku.ca
Address: Centre for Vision Research, York University, Toronto, Canada. 
References
Behrmann, M., Zemel, R. S., & Mozer, M. C. (1998). Object-based attention and occlusion: Evidence from normal participants and a computational model. Journal of Experimental Psychology: Human Perception and Performance, 24, 1011–1036.
Boykov, Y., Veksler, O., & Zabih, R. (2001). Fast approximate energy minimization via graph cuts. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23 (11), 1222–1239.
Brincat, S. L., & Connor, C. E. (2004). Underlying principles of visual shape selectivity in posterior inferotemporal cortex. Nature Neuroscience, 7 (8), 880–886.
Cacciamani, L., Scalf, P. E., & Peterson, M. A. (2015). Neural evidence for competition-mediated suppression in the perception of a single object. Cortex, 72, 124–139.
Cass, J. R., & Spehar, B. (2005). Dynamics of collinear contrast facilitation are consistent with long-range horizontal striate transmission. Vision Research, 45, 2728–2739.
Chen, M., Yan, Y., Gong, X., Gilbert, C. D., Liang, H., & Li, W. (2014). Incremental integration of global contours through interplay between visual cortical areas. Neuron 82 (3), 682–694.
Connor, C., Brincat, S., & Pasupathy, A. (2007). Transformation of shape information in the ventral pathway. Current Opinion in Neurobiology, 17, 140–147.
Corcoran, P., Mooney, P., & Tilton, J. (2011). Convexity grouping of salient contours. In Jiang, X. Ferrer, M. & Torsello A. (Eds.), Proceedings of the International Workshop on Graph Based Representations in Pattern Recognition (Vol. 6658 of Lecture Notes in Computer Science, pp. 235–244). Berlin, Germany: Springer.
Craft, E., Schutze, H., Niebur, E., & von der Heydt, R. (2007). A neural model of figure-ground organization. Journal of Neurophysiology, 97, 4310–4326.
de Beeck, H. O., Wagemans, J., & Vogels, R. (2001). Inferotemporal neurons represent low-dimensional configurations of parameterized shapes. Nature Neuroscience, 4 (12), 1244–1252.
Dobbins, A., Zucker, S. W., & Cynader, M. S. (1987). Endstopped neurons in the visual cortex as a substrate for calculating curvature. Nature, 329 (6138), 438–441.
Drewes, J., Goren, G., Zhu, W., & Elder, J. (2016). Recurrent processing in the formation of shape percepts. Journal of Neuroscience, 36 (1), 185–192.
Elder, J. H. (2013). Perceptual organization of shape. In Dickinson, S. & Pizlo Z. (Eds.), Shape perception in human and computer vision: An interdisciplinary perspective (pp. 71–83). London, UK: Springer.
Elder, J. H. (2015). Bridging the dimensional gap: Perceptual organization of contour into two-dimensional shape. In Wagemans J. (Ed.), Oxford handbook of perceptual organization (pp. 207–235). Oxford, UK: Oxford University Press.
Elder, J. H. (2018). Shape from Contour: Computation and Representation. Annual Review of Vision Science, 4 423–450.
Elder, J. H., & Goldberg, R. M. (2002). Ecological statistics of Gestalt laws for the perceptual organization of contours. Journal of Vision, 2 (4): 5, 324–353, https://doi.org/10.1167/2.4.5. [PubMed] [Article]
Elder, J. H., Krupnik, A., & Johnston, L. A. (2003). Contour grouping with prior models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25 (25) 661–674.
Elder, J. H., & Velisavljević, L. (2009). Cue dynamics underlying rapid detection of animals in natural scenes. Journal of Vision, 9 (7): 7, 1–20, https://doi.org/10.1167/9.7.7. [PubMed] [Article]
Elder, J. H., & Zucker, S. W. (1993). The effect of contour closure on the rapid discrimination of two-dimensional shapes. Vision Research, 33 (7), 981–991.
Elder, J. H., & Zucker, S. W. (1994). A measure of closure. Vision Research 34 (24), 3361–3370.
Elder, J. H., & Zucker, S. W. (1996). Computing contour closure. In Buxton, B. F. & Cipolla, R. (Eds.), Proceedings of the 4th European Conference on Computer Vision (pp. 399–412). New York, NY: Springer Verlag.
Enns, J., & Di Lollo, V. (1997). Object substitution: A new form of masking in unattended visual locations. Psychological Science, 8, 135–139.
Enns, J., & Di Lollo, V. (2000). What's new in visual masking? Trends in Cognitive Sciences, 4, 345–352.
Feldman, J. (2007). Formation of visual “objects” in the early computation of spatial relations. Perception and Psychophysics, 69 (5), 816–827.
Field, D., Hayes, A., & Hess, R. F. (1993). Contour integration by the human visual system: Evidence for a local “association field.”: Vision Research, 33 (2), 173–193.
Freud, E., Plaut, D. C., & Behrmann, M. (2016). ‘What' is happening in the dorsal visual pathway. Trends in Cognitive Sciences, 20 (10), 773–784.
Garrigan, P. (2012). The effect of contour closure on shape recognition. Perception, 41, 221–235.
Geisler, W. S., & Perry, J. S. (2009). Contour statistics in natural images: Grouping across occlusions. Visual Neuroscience, 26 (1), 109–121.
Geisler, W. S., Perry, J. S., Super, B. J., & Gallogly, D. P. (2001). Edge co-occurence in natural images predicts contour grouping performance. Vision Research, 41 (6), 711–724.
Gelman, A. (1996). Inference and monitoring convergence. In Gilks, W. R. Richardson, S. & Spiegelhalter D. J. (Eds.). Markov chain Monte Carlo in practice (pp. 131–143). Boca Raton, FL: Chapman & Hall.
Gilbert, C. D., & Li, W. (2013). Top-down influences on visual processing. Nature Reviews Neuroscience, 14 (5), 350–363.
Gilbert, C. D., & Sigman, M. (2007). Brain states: Top-down influences in sensory processing. Neuron, 54 (5) 677–696.
Gilbert, C. D., & Wiesel, T. N. (1989). Columnar specificity of intrinsic horizontal and corticocortical connections in cat visual cortex. Journal of Neuroscience, 9 (7), 2432–2443.
Goldin, I., Delosme, J.-M., & Bruckstein, A. M. (2010). Vesicles and amoebae: On globally constrained shape deformation. Journal of Mathematical Imaging and Vision, 37, 112–131.
Habak, C., Wilkinson, F., & Wilson, H. (2006). Dynamics of shape interaction in human vision. Vision Research, 46, 4305–4320.
Halgren, E., Mendola, J., Chong, C., & Dale, A. (2003). Cortical activation to illusory shapes as measured with magnetoencephalography. NeuroImage, 18, 1001–1009.
Hastings, W. K. (1970). Monte Carlo sampling methods using Markov chains and their applications. Biometrika, 57 (1), 97–109.
Heitger, F., Rosenthaler, L., von der Heydt, R., Peterhans, E., & Kulber, O. (1992). Simulation of neural contour mechanisms: From simple to end-stopped cells. Vision Research, 32 (5), 963–981.
Hess, R. F., & Dakin, S. C. (1999). Contour integration in the peripheral field. Vision Research, 39 (5), 947–959.
Hubel, D. H., & Wiesel, T. N. (1968). Receptive fields and functional architecture of monkey striate cortex. Journal of Physiology, 195, 215–243.
Jacobs, D. (1996). Robust and efficient detection of salient convex groups. IEEE Transactions on Pattern Analysis and Machine Intelligence 18 (1), 23–37.
Jacobs, D. (2003). What makes viewpoint-invariant properties perceptually salient? Journal of the Optical Society of America A, 20 (7), 1304–1320.
Jepson, A., Richards, W., & Knill, D. (1996). Modal structure and reliable inference. In Knill, D. & Richards W. (Eds.), Perception as Bayesian inference (pp. 63–92). Cambridge, UK: Cambridge University Press.
Koffka, K. (1935). Principles of Gestalt psychology. New York, NY: Harcourt, Brace & World.
Kourtzi, Z., & Kanwisher, N. (2000). Cortical regions involved in perceiving object shape. Journal of Neuroscience, 20 (9), 3310–3318.
Kourtzi, Z., & Kanwisher, N. (2001, August 24). Representation of perceived object shape by the human lateral occipital complex. Science, 293 (5534), 1506–1509.
Kovacs, I., & Julesz, B. (1993). A closed curve is much more than an incomplete one: Effect of closure in figure-ground discrimination. Proceedings of the National Academy of Science, USA, 90, 7495–7497.
Kravitz, D. J., Saleem, K. S., Baker, C. I., Ungerleider, L. G., & Mishkin, M. (2013). The ventral visual pathway: An expanded neural framework for the processing of object quality. Trends in Cognitive Sciences, 17 (1), 26–49.
Kruger, N. (1998). Collinearity and parallelism are statistically significant second order relations of complex cell responses. Neural Processing Letters, 8, 117–129.
Kuai, S.-G., Li, W., Yu, C., & Kourtzi, Z. (2017). Contour integration over time: Psychophysical and fMRI evidence. Cerebral Cortex, 27, 3042–3051.
Lamme, V. A. F. (1995). The neurophysiology of figure-ground segregation in primary visual cortex. Journal of Neuroscience, 15 (12), 1605–1615.
Lamme, V. A. F., Super, H., & Spekreijse, H. (1998). Feedforward, horizontal, and feedback processing in the visual cortex. Current Opinion in Neurobiology, 8 (4), 529–535.
Ledgeway, T., Hess, R., & Geisler, W. (2005). Grouping local orientation and direction signals to extract spatial contours: Empirical tests of “association field” models of contour integration. Vision Research, 45 (19), 2511–2522.
Lee, T. S., & Mumford, D. (2003). Hierarchical Bayesian inference in the visual cortex. Journal of the Optical Society of America A, 20 (7), 1434–1448.
Lee, T. S., Mumford, D., Romero, R., & Lamme, V. A. (1998). The role of the primary visual cortex in higher level vision. Vision Research, 38 (15), 2429–2454.
Li, W., & Gilbert, C. (2002). Global contour saliency and local colinear interactions. Journal of Neurophysiology, 88 (5), 2846–2856.
Li, W., Piech, V., & Gilbert, C. D. (2006). Contour saliency in primary visual cortex. Neuron, 50 (6), 951–962.
Li, W., Piech, V., & Gilbert, C. D. (2008). Learning to link visual contours. Neuron, 57 (3), 442–451.
Li, Z. (1998). A neural model of contour integration in the primary visual cortex. Neural Computation, 10 (4), 903–940.
Liu, Z., Jacobs, D. W., & Basri, R. (1999). The role of convexity in perceptual completion. Vision Research, 39 (25), 4244–4257.
Lowe, D. G. (1985). Perceptual organization and visual recognition. Boston, MA: Kluwer.
Machilsen, B., Pauwels, M., & Wagemans, J. (2009). The role of vertical mirror symmetry in visual shape detection. Journal of Vision, 9 (12): 11, 1–11, https://doi.org/10.1167/9.12.11. [PubMed] [Article]
McManus, J. N. J., Li, W., & Gilbert, C. D. (2011). Adaptive shape processing in primary visual cortex. Proceedings of the National Academy of Sciences, USA, 108 (24), 9739–9746.
Metropolis, N., Rosenbluth, A. W., Rosenbluth, M. N., Teller, A. H., & Teller, E. (1953). Equation of state calculations by fast computing machines. Journal of Chemical Physics, 21 (6), 1087–1092.
Mohan, R., & Nevatia, R. (1992). Perceptual organization for scene segmentation and description. IEEE Transactions on Pattern Analysis and Machine Intelligence, 14 (6), 616–635.
Movahedi, V., & Elder, J. H. (2013). Combining local and global cues for closed contour extraction. In Burghardt, T. Damen, D. Mayol-Cuevas, W. & Mirmehdi M. (Eds.), Proceedings of the British Machine Vision Conference. Durham, UK: BMVA Press.
Op de Beeck, H. P., Torfs, K., & Wagemans, J. (2008). Perceived shape similarity among unfamiliar objects and the organization of the human object vision pathway. Journal of Neuroscience, 28 (40), 10111–10123.
Parent, P., & Zucker, S. W. (1989). Trace inference, curvature consistency, and curve detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 11, 823–839.
Pasupathy, A., & Connor, C. E. (1999). Responses to contour features in Macaque area V4. Journal of Neurophysiology, 82, 2490–2502.
Peterson, M. A. (1994). Object recognition processes can and do operate before figure–ground organization. Current Directions in Psychological Science, 3 (4), 105–111.
Peterson, M. A., & Cacciamani, L. (2013). Toward a dynamical view of object perception. In Pizlo, Z. & Dickinson S. (Eds.), Shape Perception in Human and Computer Vision: An Interdisciplinary Perspective. New York, NY: Springer.
Pettet, M. W. (1999). Shape and contour detection. Vision Research, 39, 551–557.
Pettet, M. W., Mckee, S., & Grzywacz, N. (1998, March). Constraints on long range interactions mediating contour detection. Vision Research, 38 (6), 865–879.
Piech, V., Li, W., Reeke, G. N., & Gilbert, C. D. (2013). Network model of top-down influences on local gain and contextual interactions in visual cortex. Proceedings of the National Academy of Sciences, USA, 110 (43), E4108–E4117.
Poort, J., Self, M. W., van Vugt, B., Malkki, H., & Roelfsema, P. R. (2016). Texture segregation causes early figure enhancement and later ground suppression in areas V1 and V4 of visual cortex. Cerebral Cortex 26 (10), 3964–3976.
Pradhan, R., Kumar, S., Agarwal, R., Pradhan, M. P., & Ghose, M. K. (2010). Contour line tracing algorithm for digital topographic maps. International Journal of Image Processing, 4 (2), 156–163.
Raftery, A. E., & Lewis, S. M. (1996). Implementing MCMC. In Gilks, W. R. Richardson, S. & Spiegelhalter D. J. (Eds.), Markov chain Monte Carlo in practice (Vol. 1, pp. 115–130). Boca Raton, FL: Chapman & Hall.
Ramalingam, N., McManus, J. N. J., Li, W., & Gilbert, C. D. (2013). Top-down modulation of lateral interactions in visual cortex. Journal of Neuroscience, 33 (5), 1773–1789.
Ren, X., Fowlkes, C., & Malik, J. (2008). Learning probabilistic models for contour completion in natural images. International Journal of Computer Vision, 77, 47–63.
Roelfsema, P., & de Lange, F. (2016). Early visual cortex as a multiscale cognitive blackboard. Annual Review of Vision Science, 2, 131–151.
Sasaki, Y. (2007). Processing local signals into global patterns. Current Opinion in Neurobiology, 17 (2), 132–139.
Sigman, M., Cecchi, G. A., Gilbert, C. D., & Magnasco, M. O. (2001). On a common circle: Natural scenes and Gestalt rules. Proceedings of the National Academy of Sciences, USA, 98 (4), 1935–1940.
Stahl, J., & Wang, S. (2008). Globally optimal grouping for symmetric closed boundaries by combining boundary and region information. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30 (3), 395–411.
Stettler, D., Das, A., Bennett, J., & Gilbert, C. (2002). Lateral connectivity and contextual interactions in macaque primary visual cortex. Neuron, 36 (4), 739–750.
Thorpe, S., Fize, D., & Marlot, C. (1996, June 6). Speed of processing in the human visual system. Nature, 381 (6582), 520–522.
Tierney, L. (1996). Introduction to general state-space Markov chain theory. In Gilks, W. R. Richardson, S. & Spiegelhalter D. J. (Eds.), Markov chain Monte Carlo in practice (Vol. 1, chapter 4; pp. 59–74. Boca Raton, FL: Chapman & Hall.
Tversky, T., Geisler, W. S., & Perry, J. S. (2004). Contour grouping: Closure effects are explained by good continuation and proximity. Vision Research, 44, 2769–2777.
Ungerleider, L. (1995, November 3). Functional brain imaging studies of cortical mechanisms for memory. Science, 270 (5237), 769–775.
Uyar, F., Shomstein, S., Greenberg, A. S., & Behrmann, M. (2016). Retinotopic information interacts with category selectivity in human ventral cortex. Neuropsychologia, 92, 90–106.
Wagemans, J. (1995). Detection of visual symmetries. Spatial Vision, 9 (1), 9–32.
Wassermann, L. (2006). All of nonparametric statistics. New York, NY: Springer.
Watson, A. B., & Pelli, D. G. (1983). Quest: A Bayesian adaptive psychometric method. Perception and Psychophysics, 33 (2), 113–120.
Watt, R., Ledgeway, T., & Dakin, S. (2008). Families of models for Gabor paths demonstrate the importance of spatial adjacency. Journal of Vision, 8 (7): 23, 1–19, https://doi.org/10.1167/8.7.23. [PubMed] [Article]
Wertheimer, M. (1938). Laws of organization in perceptual forms. In Ellis W. D. (Ed.), A sourcebook of Gestalt psychology (pp. 71–88). London, UK: Routledge and Kegan Paul.
Wilder, J., Feldman, J., & Singh, M. (2016). The role of shape complexity in the detection of closed contours. Vision Research, 126, SI220–231.
Wokke, M. E., Vandenbroucke, A. R. E., Scholte, H. S., & Lamme, V. A. F. (2013). Confuse your illusion: Feedback to early visual cortex contributes to perceptual completion. Psychological Science, 24 (1), 63–71.
Yau, J., Pasupathy, A., Brincat, S., & Connor, C. (2013). Curvature processing dynamics in Macaque Area V4. Cerebral Cortex, 23, 198–209.
Yen, S., & Finkel, L. (1998). Extraction of perceptually salient contours by striate cortical networks. Vision Research, 38 (5), 719–741.
Yoshino, A., Kawamoto, M., Yoshida, T., Kobayashi, N., & Shigemura, J. (2006). Activation time course of responses to illusory contours and salient region: A high-density electrical mapping comparison. Brain Research, 1071 (1), 137–144.
Zhu, S. C. (1999). Embedding Gestalt laws in Markov random fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 21 (11), 1170–1187.
Zipser, K., Lamme, V. A., & Schiller, P. H. (1996). Contextual modulation in primary visual cortex. Journal of Neuroscience, 16 (22), 7376–7389.
Zisserman, A., Mundy, J., Forsyth, D., Lui, J., Pillow, N., Rothwell, C., & Utcke, S. (1995). Class-based grouping in perspective images. In Proceedings of the International Conference on Computer Vision (pp. 183–188). Cambridge, MA: IEEE Computer Society.
Footnotes
1  Thresholds are somewhat lower than for Experiments 1 and 2. This may be due to the fact that targets in Experiments 1 and 2 were slightly less eccentric on average than in Experiment 3, due to an overly conservative rule used to ensure targets remained within the 20° stimulus window. Also, in Experiments 1 and 2 the stimulus region was defined by a midgray disk within the black background of the screen, whereas in Experiment 3 the entire background was black. The absence of a clear boundary for the stimulus region may have created more spatial uncertainty for the observer.
Appendix: Construction of closed metamer stimuli
From each of the high-resolution animal shape polygons, we generated a coarser, approximately equilateral polygonal stimulus with only 120 vertices. Each of these simpler stimuli was formed by partitioning the sequence of high-resolution vertices representing a shape into a sequence of 120 disjoint, contiguous subsequences, and then representing the shape by vertices located at the 120 arc-length mean locations for each of these subsequences. 
To determine an approximately equilateral polygonal model, we searched the exponential space of possible partitions to find the partition minimizing the variance in side length, using an expansion moves algorithm (Boykov, Veksler, & Zabih, 2001). The resulting polygons were approximately equilateral, with a mean standard deviation in length of 2.0%. 
Maximum entropy sampling
The creation of closed metamer stimuli depends upon a fair sampling from the space Display Formula\({{\cal S}_N}\) of all simple (nonintersecting) polygons of length N. To achieve this, we used a random walk over polygons inspired by the work of Goldin et al. (2010). In particular, we devised an isoperimetric procedure for incrementally changing the turning angles of a shape while not altering any of the segment lengths, so that the polygon remains equilateral. Further, the resulting shape must not self-intersect. Note that almost all angle updates will not satisfy these requirements. For example, changing only one turning angle will generally require the length of at least one segment to change. 
What is the minimum number of turning angles that must be changed in order to preserve the equilateral property? Consider the triangle defined by a random triplet of points (generally not neighboring points) on the contour. Note that moving any of these three points relative to each other must entail a length change in at least one of the triangle sides, as the side lengths uniquely determine the triangle. But now consider a tetragon defined by a quartet of points on the contour. In this case, there is a one-dimensional family of tetragons with exactly the same sequence of side lengths. Let us express the original contour as a concatenation of four fragments, each completed by one of the four tetragon segments (Figure A1). Imagine continuously rotating through this one-dimensional family, keeping each contour fragment rigidly attached to its associated tetragon segment, but allowing it to rotate relative to the adjacent fragments. Since the sides of the tetragon remain fixed in length, all of the sides of the associated contour fragments will also remain fixed in length; only four of the turning angles will change, and these four changes must sum to 0. In this way, an isoperimetric shape change preserving the equilateral nature of the contour has been induced by a change to just four of the turning angles. Note that perceptually the change in shape can be dramatic if the sampled quartet of points is widely spaced. 
Figure A1
 
Constructing the next random shape in the sequence. (a) Random shape (solid) and associated tetragon (dashed) at Iteration i. (b) Random isoperimetric deformation of the tetragon, preserving side lengths. (c) Deformed shape induced by reattaching four shape fragments to rotated tetragon segments.
Figure A1
 
Constructing the next random shape in the sequence. (a) Random shape (solid) and associated tetragon (dashed) at Iteration i. (b) Random isoperimetric deformation of the tetragon, preserving side lengths. (c) Deformed shape induced by reattaching four shape fragments to rotated tetragon segments.
Higher order inscribing polygons could also be used. For example, if a quintet of points is sampled, the result will be a two-dimensional family of pentagons with exactly the same side lengths. Note, however, that in general a number of shape changes must be proposed before finding one that does not induce a self-intersection. We use the tetragon method because it involves the least computation, and is thus practical even when the proportion of valid proposals is small. 
To be precise, given a nonintersecting polygon Pn, we construct a new polygon Pn+1 by first randomly selecting four vertices from Pn. These four points define a tetragon T with edges of length Display Formula\({\ell _1},{\ell _2},{\ell _3},{\ell _4}\), corresponding to four fragments p1, p2, p3, p4 of the original polygon Pn. Note that this tetragon is only one member of a one-dimensional family Display Formula\({\cal T}\) of tetragons with identical edge length sequences but different turning angles. To produce a new shape, we sample uniformly from this one-dimensional family to generate a new tetragon T and replace the edges of this tetragon with the four polygon fragments p1, p2, p3, p4 to form a new candidate polygon P. We then check P for self-intersections: If none are found, we set Pn+1 = P. Otherwise, we draw another uniform sample T from Display Formula\({\cal T}\). If a valid nonintersecting polygon has not been generated after five draws, we draw another random tetragon T and repeat this process. We initialize the sequence by setting all turning angles to the same value of 2π/120, forming a regular (nonintersecting) 120-sided polygon. 
The random walk constrained in this way is a Markov chain satisfying detailed balance for a uniform distribution on Display Formula\({{\cal S}_N}\), and it is irreducible. Therefore, the stationary distribution of the Markov chain is the uniform distribution on Display Formula\({{\cal S}_N}\) (Tierney, 1996). 
Initial samples are colored by the initial conditions and are therefore not fair samples from the uniform distribution on Display Formula\({{\cal S}_N}\), necessitating a burn-in period prior to selecting samples for analysis or as psychophysical stimuli. In addition, correlation between successive samples in the chain increases the number of samples required for parameter estimation. To minimize correlation between shapes used for parameter estimation, we apply a thinning method (Gelman, 1996), selecting samples at regularly spaced intervals along the chain. Sample size, burn-in and thinning parameters are determined using a standard method (Raftery & Lewis, 1996): We used 38,415 samples. The ratio of between-chain variance to within-chain variance was less than 0.973 for all chains, reflecting reliable convergence of our Markov-chain Monte Carlo sampling method (Gelman, 1996). 
Matching the turning angles of animal shapes
The foregoing procedure creates closed, nonintersecting polygons that are otherwise maximum entropy. To form the desired closed metamer control stimuli, we wish to add the additional constraint that the probability distribution over turning angles matches that of our animal shapes. 
To this end, we used kernel density estimation with cross validated bandwidth (Wassermann, 2006) to describe the marginal distribution of the turning angles for our animal shapes and then performed Metropolis-Hastings sampling (Metropolis et al., 1953; Hastings, 1970) with this density model as the target distribution. Given this kernel density model, we can generate samples of the desired constrained maximum entropy shapes using a modification of the random walk model. In particular, instead of accepting every candidate shape that has no intersections, we accept candidate shapes conditionally (Zhu, 1999): If f(P) ≥ f(Pn) we always set Pn+1 = P. If, however, f(P) < f(Pn), we set Pn+1 = P with probability f(P)/f(Pn). This is a special form of Metropolis-Hastings sampling (Metropolis et al., 1953; Hastings, 1970) and is guaranteed to sample fairly from the distribution associated with f in the long run. Samples size, burn-in and thinning parameters were determined automatically (Raftery & Lewis, 1996). 
Figure 1
 
The problem of contour grouping. Neurons in primary visual cortex (V1) code the local orientation of contours in the image, but without explicit representation of how they are linked together. Our percept, on the other hand, is of coherent and complete objects, segmented from their background.
Figure 1
 
The problem of contour grouping. Neurons in primary visual cortex (V1) code the local orientation of contours in the image, but without explicit representation of how they are linked together. Our percept, on the other hand, is of coherent and complete objects, segmented from their background.
Figure 2
 
(a) Local geometric cues to contour grouping. Smoothness properties of objects give rise to the classical Gestalt cues of proximity and good continuation (Wertheimer, 1938). For a pair of local oriented elements these can be expressed through a variable d encoding the separation of the elements and two angles (α,β) induced by a linear interpolation between the elements. (b) Posterior probability, derived from ecological statistics, that a local oriented element (shown in black) should be grouped directly with a second observed element, as a function of its proximity, direction, and relative orientation (Elder & Goldberg 2002). (c) Local Gestalt cues, together with a Markov independence assumption, lead to an efficient framework for extracting extended contours.
Figure 2
 
(a) Local geometric cues to contour grouping. Smoothness properties of objects give rise to the classical Gestalt cues of proximity and good continuation (Wertheimer, 1938). For a pair of local oriented elements these can be expressed through a variable d encoding the separation of the elements and two angles (α,β) induced by a linear interpolation between the elements. (b) Posterior probability, derived from ecological statistics, that a local oriented element (shown in black) should be grouped directly with a second observed element, as a function of its proximity, direction, and relative orientation (Elder & Goldberg 2002). (c) Local Gestalt cues, together with a Markov independence assumption, lead to an efficient framework for extracting extended contours.
Figure 3
 
Example animal shapes extracted from the Hemera photo-object database.
Figure 3
 
Example animal shapes extracted from the Hemera photo-object database.
Figure 4
 
Alternating visible and gap segments on polygonal target shapes.
Figure 4
 
Alternating visible and gap segments on polygonal target shapes.
Figure 5
 
Example stimuli.
Figure 5
 
Example stimuli.
Figure 6
 
Stimulus sequence.
Figure 6
 
Stimulus sequence.
Figure 7
 
(a) Open metamer construction. The lengths li of elements and the geometric relationships (di, αi, βi) between successive elements of animal contours (left) are transcribed in a random order to form first-order metamers (right). Only the transcription of the geometric relationships is shown, for clarity. (b) Example of animal and (open) metamer stimuli.
Figure 7
 
(a) Open metamer construction. The lengths li of elements and the geometric relationships (di, αi, βi) between successive elements of animal contours (left) are transcribed in a random order to form first-order metamers (right). Only the transcription of the geometric relationships is shown, for clarity. (b) Example of animal and (open) metamer stimuli.
Figure 8
 
Experiment 1 results: animal contours versus first-order metamers M1 (with convexity bias) and M2 (without convexity bias). Left: mean noise thresholds as a function of the number of oriented elements sampling the target shape (target length). Right: pairwise differences between conditions. Error bars indicate standard error of the mean.
Figure 8
 
Experiment 1 results: animal contours versus first-order metamers M1 (with convexity bias) and M2 (without convexity bias). Left: mean noise thresholds as a function of the number of oriented elements sampling the target shape (target length). Right: pairwise differences between conditions. Error bars indicate standard error of the mean.
Figure 9
 
Experiment 1 analysis: target area. (a) Mean area of the convex hull of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Histogram of areas of equalized subsets of animal and first-order metamer M1 stimuli (with convexity bias). (c) Mean noise thresholds for area-equalized animal and M1 targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 9
 
Experiment 1 analysis: target area. (a) Mean area of the convex hull of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Histogram of areas of equalized subsets of animal and first-order metamer M1 stimuli (with convexity bias). (c) Mean noise thresholds for area-equalized animal and M1 targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 10
 
Experiment 1 analysis: target elongation. (a) Mean elongation of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Mean noise thresholds for elongation-equalized animal and M1 metamer targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 10
 
Experiment 1 analysis: target elongation. (a) Mean elongation of the animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. (b) Mean noise thresholds for elongation-equalized animal and M1 metamer targets (with convexity bias) for target-length = 20 condition. Error bars indicate standard error of the mean.
Figure 11
 
Experiment 1 analysis: eccentricity and self-intersections. (a) Mean eccentricity of the centroid of the target animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. Error bars indicate standard error of the mean. (b) Mean noise thresholds for animal targets and the subset of open metamer targets that do not self-intersect. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 11
 
Experiment 1 analysis: eccentricity and self-intersections. (a) Mean eccentricity of the centroid of the target animal and first-order metamer M1 (with convexity bias) and M2 (without convexity bias) stimuli. Error bars indicate standard error of the mean. (b) Mean noise thresholds for animal targets and the subset of open metamer targets that do not self-intersect. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 12
 
Experiment 1 analysis: mean number of inflections for animal and metamer stimuli. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 12
 
Experiment 1 analysis: mean number of inflections for animal and metamer stimuli. M1: metamers with convexity bias; M2: metamers without convexity bias. Error bars indicate standard error of the mean.
Figure 13
 
Experiment 2 results: mean noise thresholds and contrasts for animal and first-order metamer M1 (with convexity bias) target stimuli, as a function of target length and degree of closure for animal targets. Error bars indicate standard error of the mean.
Figure 13
 
Experiment 2 results: mean noise thresholds and contrasts for animal and first-order metamer M1 (with convexity bias) target stimuli, as a function of target length and degree of closure for animal targets. Error bars indicate standard error of the mean.
Figure 14
 
Experiment 3. The probability density of turning angles for our closed metamer shapes is a close match to the turning angle density for the original animal shapes.
Figure 14
 
Experiment 3. The probability density of turning angles for our closed metamer shapes is a close match to the turning angle density for the original animal shapes.
Figure 15
 
Experiment 3 results. A: Animal targets; M1: first-order metamers (open, with convexity bias); M3: first-order metamers (closed). Left: mean noise thresholds. Right: pairwise contrasts. Error bars indicate standard error of the mean.
Figure 15
 
Experiment 3 results. A: Animal targets; M1: first-order metamers (open, with convexity bias); M3: first-order metamers (closed). Left: mean noise thresholds. Right: pairwise contrasts. Error bars indicate standard error of the mean.
Figure 16
 
Closure as a bridge from 1D to 2D shape (Elder & Zucker, 1993, 1994). (a) Shape discrimination is easy when good continuation and closure are strong. (b) Discrimination becomes hard when good continuation and closure are weak.
Figure 16
 
Closure as a bridge from 1D to 2D shape (Elder & Zucker, 1993, 1994). (a) Shape discrimination is easy when good continuation and closure are strong. (b) Discrimination becomes hard when good continuation and closure are weak.
Figure A1
 
Constructing the next random shape in the sequence. (a) Random shape (solid) and associated tetragon (dashed) at Iteration i. (b) Random isoperimetric deformation of the tetragon, preserving side lengths. (c) Deformed shape induced by reattaching four shape fragments to rotated tetragon segments.
Figure A1
 
Constructing the next random shape in the sequence. (a) Random shape (solid) and associated tetragon (dashed) at Iteration i. (b) Random isoperimetric deformation of the tetragon, preserving side lengths. (c) Deformed shape induced by reattaching four shape fragments to rotated tetragon segments.
Table 1
 
Experiment 1: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (with convexity bias) and M2 (without convexity bias).
Table 1
 
Experiment 1: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (with convexity bias) and M2 (without convexity bias).
Table 2
 
Experiment 2: Statistical analysis of noise threshold contrasts for animal targets with progressive degrees of closure vs. M1 metamer controls, based on two-tailed matched-sample t tests over observers.
Table 2
 
Experiment 2: Statistical analysis of noise threshold contrasts for animal targets with progressive degrees of closure vs. M1 metamer controls, based on two-tailed matched-sample t tests over observers.
Table 3
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Table 3
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Table 4
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
Table 4
 
Experiment 3: Statistical analysis of noise threshold contrasts based on two-tailed matched-sample t tests over observers. Note: Animal targets contrasted with first-order metamers M1 (open, with convexity bias) and M3 (closed).
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×