Free
Research Article  |   May 2004
The role of spatial interactions in perceptual synchrony
Author Affiliations
  • Isamu Motoyoshi
    Human and Information Science Laboratory, NTT Communication Science Laboratories, NTT Corporation, Atsugi, Japanmotoyosi@apollo3.brl.ntt.co.jp
Journal of Vision May 2004, Vol.4, 1. doi:https://doi.org/10.1167/4.5.1
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Isamu Motoyoshi; The role of spatial interactions in perceptual synchrony. Journal of Vision 2004;4(5):1. https://doi.org/10.1167/4.5.1.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

To understand how the visual system processes synchronies between visual patterns, we investigated the temporal acuity for detecting a Gabor pattern whose orientation was alternated (vertical-horizontal) in a different temporal phase from three other Gabor patterns. Thresholds of both advanced and lagged temporal-phase offsets were measured for various temporal frequencies of orientation alternation and for various spatial distances between Gabor patterns. The thresholds for advanced phase offsets were lower than those for lagged phase offsets; the target pattern whose orientation changed earlier than the others was easier to detect than the target whose orientation changed later by the same amount. It was found that the amount of this temporal asymmetry increased proportionally with the distance between patterns. The upper temporal-frequency limit of orientation alternation for detecting the target pattern also systematically decreased with the distance between patterns. These results were interpreted as reflecting the temporal dynamics of mutual interactions between local orientation detectors, which necessarily involve a greater degree of temporal blur and longer delays of interacting signals as the spatial distance between detectors increases. This explanation leads to the notion that perceptual synchrony between visual patterns is determined in a space-time relative manner.

Introduction
Synchrony is an important factor not only in motion vision but also in pattern vision. The visual system has a strong tendency to group synchronous patterns into a region while segmenting asynchronous ones as distinct regions, regardless of whether the asynchronies cause the sensation of motion. A number of psychophysical studies have investigated critical roles of synchrony and of other temporal factors in a variety of pattern-vision tasks, including figure-ground segmentation (Fahle, 1993; Leonards, Singer, & Fahle, 1996; Rogers-Ramachandran & Ramachandran, 1998; Forte, Hogben, & Ross, 1999; Lee & Blake, 1999; Motoyoshi & Nishida, 2001, 2002; Farid & Adelson, 2001), contour linking (Usher & Donnelly, 1998; Hess, Beaudot, & Mullen, 2001; Bex, Simmers, & Dakin, 2001; Beaudot, 2002; Dakin & Bex, 2002), vernier acuity (Westheimer & McKee, 1977), and object recognition (Blake & Yang, 1997; Alais, Blake, & Lee, 1998; Holcombe & Cavanagh, 2001; Motoyoshi, 2002). 
Synchrony is also one of the basic attributes in temporal vision. In general, physically synchronous events are perceived to occur simultaneously while asynchronous events are perceived to occur at different moments. However, recent evidence has revealed that physical synchrony of events does not always give rise to perceptual synchrony (Hikosaka, Miyauchi, & Shimojo, 1993; Moutoussis & Zeki, 1997; Nishida & Johnston, 2002). These findings have triggered an intense discussion concerning the structure of perceptual space-time. 
How does the visual system process synchrony/asynchrony between patterns? In most studies, it has been assumed, either explicitly or implicitly, that the synchrony between patterns is subserved by a mechanism that compares features of the patterns across space in the neural snapshot at a certain moment. This is a natural assumption considering the fact that the judgment of asynchrony/synchrony between patterns is behaviorally equivalent with the detection/non-detection of differences between patterns under a dynamic presentation. Under this assumption, if the mechanism detected a difference between patterns in the neural snapshot, which might, of course, be a temporally integrated representation of the preceding processes, the patterns are perceived as asynchronous, and if not, synchronous. Such a mechanism is conceptually equivalent to the basic models for spatial pattern discrimination (e.g., first- and second-order spatial filters: Wilson, 1993). The neural snapshot model is simple and parsimonious because it requires no explicit representation of the timing of neural signals, and only refers to well-known, pattern-vision mechanisms. Several psychophysical studies have recently challenged this view by proposing the existence of special “synchrony detectors” (Usher & Donnelly, 1998; Lee & Blake, 1999), but the validity of that evidence has been essentially rejected (Farid & Adelson, 2001; Dakin & Bex, 2002). 
If the detection of asynchrony is equivalent to spatial discrimination in the neural snapshot, it can be considered in terms of the temporal dynamics of pattern-vision mechanisms. For such dynamics, recent studies have reported the intriguing phenomenon that the temporal resolution of spatial discrimination declines as the distance between patterns increases. For example, segregation of texture regions in which the luminance contrast of elements is rapidly reversed out of phase is dramatically impaired when a spatial gap is introduced between the textures (Rogers-Ramachandran & Ramachandran, 1998; Forte et al., 1999). It is also known that the temporal resolution of texture segregation and contour linking declines with the spatial range of the stimuli to be integrated (Motoyoshi & Nishida, 2001, 2002; Hess et al., 2001). 
These spatiotemporal characteristics of pattern discrimination seem to provide an important insight to the computational basis of asynchrony detection. In addition, the relationship between simultaneity and space-time is generally a fundamental issue for understanding the structure of both the physical and the perceptual world. Using a highly simplified stimulus, the present study specifically investigated how the temporal sensitivity for asynchronies (i.e., temporal-phase differences) depends on the spatial distance between stimuli. The results showed an advantage in detection of asynchronous patterns that changes “earlier” than the others, as has been reported in a recent contour-detection study (Beaudot, 2002). The current analysis revealed that this temporal asymmetry increased proportionally with the spatial distance between patterns. The results also indicated a systematic decline of the temporal resolution (upper temporal-frequency limit) with the distance between patterns (Forte et al., 1999; Hess et al., 2001; Motoyoshi & Nishida, 2002). These spatiotemporal interactions were simply explained in terms of the temporal dynamics of mutual interactions between local pattern detectors, which necessarily involve a greater degree of temporal blur and longer delays of interacting signals as the spatial distance between detectors increases. This model indicated the possibility that perceptual synchrony/asynchrony between visual patterns is determined in a space-time relative manner. 
Methods
Stimuli
The stimulus consisted of a square array containing four local patterns whose orientations were periodically alternated in time between vertical and horizontal (Figure 1). The four patterns were separated by a variable spatial distance (inter-element distance [IED]; Figure 1a and 1b). Each pattern consisted of overlapping vertical and horizontal Gabor patches. Each Gabor patch had a spatial wavelength (λ) of 0.17 deg (spatial frequency was 6.0 c/deg) and a SD (σ) of 0.08 deg. The contrasts of the two patches were sinusoidally modulated out of phase around the mean (0.4) at a variable temporal frequency (f) and with an amplitude (m) of 1.0. The luminance-contrast profile of the element pattern (C(x, y, t)) was given as follows:  
(1)
where  
(2)
The out-of-phase contrast modulation produced a gradual horizontal-vertical alternation of the pattern orientation (Figure 1c and d). One of the four patterns (target pattern) alternated its orientation with a temporal phase offset by Δφ from the others (Figure 1c). The absolute temporal phase of the orientation alternation of all element patterns was randomly set for each presentation. All patterns had a mean luminance of 53 cd/m2, which was equal to the luminance of the uniform gray background of 13.3 (H) × 10.0 (V) deg. 
Figure 1
 
Stimuli used in the experiment. Snapshots of stimuli with an inter-element distance of 6λ (a) and 2λ (b), respectively. (c). The waveform of the out-of-phase contrast modulation (orientation alternation) of horizontal and vertical Gabor patches for the target pattern (black line) and for the other patterns (gray line). (d). Sequential images illustrating the orientation alternation of the target (top) and of the others (bottom).
Figure 1
 
Stimuli used in the experiment. Snapshots of stimuli with an inter-element distance of 6λ (a) and 2λ (b), respectively. (c). The waveform of the out-of-phase contrast modulation (orientation alternation) of horizontal and vertical Gabor patches for the target pattern (black line) and for the other patterns (gray line). (d). Sequential images illustrating the orientation alternation of the target (top) and of the others (bottom).
Procedure
Thresholds of the temporal phase offset (Δφ) were measured for detecting the target pattern by means of a spatial four-alternative forced-choice (4AFC) task. In each trial, the stimulus was presented for 1 s within a temporal rectangular window tapered by the positive half-cycle of a cosine function with a wavelength of 167 ms. The observer indicated the location of the target pattern among the four (upper left, upper right, lower left, or lower right), while steadily fixating on the cross in the center of the array. An incorrect response was followed by a feedback tone. No observers reported perceiving motion between Gabor patterns at all. The phase-offset limit was estimated by means of the double-random staircase method (one-up/one-down, step size was 8.3 ms). Each staircase terminated at the eighth reversal of the up-down sequence. The phase-offset limit, giving the 62.5% correct response, was estimated by probit analysis using all binary responses after at least five staircase measurements for each condition. 
There were two types of phase-offset limits; one was the advanced phase-offset limits (Δφ < 0°), and the other was the lagged phase-offset limits (Δφ > 0°). Both phase-offset limits were measured for various temporal frequencies and for various IEDs. For large advanced and lagged phase offsets, the upper temporal-frequency limits were measured using an analogous staircase procedure in which the temporal wavelength of the orientation alternation was varied by 8.3 ms in accordance with the observer’s response. The measurements for different temporal frequencies, phases, and IEDs were performed in separate blocks. 
Apparatus
Stimuli were generated by a VSG2/5 card (Cambridge Research Systems) hosted by a computer (Dell Dimension XPS T700r) and displayed on a 21-in CRT (Sony GDM F500R) with a refresh rate of 120 Hz, and a luminance resolution of 14 bits. The pixel resolution of the CRT was 1 min/pixel at the viewing distance of 143 cm. 
Observers
Two naïves (CH and YT) and the author (IM) served as observers. All had corrected-to-normal vision. 
Results
Figure 2 shows the relationships between the phase-offset limit (Δφ, abscissa) and the upper temporal-frequency limit (f, ordinate) of orientation alternation. A negative value of Δφ indicates the advanced phase offset of the target relative to the others, whereas a positive value indicates the lagged phase offset. It is clear that both phase-offset limits increase (on the abscissa) as the temporal frequency of orientation alternation increases and as the distance between element patterns (IED) increases. 
Figure 2
 
The phase-offset limits between the target and the other patterns (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). The data in the left half of the panel show the lagged phase-offset limits (Δφ > 0), and those in the right half show the advanced phase-offset limits (Δφ < 0). Different color of symbols shows the results for different inter-element distances (IEDs). Smooth lines represent the fitted curves described in the text. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 2
 
The phase-offset limits between the target and the other patterns (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). The data in the left half of the panel show the lagged phase-offset limits (Δφ > 0), and those in the right half show the advanced phase-offset limits (Δφ < 0). Different color of symbols shows the results for different inter-element distances (IEDs). Smooth lines represent the fitted curves described in the text. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
An inspection of Figure 2 shows that the threshold functions are asymmetric with respect to a phase offset of 180°. The functions appear horizontally shifted toward the right, especially when the IED is large. The asymmetry is more clearly shown in Figure 3, in which portions of the data are replotted as absolute values (|Δφ|°, abscissa). The advanced phase-offset limits are lower than the lagged ones for a range of temporal frequencies; the target pattern whose orientation changed earlier (later) than the others was perceived to be more asynchronous (synchronous) as compared to the target whose orientation changed later (earlier) than the others. Moreover, these differences become larger as the IED increases. 
Figure 3
 
The advanced and lagged phase-offset limits in Figure 2 replotted as absolute value (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). Closed symbols show the lagged phase-offset limits, and open symbols show the advanced phase-offset limits. The color of symbols represents the inter-element distance (IED); only the results for 2, 6, and 10λ are shown for clarity. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 3
 
The advanced and lagged phase-offset limits in Figure 2 replotted as absolute value (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). Closed symbols show the lagged phase-offset limits, and open symbols show the advanced phase-offset limits. The color of symbols represents the inter-element distance (IED); only the results for 2, 6, and 10λ are shown for clarity. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
We sought to quantify the degree of the asymmetry (difference between the advanced and lagged phase offsets) using all threshold data in Figure 2. For each IED, the threshold data were well approximated by a parabolic function (solid lines in Figure 2) given as f = aφ−180−k)2 + b, where Δφ is the phase offset (0–360°), f is the temporal frequency, k the angular shift in the phase (°), and a and b are constants. Because the approximation was found to be better in almost all cases when the asymmetries were defined in milliseconds rather than degrees, k was calculated from a constant value in millisecond (s) (i.e., k = sφ−180)2 + b). Figure 4 shows the estimated difference between the two phase-offset limits in milliseconds. The difference increases with the IED in an almost linear fashion with slopes (ms/λ) of 1.51 for observer CH, 2.15 for YT, 3.06 for IM, and 2.26 for the average across observers, respectively. Thus the temporal asymmetry is proportional to the IED. 
Figure 4
 
The estimated difference between the lagged and advanced phase-offset limits as a function of the inter-element distance. The lines are fitted linear functions.
Figure 4
 
The estimated difference between the lagged and advanced phase-offset limits as a function of the inter-element distance. The lines are fitted linear functions.
Is the temporal asymmetry proportional to the IED in absolute visual angle (deg) or on scale relative to the size of the element pattern (λ)? In other words, is the temporal asymmetry scale invariant? To answer this question, we measured the phase-offset limits at the temporal frequency of 4 Hz, using identical stimuli but now viewed at half the viewing distance (77 cm, Image not available deg); thus, the stimulus size was twice that of the original display. Figure 5 shows the difference between the advanced and lagged phase-offset limits, plotted against the IED in λ. If the difference between the two phase-offset limits depends on the IED in visual angle, it should have increased with a slope twice as steep as that obtained for the original stimuli (gray lines). However, this difference increases with a slope comparable to that for the original stimuli (black lines). This indicates that the temporal asymmetry depends not on the distance in visual angle but on the distance relative to the size of the pattern. Thus, the temporal asymmetry is scale invariant. 
Figure 5
 
The difference between the lagged and advanced phase-offset limits as a function of the inter-element distance, obtained for double-scaled stimuli with a temporal frequency of 4 Hz (green circles). Red circles represent the results for the original stimuli at 4 Hz. Lines are predictions from the results for the original stimuli (Figure 4) if the difference between the two phase-offset limits depends on the inter-element distance in visual angle (gray line) and in the relative scale to the pattern size (black line), respectively. The top panel shows the results for observer IM, and the bottom the results for YT, respectively.
Figure 5
 
The difference between the lagged and advanced phase-offset limits as a function of the inter-element distance, obtained for double-scaled stimuli with a temporal frequency of 4 Hz (green circles). Red circles represent the results for the original stimuli at 4 Hz. Lines are predictions from the results for the original stimuli (Figure 4) if the difference between the two phase-offset limits depends on the inter-element distance in visual angle (gray line) and in the relative scale to the pattern size (black line), respectively. The top panel shows the results for observer IM, and the bottom the results for YT, respectively.
The phase offset of 180° (= −180°) means that the orientation alternation of the target and of the other elements were physically completely counter phase. The upper temporal-frequency limit at this phase offset is considered to be a representative measure of the temporal resolution of asynchrony detection. Here we call it the critical temporal frequency. Figure 6 shows the critical temporal frequency plotted as a function of the IED. As suggested in the previous studies (Forte et al., 1999; Motoyoshi & Nishida, 2001, 2002), the temporal resolution declines as the IED increases. Moreover, the present data further suggest that the decline is systematic. The critical temporal frequency decreases almost linearly with the IED in log-log coordinates; the data were well approximated by a*IEDb, where [a, b] were [11.2, −0.24] for observer CH, [13.6, −0.47] for YT, [18.3, −0.57] for IM, and [14.3, −0.44] for the average across observers, respectively. 
Figure 6
 
The critical temporal frequency of orientation alternation with a phase-offset of 180° as a function of the inter-element distance. The lines are fitted linear functions on log-log coordinates.
Figure 6
 
The critical temporal frequency of orientation alternation with a phase-offset of 180° as a function of the inter-element distance. The lines are fitted linear functions on log-log coordinates.
Discussion
In the present study, we systematically investigated the temporal accuracy of human observers for detecting asynchronies between oriented patterns. The results demonstrated, as suggested by previous findings (Forte et al., 1999; Motoyoshi & Nishida, 2001, 2002; Beaudot, 2002), the decline in the temporal resolution with the spatial distance between patterns, and the asymmetry with regard to the polarity of asynchrony; observers were more sensitive to the advanced phase offsets than to the lagged ones. Beaudot (2002) has interpreted this temporal asymmetry as a particular facilitatory effect in the synchrony processing. However, we found that the degree of temporal asymmetry increased proportionally with the distance between patterns expressed on a relative scale. The space-dependent temporal dynamics seem to provide a further insight for the neural basis of synchrony processing. 
Recent neurophysiological findings suggest that spatial discrimination between oriented patterns is, at least partially, subserved by inhibitory interactions between orientation-selective neurons. The response of V1 neurons to oriented stimuli is suppressed by surrounding stimuli with the same orientation, but not by those with different orientations (Blakemore & Tobin, 1972; Gilbert & Wiesel, 1990; Knierim & Van Essen, 1992; Sillito, Grieve, Jones, Cuderio, & Davis, 1995; Zipser, Lamme, & Schiller, 1996; Lee, Mumford, Romero, & Lamme, 1998). These suppressive effects are believed to be based on inhibitory, horizontal connections between V1 neurons and/or feedback connections from higher visual areas (Gilbert & Wiesel, 1990; Knierim & Van Essen, 1992; Zipser et al., 1996; Lee et al., 1998; Li, 1999). Importantly, these neural interactions have been shown to spread gradually in time (Grinvald, Lieke, Frostig, & Hildesheim, 1994; Lee et al., 1998; but see Bair, Cavanaugh, & Movshon, 2003). 
Such temporal dynamics of neural interactions in V1 are consistent with the present psychophysical data that show a highly systematic dependency of the temporal asymmetry on the spatial distance between oriented patterns. Figure 7 illustrates a schematic model of such neural interactions. In this model, spatial differences between oriented patterns are detected via iterative interactions between first-stage, local orientation detectors, each responding to a local oriented pattern, but also inhibited by other orientation detectors tuned to the same orientation (cf. Li, 1999). The model detects the target pattern by selecting the least inhibited orientation detector, or by further comparing the detectors’ outputs in the higher stages. Because the detection of asynchrony is equivalent with the spatial discrimination between dynamic patterns as noted in the introduction, this model is responsible also for the detection of asynchronies. Thus, the difference in the output between the detectors is defined as the signal of asynchrony. In case of our stimuli, those differential outputs, when averaged temporally, are expected to increase as the phase offset between the target and the others increases and gradually peak at the phase offset of 180°, and also decrease as the temporal frequency increases. This would give rise to the rounded threshold curves as observed (Figure 2). 
Figure 7
 
(a). Schematic model of asynchrony detection between oriented patterns. The model consists of two stages. At the first stage, individual local patterns are encoded by orientation detectors having a specific orientation tuning (only vertical is shown) and specific temporal tunings. The second stage consists of inhibitory interactions between the orientation detectors. The response of each detector is inhibited by signals propagating from the other detectors via neural connections with specific temporal tunings. (b). As a result, the propagating signals are temporally blurred and mutually delayed as the distance between detectors increases. See text for details.
Figure 7
 
(a). Schematic model of asynchrony detection between oriented patterns. The model consists of two stages. At the first stage, individual local patterns are encoded by orientation detectors having a specific orientation tuning (only vertical is shown) and specific temporal tunings. The second stage consists of inhibitory interactions between the orientation detectors. The response of each detector is inhibited by signals propagating from the other detectors via neural connections with specific temporal tunings. (b). As a result, the propagating signals are temporally blurred and mutually delayed as the distance between detectors increases. See text for details.
An important feature of this model is that signals between orientation detectors propagate with a finite velocity through biological media such as axons and inter-neurons. The interactions necessarily involve temporal blur and mutual delays of signals between the detectors. Moreover, if the media have a spatially uniform temporal property, the temporal blur and mutual delays should increase proportionally to the spatial distance between the detectors. These mutual delays of signals seem responsible for the space-dependent temporal asymmetry, and the blur for the decline in the temporal resolution. 
The mutual delays of propagating signals in the network of Figure 7 indeed predict the degree of temporal asymmetry to be proportional to the spatial distance. In the case of our stimuli, consisting of one asynchronous target and three other synchronous patterns, the delayed interaction produces a temporally asymmetric effect on the detection of the target. Thus, the response to the target is more suppressed by signals from the three non-target patterns when these precede the target, whereas the three non-target patterns are suppressed only by the single target-pattern when the target precedes the three non-target patterns. This results in easier detection of the target with advanced phase offsets compared to lagged phase offsets. Moreover, given that the amount of delay is proportional to the distance of signal propagation, it is natural that the temporal asymmetry increases proportionally to the spatial distance (Figure 4). Similar dynamics of signal propagation have also been demonstrated as masking phenomena, in the context of brightness and texture filling-in (Paradiso & Nakayama, 1991; Caputo, 1998; Motoyoshi, 1999). 
The above notion was confirmed by a numerical simulation based on a simplified version of the mutual interaction model, in which the output of a local orientation detector was subtracted from each other with delays (see “Appendix”). Figure 8 shows (a) the hypothetical sensitivity of the model to the target element and (b) the predicted phase-offset versus temporal-frequency curves. Although the prediction is not quantitatively perfect because the model was extremely simple, the simulated results duplicate the overall shape of the threshold curves and the temporal asymmetry well; the simulated curves is also horizontally shifted toward the right when the IED is large. 
Figure 8
 
Results of simulations with a simple mutual-interaction model. (a). The model sensitivity to the target element when the IED was 2λ (left) and 10λ (right). (b). Predicted phase-offset versus temporal-frequency curves. Filled circles are the observed average data (Figure 2).
Figure 8
 
Results of simulations with a simple mutual-interaction model. (a). The model sensitivity to the target element when the IED was 2λ (left) and 10λ (right). (b). Predicted phase-offset versus temporal-frequency curves. Filled circles are the observed average data (Figure 2).
Whereas temporal asymmetry increased with the distance between patterns (IED) in a linear fashion, the dependence of the critical temporal frequency of orientation alternation on IED was linear only when expressed in log-log coordinates (Figure 6). This result appears to be inconsistent with the assumption of the model that temporal blur increases proportionally to the distance between the orientation detectors. It should be noted, however, that the cut-off temporal frequency is not a direct measure of the amount of blur (i.e., one would need to know the temporal-MTF of the system), and does not necessarily suggest by itself the involvement of spatial interactions. Nevertheless, the highly systematic dependency of the temporal resolution on the IED we obtained here implies a simple relationship between temporal blur and distance between detectors (see also the “Appendix”). Further investigations will be required regarding this issue. 
Are there other possibilities that can explain the space-dependent temporal asymmetry? In our stimuli, the local patterns were located at more peripheral locations in the visual field as the IED increased. The absolute sensitivity for pattern detection/discrimination is known to decline gradually as the retinal eccentricity of stimuli increases (e.g., Levi, Klein, & Aitsebaomo, 1985). Thus, it is likely that the decline in the temporal resolution in our task was at least partially caused by the increase in the eccentricity of Gabor patterns rather than the distance between them (we did not examine this possibility because the temporal “resolution” was not in the main scope of the present study). On the other hand, the increase in the stimulus eccentricity might also cause an increase in latency of the direct response for individual patterns. However, these increases should always be equal for all four patterns because they were designed to be located at the same retinal eccentricity, and thus should not predict the temporal asymmetry. 
Although the spatial-interaction model is consistent with the dynamics of neural responses of V1 cells, it seems unrealistic to assume that such a network is exactly implemented as a neural circuit in V1. First, if the temporal asymmetry is determined by the signal delays between V1 neurons, it should have depended on the cortical distance between neurons (i.e., log-IED in visual angle). This is obviously inconsistent with the fact that the temporal asymmetry is linearly related to the distance between the element patterns relative to their size. Second, it is impossible to discriminate between patterns across the vertical meridian using only the neural circuit in V1, which represents the left and right visual fields separately. Finally, recent electrophysiological analyses have revealed that the neural interactions involve spatial anisotropy (Kapadia, Westheimer, & Gilbert, 2000), is not subtractive but divisive (Cavanaugh, Bair, & Movshon, 2002), and critically involve feed-back signals from higher cortical areas (Zipser et al., 1996; Lee et al., 1998; Hupe, James, Payne, Lomber, Girard, & Bullier, 1998; Angelucci, Levitt, Walton, Hupe, Bullier, & Lund, 2002; Bair et al., 2003). These studies thus suggest that the model assumed in the present analysis is too simple. The temporal blur and delay of signals in the recurrent network might not necessarily indicate a specific neural structure in V1 but a computational principle of the entire visual system in the detection of asynchrony. From a phenomenological viewpoint, this further indicates the relativity of perceptual space-time. 
Implications for temporal localization
The temporal asymmetry in asynchrony detection demonstrates that the relationships between physical synchrony and perceptual synchrony are more complicated than previously thought. As mentioned earlier, such relationships have been reported in other psychophysical experiments also. For example, it is known that attended stimuli are perceived earlier than unattended ones (Sternberg & Knoll, 1973; Hikosaka et al., 1993), and that gradually changing visual features are perceived to lead flashed ones (Nijhawan, 1994; Whitney & Murakami, 1998; Sheth, Nijhawan, & Shimojo, 2000). It has also been shown that alternation of color (e.g., red-green) appears to lead alternation of the direction of motion (e.g., upward-downward) when the alternation rate is relatively high (Moutoussis & Zeki, 1997; Nishida & Johnston, 2002). A common explanation for these illusory-synchrony phenomena is the difference in the latency of the direct response to stimuli (but, see Nishida & Johnston, 2002). However, it is difficult to explain our results in terms of the latency difference because our stimuli were designed to cancel asymmetric structures in the temporal response to individual patterns. On the other hand, our interaction-based model of perceptual synchrony may provide a new account for these phenomena. Thus the illusory asynchronies might occur not at the stage of local feature coding but at the stage of interactions between encoded feature signals. If this is indeed the case then, according to the space-time relativity, some of the above illusions would also exhibit a systematic dependency on the spatial distance between stimuli (or on the distance between neurons in the brain). 
Relations to long-range apparent motion
Finally, we mention the relationships between the temporal asymmetry in asynchrony detection and long-range (not short-range) apparent motion because the visual computation subserving the detection of asynchrony and the detection of motion seems indistinguishable, except that the polarity of asynchrony (earlier/later) is specified in motion detection. Although no explicit motion was perceived in our stimuli, the distance dependency of the temporal asymmetry in asynchrony detection is qualitatively consistent with the phenomenological law of long-range apparent motion, so-called Korte’s third law, which states that the optimal interstimulus interval (ISI) for apparent motion is proportional to the distance between stimuli (Burt & Sperling, 1981). This implies a similar computational scheme underlying both phenomena; long-range apparent motion might also be a consequence of the same dynamic spatial interactions that account for the present results. In fact, recent psychophysical evidence demonstrated a change in the perceived speed of an array of oriented patterns depending on the patterns’ orientation, which could be modeled by lagged interactions between V1 neurons (Series, Georges, Lorenceau, & Fregnac, 2002). 
Acknowledgments
The author thanks Nicolaas Prins for comments on the draft. This study was supported by Japan Society for the Promotion of Sciences (JSPS), and was partly presented at Vision Sciences Society meeting 2002 in Sarasota, FL, USA. 
Commercial relationships: none. 
Corresponding author: Isamu Motoyoshi. Email: motoyosi@apollo3.brl.ntt.co.jp
Address: Human and Information Science Laboratory, NTT Communication Science Laboratories, NTT Corporation, 3-1 Morinosato-Wakamiya, Atsugi, Japan. 
Appendix
To examine the temporally asymmetric behavior of the mutual interaction model, a numerical simulation was conducted using a simplified version of the model that only considered the temporal delay. 
The model consisted of four neural units, each of which responded linearly to the luminance contrast of a vertical (or horizontal) component of a local element. In case of our stimuli (Equation 1 in the text), the response of the i-th unit, Ii(t), was given as follows:  
(A1)
where f is the temporal frequency, and φi the temporal phase of the i-th element. Next, these responses were assumed to be suppressed via mutual interactions. Thus, the final output of the i-th unit, Ri(t), was derived as a weighted sum of the delayed outputs of the other units, Rj(t-d/c), and the direct response, Ii(t) (no cross-orientation interaction was assumed):  
(A2)
where d is the distance between the units (λ), and c the propagation velocity (λ/ms). Af,ied is the factor related to the total temporal-frequency characteristic of the model, which decreases with the temporal frequency and the IED. The temporal-frequency characteristic was defined in such an ad hoc manner for simplicity. Strictly, it must be determined by both the temporal characteristic of the local unit itself and of the mutual interactions, which cannot both be identified from the present results as mentioned in the text. Here Af,ied was defined as a form of n-stage low-pass filter, in which the number of leaky stages was assumed to increase proportionally with the IED, because it duplicates the linear relationships between the upper-temporal frequency limit and the IED on the log-log coordinates well:  
(A3)
where τ is the time constant, and κ and β are parameters that determine the number of stages. 
The final sensitivity of the model was defined as the difference of the maximum output to the target during the stimulus presentation (1,000 ms; i.e., it was assumed that observers selected the element giving the largest response during the presentation) from the average:  
(A4)
Similar but noisier results were obtained when the temporal average of R(t) instead of the peak was employed. 
Figure 8 shows the simulation results, in which parameters were adjusted to give the best fit to the average data of the three subjects. The fitting was done by minimizing the root-mean-square error between the model sensitivity, S, and the observed sensitivity (=1) at given temporal frequencies and phase offsets. The estimated values of the parameters were [A, c, τ, κ, β] = [9.87, 1.45, 15.5, 1.23, 0.21]. 
References
Alais, D. Blake, R. Lee, S. H. (1998). Visual features that vary together over time group together over space. Nature Neuroscience, 1, 160–164. [PubMed] [CrossRef] [PubMed]
Angelucci, A. Levitt, J. B. Walton, E. J. Hupe, J. M. Bullier, J. Lund, J. S. (2002). Circuits for local and global signal integration in primary visual cortex. Journal of Neuroscience, 22, 8633–8646. [PubMed] [PubMed]
Bair, W. Cavanaugh, J. R. Movshon, J. A (2003 ). Time course and time-distance relationship for surround suppression in macaque V1 neurons. Journal of Neuroscience, 23, 7690–7701. [PubMed] [PubMed]
Beaudot, W. H. (2002). Role of onset asynchrony in contour integration. Vision Research, 42, 1–9. [PubMed] [CrossRef] [PubMed]
Bex, P. J. Simmers, A. J. Dakin, S. C. (2001). Snakes and ladders: The role of temporal modulation in visual contour integration. Vision Research, 41, 3775–3782. [PubMed] [CrossRef] [PubMed]
Blake, R. Yang, Y. (1997). Spatial and temporal coherence in perceptual binding. Proceedings of the National Academy of Sciences, 94, 7115–7119. [PubMed] [CrossRef]
Blakemore, C. Tobin, E. A. (1972). Lateral inhibition between orientation detectors in the cat visual cortex. Experimental Brain Research, 15, 439–440. [PubMed] [CrossRef] [PubMed]
Burt, P. Sperling, G. (1981). Time, distance, and feature trade-offs in visual apparent motion. Psychological Review, 88, 171–195. [PubMed] [CrossRef] [PubMed]
Caputo, G. (1998). Texture brightness filling-in. Vision Research, 38, 841–851. [CrossRef] [PubMed]
Cavanaugh, J. R. Bair, W. Movshon, J. A. (2002). Nature and interaction of signals from the receptive field center and surround in macaque V1 neurons. Journal of Neurophysiology, 88, 2530–2546. [PubMed] [CrossRef] [PubMed]
Dakin, S. C. Bex, P. J. (2002). Role of synchrony in contour binding: Some transient doubts sustained. Journal of the Optical Society of America A, 19, 678–686. [PubMed] [CrossRef]
Fahle, M. (1993). Figure-ground discrimination from temporal information. Proceedings of the Royal Society of London B, 254, 199–203. [PubMed] [CrossRef]
Farid, H. Adelson, E. H. (2001). Synchrony does not promote grouping in temporally structured displays. Nature Neuroscience, 4, 875–876. [PubMed] [CrossRef] [PubMed]
Forte, J. Hogben, J. H. Ross, J. (1999). Spatial limitations of temporal segmentation. Vision Research, 39, 4052–4061. [PubMed] [CrossRef] [PubMed]
Gilbert, C. D. Wiesel, T. N. (1990). The influence of contextual stimuli on the orientation selectivity of cells in primary visual cortex of the cat. Vision Research, 30, 1689–1701. [PubMed] [CrossRef] [PubMed]
Grinvald, A. Lieke, E. E. Frostig, R. D. Hildesheim, R. (1994). Cortical point-spread function and long-range lateral interactions revealed by real-time optical imaging of macaque monkey primary visual cortex. Journal of Neuroscience, 14, 2545–2568. [PubMed] [PubMed]
Hess, R. F. Beaudot, W. H. Mullen, K. T. (2001). Dynamics of contour integration. Vision Research, 41, 1023–1037. [PubMed] [CrossRef] [PubMed]
Hikosaka, O. Miyauchi, S. Shimojo, S. (1993). Focal visual attention produces illusory temporal order and motion sensation. Vision Research, 33, 1219–1240. [PubMed] [CrossRef] [PubMed]
Holcombe, A. O. Cavanagh, P. (2001). Early binding of feature pairs for visual perception. Nature Neuroscience, 4, 127–128. [PubMed] [CrossRef] [PubMed]
Hupe, J. M. James, A. C. Payne, B. R. Lomber, S. G. Girard, P. Bullier, J. (1998). Cortical feedback improves discrimination between figure and background by V1, V2 and V3 neurons. Nature, 394, 784–787. [PubMed] [CrossRef] [PubMed]
Kapadia, M. K. Westheimer, G. Gilbert, C. D. (2000). Spatial distribution of contextual interactions in primary visual cortex and in visual perception. Journal of Neurophysiology, 84, 2048–2062. [PubMed] [PubMed]
Knierim, J. J. Van Essen, D. C. (1992). Neuronal responses to static texture patterns in area V1 of the alert macaque monkey. Journal of Neurophysiology, 67, 961–980. [PubMed] [PubMed]
Lee, S. Blake, R. (1999). Visual form created solely from temporal structure. Science, 284, 1165–1168. [PubMed] [CrossRef] [PubMed]
Lee, T. S. Mumford, D. Romero, R. Lamme, V. A. (1998). The role of the primary visual cortex in higher level vision. Vision Research, 38, 2429–54. [PubMed] [CrossRef] [PubMed]
Leonards, U. Singer, W. Fahle, M. (1996). The influence of temporal phase differences on texture segmentation. Vision Research, 36, 2689–2697. [PubMed] [CrossRef] [PubMed]
Levi, D. M. Klein, S. A. Aitsebaomo, A. P. (1985). Vernier acuity, crowding and cortical magnification. Vision Research, 25, 963–977. [PubMed] [CrossRef] [PubMed]
Li, Z. (1999). Visual segmentation by contextual influences via intra-cortical interactions in the primary visual cortex. Network, 10, 187–212. [CrossRef] [PubMed]
Motoyoshi, I. (1999). Texture filling-in and texture segregation revealed by transient masking. Vision Research, 35, 1285–1291. [PubMed] [CrossRef]
Motoyoshi, I. Nishida, S. (2001). Temporal-resolution of orientation-based texture segregation. Vision Research, 41, 2089–2105. [PubMed] [CrossRef] [PubMed]
Motoyoshi, I. Nishida, S. (2002). Spatiotemporal interactions in detection of texture orientation modulations. Vision Research, 42, 2829–2841. [PubMed] [CrossRef] [PubMed]
Motoyoshi, I. (2002). Perceptual grouping based on illusory synchrony: On the relationship between synchrony perception and spatial interactions. The Japanese Journal of Psychonomic Science, 21, 51–52.
Moutoussis, K. Zeki, S. (1997). A direct demonstration of perceptual asynchrony in vision. Proceedings of the Royal Society of London B, 264, 393–399. [PubMed] [CrossRef]
Nijhawan, R. (1994). Motion extrapolation in catching. Nature, 370, 256–257. [PubMed] [CrossRef] [PubMed]
Nishida, S. Johnston, A. (2002). Marker correspondence, not processing latency, determines temporal binding of visual attributes. Current Biology, 12, 359–368. [PubMed] [CrossRef] [PubMed]
Paradiso, M. A. Nakayama, K. (1991). Brightness perception and filling-in. Vision Research, 31, 1221–1236. [PubMed] [CrossRef] [PubMed]
Rainville, S. J. M. Makous, W. L. (2002). The temporal mechanisms mediating synchrony perception [Abstract]. Journal of Vision, 2(7), 225a, http://journalofvision.org/2/7/225, doi:10.1167/2.7.225. [CrossRef]
Rogers-Ramachandran, D. C. Ramachandran, V. S. (1998). Psychophysical evidence for boundary and surface systems in human vision. Vision Research, 38, 71–77. [PubMed] [CrossRef] [PubMed]
Series, P. Georges, S. Lorenceau, J. Fregnac, Y. (2002). Orientation dependent modulation of apparent speed: A model based on the dynamics of feed-forward and horizontal connectivity in V1 cortex. Vision Research, 42, 2781–2797. [PubMed] [CrossRef] [PubMed]
Sheth, B. R. Nijhawan, R. Shimojo, S. (2000). Changing objects lead briefly flashed ones. Nature Neuroscience, 3, 489–495. [PubMed] [CrossRef] [PubMed]
Sillito, A. M. Grieve, K. L. Jones, H. E. Cuderio, J. Davis, J. (1995). Visual cortical mechanisms detecting focal orientation discontinuities. Nature, 378, 492–495. [PubMed] [CrossRef] [PubMed]
Sternberg, S. Knoll, R. L. (1973). The perception of temporal order: Fundamental issues and a general model. Kornblum, S. Attention and performance IV. New York: Academic Press.
Usher, M. Donnelly, N. (1998). Visual synchrony affects binding and segmentation in perception. Nature, 394, 179–182. [PubMed] [CrossRef] [PubMed]
Westheimer, G. McKee, S. P. (1977). Perception of temporal order in adjacent visual stimuli. Vision Research, 17, 887–892. [PubMed] [CrossRef] [PubMed]
Whitney, D. Murakami, I. (1998). Latency difference, not spatial extrapolation. Nature Neuroscience, 1, 656–657. [PubMed] [CrossRef] [PubMed]
Wilson, H. R. (1993). Nonlinear processes in visual pattern discrimination. Proceedings of the National Academy of Sciences, 90, 9785–9790. [PubMed] [CrossRef]
Zipser, K. Lamme, V. A. Schiller, P. H. (1996). Contextual modulation in primary visual cortex. Journal of Neuroscience, 16, 7376–7389. [PubMed] [PubMed]
Figure 1
 
Stimuli used in the experiment. Snapshots of stimuli with an inter-element distance of 6λ (a) and 2λ (b), respectively. (c). The waveform of the out-of-phase contrast modulation (orientation alternation) of horizontal and vertical Gabor patches for the target pattern (black line) and for the other patterns (gray line). (d). Sequential images illustrating the orientation alternation of the target (top) and of the others (bottom).
Figure 1
 
Stimuli used in the experiment. Snapshots of stimuli with an inter-element distance of 6λ (a) and 2λ (b), respectively. (c). The waveform of the out-of-phase contrast modulation (orientation alternation) of horizontal and vertical Gabor patches for the target pattern (black line) and for the other patterns (gray line). (d). Sequential images illustrating the orientation alternation of the target (top) and of the others (bottom).
Figure 2
 
The phase-offset limits between the target and the other patterns (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). The data in the left half of the panel show the lagged phase-offset limits (Δφ > 0), and those in the right half show the advanced phase-offset limits (Δφ < 0). Different color of symbols shows the results for different inter-element distances (IEDs). Smooth lines represent the fitted curves described in the text. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 2
 
The phase-offset limits between the target and the other patterns (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). The data in the left half of the panel show the lagged phase-offset limits (Δφ > 0), and those in the right half show the advanced phase-offset limits (Δφ < 0). Different color of symbols shows the results for different inter-element distances (IEDs). Smooth lines represent the fitted curves described in the text. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 3
 
The advanced and lagged phase-offset limits in Figure 2 replotted as absolute value (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). Closed symbols show the lagged phase-offset limits, and open symbols show the advanced phase-offset limits. The color of symbols represents the inter-element distance (IED); only the results for 2, 6, and 10λ are shown for clarity. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 3
 
The advanced and lagged phase-offset limits in Figure 2 replotted as absolute value (abscissa) as a function of the temporal frequency of orientation alternation (ordinate). Closed symbols show the lagged phase-offset limits, and open symbols show the advanced phase-offset limits. The color of symbols represents the inter-element distance (IED); only the results for 2, 6, and 10λ are shown for clarity. Each panel shows the results for a single observer, and the lower right panel shows the average. Vertical and horizontal error bars represent ±1 SE.
Figure 4
 
The estimated difference between the lagged and advanced phase-offset limits as a function of the inter-element distance. The lines are fitted linear functions.
Figure 4
 
The estimated difference between the lagged and advanced phase-offset limits as a function of the inter-element distance. The lines are fitted linear functions.
Figure 5
 
The difference between the lagged and advanced phase-offset limits as a function of the inter-element distance, obtained for double-scaled stimuli with a temporal frequency of 4 Hz (green circles). Red circles represent the results for the original stimuli at 4 Hz. Lines are predictions from the results for the original stimuli (Figure 4) if the difference between the two phase-offset limits depends on the inter-element distance in visual angle (gray line) and in the relative scale to the pattern size (black line), respectively. The top panel shows the results for observer IM, and the bottom the results for YT, respectively.
Figure 5
 
The difference between the lagged and advanced phase-offset limits as a function of the inter-element distance, obtained for double-scaled stimuli with a temporal frequency of 4 Hz (green circles). Red circles represent the results for the original stimuli at 4 Hz. Lines are predictions from the results for the original stimuli (Figure 4) if the difference between the two phase-offset limits depends on the inter-element distance in visual angle (gray line) and in the relative scale to the pattern size (black line), respectively. The top panel shows the results for observer IM, and the bottom the results for YT, respectively.
Figure 6
 
The critical temporal frequency of orientation alternation with a phase-offset of 180° as a function of the inter-element distance. The lines are fitted linear functions on log-log coordinates.
Figure 6
 
The critical temporal frequency of orientation alternation with a phase-offset of 180° as a function of the inter-element distance. The lines are fitted linear functions on log-log coordinates.
Figure 7
 
(a). Schematic model of asynchrony detection between oriented patterns. The model consists of two stages. At the first stage, individual local patterns are encoded by orientation detectors having a specific orientation tuning (only vertical is shown) and specific temporal tunings. The second stage consists of inhibitory interactions between the orientation detectors. The response of each detector is inhibited by signals propagating from the other detectors via neural connections with specific temporal tunings. (b). As a result, the propagating signals are temporally blurred and mutually delayed as the distance between detectors increases. See text for details.
Figure 7
 
(a). Schematic model of asynchrony detection between oriented patterns. The model consists of two stages. At the first stage, individual local patterns are encoded by orientation detectors having a specific orientation tuning (only vertical is shown) and specific temporal tunings. The second stage consists of inhibitory interactions between the orientation detectors. The response of each detector is inhibited by signals propagating from the other detectors via neural connections with specific temporal tunings. (b). As a result, the propagating signals are temporally blurred and mutually delayed as the distance between detectors increases. See text for details.
Figure 8
 
Results of simulations with a simple mutual-interaction model. (a). The model sensitivity to the target element when the IED was 2λ (left) and 10λ (right). (b). Predicted phase-offset versus temporal-frequency curves. Filled circles are the observed average data (Figure 2).
Figure 8
 
Results of simulations with a simple mutual-interaction model. (a). The model sensitivity to the target element when the IED was 2λ (left) and 10λ (right). (b). Predicted phase-offset versus temporal-frequency curves. Filled circles are the observed average data (Figure 2).
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×