Retinal motion of images of objects in the scene can result from observer movement (
Figure 1A), object movement (
Figure 1B), or some combination of the two (
Figure 1C). This raises a difficult problem for the brain of a moving observer to work out which components of the retinal motion arise from movement of objects in the scene. Essentially the brain needs to separate (or parse) the retinal motion field into its causal components. How might the brain achieve this?
One solution to this problem involves the use of nonvisual information about self movement. Investigations conducted by von Holst and Mittelstaedt (
1950) and Sperry (
1950) suggest that copies of the motor commands driving observer movement (efference copy) are generated in the central nervous system and used to cancel the sensory input arising due to observer movement. Similarly, other nonvisual information sources about observer movement, such as vestibular signals and proprioception, could also contribute to the cancellation process (e.g., see Donaldson,
2000).
In addition to the nonvisual solution, several investigators have suggested that a purely visual mechanism might also play a role in the assessment of object movement during observer movement (Bravo,
1998; Royden, Wolfe, & Klempen,
2001; Rushton & Warren,
2005). A visual solution is necessary because circumstances exist in which nonvisual information (efference copy and both vestibular and proprioceptive signals) provides unreliable, inappropriate, or noisy information about observer movement (Warren & Rushton,
2009a). For example, consider a passenger in a car travelling at a constant speed of 70 miles per hour. Efference copy and proprioceptive information would signal that the observer is staying relatively still and vestibular information would be limited due to the constant velocity movement. However, the observer is moving rapidly through the environment and, consequently, experiencing a complex optic flow field which must be disentangled to assess how other objects are moving in the scene.
Although the simple example given above indicates the need for a visual solution, compelling evidence for the existence of such a solution has only recently been reported (Rushton & Warren,
2005; Rushton, Bradshaw, & Warren,
2007; Warren & Rushton
2007; Warren & Rushton,
2008; Matsumiya & Ando,
2009; Warren & Rushton
2009a,
2009b; Royden & Connors,
2010; Calabro, Soto-Faraco, & Vaina,
2011).
In particular, our work has put forward the
flow parsing hypothesis (Rushton & Warren,
2005; Rushton, Bradshaw, & Warren,
2007; Warren & Rushton,
2007; Warren & Rushton,
2008; Warren & Rushton,
2009a,
2009b). Under this hypothesis the brain uses visual information about self movement in the form of optic flow to assess scene-relative object movement. Briefly, the flow parsing hypothesis suggests that the brain is able to perform something akin to a global subtraction of the optic flow arising due to observer movement.
For example, consider the scene in
Figure 2 in which an observer walks forwards down a corridor whilst a ball falls downwards under the influence of gravity. The left panel of
Figure 2 demonstrates the problem: Due to the combination of self-generated and object-generated movement the retinal trajectory of any object in the scene is not necessarily indicative of the world-relative movement of that object. The center panel of
Figure 2 provides a conceptual illustration of the operation proposed under flow parsing. It shows a cancellation field to be applied to the left panel if the observer were able to perfectly discount (subtract off) the self movement component of optic flow. The right panel of
Figure 2 shows the perceived scene motion if the global subtraction process underpinning flow parsing were perfect. The motion signals remaining after the subtraction would be comparable to those which would be present if the observer was stationary in the scene.
The flow parsing hypothesis has been tested repeatedly in a number of paradigms (Rushton & Warren,
2005; Rushton, Bradshaw, & Warren,
2007; Warren & Rushton
2007; Warren & Rushton,
2008; Warren & Rushton,
2009a,
2009b). Perhaps the most compelling demonstration of the global subtraction process at work is found in Warren and Rushton (
2009a). In this study, stationary observers viewed limited lifetime dot motion displays consistent with forwards translation through a cloud of dots (a radial optic flow field). In addition observers saw a probe on the screen moving horizontally at 2° or 4° above fixation.
Under the flow parsing hypothesis it was predicted that the radial expansion field should be globally subtracted from the retinal motion field. At the location of the probe this corresponds to subtraction of the blue vector shown in
Figure 3. Subtraction of the blue vector is equivalent to addition of the magenta vector in
Figure 3 and consequently the perceived trajectory of the probe is predicted to tilt inwards towards the center (focus of expansion, FOE) of the radial flow field (
Figure 3). A second prediction of the flow parsing account was that the size of the tilt effect should increase with probe eccentricity, because in a radial flow field motion vectors further from the focus of expansion tend to be larger (
Figure 3) and so the component to be subtracted should also be larger. A third prediction suggested that, due to the global nature of the subtraction process, the first two predictions would still hold even if the radial flow information was removed in the hemi-field containing the probe. It is worth emphasizing the strength of this prediction since it leads to the very counterintuitive claim that when the probe is moved further away from the hemi-field containing the radial flow stimulus, the illusory tilt inwards should increase. The results of Warren and Rushton (
2009a) confirmed all these predictions. Taken together with independent evidence from other investigators (Matsumiya & Ando,
2009; Royden & Connors,
2010; Calabro, Soto-Faraco, & Vaina,
2011) these studies put forward a compelling case for the existence of a visual optic flow parsing mechanism which supports the estimation of scene-relative object movement.
The flow parsing hypothesis provides a novel role for optic flow processing in the human brain. A large body of earlier research, inspired primarily by the work of Gibson (
1950), has focused instead on the ability of observers to estimate
heading, the instantaneous direction of self-movement, from optic flow fields. It has been demonstrated that observers can estimate heading from optic flow information to within a degree or two (Warren, Morris, & Kalish,
1988) in less than 150 ms (van den Berg,
1992).
Given that flow parsing and heading estimation both rely upon processing of optic flow to recover information about self movement, an obvious question is whether there is a relationship between these mechanisms. Is flow parsing dependent on a prior estimate of heading direction or are the two processes and associated judgments independent?
In this paper we exploit an interesting illusory heading phenomenon called the optic flow illusion (OFI; Duffy & Wurtz,
1993) to investigate these questions. The optic flow illusion arises when a planar flow field (consistent with eye, or to a first approximation, head rotation) is superimposed onto a radial flow field (consistent with forwards observer translation) (
Figure 4A). In such circumstances observers typically report that the perceived heading direction (and perceived FOE of the radial flow field) is shifted in the same direction as the planar flow (
Figure 4C). This is a surprising result since the effect is in the opposite direction to that predicted if the brain performed an addition or averaging operation on the overlapping vector flow fields, as might be expected intuitively (
Figure 4B). In fact the commonly perceived effect illustrated in
Figure 4C appears to be more consistent with subtraction of the planar field from the radial field.
If flow-parsing is dependent on a prior estimate of heading then we might expect perceived trajectory of a probe object presented within the OFI stimulus to also be susceptible to the illusion. Consequently we would expect perceived object trajectory to reflect global subtraction of the flow field associated with the OFI (
Figure 4C). We tested this prediction in a series of experiments.
In a
Preliminary experiment we measured the size of the OFI illusion for the stimuli used in the present study. The primary reason for doing this was to ensure that the basic stimulus used here did give rise to the OFI and to establish the approximate magnitude of the effect.
In the main experiment (
Experiment 1) we test whether perceived trajectory of a probe presented in the OFI stimulus is susceptible to the OFI. The stimuli in
Experiment 1 and possible outcomes are shown schematically in
Figure 5. The basic manipulations are seen in the left panel of
Figure 5A. Observers viewed stimuli comprising overlaid radial and planar flow fields together with a single probe object moving upwards at one of two possible onscreen locations. The right panel of
Figure 5A shows the predictions for perceived trajectory of the probe if flow-parsing is susceptible to the OFI. If the flow parsing and heading estimation mechanisms were not independent then, due to the shift in perceived heading under the OFI, we would expect the trajectory to exhibit a commensurate tilt towards the illusory FOE. Alternatively, the middle panel of
Figure 5A shows the predictions if the flow-parsing effect seen is, instead, more compatible with prior vector addition of the flow fields comprising the OFI stimulus. In this case we would expect the perceived probe trajectory to tilt towards the FOE obtained under vector addition.
In
Figures 5B and C we show hemi-field versions of the stimulus in
Figure 5A together with associated predictions. These conditions were added to explicitly test that the effects seen are not dependent on the presence of local motion information surrounding the probe. Flow parsing is a global process and, based on Warren and Rushton (
2009a), we would expect to find a similar (but reduced in magnitude) pattern of responses when only a hemi-field is present, even when the probe and background motion are on opposite sides of the display.
In
Experiment 2 we decompose the flow fields seen in
Experiment 1 so that observers reported perceived probe trajectory when it was presented amidst the individual radial and planar optic flow fields. This allowed us to assess directly whether the effects seen in
Experiment 1 could be described as a linear combination of effects found for the individual flow fields.
To anticipate the results, we demonstrate that, when presented together with the superimposed flow fields of the OFI, participants perceive the probe to move in a manner that is not consistent with prior estimation and use of heading (i.e., not consistent with the predictions in the right panels of
Figure 5). We also find that the pattern of effects obtained in
Experiment 1 is well approximated by a positively weighted vector sum of the effects seen in
Experiment 2 for the individual flow fields comprising the OFI stimulus. This result is consistent with the idea that the perceived trajectory seen in
Experiment 1 is based on vector field addition rather than subtraction (i.e., consistent with the predictions in the middle panels of
Figure 5).