Early studies of pointing and grasping movements using only one or both eyes suggested that binocular cues are critical for efficient movements in 3D space. A number of more recent studies have questioned the generality of this result. Watt and Bradshaw (
2000,
2003) have shown, for example, that monocular cues like motion parallax can by themselves support accurate scaling of hand transport velocities (but not grip aperture). Similarly, both the accuracy and the shapes of movement trajectories in the object placement task used here are similar under binocular and monocular viewing (Knill & Kersten,
2003). That it is possible to guide movements effectively with monocular in-formation is effectively illustrated by the many people who successively navigate their world without binocular vision. Several people with only one eye have even succeeded at high levels of athletics (e.g., a recent Division 1 college basketball player had lost one eye early in life). None of these observations, however, tells us about the relative contribution of binocular and monocular cues to motor control when both are present in a stimulus. We have shown that monocular cues about 3D surface orientation can contribute significantly to motor control even in the presence of binocular cues; however, visuomotor control of object placement relies much more heavily on binocular cues than does the perceptual system in tasks requiring estimates of the same surface property.
A number of authors have suggested that the brain performs different visual computations for perception and motor control (Milner & Goodale,
1995). The most commonly cited behavioral evidence for this hypothesis has come from studies that show an attenuation of illusory visual effects when measured using motor behavior rather than explicit perceptual report (Aglioti, DeSouza, & Goodale,
1995; Brenner & Smeets,
1996; Haffenden & Goodale,
1998). Recent studies, however, have cast doubt on these conclusions on methodological (Franz,
2001; Franz, Fahle, Bulthoff, & Gegenfurtner,
2001) or conceptual grounds (Smeets, Brenner, de Grave, & Cuijpers,
2002). Even if one were to reliably find that a perceptual illusion is attenuated in observed motor behavior, such an effect could be (and usually is) interpreted as reflecting differences in the representations on which perceptual judgments and motor behavior rely (e.g., object-centered vs. viewer-centered) rather than on differences in the intermediate computations used to derive the representations (Smeets et al.,
2002). Because our experiments studied the cue-integration process that gives rise to estimates of an object’s 3D properties for both perceptual judgments and for motor control, the results reflect differences in the internal computations that lead up to the representations on which both types of behavior are based.
Do our results imply that the brain processes visual depth information independently for visuomotor control and perception as suggested by Milner and Goodale (1995)? Such an account does not explain why one should obtain different cue weights for the two types of functions. The optimal cue-integration strategy should be the same for the visuomotor and perceptual tasks used here, as it depends only on the information content of the stimuli (because both types of task required estimates of viewer-centered surface slant). Thus, it would appear that if the weights we measured for one task were optimal, they would be suboptimal for the other task. What rational basis would exist for visuomotor control relying more on binocular information than on perceptual judgments?
One possibility is that different cue weights might be optimal for different tasks when one considers the specific demands of different tasks. In Bayesian decision theory, task demands are enforced by specifying cost functions associated with a task and estimating a parameter like slant to minimize the expected cost of performance errors (Maloney,
2002; Yuille & Bulthoff,
1996). In the context of motor control, the cost (or gain) associated with performance is a combination of estimation errors, motor errors, and the costs or gains associated with each possible movement. That subjects adjust their motor strategies based on the costs and gains associated with motor performance has been demonstrated in pointing tasks (Trommershauser, Maloney, & Landy,
2003a,
2003b). It is difficult, however, to construct a scenario in which changing the cost function for the task leads to significant changes in the measured cue weights used to combine cues. An optimal estimator derives its estimate by applying a cost function to the combined information from both cues. If the likelihood functions associated with slant estimates from each cue are approximately Gaussian, applying different cost functions for different tasks amounts to applying a point nonlinearity to the weighted average of slants derived from each cue separately. This has little effect on relative cue weights when a linear model is fit to the result.
Another possibility is that our results reflect more the properties of online visual control than motor planning. Even were motor planning based on the same visual estimates of slant as were perceptual judgments, the weights that we derived from contact slants might have been influenced by visual estimates of surface slant computed during the online control phase of movements (even without visual feedback from the hand). Glover and Dixon have argued that illusions influence motor planning much more than online control of hand movements, suggesting that the visual processes underlying the two stages of motor control may be distinct (S. Glover & P. Dixon,
2001; S. R. Glover & P. Dixon,
2001). Whether or not different visual computations subserve the two control phases, the tight time constraints under which online control must operate could affect the relative contributions of binocular and monocular cues during that part of a movement. If the visual system processes binocular cues more quickly than monocular cues (or at least, the monocular cues used here), we might expect the system to effectively give more weight to binocular cues during online control. In another study, we have found that for the object placement task used here, subjects do appear to process binocular cues to slant more quickly than monocular cues when making online adjustments in their movements.
The mechanisms underlying visuomotor and perceptual differences in cue weighting are necessarily a matter of speculation at this point. Nevertheless, our results suggest that task-specific computational constraints on visual mechanisms other than those imposed by the available image information influence how the brain integrates different sensory cues about the world for guiding behavior. In the terms used in the
Introduction, cue weighting is affected not only by the information in the input to the system, but also by the function for which the information is used-the system’s output.