Purchase this article with an account.
Sergei Gepshtein, Johannes Burge, Marc O. Ernst, Martin S. Banks; The combination of vision and touch depends on spatial proximity. Journal of Vision 2005;5(11):7. https://doi.org/10.1167/5.11.7.
Download citation file:
© ARVO (1962-2015); The Authors (2016-present)
The nervous system often combines visual and haptic information about object properties such that the combined estimate is more precise than with vision or haptics alone. We examined how the system determines when to combine the signals. Presumably, signals should not be combined when they come from different objects. The likelihood that signals come from different objects is highly correlated with the spatial separation between the signals, so we asked how the spatial separation between visual and haptic signals affects their combination. To do this, we first created conditions for each observer in which the effect of combination—the increase in discrimination precision with two modalities relative to performance with one modality—should be maximal. Then under these conditions, we presented visual and haptic stimuli separated by different spatial distances and compared human performance with predictions of a model that combined signals optimally. We found that discrimination precision was essentially optimal when the signals came from the same location, and that discrimination precision was poorer when the signals came from different locations. Thus, the mechanism of visual–haptic combination is specialized for signals that coincide in space.
This PDF is available to Subscribers Only