It is only recently that vision scientists have begun to experimentally evaluate the idea that “touch educates vision.” Held et al. (
2011) tested treatable, congenitally blind individuals. After sight restoration, it was found that these individuals were unable to visually match an object to a haptically sensed sample, though this ability developed within days suggesting that visual-haptic matching requires visual-haptic experience. Other researchers have studied the affects of visual-haptic experience on visual perception when vision and haptics are discrepant or when visual information is highly ambiguous. Ernst, Banks, and Bülthoff (
2000) and Atkins, Fiser, and Jacobs (
2001) found that subjects' estimates of visual depth relied more heavily on a visual cue (e.g., texture) when the cue was congruent with a haptic signal versus when it was incongruent with this signal. Atkins, Jacobs, and Knill (
2003) reported that subjects recalibrate their interpretations of a visual stereo cue so that depth-from-stereo percepts are in greater agreement with depth-from-haptic percepts when visual and haptic signals are discrepant. Adams, Graf, and Ernst (
2004) and Adams, Kerrigan, and Graf (
2010) showed that visual-haptic experience can modify the visual system's “light-from-above” assumption used when observing images in which shading information to depth is ambiguous. Within the literature on artificial intelligence, there are relatively few articles studying how touch can educate vision. An exception is the work of Pinto, Gandhi, Han, Park, and Gupta (
2016) showing that physical interactions (e.g., grasping, pushing, poking) can aid the acquisition of meaningful visual representations in a robot system.
1