When different sensory modalities simultaneously provide an estimate for the same object property, thus providing redundant information, the sensory estimates are often integrated into a single percept. In the framework of maximum likelihood estimation (MLE) we consider the integration process to be
statistically optimal when the weight that each modality estimate receives is fully determined by its relative precision. Thus, according to MLE, the less precise modality receives less weight. Importantly, this way of combining the signals results in the most precise (i.e., minimal variance) combined estimate possible. This combined estimate will be more precise than either the visual or haptic estimate alone. This type of optimal integration has been shown to apply to, for example, the combination of visual and proprioceptive position information (Reuschel, Drewing, Henriques, Roesler, & Fiehler,
2010; van Beers, Sittig, & van der Gon,
1999), visual and auditive speech perception (Alais & Burr,
2004), visual and haptic size and shape perception (Ernst & Banks,
2002; Gepshtein & Banks,
2003; Helbig & Ernst,
2007b), the numerosity of sequences of events (e.g., Bresciani, Dammeier, & Ernst,
2006; Shams, Kamitani, & Shimojo,
2002) as well as visual-vestibular heading perception (Fetsch, DeAngelis, & Angelaki,
2010).