The human eye has a rather small region on the retina where visual information can be processed with high resolution, the fovea. Therefore, humans move their eyes, head, and body to bring the image of informative stimuli on the fovea. Relevant information is then extracted from the foveal region and used for the current task. In object-related actions, relevant visual information is normally used to control eye movements as well as hand movements. For instance, when grasping an object, the eyes usually reach the location where the object has to be grasped sometime before the hand (e.g., Crawford, Medendorp, & Marotta,
2004; Desmurget, Pelisson, Rossetti, & Prablanc,
1998; Droll & Hayhoe,
2007; Droll, Hayhoe, Triesch, & Sullivan,
2005; Mennie, Hayhoe, & Sullivan,
2006). When placing an object, the eyes usually reach the placement location shortly before the object is set down (e.g., Ballard et al.,
1992; Droll & Hayhoe,
2007; Droll et al.,
2005; Epelboim et al.,
1995). The same eye–hand dynamics have been found in well-learned everyday sensorimotor tasks such as sandwich or tea making (Hayhoe, Shrivastava, Mruczek, & Pelz,
2003; Land & Hayhoe,
2001; Land, Mennie, & Rusted,
1999; Land & Tatler,
2009) and throughout the learning process of novel sequential sensorimotor tasks such as speed stacking (Foerster, Carbone, Koesling, & Schneider,
2011). Moreover, it has been shown that hand movements are executed more accurately if hand target locations have been fixated beforehand (Prablanc, Desmurget, & Gréa,
2003; Prablanc, Echallier, Komilis, & Jeannerod,
1979; Prablanc, Pélisson, & Goodale,
1986). This “eye-guides-hand” benefit may explain the robustness of the finding that the eyes lead the hands (positive eye–hand time spans) in sensorimotor tasks.