Indeed, there is much evidence to suggest that low-level feature information such as orientation is encoded retinotopically in V1 (Engel, Glover, & Wandell,
1997; Ress & Heeger,
2003); however, it is unclear whether the subsequent integration of this low-level information may also depend on similar retinotopic processes. Had the late noise model fitted better, this would unanimously suggest that pre- and postsaccadic information may be processed by a single, spatiotopic channel, and that the integration of this information would happen very early such that the predominant source of noise would occur after integration. This distinction between whether integration occurs in spatiotopic or retinotopic coordinates is interesting in the broader context of coordinates of perceptual stability in general: Many accounts of perceptual stability argue that perceptual updating occurs in a spatiotopic reference frame: for example, spatiotopic updating of information (Fabius et al.,
2016) and memory (Zerr et al.,
2017) across saccades, and adaptation after-effects (Melcher,
2005). However, other studies suggest that processes underlying perceptual stability may occur in a more retinotopic manner: For example, predictive remapping occurs in retinotopic coordinates (Duhamel, Colby, & Goldberg,
1992), visual features are preserved and remapped in retinotopic coordinates (Harrison, Retell, Remington, & Mattingley,
2013; Melcher,
2008) and transsaccadic memory and attention show retinotopic effects (Golomb, Nguyen-Phuc, Mazer, McCarthy, & Chun,
2010). These divergent results suggest that both retinotopic and spatiotopic mechanisms contribute to the perceived unity and stability of the world, and it is likely that there is a complex interplay between the two (Burr & Morrone,
2012). It is also important to consider that integration may also occur at different stages of processing, and that this may differ depending on the complexity and features of the stimuli, and the inherent physiological differences in encoding the stimuli. For instance Burr and Morrone (
2011) suggest that the integration that may occur at earlier stages of processing such as V1 should not be spatiotopic, as V1 encodes stimuli in retinotopic coordinates. At later stages of processing, however, for example for features such as form and motion which may show additional spatiotopic encoding in MT and MST (d'Avossa et al.,
2006; Crespi et al.,
2011), integration could also occur on a spatiotopic level; nevertheless, note that the issue of whether MT does in fact encode information in spatiotopic coordinates is contentious (Gardner, Merriam, Movshon, & Heeger,
2008; Knapen, Rolfs, Wexler, & Cavanagh,
2010), and it is likely that, similar to other transsaccadic processes such as the motion after-effect, there are both spatiotopic and retinotopic effects present (Ezzati, Golzar, & Afraz,
2008). The implications of this are twofold: First, different stimulus features may give rise to different forms of integration, and second, low-level integration of basic features may follow this retinotopic, early noise model, whereas higher-level integration of more complex form or motion may follow a more spatiotopically defined late noise model. As the late noise model also predicts greater integration benefits than the early noise model (
Figure 4), it would perhaps be easier to test integration in these cases, and to observe greater integration effects. The current study clearly showed that integration occurred in spatiotopic but not retinotopic coordinates and also showed that integration performance, and thus potentially perceptual stability may not rely solely on higher level perceptual effects that seem to arise spatiotopically, potentially after integration, but that the low-level retinotopic representations of the stimuli that occur before integration are also crucial. Interestingly, Fabius et al. (
2016) showed that integration of information can occur in spatiotopic coordinates, but also in retinotopic coordinates (albeit to a lesser extent). This difference from our results could reflect the different stimuli used: Whereas our experiment was an orientation detection task, Fabius et al. (
2016) used motion to induce an illusory jump in the direction of a noise patch. This difference in integration coordinates could reflect the coordinates in which the different stimuli (orientation vs. motion) are encoded.
Experiment 2 showed not only a lack of retinotopic integration, but an additional cost in transsaccadic compared to best-single performance. This could be due to the stimulus shift in the retinotopic condition: When the pre- and postsaccadic stimuli appear in different spatiotopic coordinates, they may no longer be grouped as a single object that would benefit from integration. This may point to the system having a general assumption of spatiotopic constancy, as observed for instance in saccadic suppression of displacement (Bridgeman et al.,
1975; Deubel, Schneider, & Bridgeman,
1996). Preserving separate representations of the pre- and postsaccadic stimulus might require additional resources in attention and working memory (Schneider,
2013; Poth, Herwig, & Schneider,
2015).