Open Access
Article  |   July 2021
Transsaccadic integration operates independently in different feature dimensions
Author Affiliations
  • Garry Kong
    Department of Psychology, University of Cambridge, Cambridge, UK
    kong.garry@gmail.com
  • David Aagten-Murphy
    Department of Psychology, University of Cambridge, Cambridge, UK
    david.aagtenmurphy@gmail.com
  • Jessica M. V. McMaster
    Department of Psychology, University of Cambridge, Cambridge, UK
    jm2309@cam.ac.uk
  • Paul M. Bays
    Department of Psychology, University of Cambridge, Cambridge, UK
    pmb20@cam.ac.uk
Journal of Vision July 2021, Vol.21, 7. doi:https://doi.org/10.1167/jov.21.7.7
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Garry Kong, David Aagten-Murphy, Jessica M. V. McMaster, Paul M. Bays; Transsaccadic integration operates independently in different feature dimensions. Journal of Vision 2021;21(7):7. https://doi.org/10.1167/jov.21.7.7.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Our knowledge about objects in our environment reflects an integration of current visual input with information from preceding gaze fixations. Such a mechanism may reduce uncertainty but requires the visual system to determine which information obtained in different fixations should be combined or kept separate. To investigate the basis of this decision, we conducted three experiments. Participants viewed a stimulus in their peripheral vision and then made a saccade that shifted the object into the opposite hemifield. During the saccade, the object underwent changes of varying magnitude in two feature dimensions (Experiment 1, color and location; Experiments 2 and 3, color and orientation). Participants reported whether they detected any change and estimated one of the postsaccadic features. Integration of presaccadic with postsaccadic input was observed as a bias in estimates toward the presaccadic feature value. In all experiments, presaccadic bias weakened as the magnitude of the transsaccadic change in the estimated feature increased. Changes in the other feature, despite having a similar probability of detection, had no effect on integration. Results were quantitatively captured by an observer model where the decision whether to integrate information from sequential fixations is made independently for each feature and coupled to awareness of a feature change.

Introduction
With each saccadic eye movement, incoming visual information undergoes an abrupt transformation. Not only does the shift in gaze direction create a substantially different image on the retina, but the transition between these snapshots is obscured by blur induced by the movement itself. Our smooth and continuous perceptual experience during this process makes it easy to overlook the complex computations required by the visual system to coordinate it. According to contemporary theories, a limited amount of visual information from before a saccade, primarily from the region of the intended target, is maintained and used to resolve perception after the saccade (Burr & Morrone, 2011; Cavanagh, Hunt, Afraz, & Rolfs, 2010; Melcher & Morrone, 2015), whereas changes outside this focus go unnoticed due to limitations of attention and working memory (Aagten-Murphy & Bays, 2018; Rensink, O'Regan, & Clark, 1997). 
One line of evidence for these theories has come from studies investigating the phenomenon of suppression of saccadic displacement (Bridgeman, Hendry, & Stark, 1975), whereby observers may fail to notice large shifts of a saccade target occurring during the eye movement (as much as half the saccade amplitude). This has been explained on the basis that the eye movement itself creates a large and uncertain shift of object location on the retina, making it difficult to determine whether to attribute retinal displacements to the saccade or to a change in object location (Niemeier, Crawford, & Tweed, 2003). Supporting this account, factors that would favor the hypothesis that the object has changed aid in detection of a displacement, such as an intrasaccadic change to a surface feature of the object or a brief blank interval before reappearance of the object after the saccade (Demeyer, De Graef, Wagemans, & Verfaillie, 2010; Poth, Herwig, & Schneider, 2015; Tas, Moore, & Hollingworth, 2012; Wexler & Collins, 2014). 
There is, however, even more direct evidence that presaccadic input is available and used by the visual system after the saccade. When the surface properties of an object are subtly changed during a saccade, below the threshold of detection, observers’ report their perception of the object as intermediate between its pre- and postsaccadic properties (Oostwoud Wijdenes, Marshall, & Bays, 2015). Furthermore, there is evidence that this transsaccadic integration process follows principles of optimal probabilistic inference (e.g., Ernst & Banks, 2002), in that the perception reflects a weighted average, with weights that reflect the relative reliability of pre- and postsaccadic estimates of object features. Oostwoud Wijdenes et al. (2015) showed that, when objects were viewed at different eccentricities before and after a saccade, the integrated percept was biased toward the feature value in the more foveal view. Similar shifts in perceived feature value were elicited when reliability was directly manipulated by adding external noise to the pre- or postsaccadic image. A number of studies (Ganmor, Landy, & Simoncelli, 2015; Hubner & Schutz, 2017; Stewart & Schütz, 2019a; Stewart & Schütz, 2019b; Wolf & Schutz, 2015) further demonstrated that a stable feature of an object visible both before and after a saccade can be identified more precisely than is possible from either the pre- or the postsaccadic view alone. This advantage is predicted as a result of averaging over independent noise in pre- and postsaccadic input. Although the experimental demonstrations of this have typically involved items initially viewed in the periphery before a saccade brings them to the fovea, the behavioral benefits of averaging may be even more significant for items that remain peripheral and therefore relatively poorly represented in the input over the course of several fixations. For example, such peripheral-to-peripheral integration might allow an object to be identified that otherwise would not have been, whereas the benefits of peripheral-to-foveal integration will mostly be limited to resolving fine details. 
However, our environment is rarely entirely static, and when changes occur during an eye movement information from previous saccades may become actively misleading about the present state of the world. An apparent discrepancy between pre- and postsaccadic vision of an object could be the result of inherent variability in the visual system, in which case integration would be beneficial for accurate perception, or could be due to a real change in the object, in which case it would be harmful. The process of resolving this uncertainty has been termed causal inference (for a review, see Shams & Beierholm, 2010). In the context of transsaccadic integration of spatial information, a model formulated by Atsma, Maij, Koppen, Irwin, & Medendorp (2016) considers two possible causes of a discrepancy between pre- and postsaccadic estimates of object location: an inaccurate saccade or a change in the state of the world. In the former case, the goal is to integrate the two estimates, whereas in the latter it is to segregate them and base decisions only on the postsaccadic input (Cicchini, Binda, Burr, & Morrone, 2013; Demeyer, De Graef, Wagemans, & Verfaillie, 2009; Stewart, Valsecchi, & Schütz, 2020). These possible causes and their resulting location estimates are then weighted, incorporating a prior belief favoring a stable world, so that larger intrasaccadic differences lead to responses weighted more toward the segregated postsaccadic estimate, and smaller differences lead to a percept weighted toward an integrated estimate. 
The causal model of Atsma et al. (2016) makes inferences about, and on the basis of, a single visual parameter: object location. Real-world objects are made up of multiple features, and, although internal variability may affect each feature independently, external causes will often change multiple features simultaneously. How should such cases be resolved? Given any evidence that an object has changed during an eye movement, the visual system might rely solely on the postsaccadic input; we describe such a process as being object based. Alternatively, evidence of a change in one feature could influence only the decision to integrate or segregate the pre- and postsaccadic estimates of that same feature, with independent decisions made for each object feature; we describe such a process as being feature based
In this study, we investigated transsaccadic integration under circumstances involving changes to multiple features. We made intrasaccadic changes of varying magnitude to two features of a single object. We asked observers on every trial whether they had detected any change to the object and also for an analog estimate of one of the two features that they last perceived. Importantly, although we asked for a change detection response for any feature, as the differences in magnitude between the two features were uncorrelated, we can probabilistically infer in which feature a change was detected. This method revealed a correlation between detection of changes and transsaccadic integration within a feature dimension, but no evidence for interaction between feature dimensions. 
Methods
Participants
Thirty-six participants took part in this study in total. All had normal or corrected-to-normal vision and were naïve as to the purpose of the experiment. The experiments were approved by the Cambridge Psychology Research Ethics Committee, and informed consent was obtained in accordance with the tenets of the Declaration of Helsinki. 
Apparatus and stimuli
Stimuli were presented on a 27-inch Asus ROG Swift PG279Q monitor (144-Hz refresh rate, 2560 × 1440 pixels; Asus, Taipei, Taiwan) at a distance of 60 cm from the participant. The background of the screen was black (0.3 cd/m2) throughout the experiment. Eye position was tracked online at 1000 Hz with a desk-mounted EyeLink 1000 (SR Research, Kanata, Ontario, Canada). The stimuli were generated in MATLAB (MathWorks, Natick, MA) using the Psychophysics Toolbox extensions (Kleiner, Brainard, Pelli, Ingling, Murray, & Broussard, 2007
Experiment 1
Twelve participants (five male, seven female; mean age, 23.1 years; range, 19–30) took part in this experiment. A summary of the procedure is shown in Figure 1A. Each trial began with the presentation of two dots (diameter, 0.5° of visual angle), located 6° horizontally left and right of the center of the screen. One dot was white (100 cd/m2), indicating the current location for fixation, and the other dot was gray (10 cd/m2), indicating the upcoming saccade target. 
Figure 1.
 
Experimental design. (A) Trial sequence for Experiment 1. A color patch was presented that changed in color and/or location while the participant was making a saccade. The participants reported the last color they had seen and then whether they had detected any change in the stimulus. (B) Experiment 2 followed the top path of this trial sequence (color and/or orientation change, followed by color report); Experiment 3 followed either the top or bottom path (color or orientation report) with equal frequency. Dashed red circles represent gaze fixations, and dashed red arrows represent saccades. The stimulus changed as soon as gaze crossed the vertical midline of the screen.
Figure 1.
 
Experimental design. (A) Trial sequence for Experiment 1. A color patch was presented that changed in color and/or location while the participant was making a saccade. The participants reported the last color they had seen and then whether they had detected any change in the stimulus. (B) Experiment 2 followed the top path of this trial sequence (color and/or orientation change, followed by color report); Experiment 3 followed either the top or bottom path (color or orientation report) with equal frequency. Dashed red circles represent gaze fixations, and dashed red arrows represent saccades. The stimulus changed as soon as gaze crossed the vertical midline of the screen.
When the participant had maintained gaze within 1.5° of the fixation dot for 500 ms, a presaccadic stimulus appeared, consisting of a colored disk with blurred edges (diameter, 1°) located at one of four vertical displacements (–5°, –4°, +4°, +5°) from the screen center, randomly chosen with equal probability on each trial. The color of the presaccadic stimulus was randomly drawn from a circle in CIELAB space, centered on L* = 74, a* = 0, and b* = 0, with a radius of 40. Note that stimulus parameters, including size, eccentricity, and luminance, were chosen to make the task challenging and avoid ceiling effects. After 1000 ms, the two dots switched luminance, indicating that the participant should immediately make a saccade to the new fixation location on the opposite side of the screen. As soon as gaze position was detected to have crossed the vertical midpoint of the screen, the presaccadic stimulus was replaced with a postsaccadic stimulus. 
The postsaccadic stimulus was identical to the presaccadic stimulus, except that its color was shifted clockwise or counterclockwise by 20°, 35°, or 75° (equal frequency) on the color wheel, and its location was shifted vertically either up or down (equal probability) by 0°, 1.5°, or 2.5° of visual angle (equal frequency). These values were selected based on pilot testing to avoid floor or ceiling effects in the change detection responses. 
After 300 ms of fixation at the new location, the postsaccadic stimulus was removed and a color wheel was displayed, centered on the fixation dot. The color wheel was randomly rotated from trial to trial. Participants were instructed to report the last color they had seen on the trial with a mouse click on the wheel. Moving the mouse cursor over the color wheel caused a disk to appear at fixation, with a color corresponding to the cursor location. After the color response, participants reported whether they had detected any change to the stimulus (regardless of whether the change was detected in color or location or both features) during the trial by clicking the left (no change) or right (change) mouse button. After the response, the next trial began with initial fixation at the opposite location to the previous trial. Note that the color estimation was always requested before the change detection response. This was intended to minimize any effect of the change detection response on the color estimate (e.g., a consistency bias that might have led participants to exaggerate the color difference from the presaccadic value after reporting a change). 
If the participant made any erroneous saccades, took longer than 150 ms to initiate their saccade, finished their saccade farther than 2.5° from the saccade target, or blinked before the response screen, the trial was immediately aborted. Aborted trials were presented again at the end of each block. Each block consisted of 54 successful trials. Participants performed either six blocks or 1.5 hours of the task, whichever took longer, resulting in a minimum of 324 successful trials and a maximum of 540. 
Experiment 2
Twelve participants (four male, seven female, one non-binary; mean age, 23.8 years; range, 18–31) took part in this experiment. The design was identical to Experiment 1, with the following exceptions. The pre- and postsaccadic stimuli were colored Gabor patches, 1.5° in diameter, with a spatial frequency of 4.5 cycles/degree. The orientation of the presaccadic Gabor patch was drawn from a uniform distribution over all possible orientations. The orientation of the postsaccadic stimulus was rotated clockwise or counterclockwise by 10°, 30°, or 90° (equal frequency) from that of the presaccadic stimulus. The colors of the pre- and postsaccadic stimuli varied as in Experiment 1. The pre- and postsaccadic locations were the same on each trial. As in Experiment 1, participants reported the last color they had seen and whether they had detected any change to the stimulus during the trial. Participants performed either six blocks or 1.5 hours of the task, whichever took longer, resulting in a minimum of 324 successful trials and a maximum of 540. 
Experiment 3
Twelve participants (five male, seven female; mean age, 21.3 years; range, 18–29) took part in this experiment. The design was identical to Experiment 2, except that participants were unpredictably asked to report either color or orientation on each trial (equal frequency, randomly interleaved). On orientation report trials, a gray wheel appeared instead of the color wheel in the response stage of the task. This signaled the participant to report the last orientation they had seen with a mouse click at a corresponding point on the wheel. Moving the mouse cursor over the gray wheel prompted a grayscale Gabor patch to appear at fixation, with an orientation corresponding to the cursor position on the wheel. As in the previous experiments, participants also reported whether they had detected any change to the stimulus on each trial. Participants completed either 12 blocks or 3 hours of the task, whichever took longer, resulting in a minimum of 738 successful trials and a maximum of 900. 
Analysis
For each combination of participant, magnitude of color change, and magnitude of location change (in Experiment 1; orientation change in Experiments 2 and 3), we calculated the mean proportion of trials in which a change was reported and a mean color bias. The latter was obtained by first rotating and reflecting the color response on each trial such that 0° corresponded to the presaccadic color, and positive values were in the direction of the postsaccadic color shift. The mean color bias was then calculated as the circular mean of these color values divided by the magnitude of the color change, such that 0 corresponded to the presaccadic color and 1 to the postsaccadic color. Orientation report trials in Experiment 3 were analyzed independently but identically to color report trials for the descriptive statistics and in conjunction with the color report trials for the model analysis. Statistical tests of hypotheses were conducted using Bayesian analyses of variance (ANOVAs) and Bayesian t-tests in JASP (JASP Team, 2020) with default priors. The outcomes are reported as Bayes factors (BFs); for example, for a t-test, BF10 = 5 indicates that the strength of evidence for a difference is five times greater than for no difference, and BF01 = 5 indicates the same strength of evidence favoring no difference. For the Bayesian ANOVAs, models containing each of the main effects individually, a model containing both main effects, and a null model incorporating only the intercept were compared using their Bayes factors. For the sake of brevity, where the best-fitting model contains both main effects, we report only the BFs for this model versus the null; otherwise, we report all relevant pairwise comparisons among models. 
Model
Although participants in each trial were asked only if they detected a change in either feature and to report their last perception of just one feature, which features they detected changes in and whether the responses were of integrated or segregated estimates can nonetheless be inferred using a probabilistic model. This is possible because the magnitudes of the changes in the two features were independent and controlled by the experimenter. Note that the goal of the modeling was not to provide a definitive account of transsaccadic perception but rather to incorporate only the detail necessary to distinguish feature- and object-based accounts of the integration process. As such, a number of simplifying assumptions have been made (e.g., linear approximations for circular spaces) that may lead to an overall lower quality of fit but should not meaningfully affect the reliability of the model comparisons. 
A pictorial representation of the model can be seen in Figure 2. For each trial in Experiment 1, the variables entered into the model were the presaccadic (–) and postsaccadic (+) feature values in color (C) and location (L) dimensions, \(\{ {x_C^ - ,x_L^ - ,x_C^ + ,x_L^ + } \}\); an analog color estimate, \({\hat x_C}\); and a binary judgment of whether a change had occurred in either dimension, y ∈{0, 1}. 
Figure 2.
 
Pictorial representation of the probabilistic observer model. (A) The pre- and postsaccadic color and location representations are corrupted by Gaussian noise, as represented by the bivariate Gaussians on the location and color feature spaces. (B) Two-tailed thresholds, as represented by the dotted lines, are set for the purpose of detecting a change. If the postsaccadic stimulus exceeds a feature threshold, then a change is detected in the corresponding feature. For both models, if the postsaccadic representation lies in the green region, then pre- and postsaccadic representations will be integrated. If it falls in the purple region, then the representations will be kept separate. (C) If the representations are integrated, then the analog response will be centered on a point between the pre- and postsaccadic representations, as indicated by the green probability distribution. If the representations are segregated, then the analog response will be centered on the postsaccadic representation (as indicated by the purple distribution), as the instruction was to report the last color perceived.
Figure 2.
 
Pictorial representation of the probabilistic observer model. (A) The pre- and postsaccadic color and location representations are corrupted by Gaussian noise, as represented by the bivariate Gaussians on the location and color feature spaces. (B) Two-tailed thresholds, as represented by the dotted lines, are set for the purpose of detecting a change. If the postsaccadic stimulus exceeds a feature threshold, then a change is detected in the corresponding feature. For both models, if the postsaccadic representation lies in the green region, then pre- and postsaccadic representations will be integrated. If it falls in the purple region, then the representations will be kept separate. (C) If the representations are integrated, then the analog response will be centered on a point between the pre- and postsaccadic representations, as indicated by the green probability distribution. If the representations are segregated, then the analog response will be centered on the postsaccadic representation (as indicated by the purple distribution), as the instruction was to report the last color perceived.
Model specification
We assumed that internal representations of pre- and postsaccadic features were corrupted by independent Gaussian noise (wrapped onto the circular feature space in the case of color (Figure 2A):  
\begin{eqnarray} &&{\hat x_C^ - }\, \sim\, {WN\left( {x_C^ - ,\sigma _C^{2 - }} \right)}\nonumber\\ &&{\hat x_C^ + }\, \sim\,{ WN\left( {x_C^ + ,\sigma _C^{2 + }} \right)}\nonumber\\ &&{\hat x_L^ - }\, \sim\,{N\left( {x_L^ - ,\sigma _L^{2 - }} \right)}\nonumber\\ &&{\hat x_L^ + }\, \sim\,{N\left( {x_L^ + ,\sigma _L^{2 + }} \right)} \end{eqnarray}
(1)
where N(µ, σ2) and WN(µ, σ2) are the normal and wrapped normal distributions with mean µ and variance σ2. The observer detected a change if the estimated change magnitude in either dimension exceeded the threshold for that dimension (Figure 2B):  
\begin{equation}\hat y = \left\{ {\begin{array}{@{}*{3}{l}@{}} 1&\;{{\rm{if}}\;\;{{\hat \Delta }_C} > {\theta _C} \cup {{\hat \Delta }_L} > {\theta _L}}&{}\\ 0&\;{{\rm{otherwise}}}&{} \end{array}} \right.\end{equation}
(2)
where \({\hat \Delta _C} = | {\hat x_C^ + \ominus \hat x_C^ - } |\) and \({\hat \Delta _L} = | {\hat x_L^ + - \hat x_L^ - } |\), with \(\ominus\) indicating subtraction on the circle. The observed detection response, y, reflected the internal detection state, \(\hat y\), with a lapse rate of γ:  
\begin{equation}y = \left\{ {\begin{array}{@{}*{3}{l}@{}} {\hat y}&\;{{\rm{with}}\;\;{\rm{probability}}}&{1 - \gamma }\\ {1 - \hat y}&\;{{\rm{with}}\;\;{\rm{probability}}}&\gamma \end{array}} \right.\end{equation}
(3)
 
We compared predictions of two models that differed only in the conditions determining whether integration of pre- and postsaccadic features occurred in the reported dimension. In the object-based integration model, the conditions on integration matched those on detection. If the estimated change in either dimension exceeded threshold, the observer reported the internal postsaccadic estimate; otherwise, the observer reported an integrated value based on both pre- and postsaccadic estimates (Figure 2C):  
\begin{equation}{\hat x_C} = \left\{ {\begin{array}{@{}*{3}{l}@{}} {\hat x_C^ + }&\;{{\rm{if}}\;\;{{\hat \Delta }_C} > {\theta _C} \cup {{\hat \Delta }_L} > {\theta _L}}&{}\\ {\hat x_C^I}&\;{{\rm{otherwise}}}&{} \end{array}} \right.\end{equation}
(4)
where the integrated estimate was a weighted average of pre- and postsaccadic estimates with weighting determined by their relative variances:  
\begin{eqnarray} \hat x_C^I = \hat x_C^ - \oplus w\left( {\hat x_C^ + \ominus \hat x_C^ - } \right)\quad\end{eqnarray}
(5)
 
\begin{eqnarray} w = \sigma _C^{2 - }/\left( {\sigma _C^{2 + } + \sigma _C^{2 - }} \right)\quad \end{eqnarray}
(6)
This weighting is optimal for combining normally distributed estimates (it is the minimum variance unbiased estimator) and so is also optimal for wrapped normal variables as long as their variances are not too large. 
The feature-based integration model was identical, except that whether integration occurred in the reported dimension (color) depended only on whether a change was detected in the same feature dimension:  
\begin{equation}{\hat x_C} = \left\{ {\begin{array}{@{}*{3}{l}@{}} {\hat x_C^ + }&\;{{\rm{if}}\;{{\hat \Delta }_C} > {\theta _C}}&{}\\ {\hat x_C^I}&\;{{\rm{otherwise}}}&{} \end{array}} \right.\end{equation}
(7)
 
Model fitting
The models are fully specified by the description above; however, some further analysis is needed to determine the likelihoods of model parameters. On a single trial, the probability of detecting a change in each dimension is  
\begin{eqnarray} {\rm{Pr}}\left( {{D_C}} \right)&\;=& {\rm{Pr}}({{\hat \Delta }_C} > {\theta _C}) \nonumber\\ &\;=& \left(1 - {\Phi _{WN}}\left( {{\theta _C};x_C^ + \ominus x_C^ - ,\sigma _C^{2 - } + \sigma _C^{2 + }} \right)\right)\nonumber\\ &&\;+\, {{\Phi _{WN}}\left( { - {\theta _C};x_C^ + \ominus x_C^ - ,\sigma _C^{2 - } + \sigma _C^{2 + }} \right)}\end{eqnarray}
(8)
 
\begin{eqnarray} {\rm{Pr}}\left( {{D_L}} \right) &\;=& {\rm{Pr}} ({{\hat \Delta }_L} > {\theta _L}) \nonumber\\ &\;=& \left(1 - \Phi \left( {{\theta _L};x_L^ + - x_L^ - ,\sigma _L^{2 - } + \sigma _L^{2 + }} \right)\right)\nonumber\\ &&\;+\,{ \Phi \left( { - {\theta _L};x_L^ + - x_L^ - ,\sigma _L^{2 - } + \sigma _L^{2 + }} \right)} \end{eqnarray}
(9)
where \({\Phi _{WN}}( {\theta ;\mu ,\;{\sigma ^2}} ) = \mathop \smallint \nolimits_{ - \pi }^\theta {\phi _{WN}}( {\theta ^{\prime};\mu ,\;{\sigma ^2}} )d\theta ^{\prime}\) is the cumulative density function of the wrapped normal distribution, and Φ(θ; μ,  σ2) is the usual normal cumulative density function. The joint probability of obtaining analog response \({\hat x_C}\) and not detecting a change is the same for both the object-based and feature-based models:  
\begin{equation}\begin{array}{@{}*{2}{l}@{}} {p\left( {{{\hat x}_C},\;\hat y = 0} \right)}&{ = p({{\hat x}_C}|\hat y = 0){\rm{Pr}}\left( {\hat y = 0} \right)}\\ \\ {}&{ = p\left(\hat x_C^I|{{\bar D}_C}\right){\rm{Pr}}\left( {{{\bar D}_C}} \right){\rm{Pr}}\left( {{{\bar D}_L}} \right)} \end{array}\end{equation}
(10)
where \({\bar D_ \circ }\) means no change detected in the indicated dimension, \({\rm{Pr}}( {{{\bar D}_ \circ }} ) = 1 - {\rm{Pr}}( {{D_ \circ }} )\), and the distribution of the integrated color estimate is statistically independent of detection of a location change. 
The joint probability of the analog response and a change being detected differs between models. For the object-based model it is  
\begin{eqnarray} p\left( {{{\hat x}_C},\;\hat y = 1} \right) &\;=& p\left(\hat x_C^ + |{D_C}\right){\rm{Pr}}\left( {{D_C}} \right)\nonumber\\ &&\;+ p\left(\hat x_C^ + |{\bar D_C}\right){\rm{Pr}}\left( {{{\bar D}_C}} \right){\rm{Pr}}\left( {{D_L}} \right)\quad\quad\end{eqnarray}
(11)
 
whereas for the feature-based model it is  
\begin{eqnarray} p\left( {{{\hat x}_C},\;\hat y = 1} \right) &\;=& p\left(\hat x_C^ + |{D_C}\right){\rm{Pr}}\left( {{D_C}} \right) \nonumber\\ &&\;+ p\left(\hat x_C^I|{\bar D_C}\right){\rm{Pr}}\left( {{{\bar D}_C}} \right){\rm{Pr}}\left( {{D_L}} \right)\qquad\end{eqnarray}
(12)
 
The conditional probability density for the analog response given no color change was detected can be calculated as  
\begin{eqnarray}\!\!\!\!\!\!\!\!\!\!\!\!\!\! &&\!\!\!\!\!\!\!\!\!\!\!\!\!\! {p\left(\hat x_C^I|{{\bar D}_C}\right)} \approx { p\left( {\hat x_C^I} \right)}\nonumber\\ &&\!\!\!\!\!\!\! \approx {\phi _{WN}}\Bigg( \hat x_C^I;x_C^ - \oplus w \left( {x_C^ + \ominus x_C^ - } \right),\frac{{\sigma _C^{2 + }\sigma _C^{2 - }}}{{\sigma _C^{2 + } + \sigma _C^{2 - }}} \Bigg)\qquad \end{eqnarray}
(13)
where, due to the circularity of the color space, both the independence of the integrated estimate from the detection state and its distribution as a wrapped normal with the stated variance are close approximations only. Using Bayes’ theorem,  
\begin{equation}p\left( {\hat x_C^ + {\rm{|}}{D_C}} \right) = \frac{{\Pr \left( {{D_C}{\rm{|}}\hat x_C^ + } \right)}}{{\Pr \left( {{D_C}} \right)}}p\left( {\hat x_C^ + } \right)\;\end{equation}
(14)
and  
\begin{equation} p\left(\hat x_C^ + |{\bar D_C}\right) = \frac{{1 - {\rm{Pr}}({D_C}|\hat x_C^ + )}}{{1 - {\rm{Pr}}\left( {{D_C}} \right)}}p\left( {\hat x_C^ + } \right)\end{equation}
(15)
where  
\begin{eqnarray}\!\!\!\!\!\!\!\!\!\!\!\!\! {{\rm{Pr}}\left({D_C}|\hat x_C^ + \right)}&\; =& {\rm{Pr}}\left(\hat x_C^ + \ominus \hat x_C^ - > {\theta _C} \cup \hat x_C^ + \ominus \hat x_C^ - < { - {\theta _C}} |\hat x_C^ + \right)\!\!\!\!\!\nonumber\\ & \;=& {\Phi _{WN}}\left( { - {\theta _C};\hat x_C^ + \ominus x_C^ - ,\sigma _C^{2 - }} \right) \nonumber\\ &&+ \left( {1 - {\Phi _{WN}}\left( {{\theta _C};\hat x_C^ + \ominus x_C^ - ,\sigma _C^{2 - }} \right)} \right) \end{eqnarray}
(16)
 
Pr(DC) is given in Equation 8, and \(p( {\hat x_C^ + } ) = {\phi _{WN}}( {\hat x_C^ + ,x_C^ + ,\sigma _C^{2 + }} )\). The probability of both responses is then given by  
\begin{eqnarray}p\left( {{{\hat x}_C},\;y} \right) &\;=& \left( {1 - \gamma } \right)p\left( {{{\hat x}_C},\;\hat y = y} \right) \nonumber\\ &&\;+\, \gamma p\left( {{{\hat x}_C},\;\hat y = 1 - y} \right)\qquad\end{eqnarray}
(17)
 
We used a nonlinear optimization algorithm (fminsearch in MATLAB) to find parameters of the model that maximized this likelihood over all trials, for each participant separately. The pre- and postsaccadic variances in location, \(\sigma _L^{2 - }\) and \(\sigma _L^{2 + }\), were not separable in the model so were fit with a single average parameter \({\bar \sigma} _L^2 = {{( {\sigma _L^{2 - } + \sigma _L^{2 + }} )} / 2}\). The models therefore each had six free parameters: \({\theta _C},\;{\theta _L},\;\sigma _C^{2 - },\;\sigma _C^{2 + },\;\bar \sigma _L^2,\;{\rm{and }}\;\gamma \)
The models for Experiment 2 and the color-report trials of Experiment 3 were identical, with the exception that orientation took the place of location and was treated as a circular variable, with individual estimates distributed as wrapped normals and addition and subtraction on the circle. The models for orientation-report trials of Experiment 3 incorporated these same changes, but with color interchanged with orientation. In this case, pre- and postsaccadic orientation variances were separable and were treated as two free parameters, increasing the number of parameters in each model to seven. 
All model comparisons were between models with the same number of free parameters, so we report the results as differences in log likelihood (ΔLL). Using Bayesian information criteria or Akaike's information criteria differences instead would not have changed conclusions, as these differences simply have magnitudes of twice the log likelihood difference. 
Results
In Experiment 1, we investigated whether changes in the location of an object would affect the transsaccadic integration of color information. Observers were instructed to report the color of the stimulus (the last color they saw if they believed it had changed) and whether they had detected any change in the color or location of the stimulus. 
The effects of the magnitude of the color and location change on the proportion of trials on which a change was reported are shown by the red symbols in Figures 3A and 3B, respectively. As expected, detection probability (red lines) increased with the size of the change in each feature and shows no evidence of either floor or ceiling effects. Although this may be surprising given the large magnitudes of some of the color and location changes, this can be attributed to the stimulus design parameters, which were chosen specifically to make the task difficult enough that 100% detection was rarely reached (see Methods). A two-way, repeated-measures Bayesian ANOVA found that the best fitting model incorporated both main effects of color and location (BF10 = 6.48 × 1017). Follow-up Bayesian t-tests indicated that increasing changes in color led to increasing detection, with more change reports for 70° color differences than for 35° (BF10 = 839.50) and more for 35° than for 20° (BF10 = 24.30). Similarly, there were more change reports for 2.5° location differences than 1.5° (BF10 = 25.57) and more for 1.5° than 0° (BF10 = 42.46). 
Figure 3.
 
Results of Experiment 1. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and location change (B). Curves show predictions of the best-fitting model, which was feature based. (C) Distribution of reported color relative to presaccadic color (0°) and postsaccadic color (dotted vertical line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and location change (columns: magnitude indicated at top). Error bars and error patches represent ± 1 SE.
Figure 3.
 
Results of Experiment 1. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and location change (B). Curves show predictions of the best-fitting model, which was feature based. (C) Distribution of reported color relative to presaccadic color (0°) and postsaccadic color (dotted vertical line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and location change (columns: magnitude indicated at top). Error bars and error patches represent ± 1 SE.
Having established that both the location and color manipulations were effective at modulating the probability of detecting a change to the stimulus, we examined their effects on its perceived color. We computed the bias in reported color as a value between 1 (the postsaccadic color) and 0 (the presaccadic color). Blue symbols in Figures 3A and3B show the mean color bias for each color and location change magnitude. For the smallest color changes, despite the instruction to report only the last color seen, observers’ responses fell on average approximately midway (0.60 ± 0.05) between pre- and postsaccadic values. As the magnitude of color change increased, responses were increasingly biased toward the postsaccadic color, with a mean bias of 0.90 ± 0.05 for the largest color change. 
In contrast, changes in location had minimal influence on transsaccadic integration of color information, as assessed by the mean color bias (Figure 3B, blue symbols), despite the clear effects on detection (red symbols). This qualitative difference was confirmed by a two-way Bayesian repeated-measures ANOVA, which supported the model incorporating only the main effect of color over the null hypothesis (BF10 = 3.39 × 1013), but the analysis showed that the null hypothesis was favored over a model incorporating the main effect of location (BF01 = 7.04), and a model incorporating only the main effect of color was favored over a model that incorporated both main effects (BF = 3.93). Follow-up Bayesian t-tests indicated that color estimates were proportionally more biased toward the postsaccadic color as the color difference increased, as color estimates were more biased for 70° color differences than 35° (BF10 = 7.94 × 104) and were more biased for 35° than 20° (BF10 = 2.30). 
These results suggest that, although both color and location changes were detected by observers on some trials and varying the change magnitude successfully modulated detectability over a large range for both features, only the detection of color changes had an influence on whether pre- and postsaccadic colors were integrated or segregated. To demonstrate the plausibility of this explanation and to create a direct test for it, we compared fits of two observer models to the joint distributions of color estimates and change reports (shown in Figure 3C). 
In the feature-based integration model, observers reported an integrated color estimate (green distribution in Figure 2C) if the change in color across the saccade was not detected (green region in Figure 2B) and an estimate based only on the postsaccadic input if it was detected (purple region and distribution in Figures 2B and 2C, respectively). This was contrasted with an object-based integration model, in which an integrated estimate was reported only if no change in either feature was detected. Model comparison favored the feature-based model (ΔLL = 6.58 ± 3.23; nine out of 12 participants). Predictions of this best-fitting model are shown as colored lines (with error patches) in each panel of Figure 3 (predictions for the object-based integration model can be seen in the Supplementary Materials). 
This is strong evidence that the perception of a location change is disregarded when determining the perception of the color of an object after a saccade. We considered two explanations of this finding to have roughly equal plausibility. One possibility is that the decision to integrate or segregate features across a saccade is made independently in each different feature dimension (as assumed by our feature-based integration model). Alternatively, location changes specifically might be disregarded for purposes of integration. This could be because objects tend to change their locations more often than they change other features such as color, making a change in location a less reliable marker of object continuity than other features. Indeed, changes of retinal location are an inevitable result of every saccade, and spatiotopic locations also change frequently because of physical interactions and gravity. To distinguish between these possibilities, we conducted Experiment 2. 
Experiment 2
This experiment was identical to Experiment 1, except that location changes were replaced with changes of stimulus orientation (illustrated by the top path in Figure 1B). The effects of color and orientation change magnitude on the proportion of trials on which a change was reported are shown by the red symbols in Figures 4A and 4B, respectively. As expected, detection probability increased with the size of the change in each feature. A Bayesian ANOVA found that the best-fitting model was one that incorporated both main effects of color and orientation (BF10 = 7.42 × 1010). Follow-up Bayesian t-tests indicated that increasing differences in color led to increasing detection, with more change reports for 70° color differences than for 35° (BF10 = 31.24) and more for 35° color differences than for 20° (BF10 = 7.39). Weak evidence was found indicating that greater differences in orientation also led to increasing detection, with more change reports for 35° orientation differences than for 10° (BF10 = 2.34), and ambivalent evidence for any difference in change reports between 90° and 35° orientation differences (BF01 = 1.16). 
Figure 4.
 
Results of Experiment 2. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic color (dotted line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 4.
 
Results of Experiment 2. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic color (dotted line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Because both features were effective at modulating detection of changes in the stimulus, we examined their effects on perceived color. As shown by the blue symbols in Figure 4A, color responses were biased increasingly toward the postsaccadic value as the color change magnitude increased (from 0.72 ± 0.03 to 0.98 ± 0.02). However, changes in orientation had minimal effect on the integration of color information (Figure 4B, blue symbols), despite improving detection (red symbols). This was confirmed by a Bayesian ANOVA, which found that a model incorporating only the main effect of color was more likely than the null hypothesis (BF10 = 1.33 × 1011) but that the null hypothesis was favored over a model that incorporated only the main effect of orientation (BF01 = 8.18). The best-fitting model incorporated only the main effect of color and was more likely than one that incorporated both main effects (BF = 6.26). Follow-up Bayesian t-tests indicated that color estimates were more biased toward the postsaccadic color as the color difference increased (70° vs. 35°, BF10 = 1.47 × 104; 35° vs. 20°, BF10 = 118.92). 
These results replicate those found for location changes in Experiment 1. Although the detection of changes increased with the magnitude of change in both color and orientation changes, the integration of pre- and postsaccadic colors was affected only by the detection of color changes. We tested this by comparing fits of the feature- and object-based integration models to the joint distributions of color estimates and change reports (Figure 4C). Model comparison favored the feature-based model (ΔLL = 1.01 ± 0.36; 10 out of 12 participants). 
This is evidence that the lower reliability of location as a marker of object correspondence is not the reason why it is disregarded for the purpose of integrating color information across saccades. Instead, the evidence favors the explanation that the decision of whether or not to integrate a feature across a saccade depends only on the feature that is being integrated. However, we note that both Experiments 1 and 2 required an analog report only in the color dimension, with the relevance of the other feature dimension limited to the binary change response. It is possible that this could have encouraged participants to pay more attention to color than location or orientation, which could influence feature binding (Treisman & Gelade, 1980). We conducted Experiment 3 to test this hypothesis. 
Experiment 3
In this experiment, we balanced attentional requirements by making unpredictable which feature participants would be asked to report on each trial. This is important, as the amount of attention to the stimulus has been shown to affect perception in transsaccadic integration (Stewart & Schütz, 2018; Van der Stigchel, Schut, Fabius, & Van der Stoep, 2020). Half of the trials were identical to those of Experiment 2. On the other half of trials, randomly interleaved, participants were asked to reproduce the last orientation they had seen instead of the last color. This removed the asymmetry in task demands that might have caused participants to pay more attention to the color of the stimulus. 
Color report trials
The effects of color and orientation change magnitude on the proportion of trials in which a change was reported are indicated by the red symbols in Figures 5A and 5B, respectively. As in the previous two experiments, detection rates increased with magnitude of the change in each feature. A Bayesian ANOVA confirmed this, with the best model incorporating both main effects (BF10 = 1.16 × 1017). Follow-up Bayesian t-tests indicated that increasing differences in color led to more reports of changes (70° vs. 35°, BF10 = 5.53 × 104; 35° vs. 20°, BF10 = 282.63), as did increasing differences in orientation (90° vs. 35°, BF10 = 33.14; 35° vs. 10°, BF10 = 19.73). 
Figure 5.
 
Results from the color-report trials in Experiment 3. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 5.
 
Results from the color-report trials in Experiment 3. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Having established that both features were effective at modulating the probability of detecting a change in the stimulus, we turned to the effects on perceived color. As shown by the blue symbols in Figure 5A, color reports were increasingly biased toward the postsaccadic color (from 0.66 ± 0.03 to 0.99 ± 0.02) as the magnitude of the color change increased. However, as shown by the blue symbols in Figure 5B, the magnitude of the orientation change had minimal effect on the color reports. This was confirmed by a Bayesian ANOVA, which found that the model incorporating only the main effect of color was favored over the null hypothesis (BF10 = 1.18 × 1015), whereas the null hypothesis was favored over the model incorporating only the main effect of orientation (BF01 = 5.24). Furthermore, the best-fitting model incorporated only the main effect of color, being favored over a model incorporating both main effects (BF = 2.02). Follow-up Bayesian t-tests indicated that color estimates were proportionally more biased toward the postsaccadic color as the color difference increased (70° vs. 35°, BF10 = 1.32 × 104; 35° vs. 20°, BF10 = 37.44). 
These results indicate that, even when task demands were balanced across the two feature dimensions, the independence between color bias and orientation changes observed in Experiments 1 and 2 remained. This was confirmed by fitting the joint distributions of color estimates and change reports (shown in Figure 5C) with the feature-based and object-based integration models. Model comparison again favored the feature-based model (ΔLL = 3.65 ± 2.03; 10 out of 12 participants). 
Orientation report trials
Although the choice of orientation changes in Experiment 3 was optimized for examining integration of color estimates, this experiment also gave us an opportunity to test if our conclusions generalized from color integration to integration of orientations. 
As expected, all of the effects of color and orientation change magnitude on detection that we had observed in the color report trials were also present in the orientation report trials (red symbols in Figures 6A and 6B). However, the effects of color and orientation changes on orientation estimates were the opposite of those seen for color estimates, in that orientation bias did not vary with the color change magnitude but increased with increasing orientation change magnitude (from 0.77 ± 0.06 to 1.01 ± 0.01). This was confirmed by a Bayesian ANOVA which found that a model incorporating only the main effect of orientation was favored over the null hypothesis (BF10 = 164.46), and the null hypothesis was favored over the model incorporating only the main effect of color (BF01 = 8.10). The best fitting model incorporated only orientation, being favored over the model incorporating both main effects (BF = 7.63). Follow-up Bayesian t-tests indicated that orientation estimates were more biased toward the postsaccadic orientation, as the orientation difference increased from 10° to 35° (BF10 = 20.58). There was no difference between 35° and 90° (BF01 = 2.53), presumably because the bias was already close to its ceiling value of 1 for 35° differences. 
Figure 6.
 
Results from the orientation-report trials in Experiment 3. (A, B) Mean orientation bias (cyan symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported orientation relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of orientation change (rows: magnitude indicated in top-left of each panel) and color change (columns: magnitude indicated at top). Model fits for the 90° orientation change condition are not available, as the optimal observer response for such a large change is not uniquely defined.
Figure 6.
 
Results from the orientation-report trials in Experiment 3. (A, B) Mean orientation bias (cyan symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported orientation relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of orientation change (rows: magnitude indicated in top-left of each panel) and color change (columns: magnitude indicated at top). Model fits for the 90° orientation change condition are not available, as the optimal observer response for such a large change is not uniquely defined.
These results suggest that, although detection of color changes was used to determine whether or not to integrate color information across a saccade, changes in this feature were simultaneously ignored when determining whether or not to integrate orientation information (and vice versa). To test this directly, we fit the full data combining both color and orientation report trials with feature-based and object-based integration models. Parameters were shared across trial types such that, for example, the same color variance parameters that contributed to change detection and analog response probabilities in the color report trials also contributed to response probabilities in the orientation report trials. Note that trials with a 90° orientation change (i.e., when pre- and postsaccadic orientations were orthogonal) were omitted from the modeling, as the optimal integrated estimate is not clearly defined (Murray & Morgenstern, 2010). In practice, participants appear to have always reported the postsaccadic orientation on these trials (bottom panels of Figure 6C). Model comparison again favored the feature-based integration model (ΔLL = 1.29 ± 2.96; eight out of 12 participants), in which integration in each feature dimension was determined independently of changes detected in the other feature dimension. 
Discussion
In everyday visual experience, objects that are otherwise stable tend not to abruptly change their properties during an eye movement. This makes integrating pre- and postsaccadic visual information a potentially advantageous strategy, as the combined estimate should be less influenced by perceptual noise. However, as the fidelity of presaccadic information available for transsaccadic computations is limited (Kong, Kroell, Schneegans, Aagten-Murphy, & Bays, 2021), any evidence for a mismatch with the current environment could be considered justification to discount it and rely exclusively on the current (postsaccadic) input. However, across three experiments, we have found that the visual system operates according to different rules. 
In Experiment 1, we investigated whether changes in the location of an object during a saccade altered the relative weighting given to pre- and postsaccadic colors in reports of the color of the object. We found no evidence for an influence of location changes on color integration, and our results were consistent with a model in which the decision to integrate or segregate color information was made solely on the basis of detecting color changes. In Experiments 2 and 3, we expanded this finding by investigating whether the detection of changes in the orientation of an object would influence transsaccadic integration of its color, or vice versa. Experiment 3 also provided evidence against an alternative explanation in terms of unbalanced attention to the two features. In every case, changes in a particular feature only affected integration for that same feature. Note that our experimental settings may have influenced participants to detect changes and segregate inputs more readily, as the frequency of intrasaccadic changes was much greater than usual for an object in the natural environment. However, although absolute rates of change detection may be inflated, this does not change our conclusions about the process of transsaccadic integration. 
These results indicate that the integration of features held in presaccadic memory with postsaccadic visual input can occur in parallel for different features, separately from any overall decision about “same or different object.” In other words, an object may be seen to change in one feature at the same time that the perception of another feature reflects an averaging of pre- and postsaccadic values. For example, in trials with a large change in orientation and a small change in color, our results suggest that a participant who reported seeing a change would reproduce the postsaccadic orientation if asked for an orientation response but report an integrated pre- and postsaccadic color if asked for a color response. 
This pattern of responses may indicate that the visual system has adapted to take into account situations where objects change in one visual dimension while another remains stable; for example, a ball rolling from light into shadow keeps a constant shape, but its luminance profile changes sharply. Whether a feature-based mechanism is advantageous in an ecological setting will depend on statistical regularities in the environment, in particular the frequency with which abrupt changes in different feature dimensions of an object co-occur. A truly optimal decision would also take into account the relative behavioral costs of incorrectly integrating features that have undergone a real change versus incorrectly segregating features that could properly have been integrated. 
An important point to address is the seeming discrepancy between a purely feature-based integration account, and the finding that changes in surface features aid in the detection of a displacement during a saccade (Tas et al., 2012). Our experiment was designed to test for feature interdependence in transsaccadic integration rather than change detection, meaning that we did not ask participants for separate reports of change detection in each feature dimension. Nonetheless, our results were quite accurately captured by a model in which detection occurred independently for each feature, and we found no evidence in any of the experiments for an effect of changes in the non-report feature on the analog estimates of the report feature. 
One possibility is that small dependencies between the two feature-level change detection processes were not detected in our study because they did not translate into object-level integration or segregation. However, there are also reasons why we might expect detection of intrasaccadic displacements, particularly ones parallel to the saccade direction as used by Tas et al. (2012), to be specifically sensitive to changes in other feature dimensions in a way that the features examined in the present study are not. First, unlike object colors or orientations, locations of objects in the visual field are necessarily altered by eye movements. Determining whether the location of the object in the world has changed requires (at least in the artificially sparse environments typical of suppression of saccadic displacement studies) a comparison of its new retinal location to a prediction based on its presaccadic location and the eye movement vector. The latter is uncertain due to noise in the oculomotor system—especially so in the direction parallel to the saccade. This makes detection of transsaccadic location changes uniquely uncertain, and perhaps it is only in this case that feature changes in other dimensions have an appreciable influence. 
Another possibility lies in the fact that our stimulus was distinct from and always some distance from the saccade target. Although transsaccadic integration has been observed for objects at varying locations in the visual field (Oostwoud Wijdenes et al., 2015; Stewart & Schütz, 2019a), it is well established that the saccade target is automatically allocated attention (Deubel & Schneider, 2003) and stored in visual working memory (Bays & Husain, 2008; Schut, Van der Stoep, Postma, & Van der Stigchel, 2017). Based on the present results, we cannot rule out the possibility that an object-based integration process operates only for the saccade target or in its immediate vicinity, with feature-based integration everywhere else. However, we would argue this is unlikely. A comparison between the results of Experiments 2 and 3 indicates that attention is not critical in determining whether the integration process is feature based, and another recent study (Kong et al., 2021) demonstrated that visual working memory is involved in the integration process for multiple stimuli, not just the saccade target. 
A final reason why location changes might be treated differently by the visual system has to do with the statistics of objects in the natural environment, rather than saccades. If an object changing its location is in general less predictive of a change in its surface features than the converse, then an observer optimized for that environment will be more weakly influenced by location changes in judging whether a surface feature has changed, as compared with the influence of surface feature changes on judging location. This account would not predict differences between movements orthogonal and parallel to the saccade, and it would be interesting to know whether the finding of Tas et al. (2012) generalizes to orthogonal movements. The tendency for objects to move while their other properties remain stable may also underlie the privileged role of spatial location in binding other object features in memory (Golomb, Kupitz, & Thiemann, 2014; Kovacs & Harris, 2019; Pertzov & Husain, 2014; Schneegans & Bays, 2017). 
Recently, the term serial dependence has been used to describe phenomena in which the report of a visual stimulus is influenced by a statistically unrelated stimulus presented earlier, typically on the previous trial (Fischer & Whitney, 2014). Like transsaccadic integration (Irwin, 1991; Kong et al., 2021; Oostwoud Wijdenes et al., 2015; Schut et al., 2017), it has been suggested that serial dependence involves an integration of information obtained at two different points of time based on representations in working memory (Bliss, Sun, & D'Esposito, 2017; Fritsche, Mostert, & de Lange, 2017). Importantly for the purposes of this discussion, serial dependence also appears to happen on the level of features. One study found an attractive serial dependence in facial gender but a repulsive serial dependence of facial expression, despite the two features being bound to the same face (Taubert, Alais, & Burr, 2016). 
Note that there are also important differences between the two phenomena. Transsaccadic integration involves saccadic suppression (e.g., Binda, Cicchini, Burr, & Morrone, 2009) and a translation of the retinal locations of objects. It is generally associated with a lack of awareness that a change has occurred, and effect sizes are generally greater than in serial dependence studies. For example, Fischer and Whitney (2014) found a peak bias of 8° in orientation toward the stimulus on the previous trial, which rapidly fell to zero as the difference between the previous and current stimuli increased. This result is actually something of an outlier in serial dependency studies, which have typically found even smaller effects, including peaks of 2° to 3° for Gabor patch orientation (Cicchini, Mikellidou, & Burr, 2017), 0.5° to 2° for eye gaze direction (Alais, Kong, Palmer, & Clifford, 2018), and 2° to 3° for motion direction (Fischer, Czoschke, Peters, Rahm, Kaiser, & Bledowski, 2020). Notably, the findings of Wittenberg, Bremmer, and Wachtler (2008)—where small biases were observed toward a statistically unrelated stimulus separated from the current stimulus by a blank interval in addition to a saccade—might be more closely related to observations of serial dependence studies than transsaccadic integration. 
In comparison, in Experiment 1, we found a peak effect of 11.0° in color bias in the 35° color change condition, which fell only as far as 7.9° at a 20° color change, and 6.5° at a 70° color change. If we were to confine ourselves to only those trials on which participants reported not detecting a change (which are the trials on which we believe integration took place), the effect sizes would be larger still: 9.5°, 17.9°, and 32.6°, respectively, for 20°, 35°, and 70° changes (i.e., approximately half the size of the color change). Recent work on serial dependence has found evidence to suggest that spatial location plays a distinct role as compared with other surface features (Fischer et al., 2020). However, further research is needed to determine whether and to what extent transsaccadic integration and serial dependence are expressions of a common mechanism of object correspondence. 
Acknowledgments
Supported by a grant from the Wellcome Trust (106926 to P.M.B.). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Data and code associated with this study is publicly available at https://osf.io/2fbkm/
Commercial relationships: none. 
Corresponding author: Garry Kong. 
Email: gk426@cam.ac.uk. 
Address: Department of Psychology, University of Cambridge, Cambridge, UK. 
References
Aagten-Murphy, D., & Bays, P. M. (2018). Functions of memory across saccadic eye movements. Current Topics in Behavioral Neurosciences, 41: 155–183, https://doi.org/10.1007/7854_2018_66. [CrossRef]
Alais, D., Kong, G., Palmer, C., & Clifford, C. (2018). Eye gaze direction shows a positive serial dependency. Journal of Vision, 18(4):11, 1–12, https://doi.org/10.1167/18.4.11. [CrossRef]
Atsma, J., Maij, F., Koppen, M., Irwin, D. E., & Medendorp, W. P. (2016). Causal inference for spatial constancy across saccades. PLoS Computational Biology, 12(3), e1004766, https://doi.org/10.1371/journal.pcbi.1004766. [CrossRef]
Bays, P. M., & Husain, M. (2008). Dynamic shifts of limited working memory resources in human vision. Science, 321(5890), 851–854, https://doi.org/10.1126/science.1158023. [CrossRef]
Binda, P., Cicchini, G. M., Burr, D. C., & Morrone, M. C. (2009). Spatiotemporal distortions of visual perception at the time of saccades. The Journal of Neuroscience, 29(42), 13147–13157, https://doi.org/10.1523/jneurosci.3723-09.2009. [CrossRef]
Bliss, D. P., Sun, J. J., & D'Esposito, M. (2017). Serial dependence is absent at the time of perception but increases in visual working memory. Scientific Reports, 7(1), 14739–14739, https://doi.org/10.1038/s41598-017-15199-7. [CrossRef]
Bridgeman, B., Hendry, D., & Stark, L. (1975). Failure to detect displacement of the visual world during saccadic eye movements. Vision Research, 15(6), 719–722, https://doi.org/10.1016/0042-6989(75)90290-4. [CrossRef]
Burr, D. C., & Morrone, M. C. (2011). Spatiotopic coding and remapping in humans. Transactions of the Royal Society of London. Series B, Biological Sciences, 366(1564), 504–515, https://doi.org/10.1098/rstb.2010.0244. [CrossRef]
Cavanagh, P., Hunt, A. R., Afraz, A., & Rolfs, M. (2010). Visual stability based on remapping of attention pointers. Trends in Cognitive Sciences, 14(4), 147–153, https://doi.org/10.1016/j.tics.2010.01.007. [CrossRef]
Cicchini, G. M., Binda, P., Burr, D. C., & Morrone, M. C. (2013). Transient spatiotopic integration across saccadic eye movements mediates visual stability. Journal of Neurophysiology, 109(4), 1117–1125, https://doi.org/10.1152/jn.00478.2012. [CrossRef]
Cicchini, G. M., Mikellidou, K., & Burr, D. (2017). Serial dependencies act directly on perception. Journal of Vision, 17(14):6, 1–9, https://doi.org/10.1167/17.14.6. [CrossRef]
Demeyer, M., De Graef, P., Wagemans, J., & Verfaillie, K. (2009). Transsaccadic identification of highly similar artificial shapes. Journal of Vision, 9(4):28, 1–14, https://doi.org/10.1167/9.4.28. [CrossRef]
Demeyer, M., De Graef, P., Wagemans, J., & Verfaillie, K. (2010). Object form discontinuity facilitates displacement discrimination across saccades. Journal of Vision, 10(6):17, 1–14, https://doi.org/10.1167/10.6.17. [CrossRef]
Deubel, H., & Schneider, W. X. (2003). Delayed saccades, but not delayed manual aiming movements, require visual attention shifts. Annals of the New York Academy of Sciences, 1004(1), 289–296, https://doi.org/10.1196/annals.1303.026. [CrossRef]
Ernst, M. O., & Banks, M. S. (2002). Humans integrate visual and haptic information in a statistically optimal fashion. Nature, 415(6870), 429–433, https://doi.org/10.1038/415429a. [CrossRef]
Fischer, C., Czoschke, S., Peters, B., Rahm, B., Kaiser, J., & Bledowski, C. (2020). Context information supports serial dependence of multiple visual objects across memory episodes. Nature Communications, 11(1), 1932, https://doi.org/10.1038/s41467-020-15874-w. [CrossRef]
Fischer, J., & Whitney, D. (2014). Serial dependence in visual perception. Nature Neuroscience, 17(5), 738–743, https://doi.org/10.1038/nn.3689. [CrossRef]
Fritsche, M., Mostert, P., & de Lange, F. P. (2017). Opposite effects of recent history on perception and decision. Current Biology, 27(4), 590–595, https://doi.org/10.1016/j.cub.2017.01.006. [CrossRef]
Ganmor, E., Landy, M. S., & Simoncelli, E. P. (2015). Near-optimal integration of orientation information across saccades. Journal of Vision, 15(16):8, 1–12, https://doi.org/10.1167/15.16.8. [CrossRef]
Golomb, J. D., Kupitz, C. N., & Thiemann, C. T. (2014). The influence of object location on identity: a “spatial congruency bias”. Journal of Experimental Psychology: General, 143(6), 2262–2278, https://doi.org/10.1037/xge0000017. [CrossRef] [PubMed]
Hubner, C., & Schutz, A. C. (2017). Numerosity estimation benefits from transsaccadic information integration. Journal of Vision, 17(13):12, 1–16, https://doi.org/10.1167/17.13.12. [CrossRef]
Irwin, D. E. (1991). Information integration across saccadic eye movements. Cognitive Psychology, 23(3), 420–456, https://doi.org/10.1016/0010-0285(91)90015-g. [CrossRef]
JASP Team. (2020). JASP: A fresh way to do statistics. Retrieved from https://jasp-stats.org/.
Kleiner, M., Brainard, D., Pelli, D., Ingling, A., Murray, R., & Broussard, C. (2007). What's new in Psychtoolbox-3. Perception, 36(14), 1–16.
Kong, G., Kroell, L. M., Schneegans, S., Aagten-Murphy, D., & Bays, P. M. (2021). Transsaccadic integration relies on a limited memory resource. Journal of Vision, 21(5):24, 1–12, https://doi.org/10.1167/jov.21.5.24. [CrossRef]
Kovacs, O., & Harris, I. M. (2019). The role of location in visual feature binding. Attention, Perception, and Psychophysics, 81(5), 1551–1563, https://doi.org/10.3758/s13414-018-01638-8. [CrossRef]
Melcher, D., & Morrone, M. C. (2015). Nonretinotopic visual processing in the brain. Visual Neuroscience, 32, E017, https://doi.org/10.1017/s095252381500019x. [CrossRef]
Murray, R. F., & Morgenstern, Y. (2010). Cue combination on the circle and the sphere. Journal of Vision, 10(11):15, 1–11.
Niemeier, M., Crawford, J. D., & Tweed, D. B. (2003). Optimal transsaccadic integration explains distorted spatial perception. Nature, 422(6927), 76–80. [CrossRef]
Oostwoud Wijdenes, L., Marshall, L., & Bays, P. M. (2015). Evidence for optimal integration of visual feature representations across saccades. The Journal of Neuroscience, 35(28), 10146–10153, https://doi.org/10.1523/jneurosci.1040-15.2015. [CrossRef]
Pertzov, Y., & Husain, M. (2014). The privileged role of location in visual working memory. Attention, Perception, and Psychophysics, 76(7), 1914–1924, https://doi.org/10.3758/s13414-013-0541-y. [CrossRef]
Poth, C. H., Herwig, A., & Schneider, W. X. (2015). Breaking object correspondence across saccadic eye movements deteriorates object recognition. Frontiers in Systems Neuroscience, 9, 176, https://doi.org/10.3389/fnsys.2015.00176. [CrossRef]
Rensink, R. A., O'Regan, J. K., & Clark, J. J. (1997). To see or not to see: the need for attention to perceive changes in scenes. Psychological Science, 8(5), 368–373, https://doi.org/10.1111/j.1467-9280.1997.tb00427.x. [CrossRef]
Schneegans, S., & Bays, P. M. (2017). Neural architecture for feature binding in visual working memory. The Journal of Neuroscience, 37(14), 3913–3925, https://doi.org/10.1523/JNEUROSCI.3493-16.2017. [CrossRef]
Schut, M. J., Van der Stoep, N., Postma, A., & Van der Stigchel, S. (2017). The cost of making an eye movement: A direct link between visual working memory and saccade execution. Journal of Vision, 17(6):15, 1–20, https://doi.org/10.1167/17.6.15. [CrossRef]
Shams, L., & Beierholm, U. R. (2010). Causal inference in perception. Trends in Cognitive Sciences, 14(9), 425–432, https://doi.org/10.1016/j.tics.2010.07.001. [CrossRef]
Stewart, E. E. M., & Schütz, A. C. (2018). Attention modulates trans-saccadic integration. Vision Research, 142, 1–10, https://doi.org/10.1016/j.visres.2017.11.006. [CrossRef]
Stewart, E. E. M., & Schütz, A. C. (2019a). Transsaccadic integration benefits are not limited to the saccade target. Journal of Neurophysiology, 122(4), 1491–1501, https://doi.org/10.1152/jn.00420.2019. [CrossRef]
Stewart, E. E. M., & Schütz, A. C. (2019b). Transsaccadic integration is dominated by early, independent noise. Journal of Vision, 19(6):17, 1–19, https://doi.org/10.1167/19.6.17. [CrossRef]
Stewart, E. E. M., Valsecchi, M., & Schütz, A. C. (2020). A review of interactions between peripheral and foveal vision. Journal of Vision, 20(12):2, 1–35, https://doi.org/10.1167/jov.20.12.2. [CrossRef]
Tas, A. C., Moore, C. M., & Hollingworth, A. (2012). An object-mediated updating account of insensitivity to transsaccadic change. Journal of Vision, 12(11):18, 1–13, https://doi.org/10.1167/12.11.18. [CrossRef]
Taubert, J., Alais, D., & Burr, D. (2016). Different coding strategies for the perception of stable and changeable facial attributes. Scientific Reports, 6(1), 32239, https://doi.org/10.1038/srep32239. [CrossRef]
Treisman, A., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12(1), 97–136. [CrossRef]
Van der Stigchel, S., Schut, M. J., Fabius, J., & Van der Stoep, N. (2020). Transsaccadic perception is affected by saccade landing point deviations after saccadic adaptation. Journal of Vision, 20(9):8, 1–12, https://doi.org/10.1167/jov.20.9.8. [CrossRef]
Wexler, M., & Collins, T. (2014). Orthogonal steps relieve saccadic suppression. Journal of Vision, 14(2):13, 1–9, https://doi.org/10.1167/14.2.13. [CrossRef]
Wittenberg, M., Bremmer, F., & Wachtler, T. (2008). Perceptual evidence for saccadic updating of color stimuli. Journal of Vision, 8(14):9, 1–9, https://doi.org/10.1167/8.14.9. [CrossRef]
Wolf, C., & Schutz, A. C. (2015). Trans-saccadic integration of peripheral and foveal feature information is close to optimal. Journal of Vision, 15(16):1, 1–18, https://doi.org/10.1167/15.16.1. [CrossRef]
Figure 1.
 
Experimental design. (A) Trial sequence for Experiment 1. A color patch was presented that changed in color and/or location while the participant was making a saccade. The participants reported the last color they had seen and then whether they had detected any change in the stimulus. (B) Experiment 2 followed the top path of this trial sequence (color and/or orientation change, followed by color report); Experiment 3 followed either the top or bottom path (color or orientation report) with equal frequency. Dashed red circles represent gaze fixations, and dashed red arrows represent saccades. The stimulus changed as soon as gaze crossed the vertical midline of the screen.
Figure 1.
 
Experimental design. (A) Trial sequence for Experiment 1. A color patch was presented that changed in color and/or location while the participant was making a saccade. The participants reported the last color they had seen and then whether they had detected any change in the stimulus. (B) Experiment 2 followed the top path of this trial sequence (color and/or orientation change, followed by color report); Experiment 3 followed either the top or bottom path (color or orientation report) with equal frequency. Dashed red circles represent gaze fixations, and dashed red arrows represent saccades. The stimulus changed as soon as gaze crossed the vertical midline of the screen.
Figure 2.
 
Pictorial representation of the probabilistic observer model. (A) The pre- and postsaccadic color and location representations are corrupted by Gaussian noise, as represented by the bivariate Gaussians on the location and color feature spaces. (B) Two-tailed thresholds, as represented by the dotted lines, are set for the purpose of detecting a change. If the postsaccadic stimulus exceeds a feature threshold, then a change is detected in the corresponding feature. For both models, if the postsaccadic representation lies in the green region, then pre- and postsaccadic representations will be integrated. If it falls in the purple region, then the representations will be kept separate. (C) If the representations are integrated, then the analog response will be centered on a point between the pre- and postsaccadic representations, as indicated by the green probability distribution. If the representations are segregated, then the analog response will be centered on the postsaccadic representation (as indicated by the purple distribution), as the instruction was to report the last color perceived.
Figure 2.
 
Pictorial representation of the probabilistic observer model. (A) The pre- and postsaccadic color and location representations are corrupted by Gaussian noise, as represented by the bivariate Gaussians on the location and color feature spaces. (B) Two-tailed thresholds, as represented by the dotted lines, are set for the purpose of detecting a change. If the postsaccadic stimulus exceeds a feature threshold, then a change is detected in the corresponding feature. For both models, if the postsaccadic representation lies in the green region, then pre- and postsaccadic representations will be integrated. If it falls in the purple region, then the representations will be kept separate. (C) If the representations are integrated, then the analog response will be centered on a point between the pre- and postsaccadic representations, as indicated by the green probability distribution. If the representations are segregated, then the analog response will be centered on the postsaccadic representation (as indicated by the purple distribution), as the instruction was to report the last color perceived.
Figure 3.
 
Results of Experiment 1. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and location change (B). Curves show predictions of the best-fitting model, which was feature based. (C) Distribution of reported color relative to presaccadic color (0°) and postsaccadic color (dotted vertical line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and location change (columns: magnitude indicated at top). Error bars and error patches represent ± 1 SE.
Figure 3.
 
Results of Experiment 1. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and location change (B). Curves show predictions of the best-fitting model, which was feature based. (C) Distribution of reported color relative to presaccadic color (0°) and postsaccadic color (dotted vertical line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and location change (columns: magnitude indicated at top). Error bars and error patches represent ± 1 SE.
Figure 4.
 
Results of Experiment 2. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic color (dotted line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 4.
 
Results of Experiment 2. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic color (dotted line in each panel), plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 5.
 
Results from the color-report trials in Experiment 3. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 5.
 
Results from the color-report trials in Experiment 3. (A, B) Mean color bias (blue symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported color relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of color change (rows: magnitude indicated in top-left of each panel) and orientation change (columns: magnitude indicated at top).
Figure 6.
 
Results from the orientation-report trials in Experiment 3. (A, B) Mean orientation bias (cyan symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported orientation relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of orientation change (rows: magnitude indicated in top-left of each panel) and color change (columns: magnitude indicated at top). Model fits for the 90° orientation change condition are not available, as the optimal observer response for such a large change is not uniquely defined.
Figure 6.
 
Results from the orientation-report trials in Experiment 3. (A, B) Mean orientation bias (cyan symbols, left y-axis) and mean frequency of detecting a change (red symbols, right y-axis) are plotted as a function of the magnitude of color change (A) and orientation change (B). (C) Distribution of reported orientation relative to presaccadic (0°) and postsaccadic (dotted line in each panel) values, plotted separately for trials where a change was detected (purple) and not detected (green). Areas under each distribution reflect the frequency of detection. Each panel corresponds to a different pairing of orientation change (rows: magnitude indicated in top-left of each panel) and color change (columns: magnitude indicated at top). Model fits for the 90° orientation change condition are not available, as the optimal observer response for such a large change is not uniquely defined.
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×