In order to make sense of our visual environment, we constantly move our eyes. This allows us to observe fixed and moving objects with high foveal resolution and to inspect regions of interest while ignoring many others (
Gegenfurtner, 2016). To explain this selection process, models of attentional guidance have tried to predict gaze behavior based on, for example, task relevance (
Fecteau & Munoz, 2006); low-level image features, such as local contrast for color, intensity, and orientation (
Harel, Koch, & Perona, 2006;
Itti & Koch, 2000); and high-level, semantic features, such as faces or text (
Xu, Jiang, Wang, Kankanhalli, & Zhao, 2014). What these models all have in common is that they try to predict a typical observer and treat individual differences as noise.
More recent findings, however, show stable, trait-like differences in eye movements; for example, oculomotor measures such as mean saccade amplitude, smooth pursuit duration, or mean fixation duration vary reliably among people (
Bargary, Bosten, Goodbourn, Lawrance-Owen, Hogg, & Mollon, 2017;
Castelhano & Henderson, 2008;
Henderson & Luke, 2014). Moreover, recent twin studies have provided evidence for a strong genetic component in individual gaze (
Constantino et al., 2017;
Kennedy, D'Onofrio, Quinn, Bölte, Lichtenstein, & Falck-Ytter, 2017).
Most relevant to the current study, observers freely viewing hundreds of complex scenes showed large individual differences in the number of fixated objects and in fixation tendencies toward objects from six semantic categories (
de Haas, Iakovidis, Schwarzkopf, & Gegenfurtner, 2019). These differences were consistent across images and time and extended to first fixations after image onset, suggesting a bottom–up component.
Such semantic salience biases may be useful in the study of neurobiological mechanisms of attentional gaze control (
de Haas et al., 2019) and form a crucial baseline for evaluating the diagnostic potential of gaze behavior for neurodevelomental and clinical conditions. In fact, research investigating eye movements in autism spectrum disorder (ASD) has demonstrated reduced social visual engagement in infants and adults with ASD compared to healthy controls (
Constantino et al., 2017;
Jones & Klin, 2013;
Wang, Jiang, Duchesne, Laugeson, Kennedy, Adolphs, & Zhao, 2015). Others have found evidence for abnormal eye movements in patients with major depression (
Armstrong & Olatunji, 2012). Further, patients with schizophrenia show a reduced number and spatial dispersion of fixations (
Benson, Beedie, Shephard, Giegling, Rujescu, & St. Clair, 2012). However, the free viewing paradigm in
de Haas et al. (2019) used the full stimulus set of the Object and Semantic Images and Eye-tracking (OSIE) dataset, comprised of 700 images (
Xu et al., 2014). For practical purposes, it would be desirable to estimate individual gaze biases with a more economical test.
The present study had two main objectives. The first was to replicate the findings of stable individual gaze differences along semantic dimensions found in
de Haas et al. (2019). Therefore, we tested whether the proportion of cumulative fixation times for objects of the categories
Text and
Faces, objects with implied
Motion, objects with a characteristic
Taste, or
Touched objects vary reliably between observers. Further, we tested whether the proportion of first fixations varies reliably for objects of the categories
Text,
Faces, and
Touched. We chose and preregistered these object dimensions, because they yielded large and consistent individual gaze differences (
r > 0.6) in the original study by
de Haas et al. (2019). Note, we combined neutral and emotional
Faces, given the high covariance in fixation tendencies toward both. Moreover, we tried to replicate the findings of stable individual differences in visual exploration, as indicated by the number of objects an observer fixated. Second, we tested whether a smaller stimulus set can reliably estimate these differences. For this purpose, we selected subsets of OSIE images (OSIE
40, OSIE
100, and OSIE
200), which a search algorithm predicted would yield high congruence with individual differences observed for the full set.
Two eye-tracking sessions were completed by 103 participants on separate days; one day they free-viewed the full OSIE set (700 images), and another day they free-viewed the smaller subsets. We then computed the correlation among individual fixation tendencies between the full and smaller sets to determine the minimum set size required for reliable estimates. Additionally, we repeated these analyses for gaze data truncated to the first 1 and 2 seconds of each trial (from a total of 3 seconds). Results showed that the most prominent individual fixation biases could be estimated reliably from just 40 images shown for 2 seconds each.