Previous studies have considered the integration of moving audio–visual targets in stationary observers and consistently found evidence for optimal integration (
Alais & Burr, 2004a;
Meyer & Wuerger, 2001;
Soto-Faraco et al., 2004;
Wuerger et al., 2010). For example, variability in the estimate of moving object arrival times is reduced when both audio and visual cues are present, consistent with maximum likelihood principles (
Wuerger et al., 2010). Similarly, detection thresholds for motion perception are significantly improved in the presence of audio–visual stimuli together, rather than individual modalities alone (
Alais & Burr, 2004a). By contrast, few studies have considered the impact of self-motion on the process of integration, although several studies have demonstrated that self-movement can impact the perception of auditory and visual stimulus motion more generally. For example, rotation and translation of the body and/or head can impact the localization of both auditory and visual targets (
Carriot, Bryan, DiZio, & Lackner, 2011;
Cooper, Carlile, & Alais, 2008;
Lackner & DiZio, 2010;
Teramoto, Cui, Sakamoto, & Gyoba, 2014), and stationary stimuli are perceived as moving in the opposite direction to head and/or eye movements (
Freeman, 2007;
Freeman et al., 2017). This latter so-called Filehne illusion is also present in our current study, reflected in the biases obtained for audio, visual, and audio–visual conditions. Curiously, although we found visual biases similar to those previously reported for smooth eye pursuit (
Filehne, 1922;
Furman & Gur, 2012), the auditory biases were much smaller than in an earlier study from our lab that used similar auditory stimuli (
Freeman et al., 2017). The smaller biases may have arisen because here we presented the cues to be judged during a single sweep of the head movement, instead of continuously as in
Freeman et al. (2017). We note, too, that the Filehne illusion depends on basic stimulus properties such as spatial frequency (
Freeman & Banks, 1998;
Wertheim, 1987), which determine the size of the image-motion estimate to which the self-motion signal is compared. Hence, the Filehne illusions found for the visual and auditory conditions will also depend on the specific auditory and visual stimuli used. As such, it is possible that we found a larger visual versus auditory Filehne illusion in the present study due to differences in stimulus parameters, such as the standard deviation of the stimuli. Our results build upon existing literature, proposing a mechanism through which self-movement, visual motion, and auditory motion are integrated to help us perceive movement.