It has been proposed that judgments about space and time rely on a generalized magnitude processing mechanism (Cai & Connell,
2015; Walsh,
2003), an idea that is in agreement with the numerous reports of interaction between space and time in perception. However, this theory is not constrained by any specific timekeeping mechanism (Bueti & Walsh,
2009), and the brain mechanisms subserving time estimation remain largely debated (Finnerty, Shadlen, Jazayeri, Nobre, & Buonomano,
2015; Merchant, Harrington, & Meck,
2013). One view is that there are modality-specific mechanisms for perceived duration of sensory events (e.g., visual and auditory; Burr & Morrone,
2006; Burr, Tozzi, & Morrone,
2007; Johnston, Arnold, & Nishida,
2006; Yuasa & Yotsumoto,
2015) that could be based, for example, on mechanisms for coding speed (Gorea & Kim,
2015; Kaneko & Murakami,
2009) or temporal frequency (Kanai et al.,
2006; Linares & Gorea,
2015). A common characteristic of these studies is that they relate perceived duration to the strength of the neural response evoked by the stimulus (Eagleman & Pariyadath,
2009; Pariyadath & Eagleman,
2007). The present results, however, demonstrate that, provided that information about viewing distance is available, the brain can correctly compare durations of similar events even when these events are placed at different viewing distances and therefore evoke very different neural responses (i.e., they result in different proximal stimuli). Although it has been shown that viewing distance (in the form of linear perspective cues) can rescale the spatial extent of neural activity as early as the primary visual cortex (He, Mo, Wang, & Fang,
2015; Murray, Boyaci, & Kersten,
2006; Ni, Murray, & Horwitz,
2014), we argue that this modulation cannot fully account for the present results: While the degree of rescaling seems limited to a fraction of the objects' sizes, presumably due to feed-forward inputs to V1 (Murray et al.,
2006), the spatial extent of the visual events that were perceived as having equal durations in the perspective (but not flat) conditions of our experiments could vary up to a factor of 5.5 (ratio between the farthest and nearest trajectory lengths and ball sizes). Our results therefore call for a more general mechanism (or brain system) for time perception that not only collects information from modality-specific brain areas to assess the duration of an event (Merchant et al.,
2013) but also weights these inputs according to ecological contextual cues, such as viewing distance.