Abstract
An important task of perceptual processing is to parse incoming information from the external world into distinct units and to subsequently keep track of those units over time as the same, persisting internal representations. Within the realm of visual perception, this concept of maintaining persisting object representations has been theorized as being mediated by “object files” - episodic representations that store (and update) information about objects' properties and track objects over time and motion via spatiotemporal information (e.g., Kahneman et al., 1992). Although object files are typically conceptualized as visual representations, here, we demonstrate that object-file correspondence can be computed across sensory modalities. We employed a novel version of the object-reviewing paradigm: Line-drawn pictures (e.g., a phone and a dog) were briefly presented within two distinct objects in a preview display. Once the pictures disappeared, the objects moved (to decouple objecthood from location) and then a sound (e.g., a dog bark) occurred. The sound was localized to the left or right of the display, corresponding to the end locations of the two objects. Participants were instructed to indicate whether the sound matched either preview picture or whether it was novel (e.g., a dog bark would “match” if either preview picture was a dog). Participants were significantly faster to respond when the sound occurred with the object originally containing the associated picture compared to when the sound occurred with the other object. This significant response time benefit provides the first evidence for visual and auditory information working in tandem to underlie object-file correspondence. An object file can be initially formed with visual input and later accessed with corresponding auditory information. Object files may thus operate at a highly abstract level of perceptual processing that is not tied to specific modalities.
This research was supported by a R03 MH080849-01 grant to Stephen Mitroff.