Purchase this article with an account.
Garry Kong, Daryl Fougnie; Visual Search Within Working Memory. Journal of Vision 2018;18(10):188. doi: https://doi.org/10.1167/18.10.188.
Download citation file:
© ARVO (1962-2015); The Authors (2016-present)
Attention and working memory are two key pillars of cognition. Despite much research, there are important aspects about the relationship between the two constructs that are not well understood. Here we explore the similarity in the mechanisms that select and update working memory to those that guide attention during perception, such as in visual search. We use a novel memory search task where participants memorize a display of objects on a 4x4 grid. During memory maintenance, participants are instructed to update the spatial positions of a subset of objects. This updating process is self-paced—its speed reflecting the accessibility of the to-be-updated subset. Using this task, we explored whether landmark findings in visual search would hold true for memory search. In Experiment 1 (n=12), we found a search asymmetry—it was easier to access memory representations defined by a feature than defined by the lack of a feature, t(11) = 4.13, p = .002. In Experiment 2 (n=12) we found target-distractor similarity effects—updating a single target was easier when the distractors were farther away in feature space, t(11) = 3.29, p = .007. In Experiment 3 (n=12) we found a feature versus conjunction benefit—access times were much faster for instructions to move objects defined by only one feature (e.g., all triangles) as opposed to a conjunction of features (e.g., all red triangles), t(11) = 2.66, p = .022. Taken together, our results suggest a common coding and selection scheme for working memory and perceptual representations.
Meeting abstract presented at VSS 2018
This PDF is available to Subscribers Only