September 2018
Volume 18, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2018
Visual Search Within Working Memory
Author Affiliations
  • Garry Kong
    Science Division, New York University Abu Dhabi
  • Daryl Fougnie
    Science Division, New York University Abu Dhabi
Journal of Vision September 2018, Vol.18, 188. doi:
  • Views
  • Share
  • Tools
    • Alerts
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Garry Kong, Daryl Fougnie; Visual Search Within Working Memory. Journal of Vision 2018;18(10):188.

      Download citation file:

      © ARVO (1962-2015); The Authors (2016-present)

  • Supplements

Attention and working memory are two key pillars of cognition. Despite much research, there are important aspects about the relationship between the two constructs that are not well understood. Here we explore the similarity in the mechanisms that select and update working memory to those that guide attention during perception, such as in visual search. We use a novel memory search task where participants memorize a display of objects on a 4x4 grid. During memory maintenance, participants are instructed to update the spatial positions of a subset of objects. This updating process is self-paced—its speed reflecting the accessibility of the to-be-updated subset. Using this task, we explored whether landmark findings in visual search would hold true for memory search. In Experiment 1 (n=12), we found a search asymmetry—it was easier to access memory representations defined by a feature than defined by the lack of a feature, t(11) = 4.13, p = .002. In Experiment 2 (n=12) we found target-distractor similarity effects—updating a single target was easier when the distractors were farther away in feature space, t(11) = 3.29, p = .007. In Experiment 3 (n=12) we found a feature versus conjunction benefit—access times were much faster for instructions to move objects defined by only one feature (e.g., all triangles) as opposed to a conjunction of features (e.g., all red triangles), t(11) = 2.66, p = .022. Taken together, our results suggest a common coding and selection scheme for working memory and perceptual representations.

Meeting abstract presented at VSS 2018


This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.