September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Find the Orange: How rich and accurate is the visual percept that guides action?
Author Affiliations & Notes
  • Aryan Zoroufi
    Massachusetts Institute of Technology
  • Nishad Gothoskar
    Massachusetts Institute of Technology
  • Josh Tenenbaum
    Massachusetts Institute of Technology
  • Nancy Kanwisher
    Massachusetts Institute of Technology
  • Footnotes
    Acknowledgements  Thanks to NSF NCS Project 6945933 for funding the study
Journal of Vision September 2024, Vol.24, 1120. doi:https://doi.org/10.1167/jov.24.10.1120
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Aryan Zoroufi, Nishad Gothoskar, Josh Tenenbaum, Nancy Kanwisher; Find the Orange: How rich and accurate is the visual percept that guides action?. Journal of Vision 2024;24(10):1120. https://doi.org/10.1167/jov.24.10.1120.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Are the visual representations that guide our online interactions with the world sparse and impoverished, or richly detailed, including the 3D shape of objects and their spatial and physical relationship to each other? We address this question using a naturalistic virtual reality environment in which participants (N=10) are asked to find an occluded target object (orange) on a tabletop environment as quickly as possible, by pressing a button to indicate which occluder should be moved first, or by reaching directly for the occluder. The occluders differ in width, orientation, 3D shape, and the presence of holes (which enable the participant to see through parts of the occluder). As instructed, people launch their actions quickly, within 500 ms after stimulus onset. We find that the decisions about which of two occluder objects to move first are guided by fairly accurate estimates of the area behind occluders that take into account 1) the 3D structure of the scene (not just the 2D pixel area of the occluders) and 2) the relative size of the hidden object. We also find that the decisions are similarly fast and accurate whether participants explicitly report or move the objects. Overall, these results suggest that a fast and accurate 3D representation of both visible and occluded parts of a scene are rapidly available to guide rational search in naturalistic environments. Future work using this framework will investigate whether the information that is rapidly available during naturalistic viewing includes not only geometric but also physical properties of the scene.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×