September 2021
Volume 21, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2021
Visual Cues Reduce Spatial Uncertainty in Multi-Talker Situations
Author Affiliations & Notes
  • Ying-Zi Xiong
    University of Minnesota
  • Nam-Anh Nguyen
    University of Minnesota
  • Peggy Nelson
    University of Minnesota
  • Gordon Legge
    University of Minnesota
  • Footnotes
    Acknowledgements  Supported by Grant from the National Institutes of Health 1K99EY030145-01A1 to Y-ZX
Journal of Vision September 2021, Vol.21, 2614. doi:
  • Views
  • Share
  • Tools
    • Alerts
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ying-Zi Xiong, Nam-Anh Nguyen, Peggy Nelson, Gordon Legge; Visual Cues Reduce Spatial Uncertainty in Multi-Talker Situations. Journal of Vision 2021;21(9):2614. doi:

      Download citation file:

      © ARVO (1962-2015); The Authors (2016-present)

  • Supplements

In social situations where multiple talkers speak from different locations, tracking and segregating conversations can be challenging. We hypothesized that 1) the spatial uncertainty of the talker locations may contribute to listening difficulty; and 2) providing visual cues at the target talker location can facilitate speech recognition by reducing the spatial uncertainty. Subjects with normal hearing and vision (N = 22, 18 to 29 years) listened to simultaneously spoken sentences by three talkers with 10° or 20° separation, from different directions in the horizontal plane. In each trial, subjects attempted to repeat the sentence of a target talker indicated by a fixed starting word. The accuracy decreased from frontal (straight-ahead) to peripheral locations (left or right), and was higher with 20° separations. Word mislocation appeared to be the primary error. When the target was located at the center position among the three talkers, the mislocation error was the highest. When the target location was indicated by a brief visual pre-cue, the mislocation errors reduced and the accuracy increased. This cue benefit was only significant for the 20° separations. To model the effect of spatial uncertainty, subjects were asked to localize 200ms auditory noise (0.2-8 kHz) or visual white disks (3°) presented at random horizontal directions. Errors in localization (bias and precision) modeled by individual Gaussian functions represented spatial uncertainties for vision and audition at each azimuth. The probability of correctly locating the target talker in the multi-talker task was predicted from three auditory Gaussian functions corresponding to each talker direction. Visual cue effect was modelled by applying a visual Gaussian function to the three auditory Gaussians. This simple model provided close predictions for the multi-talker performance and visual cue benefit. Our empirical data and modeling approach revealed the important roles of both auditory and visual spatial uncertainty in multi-talker situations.


This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.