September 2015
Volume 15, Issue 12
Free
Vision Sciences Society Annual Meeting Abstract  |   September 2015
Spatial integration and nonlinear transduction of emotional expression
Author Affiliations
  • Katie Gray
    Department of Psychology, University of Reading, UK
  • Tessa Flack
    Department of Psychology, University of York, UK
  • Daniel Baker
    Department of Psychology, University of York, UK
Journal of Vision September 2015, Vol.15, 1373. doi:https://doi.org/10.1167/15.12.1373
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Katie Gray, Tessa Flack, Daniel Baker; Spatial integration and nonlinear transduction of emotional expression. Journal of Vision 2015;15(12):1373. https://doi.org/10.1167/15.12.1373.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Basic visual features such as contrast are processed in a highly nonlinear fashion, resulting in ‘dipper’ shaped functions in discrimination experiments. Previous work has applied a similar paradigm to investigate the representation of higher level properties such as facial identity (Dakin & Omigie, 2009, Vision Res, 49, 2285-2296). Here we ask whether emotional expressions are processed nonlinearly by measuring discrimination thresholds for six emotions (angry, happy, sad, fear, disgust, surprise) morphed along a continuum relative to neutral. Using a 2IFC paradigm, we estimated discrimination thresholds relative to six ‘pedestal’ morph levels between 0% and 75%. The participants’ (N=5) task was to indicate which of two faces (pedestal, or pedestal plus increment) conveyed the strongest expression. We found evidence of facilitation at low morph levels (~15%) and masking at higher levels (>60%), indicating the existence of a nonlinearity in the neural representation of expression, comparable to that reported for lower level visual features. We then asked whether facial features are integrated across the face before or after this nonlinearity by keeping the expression in one half of the face (top or bottom) fixed at neutral, and applying the pedestal and increment expressions to the other half. Sensitivity decreased by around a factor of two along the entire dipper function, relative to the whole-face condition, suggesting that facial expressions are integrated before nonlinear transduction. Finally we assessed the amount of interference between the two halves of the face, by fixing the expression in one half at a given level (the ‘mask’), and applying the target increment to the other half. This produced a strong masking effect, such that target expressions needed to approach the level of the mask to be detected. This is evidence for competition between the neural representations of different facial features.

Meeting abstract presented at VSS 2015

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×