September 2019
Volume 19, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2019
Visual context in emotion recognition is more powerful, prevalent and efficient than we thought
Author Affiliations & Notes
  • Zhimin Chen
    Department of Psychology, University of California Berkeley, Berkeley, California 94720, USA
  • David Whitney
    Department of Psychology, University of California Berkeley, Berkeley, California 94720, USA
    Vision Science Program, University of California, Berkeley, Berkeley, CA 94720, USA
    Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA 94720, USA
Journal of Vision September 2019, Vol.19, 182b. doi:https://doi.org/10.1167/19.10.182b
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Zhimin Chen, David Whitney; Visual context in emotion recognition is more powerful, prevalent and efficient than we thought. Journal of Vision 2019;19(10):182b. https://doi.org/10.1167/19.10.182b.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Emotion recognition is a core perceptual ability critical for social functioning. It is widely assumed that recognizing facial expressions predominantly determines perceived emotion, and contextual influences only coarsely modulate or disambiguate interpreted faces. Using a novel method, ‘Inferential Emotion Tracking’, we isolated and quantified the contribution of visual context versus face and body information in dynamic emotion recognition. Independent groups of observers tracked the valence and arousal of a chosen character in a silent Hollywood movie clip in 3 different conditions in real-time: 1) fully-informed, where everything in the video was visible. 2) character-only, where the face and body of the target character were visible but the visual context (everything else except the target) was masked and invisible. 3) context-only, where the face and body of the target were masked and invisible but the context was visible. Using regression models, we found that context-only affect ratings contributed a significant amount of unique variance (14.6%) to fully-informed ratings, comparable to that by character-only ratings (20.5%; p > 0.05). Importantly, we replicated the essential contribution of context with a different set of movie clips in which no other character is present (context: 14.4%, character: 20.5%), and another different set of non-Hollywood videos (e.g. home videos, documentaries; context: 23.2%, character: 17.8%). A meta-analysis over all 34 videos shows that context contributed as much as the face and body (p > 0.05). Strikingly, temporal cross-correlation analyses (with milliseconds precision) revealed that emotion information from context is available just as fast as emotion from faces. Furthermore, when observers tracked the emotion of the target using discrete emotion categories, the context still contributed significantly (12%), although less than the character did (35%; p < 0.01). Our results demonstrate that emotion recognition is, at its heart, an issue of context as much as it is about faces.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×