August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
In Silico Approach for Understanding the Associations Between Vision and Emotions Underlying the Uncanny Valley Effect
Author Affiliations & Notes
  • Takuya Igaue
    The University of Tokyo
    National Institute of Advanced Industrial Science and Technology
  • Ryusuke Hayashi
    National Institute of Advanced Industrial Science and Technology
  • Footnotes
    Acknowledgements  This research is supported by JST Moonshot Research & Development Program JPMJMS2012-2-1-2 and the commissioned research program (No. 22301) by NICT, Japan.
Journal of Vision August 2023, Vol.23, 4717. doi:https://doi.org/10.1167/jov.23.9.4717
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Takuya Igaue, Ryusuke Hayashi; In Silico Approach for Understanding the Associations Between Vision and Emotions Underlying the Uncanny Valley Effect. Journal of Vision 2023;23(9):4717. https://doi.org/10.1167/jov.23.9.4717.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

[Introduction] People typically feel uneasy when observing robots and computer graphic characters that resemble humans but are not perfectly human-like, an effect known as the “uncanny valley.” Several empirical studies examining affective responses to morphed images between human and non-human categories suggested that visual cues from two different categories elicit conflicting inferences about the entity, leading to feelings of eeriness. However, the detailed relationship between visual representations and emotional responses remains unclear. Artificial neural networks (ANNs), which can predict the relevant text description given an image, are promising models for providing insight into the processes underlying human cognition by exploring vast instances of human affective responses to visual concepts. In this study, we investigated how an ANN evaluates the matching of the morphed images to affective words used to describe uncanny valley effects in previous studies. [Methods] We created stimulus images by morphing between human faces and non-human objects at five morph levels and assessed the score of the images’ matching to words using CLIP (Contrastive Language–Image Pre-training), a state-of-the-art ANN that estimates semantic matching between an image and a caption. Ho and MacDorman proposed the indices of humanness, eeriness, and attractiveness using a semantic differential scale for evaluating the affective responses of human observers in studies of the uncanny valley. We calculated CLIP scores for the adjectives comprising the three indices and examined how these indices changed across morph levels. [Results and Conclusions] The eeriness index was highest at the midpoint of the morph continuum, where visual cue conflicts were maximal. This result indicates that CLIP associates visual cue conflicts in images with eerie impressions through training on an enormous amount of data covering our daily visual experiences. The current study explored how visual representations are related to human observers’ sentiment using ANN.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×