December 2022
Volume 22, Issue 14
Open Access
Vision Sciences Society Annual Meeting Abstract  |   December 2022
The Uncanniness of Face Swaps
Author Affiliations
  • Ethan Wilson
    University of Florida
  • Aidan Persaud
    University of Florida
  • Nicholas Esposito
    University of Florida
  • Sophie Joerg
    Clemson University
  • Rohit Patra
    University of Florida
  • Frederick Shic
    Seattle Children’s Research Institute
  • Jenny Skytta
    Seattle Children’s Research Institute
  • Eakta Jain
    University of Florida
Journal of Vision December 2022, Vol.22, 4225. doi:https://doi.org/10.1167/jov.22.14.4225
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ethan Wilson, Aidan Persaud, Nicholas Esposito, Sophie Joerg, Rohit Patra, Frederick Shic, Jenny Skytta, Eakta Jain; The Uncanniness of Face Swaps. Journal of Vision 2022;22(14):4225. https://doi.org/10.1167/jov.22.14.4225.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Face swapping algorithms, popularly known as "deep fakes", generate synthetic faces whose movements are driven by an actor's face. To create face swaps, users construct training datasets consisting of the two faces being applied and replaced. Despite the availability of public code bases, creating a compelling, convincing face swap remains an art rather than a science because of the parameter tuning involved and the unclear consequences of parameter choices. In this paper, we investigate the effect of different dataset properties and how they influence the uncanny, eerie feeling viewers experience when watching face swaps. In one experiment, we present participants with video from the FaceForensics++ Deep Fake Detection dataset. We ask them to score the clips on bipolar adjective pairs previously designed to measure the uncanniness of computer-generated characters and faces within three categories: humanness, eeriness, and attractiveness. We find that responses to face swapped clips are significantly more negative than to unmodified clips. In another experiment, participants are presented with video stimuli of face swaps generated using deepfake models that have been trained on deficient data. These deficiencies include low resolution images, lowered numbers of images, deficient/mismatched expressions, and mismatched poses. We find that mismatches in resolution, expression, and pose and deficient expressions all induce a higher negative response compared to using an optimal training dataset. Our experiments indicate that face swapped videos are generally perceived to be more uncanny than original videos, but certain dataset properties can increase the effect, such as image resolution and quality characteristics including expression/pose match-ups between the two faces. These insights on dataset properties could be directly used by researchers and practitioners who work with face swapping to act as a guideline for higher-quality dataset construction. The presented methods additionally open up future directions for perceptual studies of face swapped videos.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×