August 2023
Volume 23, Issue 9
Open Access
Vision Sciences Society Annual Meeting Abstract  |   August 2023
Artifact magnification on deepfake videos increases human detection and subjective confidence
Author Affiliations & Notes
  • Emilie Josephs
    CSAIL, MIT
  • Camilo Fosco
    CSAIL, MIT
  • Aude Oliva
    CSAIL, MIT
  • Footnotes
    Acknowledgements  This work was partially funded by an Ignite grant from SystemsThatLearn@CSAIL, and made possible by generous funding from Eric Yuan, CEO and founder of Zoom Video Communications
Journal of Vision August 2023, Vol.23, 5327. doi:https://doi.org/10.1167/jov.23.9.5327
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Emilie Josephs, Camilo Fosco, Aude Oliva; Artifact magnification on deepfake videos increases human detection and subjective confidence. Journal of Vision 2023;23(9):5327. https://doi.org/10.1167/jov.23.9.5327.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Deefake videos, where faces are modified using deep learning, are becoming easier to make and more convincing, creating online security risks. Machine learning models are being developed to detect deepfakes, but human users often disregard model predictions (Groh et al, 2021). Here, we explore a novel visual indicator called Deepfake Caricatures, in which video artifacts are magnified, and show that it is both effective and subjectively convincing. In Experiment 1, participants viewed silent videos of people, and indicated whether each was real or fake. Deepfakes were either non-signaled, or signaled with the Caricature technique (between-subjects, N=180). Detectability was measured across conditions known to affect visual search and signal detection performance: a Baseline condition (E1a, 50% prevalence, no time limit), Low Prevalence (E1b, 20% prevalence), Time Pressure (E1c, 2 second time limit), and Noisy Visual Input (E1d, video compression noise). Across all conditions, sensitivity was much higher for Caricatures than non-signaled deepfakes (E1a: t = 30.08, p<0.001; E1b: t = 32.28, p<0.001; E1c: t = 39.62, p<0.001; E1d: t=30.58, p<0.001 ; unpaired t-tests), showing that the Caricatures manipulation is effective. Additionally, we compared the detectability of non-signaled deepfakes across experiments, and found that all conditions significantly reduced deepfake detection relative to Baseline, suggesting that the present literature overestimates deepfake detectability in the wild. In Experiment 2, participants (N = 298) viewed videos and rated their confidence that they were real or fake using a slider interface, then viewed a model’s prediction displayed either as text or in Caricatures form, and finally were allowed to update their responses. Participants updated their responses more often following a Caricature than a text-based indicator (t=4.30, p<0.001), showing that Caricatures is a more convincing visual indicator than text.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×