September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
From Perception to Algorithm: Quantifying Facial Distinctiveness with a Deep Convolutional Neural Network
Author Affiliations & Notes
  • Isabelle Boutet
    University of Ottawa
  • Artem Pilzak
    University of Ottawa
  • Maxime Ouimet
    University of Ottawa
  • Alice J. O'Toole
    University of Texas at Dallas
  • Footnotes
    Acknowledgements  This research is funded by a Natural Sciences and Engineering Research Council of Canada Discovery grant (2022-03998) to IB.
Journal of Vision September 2024, Vol.24, 544. doi:https://doi.org/10.1167/jov.24.10.544
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Isabelle Boutet, Artem Pilzak, Maxime Ouimet, Alice J. O'Toole; From Perception to Algorithm: Quantifying Facial Distinctiveness with a Deep Convolutional Neural Network. Journal of Vision 2024;24(10):544. https://doi.org/10.1167/jov.24.10.544.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Background: Face distinctiveness is a pivotal concept in face recognition, due to long-standing findings indicating that the rated typicality of a face is inversely related to its recognizability (e.g., Light et al., 1979). In traditional face space models (Valentine et al., 2016), distinctive faces are perceptually distant from the average or prototype face. Objectively measurable face spaces are available from deep convolutional neural network (DCNN), which are highly accurate at face recognition. Here we test whether DCNN models capture human-rated perceptions of face typicality, as well as other human-ratings (memorability, attractiveness, etc.). Method: We utilized FaceNet (Schroff et al., 2015, 2018), a pre-trained DCNN, to derive a distinctiveness index based on the average distance of a face to other faces in the DCNN space. First, we quantified distinctiveness for 418 male and 631 female faces using FaceNet's 512-dimensional embedding space and the cosine as a measure of distances. Second, we computed correlations between ach face’s distinctiveness and various human ratings of the same faces (Bainbridge et al., 2013). Results: Male faces classified as more distinctive are rated as less common (r = 0.27) and atypical (r = 0.21). Female faces classified as more distinctive are rated as less common (r = -0.35), more attractive (r = 0.37), more egotistical (r = 0.22), and more confident (r = 0.25) (all p < 0.001). Conclusion: DCNN provides a model of memory-related traits that relate to face distinctiveness across both genders. For female faces, distinctiveness also strongly correlates with personality and social attributes. This underscores the utility of DCNN models for understanding distinctiveness and face recognition. Current efforts focus on extending these findings using an alternate face distinctiveness model and another stimulus set to bolster the generalizability of our results.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×