December 2022
Volume 22, Issue 14
Open Access
Vision Sciences Society Annual Meeting Abstract  |   December 2022
Representation of Face Shape and Surface Reflectance in Deep Convolutional Neural Networks
Author Affiliations & Notes
  • Matthew Hill
    The University of Texas at Dallas
  • Carlos Castillo
    Johns Hopkins University
  • Alice O'Toole
    The University of Texas at Dallas
  • Footnotes
    Acknowledgements  National Eye Institute grant R01EY029692-03 to AOT
Journal of Vision December 2022, Vol.22, 4369. doi:https://doi.org/10.1167/jov.22.14.4369
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Matthew Hill, Carlos Castillo, Alice O'Toole; Representation of Face Shape and Surface Reflectance in Deep Convolutional Neural Networks. Journal of Vision 2022;22(14):4369. https://doi.org/10.1167/jov.22.14.4369.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Deep Convolutional Neural Networks (DCNNs) recognize faces over image/appearance variation (e.g. viewpoint, illumination, expression) while retaining information about this variation (Hill et al., 2019). Here, we examined the extent to which DCNNs encode the 3D shape and surface reflectance properties of a face in the presence of challenging image variability. We generated synthetic face images using parametric models of 3D face shape and surface reflectance. The FLAME model uses a linear shape space to generate faces by varying 3D shape and expression parametrically (Li et al., 2017). The reflectance model was based on the Basel Face Model (Paysan et al., 2009). We generated five face shapes and five reflectance maps to create 25 unique faces by combining all possible pairs of shapes and reflectance maps. The stimulus set (n=1,125) consisted of images of these faces rendered at 9 viewpoints (yaw angle: 0°, 30°, and 60°; pitch angle: -30°, 0°, 30°) with 5 expressions. Each image was processed by a DCNN trained for face identification (ArcFace: Deng et al., 2019) to generate a vector representation. Cosine similarity was calculated between all possible pairs of image representations. Across this set of “identities”, shape and reflectance are confounded, with each identity sharing both its shape and reflectance with many other identities. We asked whether the network's reflectance code is tolerant to changes in viewpoint and expression by splitting the cosine similarity scores into distributions of same-reflectance and different-reflectance pairs. Results showed that the DCNN encodes reflectance over viewpoint and expression variation (Area Under the ROC Curve, AUC=0.66). Analogously, the distributions for same-shape and different-shape pairs yielded an AUC of 0.58, indicating that the shape properties of the face are encoded also. Therefore, DCNNs can separately encode the fundamental dimensions of shape and reflectance in the face, despite wide variations in viewpoint and expression.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×