September 2024
Volume 24, Issue 10
Open Access
Vision Sciences Society Annual Meeting Abstract  |   September 2024
Comparing Human Face and Body Recognition at Various Distance and Rotation Viewing Conditions.
Author Affiliations & Notes
  • Matthew Groth
    Massachusettes Institute of Technology
  • Michal Fux
    Massachusettes Institute of Technology
  • Hojin Jang
    Massachusettes Institute of Technology
  • Suayb S. Arslan
    Massachusettes Institute of Technology
  • Walt Dixon
    GE
  • Joydeep Munshi
    GE
  • Pawan Sinha
    Massachusettes Institute of Technology
  • Footnotes
    Acknowledgements  This research is supported by ODNI, IARPA. The views are of the authors and shouldn't be interpreted as representing official policies of ODNI, IARPA, or the U.S. Gov., which is authorized to reproduce & distribute reprints for governmental purposes notwithstanding any copyright annotation therein.
Journal of Vision September 2024, Vol.24, 1365. doi:https://doi.org/10.1167/jov.24.10.1365
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Matthew Groth, Michal Fux, Hojin Jang, Suayb S. Arslan, Walt Dixon, Joydeep Munshi, Pawan Sinha; Comparing Human Face and Body Recognition at Various Distance and Rotation Viewing Conditions.. Journal of Vision 2024;24(10):1365. https://doi.org/10.1167/jov.24.10.1365.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Deep networks trained on large face datasets have achieved impressive performance on recognition tasks. However, as we reported in last year’s meeting (Fux et al., 2023), humans still outperform DNNs when viewing conditions are challenging as with large distances, non-frontal regard, and atmospheric turbulence. In the current study, we investigate recognition performance of humans and deep networks with images of whole bodies. This task is akin to the ‘person re-identification’ challenge of great interest to the machine vision community. We worked with a large database of images acquired at a variety of distances and from multiple yaw/pitch angles. We ran an online behavioral study in which participants were asked to rank whole-body images of people from most to least likely to be the same identity as the person in the query image. The query image depicts individuals in three conditions: Whole body, head occluded, and body occluded. Distance to the camera ranged from 10m to 500m. The results enable an analysis of the relative contribution of the head and body to recognition as a function of viewing distance. A comparative analysis of humans against a whole-body trained DNN establishes a clear superiority of human performance across all tested distances. Moreover, this comparison reveals differences in the relative importance of head and body regions, with humans deriving significant identity information from the head region, unlike the DNN. These preliminary results indicate potential divergent strategies employed by humans and DNNs, offering insights into distant person identification and implications for the design of future machine models.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×