Purchase this article with an account.
Cynthia Roy, Sylvain Roy, Daniel Fiset, Zakia Hammal, Caroline Blais, Pierre Rainville, Frédéric Gosselin; Recognizing static and dynamic facial expressions of pain : Gaze-tracking and Bubbles experiments. Journal of Vision 2008;8(6):710. doi: https://doi.org/10.1167/8.6.710.
Download citation file:
© ARVO (1962-2015); The Authors (2016-present)
Facial expression is considered to be the most reliable source of information when judging on the pain intensity experienced by another (Poole & Craig 1992). Nonetheless, observers in this situation show a systematic under-estimation bias (Harrison, 1993; Kappesser & Williams, 2002). In the medical domain, this bias results in under-treatment, which leads to an insufficient pain relief for suffering patients. Despite the important impact of pain identification on patient well-being, the visual processes involved in the recognition of the facial expression of pain remain unknown. In this study, we used gaze-tracking and Bubbles (Gosselin & Schyns, 2001) to investigate the visual information used for the recognition of static and dynamic facial expressions of pain. Observers were required to categorize 80 dynamic or static facial expressions (the 6 basic emotions, pain and neutral) from the STOIC database (Roy et al., 2007). In the gaze-tracking experiment, twenty observers saw 6 times each of the 80 static and dynamic emotionnal faces. Gaze position was recorded while the stimuli were presented; heat maps were computed. The results for pain expressions will be discussed. In the Bubbles experiments, 5,000 sparse versions of these static and dynamic stimuli were created by sampling facial information at random spatial locations at five one-octave non-overlapping spatial frequency bands for the static stimuli, as well as in space-time for the dynamic stimuli (see Vinette, Gosselin & Schyns, 2004). Online calibration of sampling density ensured 75% overall accuracy. We performed mulitple linear regressions on sample space or space-time locations and on accuracy to reveal the information effectively used to recognize pain. Preliminary findings with static stimuli reveal that optimal information for the recognition of pain partly overlap with sadness and disgust ones. Preliminary results with dynamic stimuli indicate that motion contributes to the decoding of facial expressions of pain.
This PDF is available to Subscribers Only