The idea behind the following experiments is straightforward. First, we train subjects to respond either to facial identity or facial expression. We then measure the strength of identity adaptation and expression adaptation upon test morph sequences that vary simultaneously in both of these attributes. If the direction of the aftereffect is determined by that of the test sequences, then we should measure distortions of both identity and expression in both adaptation conditions. If, on the other hand, the direction of the aftereffect is independent of the direction of the test sequence, then we should see only identity aftereffects under identity adaptation and expression aftereffects under expression adaptation.
To build the stimuli employed in our experiment we used two actors
A and
B, each displaying two facial expressions, happy and sad (see
Figure 1); the collection of these images is described elsewhere (Benton et al.,
2007). We morph between these images (Tiddeman, Burt, & Perrett,
2001) to produce unidimensional morphs (the ‘edges’ in
Figure 1) and multidimensional morphs (the ‘diagonals’). We have used the term unidimensional to refer to a morph sequence in which only one attribute (either expression or identity) changes, and the term multidimensional to refer to a morph sequence where both expression and identity change.
To describe the stimuli used in our experiment we use a capital letter to indicate identity (
A or
B) followed by a lowercase letter to indicate expression (
h or
s). We use the uppercase letter ‘
M’ to indicate the identity midpoint and, similarly, the lowercase letter ‘
m’ to indicate the expression midpoint (see
Figure 1). Our
test sequences (used to measure the strength of adaptation) are
Ah→
Bs and
As→
Bh. For adaptation stimuli we used the midpoints of the unidimensional morphs:
Mh, Ms, Am, and
Bm. We split these into two conditions, an identity adaptation condition (with adaptation stimuli
Am and
Bm) and an expression adaptation condition (with adaptors
Mh and
Ms). We measure the strength of adaptation as the difference in balance point between the two adaptors within each of these conditions for each test sequence (Benton et al.,
2006,
2007). Note that each test sequence varies identity, shifting from
A to
B. However in
Sequence 1 this identity shift is accompanied by a change in facial expression from
happy to
sad, in
Sequence 2 it is accompanied by a shift from
sad to
happy.
The balance point is defined as the point along a morph sequence, which is equally likely to be judged as displaying either of the two target attributes. We measured balance points using an adaptive method of constants procedure (Watt & Andrews,
1981) in which subjects view images from our test sequences and classify these as either identity
A or identity
B or as
happy or
sad (dependent upon response condition, see below). For each test sequence we use responses from 64 image presentations (or
trials) to estimate the balance point by fitting a cumulative Gaussian to the resultant data (Wichmann & Hill,
2001a). We refer to the group of 64 trials used to measure a balance point as a
run. We term a single run presented individually, or an interleaved group of runs, as a
session.
Our eight subjects were evenly divided into two groups, an identity response group and an expression response group. Those in the identity group were instructed only to respond to identity, those in the expression group were instructed only to respond to expression. Rather than choosing the 50% midpoints of the unidimensional morphs as our adaptation stimuli we estimated their perceptual midpoints psychophysically for each subject and used these as their adaptation stimuli. A fixed choice across subjects might have added an unwanted identity or expression bias into, respectively, the expression or identity adaptation.
During the experiments, participants were presented with three types of image. Adaptation images, test images (to which subjects respond), and comparison images. The latter were essentially for training purposes and displayed the categories to which the subjects responded. The average interocular distance for the adaptation and test images was 2° (91 pixels). In order to prevent retinotopic adaptation, adapt and test images rotated around a fixation point once every 5 seconds in a circular trajectory of diameter 1°. Trajectory start position and direction were randomly determined for each stimulus. Except during presentation of the comparison images (see below) the fixation spot was always present.
There was a 500 msec gap between all stimuli presented within a trial. Test stimuli were presented for 1000 msec, comparison stimuli for 2000 msec. Adaptation consisted of an initial 30 second adaptation stimulus followed by a test stimulus (to which subjects respond). Subsequent adaptation trials within a session consisted of adaptation top-up (5 seconds) followed by the test stimulus. Comparison stimuli were presented prior to each test stimulus in the non-adaptation sessions. The comparison stimuli were composed of the start and end images of the morph sequence to which the test stimulus for that trial belonged. The end image was presented directly above the start image, with the ensemble presented in the middle of the screen. In those tasks where comparison images were used, subjects indicated whether intermediate morph images most closely resembled the start or end of the morph sequences. The purpose of the comparison images was therefore to cue the subjects to the appropriate response. The size of the component images within each ensemble was reduced to 75% of that of the adaptation and test images.
From the point of view of a participant, the procedure would start with gathering each of the unadapted balance points from the unidimensional morphs (to be used subsequently as adaptation stimuli). Depending on the sequence ( Ah→ Bh, As→ Bs, Ah→ As, or Bh→ Bs), subjects would respond as identity A or identity B or happy or sad by pressing the up and down arrows on a keyboard. These four runs were completed individually (i.e., not interleaved) and in random order, for each subject. Participants were then subjected to the adaptation part of the experiment. We measured the balance points on the two test sequences in response to each of the 4 predetermined adaptation stimuli. For each adaptation sequence the two test sequences were randomly interleaved so that we simultaneously measured the effects of adaptation on both. This was done in order to prevent subjects associating any particular identity with any particular facial expression. For each subject there were therefore 4 adaptation sessions; these were presented in random order.
Directly before each adaptation session, subjects carried out a training session. For subjects in the identity response condition the training session consisted of finding balance points on interleaved runs for measuring balance points for the Ah→ Bh and As→ Bs sequences. For subjects in the expression response condition the training session consisted of finding balance points on the interleaved Ah→ As and Bh→ Bs runs. From the point of view of the participant, the training sessions consisted of test images that varied in both identity and expression. The comparison images presented prior to the test images cued subjects to respond to either identity or expression as appropriate.
Stimuli were presented on a Lacie Electron Blue IV 22” monitor. Spatial resolution was 1024 × 768 pixels (23° by 17°); temporal resolution was 75 Hz. The edges of the face stimuli were blurred to display mean luminance (54 cd/m 2). The experiments took place in a darkened room where the monitor was the only strong source of illumination. Subjects sat comfortably in an armchair at a viewing distance of 100 cm from the monitor with the keyboard (used for responses) on their lap. Identity response subjects pressed the up arrow to indicate identity A, the bottom arrow to indicate identity B. Expression response subjects pressed the up arrow to indicate happy and the down arrow to indicate sad. All subjects had normal or corrected-to-normal vision. Stimulus presentation was controlled by a PC; the images were rendered using the Cogent Graphics Matlab extension developed by John Romaya at the LON at the Wellcome Department of Imaging Neuroscience.