Feature-based attention is often implicated as the mechanism by which we selectively attend to a limited set of objects in a cluttered environment and effectively ignore, or filter, distracting items from our awareness. Feature-based attention acts upon diagnostic elements in complex scenes to promote detection and recognition, often without explicit knowledge of the observer (i.e., Driver et al.,
1999; Gosselyn & Schyns,
2001; New, Cosmides, & Tooby,
2007).
A key example of this is action recognition from point–light animations, in which selective attention binds the unique trajectories of body parts into a perceptually coherent whole (Johansson,
1973; Wittinghofer, de Lussanet, & Lappe,
2012). Although seemingly effortless, integrating the local elements of point–light walkers is attentively demanding. Visual searches for point–light walkers among distractors proceed serially, as would be expected from a feature-based conjunction search (Cavanagh, Labianca, & Thornton,
2001; Mayer, Vuong, & Thornton,
2015), and discriminating point–light actions suffers in dual-task paradigms that make demands on the capacity limited attentional system (Chandrasekaran, Turner, Bulthoff, & Thornton,
2010). Moreover, point–light biological motion sequences constructed from rapidly changing local elements that defeat local motion signals are readily recognized (Beintema & Lappe,
2002; Tyler & Grossman,
2011), while those constructed from complex local elements that discourage binding the local elements are not readily recognized (Wittinghofer, de Lussanet, & Lappe,
2010).
While the importance of attention to the recognition of biological motion is clear from these studies, what remains uncertain is the nature of the features onto which visual attention is directed. Whereas some computational models argue that biological motion perception proceeds via a skeletal template matching algorithm (Beintema & Lappe,
2002; Thurman & Lu,
2013), most studies implicate characteristic dynamic body movements (Hiris,
2007; Giese & Poggio,
2003; Saunders, Suchan, & Troje,
2009; Thirkettle, Benton, & Scott-Samuel,
2009; Thurman, Giese, & Grossman,
2010; Thurman & Grossman,
2008; Troje,
2002). The two most commonly reported features are the backstroke of the feet (Saunders et al.,
2009) and the crossing of joints visible in profile views of locomotion (Giese & Poggio,
2003; Thurman & Grossman,
2008). Neurons tuned to point–light actions typically display selectivity for consistent combinations of body postures and body dynamics (Oram & Perrett,
1996; Vangeneugden, Vancleef, Jaeggli, VanGool, & Vogels,
2010; Vangeneugden et al.,
2011).
In addition to features embedded within the actor itself, a body of work exists demonstrating that the visual processes engaged when organizing point–light walkers also influence perception of the surround. Illustrated by the so-called “backscroll illusion,” observers experience illusory motion of the background when viewing a stationary point–light walker (lacking translation, as if on a treadmill), with the background perceived to be moving in the opposite direction of the walker (Fujimoto,
2003; Fujimoto & Yagi,
2008). This illusory motion is sufficiently powerful so as to capture and bias the perceived motion of random motion elements in favor of the background when superimposed on point–light sequences (Fujimoto & Yagi,
2007).
In this study, we use a novel neurophysiological approach to isolate the motion features observers selectively attend that promote point–light biological motion detection. In tasks with known targets, selective attention filters are employed in anticipation of the attended items, to bias perceptual encoding in favor of those features (Bridwell, Hecker, Serences, & Srinivasan,
2013; Bridwell & Srinivasan,
2012). Feature-based attention is also not spatially specific, with attention-mediated gain observed in neural populations in visual cortex tuned to the attended feature across visual field (Saenz, Buracas, & Boynton,
2002; Serences & Boynton,
2007). This phenomenon is also predicted from single-unit recordings of feature-based attention gain control, which is not spatially selective (Treue & Martínez Trujillo,
1999). Features and object categories can also be decoded from the neural activity in parietal cortex, an important brain region for cortical control of attention (Erez & Duncan,
2015; Liu & Hou,
2013). This finding is also consistent with early steady-state visual-evoked potential (SSVEP) studies that find parietal responses to be modulated by feature-based attention (Bridwell & Srinivasan,
2012; Bridwell et al.,
2013).
We have developed a paradigm to measure these attentionally mediated modulations for attended motion features using probe stimuli in unattended locations that are flickered to evoke a robust SSVEP (Bridwell et al.,
2013; Bridwell & Srinivasan,
2012). SSVEP signal modulation is associated with different cognitive and sensory phenomena, and in particular, selective attention (Ding, Sperling, & Srinivasan,
2006; Morgan, Hansen, & Hillyard,
1996; Müller et al.,
2006). Important for this study, frequency-tagged SSVEPs can capture the attention modulation of neural signals that share features with the attended objects,
even when the SSVEP is tagged to unattended and spatially displaced stimuli (Bridwell et al.,
2013; Bridwell & Srinivasan,
2012; Garcia, Srinivasan, & Serences,
2013; Painter, Dux, Travis, & Mattingley,
2014).
We therefore seek to use the attentionally-mediated SSVEP as a means for evaluating the directional selectivity of the attentive filters deployed when detecting a biological motion target. Given that subjects use task-relevant knowledge to anticipate diagnostic features, we reason that evidence for features-based attention should be apparent in the brain response throughout the interval during which subjects monitor for appearance of the walker. That includes prior to the actual appearance of the target when the walker is anticipated but not yet perceived. And because neurons tuned to those diagnostic features exhibit attentional gain throughout the visual field, we should be able to identify this attentional strategy in the SSVEP modulations for the flicker-tagged, unattended stimulus when it contains task-relevant features. If observers monitor for local opponent motion when detecting a point–light walker, this should be apparent in brain responses for task-irrelevant motion in both the forward and backward walking directions. If observers monitor for a global body posture, we should not observe any attention modulation on task-irrelevant motion features. Finally, if observers attend to the illusory motion of the background, we should observe attention gain in task-irrelevant motion opposite to that of the facing direction of the walker.