The neural mechanisms involved in visual processing in the human brain are relatively well understood. A considerable number of specialized visual areas have been identified, and various experimental approaches have contributed to characterizing the roles and specific functional properties of each of these areas. Research into (putatively) homologous visual areas in non-human primates has provided further insight into visual processing mechanisms at the neural level, making the visual system probably the best understood of the mammalian sensory systems. However, although the neural mechanisms underlying perceptual processing are therefore well understood, the neural processes subserving perceptual awareness remain largely unknown.
The brain achieves visual awareness extremely rapidly. In daily life, we rarely notice any delay incurred by neural processing: When we view our own limbs during the execution of an action, for example, the visual feedback appears instantaneous. Likewise, in laboratory settings, the visual information necessary to distinguish even complex visual categories such as animate or inanimate objects (Thorpe, Fize, & Marlot,
1996) and natural or man-made environments (Joubert, Rousselet, Fize, & Fabre-Thorpe,
2007) is available extremely rapidly. Humans can make manual behavioral responses to such tasks within about 400 ms (VanRullen & Thorpe,
2001a; perception). In fact, the majority of this time is spent preparing the motor output associated with the response: Neural measures demonstrate that the visual category is established substantially earlier (within ∼150 ms; Thorpe et al.,
1996). Recently, even this time window was shortened when Kirchner and Thorpe (
2006) demonstrated that observers can reliably make eye movements to an object from a specific semantic category within as little as 120 ms.
This ultrarapid visual categorization limits the number of processing stages that information might undergo on the way to awareness and suggests that visual processing involves substantial automatic feedforward mechanisms. However, not all of the information cascading through this processing pipeline necessarily reaches visual awareness. Indeed, phenomena such as inattentional and change blindness (Simons,
2000; Simons & Levin,
1997,
1998) indicate that although we might introspectively feel that we are aware of our entire visual environment, in fact only a very limited subset of incoming sensory information actually reaches awareness. One influential model proposes that the initial rapid feedforward sweep is followed (for some representations) by feedback from higher to lower association areas and that this recurrent processing is the neural hallmark of a sensory representation achieving awareness (Lamme,
2003). However, what determines which neural representations are processed recurrently and which are not? Some kind of selection process must clearly determine which information we become aware of and which is lost by the wayside.
The nature of this selection process forms the crux of substantial debate. Selection takes place at several levels in the visual hierarchy. Early visual areas are sensitive to high spatial and temporal frequency information, for example, whereas such information is not consciously visible (He & MacLeod,
2001). For the most part, the observer has no conscious control over such selection mechanisms. However, an important part of the selection process
is under conscious control. Evidence from change blindness and inattentional blindness suggests that attention plays a crucial role in this selection stage: When attention is diverted to a different stimulus or task, observers fail to notice even dramatic changes to visual stimuli (e.g., Rensink, O'Regan, & Clark,
1997). In this interpretation, attention plays a crucial role in determining which visual information is destined to reach awareness and which is to be lost by the wayside.
Further evidence that attention is involved in gating consciousness is the century-old notion of prior entry: that “the object of attention comes to consciousness more quickly than the objects which we are not attending to” (Spence & Parice,
2009; Titchener,
1908). Indeed, experiments show that attended stimuli are perceived to precede unattended stimuli by roughly 30–60 ms (see Spence & Parice,
2009, for a review). Behavioral responses are similarly faster for attended than unattended locations, objects, or features (e.g., Posner,
1980).
Comparable effects of attention on the latency of neural processes however remain contentious. ERP components evoked by attended and unattended stimuli, although often exhibiting different amplitudes, generally have similar latencies (McDonald, Teder-Salejarvi, Di Russo, & Hillyard,
2005). Where systematic difference have been demonstrated, these are typically an order of magnitude smaller than perceptual effects (i.e., <10 ms; Vibell, Klinge, Zampini, Spence, & Nobre,
2007). ERP components whose latency does covary with overt behavioral measures (such as reaction time) have been reported, but these components (e.g., P300; Kutas, McCarthy, & Donchin,
1977, and the sustained posterior contralateral negativity, SPCN, Brisson & Jolicoeur,
2008) are generally relatively late (300–500 ms post-stimulus), putting them well outside the ∼200-ms time window in which perceptual awareness can be achieved (e.g., Thorpe et al.,
1996). As such, a neural correlate of the mechanism that guards access to conscious awareness remains elusive.
Recently, Gosselin and Schyns (
2001) provided a number of important insights into the path to visual awareness of faces using a reverse correlation approach they termed Bubbles. By recording ERPs during the viewing of random subsets of visual information under different task sets and correlating these to behavior, they were able to describe the time course of neural processing and describe at which time points the brain is responsive to which visual features (Schyns, Petro, & Smith,
2007; Smith, Fries, Gosselin, Goebel, & Schyns,
2009). Using a similar trial-by-trial analysis strategy, Philiastides and Sajda (
2006) explored the time course of decision making on the basis of visual information. Both groups identified EEG activity in two time windows: an early time window around 170 ms during which sensory processing takes place and a later time window around 300 ms, which showed task-specific modulations (Smith et al.,
2009). Whereas the latency of the first component is constant, the latency of the second is variable and depends on the amount of available evidence (Philiastides & Sajda,
2006). Altogether, these findings support a distinction between neural activity related to automatic, feedforward processing and neural activity related to information selected for conscious awareness (see also Johnson & Olshausen,
2003; VanRullen & Thorpe,
2001b).
From a similar temporal perspective, one important empirical implication of a gateway mechanism to awareness has to date remained largely unexploited: If selection determines what reaches visual awareness, then when selection occurs will also determine when the selected stimulus reaches awareness. Logically, the speed of selection determines when downstream processes involved in generating awareness can first become active. As a result, the latency of these neural processes should depend on the latency of selection.
Given an appropriate method to estimate when attentional selection takes place, the latency of neural processes can therefore provide a way to identify the neural correlates of awareness. One approach is to use reaction time (RT) as an index (e.g., Amano et al.,
2006; Kutas et al.,
1977; McCarthy & Donchin,
1981; Renault, Ragot, Lesevre, & Remond,
1982). However, RT can be affected by a number of factors unrelated to selection per se, including for example processes involved in decision making and motor control. Second, RT requires an overt behavioral response. The corresponding neural activity will overlap and potentially disguise the true neural correlates of awareness, particularly in neuroimaging approaches with limited spatial resolution, such as electroencephalography (EEG).
We recently developed a behavioral task to estimate the latency of attentional selection without requiring observers to make speeded overt responses (Carlson, Hogendoorn, & Verstraten,
2006). In this paradigm, the observer views an array of running analog clocks. At a given moment, one of the clocks is flashed. At the end of the trial, the observer reports what the time was on that clock, when it was flashed. The difference between the reported and actual time on the flashed clock is then taken as an estimate of attentional selection latency. Because the observer simply reports what he or she
perceived (similar to the attentional gating approach; Reeves & Sperling,
1986), it directly probes what the time was when the target clock was selected by attention. Importantly, because no response is required until after the trial, it evokes minimal unrelated neural activity during the crucial part of the trial.
Here, we recorded event-related potentials (ERPs) while observers carried out this behavioral task. We show that trial-by-trial behavioral variability in attentional selection latency positively correlates to trial-by-trial variability in ERP latency in three scalp regions in particular. This suggests that these three locations form crucial waypoints along which information flows on the way to visual awareness.