Is it possible to infer a person's goal or intention by analyzing their eye fixations? There have been tremendous advances in techniques for the discovery of a person's perceptions, thoughts, or goals—a topic referred to in the popular press as “mind reading” (e.g., Grandoni,
2012). Most mind reading research has used a technique called
neural decoding that infers a person's thoughts or percepts based on only their neural activity (e.g., Kay, Naselaris, Prenger, & Gallant,
2008; Tong, & Pratte,
2012). We introduce the technique of
behavioral decoding, the use of fine-grained behavioral measures to similarly infer a person's thoughts, goals, or mental states.
We attempted behavioral decoding in the context of a visual search task, with the decoding goal being the category of a person's search target. The behaviors that we decoded were the objects fixated during search. The many fixations on non-target objects, or
distractors, made during search are not random—the more similar these objects are to the target category, the more likely they are to be fixated first (Eckstein, Beutter, Pham, Shimozaki, & Stone,
2007) or fixated longer (Becker,
2011) compared to less target-similar objects. In this sense, categorical search obeys the same rules known to govern target-specific search (Alexander & Zelinsky,
2012), with one important caveat—the expression of target-distractor similarity relationships in fixation behavior is proportional to the specificity of the target cue (Malcolm & Henderson,
2009; Schmidt & Zelinsky,
2009). Specific cues, such as the picture previews used in most search studies (Wolfe,
1998), often lead to an efficient direction of gaze to the target (Zelinsky,
2008), but are highly unrealistic—outside of the laboratory one does not often know exactly how a target will appear in a search context. The vast majority of real-world searches are categorical, with the information used to cue a target being far less reliable and specific. It is not known whether these less specific categorical cues generate sufficient information in fixation behavior to decode the category of a search target.
This study answers four questions. First, does the cueing of different categorical targets produce different patterns of fixations during search? If such behavioral differences do not exist, behaviorally decoding the target category will not be possible. Second, are these behavioral differences sufficient for decoding—can one person read another person's mind, inferring their search target by decoding their fixations on random category distractors? Third, what are the limits of behavioral decoding in this task—is it possible to decode a target from the objects fixated on a single trial, and does decoding success vary with target-distractor similarity? Finally, how does a person's mind reading ability compare to that of computer vision classifiers? Machine decoders are valuable in that they make explicit the information from fixations used to decode search targets. To the extent that human and machine decoders agree, this agreement may suggest that human decoders use similar information. This latter goal makes our approach conceptually related to other quantitative techniques for capturing and revealing information embedded in human behavior (e.g., Baddeley & Tatler,
2006; Caspi, Beutter, & Eckstein,
2004; Eckstein & Ahumada,
2002; Gosselin & Schyns,
2001; Tavassoli, van der Linde, Bovik, & Cormack, (
2009), with a difference being that our approach uses computer vision methods to decode from this information the identity of real-world object categories.