In studies involving mobile subjects, two approaches have been used for gaze estimation: tracking of eye-in-head orientation with registration of the eye position on scene images from a head mounted camera (Fotios, Uttley et al.,
2014; Li, Munn et al.,
2008), or tracking the gaze orientation in a confined environment using multiple cameras (Land,
2004). In the former approach, the head orientation is typically not measured, and interpretation of gaze behaviors requires visual examination of the recorded scene videos and manual data entry of head orientation. Using the latter approach, tracking both the eye position (in head) and the head orientation, makes it possible to measure the gaze point in real world coordinates. However, this is usually performed only in confined indoor or within-car environments, where the head orientation (relative to the car) can be tracked by the sensing system set up in the enclosed environment (Barabas, Goldstein et al.,
2004; Bowers, Ananyev et al.,
2014; Cesqui, de Langenberg et al.,
2013; Essig, Dornbusch et al.,
2012; Grip, Jull et al.,
2009; Imai, Moore et al.,
2001;
Kugler, Huppert et al.; Land
2009; Lin, Ho et al.,
2007; MacDougall & Moore
2005; Proudlock, Shekhar et al.,
2003). Due to technical difficulties in outdoor gaze tracking, there have been only a few behavioral studies (Geruschat, Hassan et al.,
2006; Hassan, Geruschat et al.,
2005; 't Hart & Einhauser,
2012) that investigated gaze behaviors while walking in uncontrolled outdoor open spaces. Furthermore, analyses of these studies were limited by the challenge of manual processing of large volumes of gaze data. For example, in studies of gaze behaviors during street crossings at intersections (Hassan, Geruschat et al.,
2005), the head mounted scene videos were visually examined and gaze behaviors were manually classified coarsely as left, center, or right relative to the crosswalk. The quantitative analysis of a large amount of gaze tracking data is impractical with such a technique.