The estimation of any visual property (e.g., position, orientation, motion) requires that information be pooled over an extended image region. Such pooling is only useful, however, if the selected region is appropriate to the estimation task. In estimating the motion of an object, for instance, pooling is necessary because of the inherent ambiguity of local motion signals (the
aperture problem). However, such pooling can also yield highly biased estimates if motion signals from two different objects are mixed—e.g., if along with the moving object of interest, the pooling also includes motion signals from a stationary background or a nearby object moving in the opposite direction (e.g., McDermott & Adelson,
2004; McDermott, Weiss, & Adelson,
2001). Similarly, object localization (e.g., for guiding saccades) requires that the visual information for one object be segregated from visual information corresponding to other objects before an estimate of location is computed (Cohen, Schnitzer, Gersch, Singh, & Kowler,
2007; Denisova, Singh, & Kowler,
2006; Melcher & Kowler,
1999; Vishwanathan, Kowler, & Feldman,
2000). Thus, the estimation of any visual property presupposes perceptual segmentation, the division into “perceptual groups.”