Vision is initiated by the light rays entering the pupil from three-dimensional (3D) scenes. The cornea and lens (physiological optics) transform these rays to form a two-dimensional (2D) spectral irradiance image at the retinal photoreceptor inner segments. How the physiological optics transforms these rays, and how the photoreceptors encode the light, limits certain aspects of visual perception and performance. These limits vary both within an individual over time and across individuals, according to factors such as eye size and shape, pupil size, lens accommodation, and wavelength-dependent optical aberrations (Wyszecki & Stiles,
1982).
Physiological optics are accounted for in vision science and engineering by a diverse array of models. In certain cases, the application is limited to central vision and flat (display) screens, and in these cases, the optical transformation is approximated using a simple formula: convolution with a wavelength-dependent point spread function (PSF; Wandell,
1995). This approximation is valid when the stimulus is at most 5° away from the fovea, because the PSF changes with eccentricity (Navarro, Artal, & Williams,
1993). To generate a retinal image spanning a large range of eccentricities, one needs PSF data for a full range of eccentricities, a method of interpolating them continuously across the field, and a way to apply wavelength-dependent translation to model transverse chromatic aberrations (TCAs).
The approximation is also valid only when the stimulus is distant. When viewing nearby 3D objects, the optical transformation is more complex and depends on the 3D geometry and spatial extent of the scene. For example, in the near field, the PSF is depth dependent, and shift-invariant convolution produces incorrect approximations of retinal irradiance at image points near depth occlusions. Accurately calculating the optical transformation for scenes within 1 to 2 m of the eye, which are important for understanding depth perception and vergence, requires more complex formulae and computational power.
This article describes Image Systems Engineering Toolbox–3D (ISET3d), a set of software tools that simulate the physiological optics transformation from a 3D spectral radiance into a 2D retinal image. The software is integrated with Image Systems Engineering Toolbox–Bio (ISETBio), an open-source package that includes a number of computations related to the initial stages of visual encoding (Brainard et al.,
2015; Cottaris, Jiang, Ding, Wandell, & Brainard,
2019; Kupers, Carrasco, & Winawer,
2019). The initial ISETBio implementations modeled image formation for planar or distant scenes with wavelength-dependent PSFs (Farrell, Catrysse, & Wandell,
2012; Farrell, Jiang, Winawer, Brainard, & Wandell,
2014). ISET3d uses quantitative computer graphics to model the depth-dependent effects of the physiological optics and enables the user to implement different schematic eye models for many different 3D scenes. For some of these models, the software can calculate retinal irradiance for a range of accommodative states, pupil sizes, and retinal eccentricities.
ISETBio uses the retinal irradiance to estimate the excitations in the cone spatial mosaic. These estimated cone excitations can be helpful in understanding the role of the initial stages of vision in limiting visual performance, including accommodation as well as judgments of depth and size. In addition, these calculations can support the design and evaluation of 3D display performance (e.g., light fields, augmented reality, virtual reality), which benefit from a quantitative understanding of how display design parameters affect the retinal image and cone excitations. For example, ISET3d may help develop engineering metrics for novel volumetric displays that render rays as if they arise from a 3D scene or for multiplanar displays that use elements at multiple depths to approximate a 3D scene (Akeley, Watt, Girshick, & Banks,
2004; MacKenzie, Hoffman, & Watt,
2010; Mercier et al.,
2017; Narain et al.,
2015). In summary, we describe ISET3d as a tool that can be helpful to various different specialties within vision science and engineering.