Open Access
Article  |   May 2025
The effect of illumination cues on color constancy in simultaneous identification of illumination and reflectance changes
Author Affiliations
  • Lari S. Virtanen
    Department of Psychology, Faculty of Medicine, University of Helsinki, Helsinki, Finland
    [email protected]
  • Maria Olkkonen
    Department of Psychology, Faculty of Medicine, University of Helsinki, Helsinki, Finland
    [email protected]
  • Toni P. Saarela
    Department of Psychology, Faculty of Medicine, University of Helsinki, Helsinki, Finland
    [email protected]
Journal of Vision May 2025, Vol.25, 4. doi:https://doi.org/10.1167/jov.25.6.4
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Lari S. Virtanen, Maria Olkkonen, Toni P. Saarela; The effect of illumination cues on color constancy in simultaneous identification of illumination and reflectance changes. Journal of Vision 2025;25(6):4. https://doi.org/10.1167/jov.25.6.4.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

To provide a stable percept of the surface color of objects, the visual system needs to account for variation in illumination chromaticity. This ability is called color constancy. The details of how the visual system disambiguates effects of illumination and reflectance on the light reaching the eye are still unclear. Here we asked how independent illumination and reflectance judgments are of each other, whether color constancy depends on explicitly identifying the illumination chromaticity, and what kinds of contextual cues support this identification. We studied the simultaneous identification of illumination and reflectance changes with realistically rendered, abstract three-dimensional scenes. Observers were tasked to identify both of these changes between sequentially presented stimuli. The stimuli included a central object whose reflectance could vary and a background that only varied due to changes in illumination chromaticity. We manipulated the visual cues available in the background: local contrast and specular highlights. We found that identification of illumination and reflectance changes was not independent: While reflectance changes were rarely misidentified as illumination changes, illumination changes clearly biased reflectance judgments. However, correct identification of reflectance changes was also not fully dependent on correctly identifying the illumination change: Only when there was no illumination change in the stimulus did it lead to better color constancy, that is, correctly identifying the reflectance change. Discriminability of illumination changes did not vary based on available visual cues of local contrast or specular highlights. Yet discriminability of reflectance changes was improved with local contrast and, to a lesser extent, with specular highlights, in the stimulus. We conclude that a failure of color constancy does not depend on a failure to identify illumination changes, but additional visual cues still improve color constancy through better disambiguation of illumination and reflectance changes.

Introduction
In photopic light levels, a prominent feature of our visual world is color, a psychological correlate for variations in the relative wavelength composition of light. Determining the colors of objects in our visual environment is complicated by the fact that the spectral power distribution of reflected light depends on the spectral reflectance of the surface and the spectral power distribution of illumination. Furthermore, the illumination in the environment varies substantially with the time of day, weather, and whether the attended location is in direct sunlight or shade. Thus, to infer the surface colors of objects, the visual system needs to constantly disambiguate illumination from reflectance (von Helmholtz, 1924), and its remarkable ability to perform this function is called color constancy
The basic mechanisms underlying human color perception, such as trichromacy and cone-opponency, are well known (e.g., Gegenfurtner & Kiper, 2003). Yet, even after decades of research efforts, we do not fully understand how the visual system produces color constancy (for reviews, see Brainard & Radonjić, 2014; Foster, 2011; Olkkonen & Ekroll, 2016; Witzel & Gegenfurtner, 2018). Early accounts included reliance on chromatic adaptation, where the illumination chromaticity sets the new perceptually achromatic point through sensory adaptation (see, e.g., Hering, 1920). The independent effects of sensory adaptation were later found to be much more limited and supplemented by strong influences of stimulus surroundings (Jameson & Hurvich, 1959; Jameson & Hurvich, 1961; Jameson & Hurvich, 1964; Wallach, 1948). Chromatic adaptation mechanisms can be seen to operate in both slow and fast time scales (Fairchild & Reniff, 1995) with an instantaneous component that specifically influences color appearance (Rinner & Gegenfurtner, 2000). The instantaneous effect is attributed to color contrast by an object’s immediate background, also referred to as local contrast (see, e.g., Hurlbert & Wolf, 2004). 
Another line of study focuses on different heuristics based on image properties that the visual system could employ in producing color constancy. For example, early computational algorithms used assumptions about image statistics to estimate the illumination chromaticity and to discount its influence to produce surface color estimates (white patch and edge contrasts: Land & McCann, 1971; average color: Buchsbaum, 1980). However, observers show poorer accuracy in estimating illumination compared to surface color, and the two estimates are not well correlated (e.g., Granzier, Brenner, & Smeets, 2009; Rutherford & Brainard, 2002). Other scene statistics have been proposed to help disambiguate changes of illumination and reflectance chromaticity while not providing a direct estimate of illumination chromaticity (mean cone excitation ratios: Foster & Nascimento, 1994; luminance–redness correlation: Golz & MacLeod, 2002). Additionally, the regularities of spectral variation in illumination and surfaces of natural environments can provide heuristics to constrain the problem of color constancy (e.g., Foster, Amano, & Nascimento, 2006; Foster, Amano, Nascimento, & Foster, 2006; Lee, 1986; Maloney & Wandell, 1987; Morimoto, Fukuda, & Uchikawa, 2016; Morimoto, Kusuyama, Fukuda, & Uchikawa, 2021). 
Which cues within a visual scene contribute to color constancy remains actively studied (see, e.g., Brainard, Cottaris, & Radonjic, 2018; Foster, 2011). Adaptation within a smaller area around the stimulus and across the whole visual scene has an influence at different time scales (Smithson & Zaidi, 2004). However, neither local adaptation, global adaptation, nor adaptation to the brightest spot can fully account for observers’ color constancy (Kraft & Brainard, 1999). Instead, slower sensory adaptation may support adjustment to the average illumination chromaticity in the environment, while faster, regional mechanisms compensate for illumination differences locally (Werner, 2014). Local contrast between adjacent areas has an especially robust effect on how color (and lightness) is perceived (Helson, 1943; Shevell, 1978; Wallach, 1948). Nevertheless, providing an inaccurate illumination cue in the immediate background reduces color constancy but does not eliminate it, pointing to a role for other cues (Delahunt & Brainard, 2004; Kraft, Maloney, & Brainard, 2002). Other stimulus properties, such as three-dimensionality of stimuli, show smaller and less consistent effects (see, for a review, Mizokami, 2019). For example, Xiao, Hurst, MacIntyre, and Brainard (2012) found color constancy to be better for flat disks than three-dimensional (3D) spheres with consistent illumination cues in the image, but the situation reversed with an inconsistent cue in the immediate background, showing better constancy for 3D spheres. Surface properties, such as glossiness, may also support color constancy by providing information about illumination chromaticity, as suggested by computational models (D’Zmura & Lennie, 1986; Lee, 1986; Lehmann & Palm, 2001). Experimental results show that this may be the case, at least when cues from local surround are weak or missing (Granzier, Vergne, & Gegenfurtner, 2014; Lee & Smithson, 2016; Wedge-Roberts et al., 2020). 
A common method for studying color constancy is asymmetric matching, in which the accuracy of matching surface colors across different illuminations is measured (Arend & Reeves, 1986; Brainard, Brunt, & Speigle, 1997; Burnham, Evans, & Newhall, 1957). Operational color constancy (Craven & Foster, 1992), on the other hand, refers to the ability to discriminate between changes in illumination and surface reflectance (see, e.g., Foster, Amano, & Nascimento, 2001; Foster, Nascimento, et al., 2001; Linnell & Foster, 1996). Recently, Ennis and Doerschner (2019) studied how observers perceive and estimate simultaneous changes in illumination and object reflectance when shown both an object and a background or either one alone, and how much these estimates rely on globally and locally calculated image statistics. They found that observers are able to resolve both illumination and reflectance when shown the whole stimulus scene but not from the object or background alone. Of the statistics Ennis and Doerschner tested, luminance–redness correlation, mean cone excitation ratios, white patch, and average color could each account for some of the variance in observer responses for either reflectance or illumination. None of the statistics alone was a good general predictor. The authors concluded that the visual system can retain explicit information of illumination changes when a background is present and that observers are likely to use various sources of information to achieve color constancy. Their experiment did not, however, probe the interdependence of illumination and surface reflectance estimation. Furthermore, the scene surrounding the object was only manipulated between a highly diagnostic white background and no illuminated background at all. 
In this study, we examine the possible interdependence of illumination and reflectance judgments in operational color constancy. Specifically, even if explicit estimates of illumination chromaticity are inaccurate as previous studies suggest (Granzier, Brenner, & Smeets, 2009; Granzier & Valsecchi, 2014; Khang & Zaidi, 2004), does correct identification of a surface reflectance change depend on explicitly identifying the change in illumination, and does this, in turn, depend on available visual cues (local contrast or specular highlights)? Observers performed simultaneous identification of illumination and reflectance changes with sequentially presented, realistically rendered stimuli. We manipulated available cues for color constancy from local contrast and specular highlights. We found that reflectance judgments were biased by illumination changes. The identification of reflectance changes did not fully depend on explicit identification of illumination changes, however, only showing a clear difference in discriminability when there was no illumination change in the stimulus. Identification of reflectance changes was poorer without local contrast cues and only improved slightly with specular highlights. Identification was also strongly dependent on whether reflectance or illumination chromaticity changed toward or away from neutral. 
Methods
Observers
Three observers took part in a pilot study, and 10 new observers took part in the main study. Of the 10 observers, 9 completed all measurements and were included in data analysis (7 female and 2 male, aged 19–31 with median age of 20). All observers reported having normal or corrected-to-normal visual acuity. Color vision was assessed to be normal with Ishihara color plates (Ishihara, 1987). Prior to the experiment, all observers were given a brief explanation about the interaction of surface reflectance and illumination in producing the light reflected to the eyes. Observers were also told that we were interested in the identification of chromatic surface reflectance and illumination changes in the scene, but they were otherwise naive to the goals of the study. The study followed the Declaration of Helsinki and was approved by the University of Helsinki Ethical Review Board in Humanities and Social and Behavioural Sciences. 
Apparatus
The experiments were carried out in a darkened laboratory room at the University of Helsinki. The experiments were run with custom code on an HP Z230 Desktop PC using MATLAB (Version R2016b, build 9.1.0.441655) with the PsychToolBox-3 extensions (Brainard, 1997; Kleiner, Brainard, & Pelli, 2007; Pelli, 1997). The display was a 23-inch ViewPixx 3D Lite, controlled by an Nvidia Quadro K620 graphics card. The spectra and gamma functions of the display primaries were measured using a SpectraScan PR-655 spectroradiometer for calibrated color reproduction. Display luminance was linearized by interpolating measured gamma functions. Display white point was set to be metameric to D65, and stimuli were displayed using a 100-Hz refresh rate and 10 bits per primary channel. The high bit-depth display mode that was used for increased color resolution halved the number of horizontal pixels of the native 1,920 × 1,080 resolution, making the effective display resolution 960 × 1,080 pixels. Viewing distance was held constant at 100 cm using a chin rest, and observers gave their responses with a gamepad. 
Stimuli
Stimuli were simple, abstract, 3D-rendered scenes (see Figure 1). The objects in the scenes were created by adding irregular Gaussian perturbations to a sphere using ShapeToolbox (Saarela, 2018) and will be referred to as “blobs.” All stimuli included a blob in the middle of the screen, extending approximately 2° in visual angle on average, and we will refer to this central blob as the “target.” Two types of stimuli comprised only a target object with an otherwise black screen (referred to as “object stimuli;” Figure 1a): (1) target flattened in the direction of viewing, “flat object” and (2) matte 3D target, “3D object.” Three types of stimuli comprised a target with different surroundings (referred to as “scene stimuli”; Figure 1b): (1) flat target with a flat gray background, “flat scene”; (2) matte 3D target with four surrounding colored matte objects, “3D matte scene”; and (3) matte 3D target with four surrounding colored glossy objects, “3D glossy scene.” 
Figure 1.
 
Illustrations of the five different types of stimuli used in the experiment. (a) Object stimuli only include a target with an otherwise black screen. (b) Scene stimuli include a target and a background affected by changing scene illumination, either as a gray flat surface or additional, colored blobs.
Figure 1.
 
Illustrations of the five different types of stimuli used in the experiment. (a) Object stimuli only include a target with an otherwise black screen. (b) Scene stimuli include a target and a background affected by changing scene illumination, either as a gray flat surface or additional, colored blobs.
The stimulus scenes were created with spectral, physically based rendering with plain path tracing in Mitsuba (Jakob, 2010) using RenderToolbox3 (Heasly, Cottaris, Lichtman, Xiao, & Brainard, 2014). In all scenes except the 3D glossy scene, all surfaces were modeled with a bidirectional scattering distribution function of an ideally diffuse (Lambertian) material with no specular reflections. In the 3D glossy scene, the target was an ideally diffuse material, while the colored blobs surrounding the target were defined as a plastic-like material with specular and diffuse components (Mitsuba's default “rough plastic” material). The material modeled a diffuse surface under a thin dielectric layer using Fresnel reflection and transmission coefficients with an interior index of refraction of 1.49 (matching polypropylene). Additionally, the material produced surface roughness by modeling microfacet normals as a Beckmann distribution with root mean square slope of 0.1. Illumination with spatial variation in intensity was created using a freely available high dynamic range light probe as an environment map (“Overcast Day at Techgate Donaucity” by Bernhard Vogl: http://dativ.at/lightprobes/). The luminance component of the environment map was kept intact while the chromatic component was replaced with the desired uniform chromaticity. 
For rendering, we defined discrete spectral distributions between 300 and 700 nm in 10-nm steps for illuminant radiance and the diffuse reflectance component of surfaces (which we will later simply refer to as “reflectance”). First, however, we defined chromaticities for different illuminants and reflectances in CIELUV color space using the uv′-coordinates (the uv′-coordinates for spectral reflectance were defined as the coordinates of the reflected light when a surface having that spectral reflectance is illuminated by an equal-energy white illuminant). As a measure of chromatic difference between colors, we used Euclidian distance in uv′-coordinates (Δuv′). Note that although CIELUV is only approximately uniform (see, e.g., Mahy, Van Eycken, & Oosterlinck, 1994; Pointer, 1981), we deemed that equal color differences in Δuv′ were sufficiently accurate for the purposes of the experiment. D65 chromaticity was always chosen as the “middle” illuminant, and six chromaticities with equal steps in Δuv′ were chosen along the daylight locus (D-illuminant) in both the lower color temperature (yellower) and the higher color temperature (bluer) directions (Figure 2). For reflectance, 13 chromaticities were chosen so that steps in illumination and reflectance chromaticities matched in the resulting reflected light chromaticity (details below). To create the spectral distributions from uv′ chromaticity coordinates, we used spectral upscaling from XYZ values with basis vectors. For illuminants, we used the CIE D-illuminant components (CIE, 2018). For reflectances, we created custom basis vectors by taking the first five principal components from the spectral measurements of Munsell chips by Parkkinen, Hallikainen, and Jaaskelainen (1989)
Figure 2.
 
Chromaticities of one of the seven stimulus sets used in stimulus rendering displayed on CIE uv′ chromaticity diagram. The u′-dimension is on the x-axis and v′-dimension on the y-axis. Red diamonds, blue crosses, and black circles show the 13 different illuminants, 13 different reflectances, and reflected light from all their combinations, respectively (note that multiple circles overlap). Green crosses show the reflectances of the four “context” blobs and green circles their reflected lights with the middle illuminant (D65).
Figure 2.
 
Chromaticities of one of the seven stimulus sets used in stimulus rendering displayed on CIE uv′ chromaticity diagram. The u′-dimension is on the x-axis and v′-dimension on the y-axis. Red diamonds, blue crosses, and black circles show the 13 different illuminants, 13 different reflectances, and reflected light from all their combinations, respectively (note that multiple circles overlap). Green crosses show the reflectances of the four “context” blobs and green circles their reflected lights with the middle illuminant (D65).
We wanted the reflected target color change to be the same for both illumination and reflectance changes. This was done in order to force reflectance estimation to depend on accounting for possible illumination changes, so that target color alone cannot be used to disentangle illumination and reflectance changes. We chose illuminant and reflectance chromaticities so that a given number of steps leads to the same reflected target color, whether they are illumination or reflectance steps, or both (see Figure 4). Conversely, this also means that opposite changes in illumination and reflectance lead to no change in reflected color. Note that this would not be possible with specular highlights because they reflect the chromaticity of the illumination, which is why we chose to use matte targets in all experimental conditions. To create the 3D glossy scene, we used specular background objects instead. 
Figure 3.
 
The course of a single trial. The reference stimulus was displayed for 1,000 ms, followed by a 250-ms ISI, and the comparison stimulus for 1,000 ms. The observer was then shown one or two indicators with written responses alternatives. When the observer gave their response, the experiment moved to the next trial after a 500-ms ITI.
Figure 3.
 
The course of a single trial. The reference stimulus was displayed for 1,000 ms, followed by a 250-ms ISI, and the comparison stimulus for 1,000 ms. The observer was then shown one or two indicators with written responses alternatives. When the observer gave their response, the experiment moved to the next trial after a 500-ms ITI.
Figure 4.
 
All possible changes in stimulus colors between the reference and the comparison stimulus in scene blocks. In this example, the reference stimulus would be identical to the stimulus in the middle. The comparison stimulus could be one of the nine possibilities in the figure. Changes in illumination are on different columns, and changes in target reflectance are on different rows. The red lines illustrate that similar changes in the illumination or reflectance result in identical changes in the reflected light from the target.
Figure 4.
 
All possible changes in stimulus colors between the reference and the comparison stimulus in scene blocks. In this example, the reference stimulus would be identical to the stimulus in the middle. The comparison stimulus could be one of the nine possibilities in the figure. Changes in illumination are on different columns, and changes in target reflectance are on different rows. The red lines illustrate that similar changes in the illumination or reflectance result in identical changes in the reflected light from the target.
To estimate suitable chromaticities, we created a custom optimization procedure to minimize errors using MATLAB's fminsearch-function. First, we calculated 13 chromaticity coordinates for reflected light using a reflectance equal to D65 chromaticity with each of the 13 illuminants. We calculated the reflected light by using the spectral upscaling detailed above and element-wise multiplication of illuminant and reflectance spectral distributions. Using reflectances with chromaticity coordinates equal to those of our 13 illuminants as starting points, we searched for reflectance uv′-coordinates that would minimize the difference in uv′ coordinates of corresponding reflected lights compared to the 13 previously calculated reflected lights, as well as Δuv′ between all adjacent reflectances. Second, we let the u′, v′, and Y′ (luminance) values of all 13 illuminants and all 13 reflectances vary. We searched for chromaticity coordinates that would minimize the difference between all corresponding reflected lights, the difference in illuminant step sizes (Δuv′), the difference in reflectance step sizes (Δuv′), and difference in Y between different illuminants and between different reflectances. An example of the resulting chromaticity coordinates is illustrated in Figure 2
Eight sets of such coordinates were created, each with a different step size (approximately 0.0039, 0.0055, 0.0064, 0.0073, 0.0085, 0.0096, 0.0109, and 0.0123 in Δuv′). The set with the smallest step size was only used for the preliminary measurement while the other seven were options for the rest of the experiment (chosen based on the observer’s performance). Finally, chromaticities for the four background blobs were selected by picking coordinates with equal distance as the furthest reflectance steps from D65 in −u′, +u′, −v′, and +v′, respectively. For the gray, flat background, reflectance spectrum was always uniform with 0.33 reflectance. 
For stimuli without a background, scenes were rendered with the D65 illuminant for all different reflectances, 13 different stimuli for each stimulus set. For stimuli with a background, scenes were rendered with all combinations of illuminant and reflectance spectra, 169 different stimuli for each stimulus set. Additionally, with four surrounding blobs as the background, the blobs were rendered in six different rotations. This was done to discourage observers from consistently using specific areas of the blobs with specular reflections to determine illumination chromaticity, which could have biased spatial attention in the 3D glossy condition. 
Procedure
The experiment was divided into three sessions held on different days. The first session included a preliminary discrimination measurement, demonstration (henceforth demo) and practice trials, and one run of the main experiment. Sessions 2 and 3 included a demo and practice set, followed by two runs of the main experiment. We first describe the tasks and then explain in more detail the content of the sessions. 
The aim of the preliminary measurement was to determine a discrimination threshold for stimulus color differences. Observers were shown two flat target stimuli sequentially with no background in 500-ms intervals, separated by a 250-ms interstimulus interval (ISI). The observer responded whether the stimulus in the second interval was yellower or bluer than in the first interval and received feedback for their response. One of the two stimuli had the median chromaticity, and the other stimulus had one of the 13 chromaticities in the stimulus set. Each chromaticity pair was repeated 13 times, and a cumulative Gaussian psychometric function was fit to the proportion of responses indicating the varying stimulus as bluer. The stimulus set with Δuv′ closest to 1.5 times the observer's psychometric function standard deviation was chosen for the first practice set. 
The main experiment consisted of the stimulus conditions presented in Figure 1, with each condition run in a separate block of trials. Each trial consisted of the presentation of a reference stimulus for 1,000 ms and a comparison stimulus for 1,000 ms, with a 250-ms ISI in between (Figure 3). The observer's task was then to respond how the stimulus colors changed from the reference to the comparison stimulus by moving visual response indicators to their choices with the gamepad's control sticks and confirming with a button press. In object blocks (Figure 1a), observers responded using a single indicator whether the target surface color had a large change toward yellow, small change toward yellow, no change, small change toward blue, or large change toward blue. In scene blocks (Figure 1b), observers judged both target surface color and scene illumination and responded using separate visual indicators if there was a change toward yellow, no change, or a change toward blue. See Figure 3 for an illustration of the screen with two visual indicators. The order of the indicators (left/right), as well as whether blue or yellow responses were up or down, was alternated between observers. Response time was not limited. After a response, observers received feedback with colored response indicators displayed for 1,000 ms in practice and dummy trials but not in main experiment trials. The experiment proceeded to the next trial after a 500-ms intertrial interval (ITI). 
In object blocks, the reference stimulus was chosen from the middle 9 of the 13 reflectance chromaticities (9 different references in total) and the comparison stimulus −2, −1, 0, +1, or +2 chromaticity steps from reference (5 different changes in total). In scene blocks, both illuminant and target reflectance chromaticities were chosen from the middle 9 of the 13 respective chromaticities (81 different references in total) and the comparison stimulus illuminant and reflectance −2, 0, or +2 chromaticity steps from reference (9 different changes in total; see Figure 4). 
After the preliminary measurement at the beginning of the first session, the rest of the first session and the following two sessions were used to collect data for the main experiment. Each of the three sessions followed the same pattern: first demo trials followed by practice trials for each block in a fixed order, then main experiment trials once for each block in a random order (see Figure 5a). In the second and third sessions, main experiment trials were repeated a second time for each block in a random order. In the first session, average d′ (d′ or marginal d′, see below) across all blocks and stimulus changes were calculated from practice trials, and this was used to select the stimulus set for the rest of the experiment. Average d′ between 1 and 1.5 led to no change in stimulus set; d′ below 1, 0.75, and 0.5 led to a stimulus set 1, 2, and 3 steps higher, respectively; and d′ above 1.5, 2, and 2.5 led to a stimulus set 1, 2, and 3 steps lower, respectively. 
Figure 5.
 
Experiment protocol. (a) Measurement Session 1 included a preliminary measurement used to choose an initial stimulus set. The following practice trials were analyzed to adjust the stimulus set toward a target level of performance. In Sessions 2 and 3, all main experiment trial blocks were repeated twice. (b) In demo and practice blocks, the practice trials were preceded by demo trials and dummy trials. (c) In main experiment blocks, sets of main experiment trials (without feedback) were interleaved with sets of dummy trials (with feedback).
Figure 5.
 
Experiment protocol. (a) Measurement Session 1 included a preliminary measurement used to choose an initial stimulus set. The following practice trials were analyzed to adjust the stimulus set toward a target level of performance. In Sessions 2 and 3, all main experiment trial blocks were repeated twice. (b) In demo and practice blocks, the practice trials were preceded by demo trials and dummy trials. (c) In main experiment blocks, sets of main experiment trials (without feedback) were interleaved with sets of dummy trials (with feedback).
Blocks of demo and practice trials (see Figure 5b) began with the demo trials where observers were first told in writing what kind of stimulus color changes would be demonstrated and were then shown two examples of such trials with different reference stimuli. This was repeated for all possible changes (object blocks) or combinations of changes (scene blocks) in a preset order. The first session had considerably more practice trials than the second and third sessions. In the first session, practice object blocks included 10 trials for each change (50 trials in total). Practice scene blocks included 15 trials for each illumination and reflectance change, shuffled into random combinations (45 trials in total). In the second and third sessions, practice included only two trials for each comparison in object blocks and three trials for each comparison in scene blocks. At the beginning of each practice block, three dummy trials selected randomly from the list of trials were presented. 
Main experiment blocks (see Figure 5c) began with and were interleaved by dummy trials with feedback. Object blocks included two dummy sets with one repetition of each change. Scene blocks included three dummy sets with two repetitions for each illumination and reflectance change, shuffled into random combinations. In total, there were five repetitions for each main experiment block. Thus, each change (object blocks) or combination of changes (scene blocks) had 40 main experiment trials used in analysis. 
Data analysis and modeling
First, all dummy trials were discarded from analysis. For overall comparisons between blocks, data were pooled across all references. Based on signal detection theory (Green & Swets, 1966), the object blocks were considered a one-dimensional five-alternative identification task, and scene blocks were considered a two-dimensional three-alternative identification task. Average of d′ (or average marginal d′ in scene blocks) over change levels was used as a measure of discriminability. Paired t-tests were used to compare discrimination between different reference-comparison changes within blocks and to compare the same reference-comparison changes between blocks. All p-values of all statistical tests mentioned were Holm-corrected (Holm, 1979). 
To investigate the relationship between identification of illumination and reflectance changes, we pooled the data across observers and stimulus changes within scene blocks and performed chi-square tests of independence for correct illumination and reflectance responses. Additionally, we pooled the data for each scene stimulus across observers only and calculated average marginal d′ to identify reflectance changes separately for trials where illumination change was identified correctly and incorrectly. Their equality was tested with significance tests for one-parameter signal detection hypotheses (Gourevitch & Galanter, 1967; Marascuilo, 1970). 
In order to explore if identifying changes toward and away from the “neutral” differ, we again used the data pooled across observers. Trials from each of the three scene blocks were binned according to whether the reference stimulus had illumination or reflectance from the yellowest, bluest, or neutral third in the stimulus pool. Within the six trial groups, criterion (c) and discriminability (d′) were calculated separately for changes toward yellow or blue, and their equality was tested with the aforementioned one-parameter tests (Gourevitch & Galanter, 1967; Marascuilo, 1970). 
Finally, we quantified the relationship between identification of illumination and reflectance changes by fitting a two-dimensional signal-detection model to the data from the scene blocks. The model observer makes its decision by comparing noise-corrupted responses to response criteria in a two-dimensional decision space, where the response to a stimulus is given by two hypothetical mechanisms: a “reflectance mechanism” and an “illumination mechanism.” This model is not intended to be a model of actual mechanisms of color constancy; rather, it is used as an analysis tool to quantify the amount of cross-talk in the processing of the reflectance and illumination cues. 
We represent the stimulus by a vector \(\boldsymbol {s} = [{\begin{array}{*{10}c}s_{\mathrm{r}} & s_{\mathrm{i}} \end{array}}]^\top\), where the reflectance value sr is negative for a yellow shift, zero for no change, and positive for a blue shift in reflectance. The same holds for illumination changes for the illumination value si. The sensitivity of the two mechanisms to illuminance and reflectance changes is given by the vectors \(\boldsymbol {m}_{\mathrm{r}}\) and \(\boldsymbol {m}_{\mathrm{i}}\), put in the rows of matrix M:  
\begin{eqnarray} \boldsymbol {M} = \left[\begin{array}{@{}c@{}}\boldsymbol {m}_{\mathrm{r}} \\ \boldsymbol {m}_{\mathrm{i}}\end{array}\right] = \left[\begin{array}{@{}c@{\quad}c@{}} m_{\mathrm{rr}} & m_{\mathrm{ri}} \\ m_{\mathrm{ir}} & m_{\mathrm{ii}} \end{array}\right], \qquad \end{eqnarray}
(1)
where the first subindex is for the mechanism and the second for the stimulus feature (so that, for example, mri gives the sensitivity of the reflectance mechanism to illumination changes). If illumination and reflectance cues in our stimuli were processed independently of each other, matrix M would be diagonal. The noisy responses of the mechanisms are then computed as  
\begin{equation} \boldsymbol {\psi } = \boldsymbol {M}\boldsymbol {s} + \boldsymbol {\epsilon } \end{equation}
(2)
 
\begin{equation} \hphantom{\boldsymbol {\psi }}= \left[\begin{array}{@{}c@{}} \psi _{\mathrm{r}} \\ \psi _{\mathrm{i}} \end{array}\right] + \left[\begin{array}{@{}c@{}} \epsilon _{\mathrm{r}} \\ \epsilon _{\mathrm{i}}\end{array}\right], \end{equation}
(3)
where the noise \(\boldsymbol {\varepsilon }\) is zero-mean and normally distributed:  
\begin{equation} \boldsymbol {\varepsilon } \sim \mathrm{MVN}(\boldsymbol {0}, \boldsymbol {\Sigma }), \end{equation}
(4)
 
\begin{equation} \boldsymbol {\Sigma } = \left[\begin{array}{@{}c@{\quad}c@{}} \sigma ^2_{\mathrm{r}} & 0 \\ 0 & \sigma ^2_{\mathrm{i}} \end{array}\right]. \end{equation}
(5)
 
We fixed the noise covariance to zero, because the possible cross-talk between the two stimulus dimensions is accounted for by the mechanism sensitivities (in matrix \(\boldsymbol {M}\)). We fixed the noise variances to 1 for zero response and assumed that the variance increases linearly with the mechanism response (σ2 = 1 + kψ, where k is a constant, for both responses). The model observer makes the final identification decision by comparing the noisy response to two response criteria along each stimulus dimension. 
We thus fit 10 parameters in total: 2 parameters to scale the variances, 4 parameters for decision criteria, and 4 parameters for the response mechanism sensitivities. Parameters were fit by maximum likelihood (in practice by minimizing the negative log-likelihood of the parameters given the observed responses). Judged by the Bayesian information criterion, this model gave a better fit than a full model where the variances and mean responses are free parameters. 
The model allows for perceptual independence to fail between the two stimulus dimensions (because the two mechanisms are not necessarily orthogonal), but it assumes decisional separability (because the decision criteria are parallel to the axes of the decision space). Note that the effect of nonorthogonality of the mechanisms could also be modeled as having orthogonal dimensions but correlated noise (Ashby & Townsend, 1986). We chose to have potentially nonorthogonal mechanisms to better illustrate the dependencies in the processing of the two cues. 
Results
Object block response counts summed across observers (Figures 6a, 6b) show that the correct option had the most responses in all cases except for the large change toward yellow for the flat object. Nearly all of the incorrect responses were in the response options adjacent to the correct option. There was some tendency to report less of a change in the yellow direction than what is shown in the stimulus, both in general and in comparison to changes toward blue. On average, discriminability of color changes in the stimuli (Figure 6c) did not show a statistically significant difference in paired t-tests between the two object blocks or between yellower and bluer changes within blocks. 
Figure 6.
 
Change identification results from object blocks. Heatmaps of response counts across all observers for (a) flat object and (b) 3D object. Each stimulus change is shown on a separate row. Left-hand icons indicate the correct color change response, and bottom icons indicate observers’ color change responses. (c) Plot discriminability measured by average d between the neutral and the “large” color differences. Different object blocks are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bar indicates ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average.
Figure 6.
 
Change identification results from object blocks. Heatmaps of response counts across all observers for (a) flat object and (b) 3D object. Each stimulus change is shown on a separate row. Left-hand icons indicate the correct color change response, and bottom icons indicate observers’ color change responses. (c) Plot discriminability measured by average d between the neutral and the “large” color differences. Different object blocks are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bar indicates ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average.
Figures 7a–c shows the response counts from the scene blocks for each of the 3 × 3 = 9 possible stimulus changes. The correct option again had the most responses in most cases, and incorrect responses were mainly for response options adjacent to the correct option. There was, however, also interaction between the responses to the two types of stimulus change. Most notably, observers were most accurate when there was either no change or a simultaneous change in both illumination and reflectance toward the same color. In contrast, observers performed worst when both illumination and reflectance changed toward opposite colors (and the reflected color thus stayed the same), often reporting no change in reflectance in this case. For comparisons of average discriminability in scene blocks (Figure 7d), observers were better at discriminating illumination changes than reflectance changes (flat: t(8) = 2.42, p = 0.046; 3D matte: t(8) = 4.41, p = 0.007; 3D glossy: t(8) = 2.81, p = 0.046). Discriminability of reflectance changes was better for the flat scene compared to the 3D matte scene (t(8) = 4.23, p = 0.009). Discriminability of illumination changes did not show a significant difference between scene blocks. 
Figure 7.
 
Change identification results from scene blocks. Heatmaps of response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene. Each stimulus change is shown in a separate heatmap. Top icons indicate the correct reflectance change response and right icons the correct illumination change response. Left icons indicate observers’ illumination responses and bottom icons indicate observers’ reflectance responses. (d) Plot discriminability measured by average d′ across stimulus reflectance and illumination changes, respectively. Different scene blocks with reflectance and illumination plotted separately are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bars indicate ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average. **p < 0.01, *p < 0.05.
Figure 7.
 
Change identification results from scene blocks. Heatmaps of response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene. Each stimulus change is shown in a separate heatmap. Top icons indicate the correct reflectance change response and right icons the correct illumination change response. Left icons indicate observers’ illumination responses and bottom icons indicate observers’ reflectance responses. (d) Plot discriminability measured by average d′ across stimulus reflectance and illumination changes, respectively. Different scene blocks with reflectance and illumination plotted separately are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bars indicate ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average. **p < 0.01, *p < 0.05.
To illustrate how correct identification of reflectance change depends on the correct identification of illumination change, cross-tabulated responses of scene blocks are presented in Figures 8a–c. Chi-square tests clearly rejected the null hypothesis of independence for the 3D matte (χ2(3,240) = 14.75, p < 0.001) and 3D glossy (χ2(3,240) = 35.06, p < 0.001) scene blocks and, less decidedly, for the flat scene block (χ2(3,240) = 4.41, p = 0.036). The independence of correct identifications varied between different color change comparisons: Across all scene blocks, independence was not rejected for comparisons where illumination change was accompanied with either a change toward blue or no change in reflectance, but it was rejected for all other comparisons (all p < 0.05). Overall, the chi-square tests rejected independence more often than not. However, on average, when the illumination change was not correctly identified, the reflectance change was still correctly identified half of the time—compared to the chance level of one third. Additionally, average conditional d′ are presented in Figure 9. In all scenes, discriminability was better with correct identification of illumination change, but only if there was no illumination change in the stimulus (flat: z = 3.33, p < 0.001; 3D matte: z = 3.60, p < 0.001; 3D glossy: z = 3.85, p < 0.001). Otherwise, discriminability did not show a statistically significant difference, and performance was always clearly above chance level. 
Figure 8.
 
Identification of reflectance changes in scene blocks depending on identification of illumination changes. Response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene, cross-tabulated based on the correctness of the illumination response (row) and reflectance response (column).
Figure 8.
 
Identification of reflectance changes in scene blocks depending on identification of illumination changes. Response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene, cross-tabulated based on the correctness of the illumination response (row) and reflectance response (column).
Figure 9.
 
Discriminability of reflectance changes in scene blocks measured by average d in trials with correct and incorrect illumination change identification, respectively. Correct and incorrect illumination change identification for yellow, no, and blue illumination changes within flat, 3D matte, and 3D glossy scene blocks are on the x-axis with d on the y-axis. Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. ***p < 0.001.
Figure 9.
 
Discriminability of reflectance changes in scene blocks measured by average d in trials with correct and incorrect illumination change identification, respectively. Correct and incorrect illumination change identification for yellow, no, and blue illumination changes within flat, 3D matte, and 3D glossy scene blocks are on the x-axis with d on the y-axis. Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. ***p < 0.001.
Next, we binned the trials based on the reference reflectance and illumination into three categories: the yellow, neutral, and blue categories. We then computed discriminabilities d′ and response criteria c for reflectance and illumination changes. When trials were split by reference illumination, the decision criterion (Figure 10a) was lower for illumination changes away from neutral than toward neutral (yellow reference: z = −12.52, p < 0.001; blue reference z = 15.80, p < 0.001). Note that as all combinations of reference value and stimulus change were analyzed separately, the response criteria for blue and yellow changes are expected to change in opposite directions: For a yellow reference value, a yellow change was away from neutral and a blue change toward neutral, and vice versa for a blue reference. A response criterion of 1 corresponds an optimal or unbiased criterion, and a criterion below 1 indicates a bias to respond with the respective color change. The observed criterion shifts thus mean that observers were more likely to respond that the illumination change was away from the neutral color than toward it. In conjunction, discriminability for illumination change (Figure 10c) was higher for changes away from neutral than toward neutral (yellow reference: z = −5.35, p < 0.001; blue reference z = −3.54, p < 0.001). In other words, observers were not only more likely to respond that the illumination change was away from neutral but also more sensitive to these changes. The response criteria for reflectance (Figure 10b) followed a similar pattern as a function of reference illumination (yellow reference: z = −3.40, p = 0.001; blue reference z = 3.96, p < 0.001), but there was no significant difference in sensitivity for reflectance changes (Figure 10d). 
Figure 10.
 
Marginal decision criteria and discrimination indices calculated across all observers for data binned by reference stimulus colors and reference-comparison color changes. Trials were binned according to whether the reference had illumination (a–d), or reflectance (e–h), from the yellowest, bluest, or neutral third in the stimulus pool (x-axis). Panels (a–b) and (e–f) show criterion (c, y-axis), and panels (c–d) and (g–h) show discriminability (d′, y-axis) for identification of illumination and reflectance changes, respectively. d′ and c were calculated between identification of no change and yellow or blue change (yellow and blue lines). Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. Note that criteria are displayed in absolute values, and a criterion below 1 means a bias to respond with the change indicated by the line color. ***p < 0.001, **p < 0.01.
Figure 10.
 
Marginal decision criteria and discrimination indices calculated across all observers for data binned by reference stimulus colors and reference-comparison color changes. Trials were binned according to whether the reference had illumination (a–d), or reflectance (e–h), from the yellowest, bluest, or neutral third in the stimulus pool (x-axis). Panels (a–b) and (e–f) show criterion (c, y-axis), and panels (c–d) and (g–h) show discriminability (d′, y-axis) for identification of illumination and reflectance changes, respectively. d′ and c were calculated between identification of no change and yellow or blue change (yellow and blue lines). Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. Note that criteria are displayed in absolute values, and a criterion below 1 means a bias to respond with the change indicated by the line color. ***p < 0.001, **p < 0.01.
When trials were binned by reference reflectance, the criteria for reflectance (Figure 10f) were higher for changes toward neutral than away from neutral (yellow reference: z = −13.96, p < 0.001; blue reference z = 11.44, p < 0.001). For illumination changes, the pattern was different (Figure 10e): The criterion was higher for illumination changes toward yellow in the yellow reference group (z = 9.11, p < 0.001). The observers were thus more likely to report blue illumination changes when the reference reflectance was yellow. There were no significant differences in sensitivity for either reflectance or illumination changes (Figures 10g, 10h). 
Finally, we quantified the amount of cross-talk in the processing of reflectance and illumination cues by fitting a two-dimensional signal-detection model to the data. The model has two hypothetical mechanisms, a “reflectance mechanism” and an “illumination mechanism,” whose directions in stimulus space determine how the two cues affect the responses. The raw data for the model fits were the response counts from the scene blocks, shown in Figures 7a–c. Note that this model does not take into account the asymmetries demonstrated in Figure 10, nor does it predict the specific pattern of conditional d′-values in Figure 9. The vectors depicting the mechanisms in stimulus space are shown in Figure 11a for the three scene conditions. The axes represent the two possible stimulus changes, a reflectance change and an illumination change, with a blue change as positive and a yellow change as negative. If there was no cross-talk in the identification of the two types of change, the vectors would lie on the corresponding axes. 
Figure 11.
 
(a) The vectors depicting the reflectance and illumination mechanisms in stimulus space for the three scene conditions: flat, 3D matte, and 3D glossy. The more aligned a given vector is with the corresponding stimulus axis, the better the model observer is at identifying the changes, and the fewer confusions between the two types of stimulus change occur. Dots indicate the nine stimulus changes used in the experiment. (b) The angle of the mechanism vectors with respect to the reflectance axis. (c) Average d′-values of the model observer for identifying reflectance and illumination changes. (d–f) Response distributions for the nine types of stimulus change plotted in the decision space of the model (the axes correspond to the mechanism responses) in the flat (d), 3D matte (e), and 3D glossy (f) conditions. The vertical and horizontal lines show the response criteria. Ellipses signify half a standard deviation of the response distribution around the mean. First letter in each distribution indicates stimulus illumination change; second letter reflectance change. Color of the circle indicates absolute color change of the target.
Figure 11.
 
(a) The vectors depicting the reflectance and illumination mechanisms in stimulus space for the three scene conditions: flat, 3D matte, and 3D glossy. The more aligned a given vector is with the corresponding stimulus axis, the better the model observer is at identifying the changes, and the fewer confusions between the two types of stimulus change occur. Dots indicate the nine stimulus changes used in the experiment. (b) The angle of the mechanism vectors with respect to the reflectance axis. (c) Average d′-values of the model observer for identifying reflectance and illumination changes. (d–f) Response distributions for the nine types of stimulus change plotted in the decision space of the model (the axes correspond to the mechanism responses) in the flat (d), 3D matte (e), and 3D glossy (f) conditions. The vertical and horizontal lines show the response criteria. Ellipses signify half a standard deviation of the response distribution around the mean. First letter in each distribution indicates stimulus illumination change; second letter reflectance change. Color of the circle indicates absolute color change of the target.
Instead, in all three conditions, neither mechanism vector is aligned with the axis. This means that illumination cues caused a response in the reflectance mechanism and vice versa. The illumination mechanism was, however, much more closely aligned with the illumination change axis in the stimulus space. In other words, the reflectance mechanism was more excited by illumination changes than the illumination mechanism was by reflectance changes. Thus, the illumination responses are much more independent of reflectance color changes than the other way around. In the 3D matte scene, the reflectance mechanism is close to having a 45° angle with the axes, which would indicate that both cues were equally likely to result in a perceived reflectance change. We tested the separation of the reflectance and illumination mechanisms with permutation tests. Compared to the 3D matte scene, the angle between the two mechanisms was larger in the presence of local contrast (flat scene, p < 0.01) or specular highlights (3D glossy scene, p < 0.05), leading to better separation of the two dimensions. The angles between the mechanism vectors and the reflectance axis (x-axis) are plotted in Figure 11b. The average d′-values are plotted in Figure 11c. Although grouped differently, these correspond to the average d′-values in Figure 7d and show a similar pattern, with poorest color constancy for the 3D matte stimulus. 
Figures 11d–f shows the response distributions for each stimulus change in decision space. The decision criteria are shown as vertical and horizontal lines. These distributions further illustrate how reflectance responses are much more dependent on illumination changes than illumination responses are on reflectance changes. 
Discussion
We examined whether operational color constancy relies on explicit identification of illumination changes and the contribution of different cues to color constancy. Comparing average discriminability in scene blocks (Figure 7d), we found that illumination changes were more reliably discriminated than reflectance changes and that local contrast improved color constancy of reflectance percepts. Discriminability suggested that specular highlights might also produce a small improvement in color constancy, and this was seen in better separation of modeled mechanism vectors (Figure 11). When we examined discriminability of reflectance changes depending on correct identification of illumination changes (Figure 9), we found that identification of a reflectance change was improved if observers correctly reported an illumination change only when there was no illumination change in the stimulus. When there was an illumination change in the stimulus, discriminability of reflectance changes was independent of correct identification of the illumination change. By comparing discriminability of color changes for different reference colors (Figure 10), we found that there was a bias to more readily report changes away from the neutral point, both with illuminations and reflectances, and also better discriminability of illumination changes away from the neutral point. Broadly, local contrast and specular highlights seem to improve color constancy—that is, they help to disambiguate illumination and surface reflectance. However, they do so without providing any better explicit judgment of illumination chromaticity. 
In our experiment, identification of illumination changes was more accurate than identification of reflectance changes (Figure 7d), which is in contrast with previous studies finding poorer judgments of illumination than reflectance chromaticity (Granzier, Brenner, & Smeets, 2009; Rutherford & Brainard, 2002). This is probably caused by the experimental design; illumination change could be identified directly from a global chromaticity shift alone (ignoring the target), while identification of reflectance change required comparison between target and context. In contrast to our findings, in the study by Ennis and Doerschner (2019), observers' responses of color change followed real stimulus changes approximately equally well for illumination and reflectance. This is likely due to the type of chromaticity changes employed. As we chose illumination and reflectance chromaticity changes that would lead to the same reflected chromaticities, the step sizes for both changes had to be identical. Instead, Ennis and Doerschner (2019) used significantly larger chromaticity changes for reflectance than for illumination. The changes were also not exactly matched for hue and included combinations of illumination and reflectance changes along different axes (red-green, blue-yellow). 
We deliberately chose illumination and reflectance chromaticity changes that can cancel each other out in the light reflected from the target object to study the dependence of operational color constancy on illumination change judgments. Correct reflectance change judgments are in theory conditional on correct illumination change judgments by the experiment design. However, unlike in previous studies, which measured the accuracy of illumination chromaticity estimates (Granzier, Brenner, & Smeets, 2009) or constancy by illumination chromaticity adjustments (Rutherford & Brainard, 2002), here the task only required identifying if a possible change in illumination was toward yellow or blue. Despite these design choices, we found that in most cases, discriminability of reflectance changes did not depend on correctly identifying the illumination change (Figure 9). Irrespective of the stimulus scene, only trials with no illumination change showed a significant dependence. It is not clear what causes this pattern of results. In the task, observers could identify the illumination change directly from the context, without having to consider the target at all. As such, if there was no illumination change, an incorrect response would require that a change in the target only is misattributed as a change across the whole stimulus or a momentary lapse in performing the task. It may be that misattributing reflectance changes as illumination changes represents a more fundamental failure in parsing chromatic changes in the scene, leading to an accumulation of trials where performance was overall poor. Regardless, discriminability of reflectance changes clearly did not fully depend on illumination change judgments. 
The illumination-estimation hypothesis posits that the visual system estimates scene illumination to determine surface color (see, e.g., Maloney & Yang, 2003), and errors in reflectance estimates could be a consequence of errors in illumination estimates (Brainard, 1998; Brainard, Brunt, & Speigle, 1997). Studies characterizing explicit illumination judgments, however, do not fully support this view (Granzier, Brenner, & Smeets, 2009; Rutherford & Brainard, 2002; Weiss, Witzel, & Gegenfurtner, 2017). It is still plausible that the full computation to estimate illumination is implicit, and as the visual system attempts to effectively discount scene illumination, perceived illumination chromaticity is consistently desaturated (as seen in Granzier, Brenner, & Smeets, 2009). However, our results suggest that color constancy is mostly disconnected from even a rough explicit illumination estimate in the form of identifying a possible illumination change toward yellow or blue. 
To examine the role of local contrast and specular highlights in disentangling illumination and reflectance changes, these cues were varied in scene blocks by manipulating the scene background. The flat scene included a background behind the target, creating local contrast at the object outline. Instead, with the 3D matte scene, illumination changes had to be interpreted across a larger spatial distance. The lack of local contrast was produced by having the objects, as explained to observers, “float in a black void.” Note that this is different from studies that manipulated local contrast cues by introducing a conflicting cue instead (Delahunt & Brainard, 2004; Kraft, Maloney, & Brainard, 2002; Xiao et al., 2012). An invalid local contrast cue does not only eliminate the cue but also is likely to interfere with other cues. In the flat scene, the uniformly gray background reflected illumination chromaticity unaltered. In contrast to the 3D matte scene in which illumination changes had to be interpreted from an average chromaticity change in the scene, the glossy scene included specular highlights that could be used as a direct cue to illumination chromaticity similarly to the flat scene. However, specular highlights had to be spotted from specific locations across larger spatial distances. These manipulations do not provide a full factorial breakdown of the visual background cues in our stimuli but are aimed toward two comparisons specifically: local contrast versus no local contrast and accurate illumination chromaticity cue versus averaging. 
Our results show that discriminability of illumination changes did not differ between the scene blocks (Figure 7d). This is somewhat surprising. Color constancy is improved with the inclusion of visual cues such as local contrast (Delahunt & Brainard, 2004; Kraft & Brainard, 1999; Kraft, Maloney, & Brainard, 2002) and specular highlights (Granzier, Vergne, & Gegenfurtner, 2014; Lee & Smithson, 2016; Wedge-Roberts et al., 2020). The visual cues can be thought to provide the visual system with more information of scene illumination (e.g., Boyaci, Doerschner, & Maloney, 2006; Yang & Maloney, 2001), which the visual system may further combine for more accurate and reliable illumination estimates (Maloney, 2002). Yet we did not find a significant benefit in discrimination of illumination changes with either local contrast or specular highlights. Again, explicit identification of illumination changes, even with more reliable cues, seems limited. 
What differed between the scene blocks was the discriminability of reflectance changes, being highest for the flat scene and lowest for the 3D matte scene (Figure 7d). Previous studies have demonstrated the importance of local contrast for accurate color constancy (Delahunt & Brainard, 2004; Kraft & Brainard, 1999; Kraft, Maloney, & Brainard, 2002), but by introducing a conflicting cue, they produced a dramatic deterioration in constancy. Instead, we found only a medium–large benefit with inclusion of local contrast (reflectance discriminability between flat scene and 3D matte scene, Cohen’s d = 0.76). Glossiness, or specular highlights, also supports color constancy, especially when cues are limited (Granzier, Vergne, & Gegenfurtner, 2014; Lee & Smithson, 2016; Wedge-Roberts et al., 2020). For specular highlights, we only found a small–medium benefit (reflectance discriminability between 3D glossy scene and 3D matte scene, Cohen’s d = 0.43, not statistically significant). These findings suggest that the visual system does not need to rely on local contrast or direct illumination chromaticity cues but can achieve a decent degree of color constancy even by averaging and/or comparison across spatially disjointed elements. 
The observed data were also well accounted for by modeling hypothetical mechanisms for illumination and reflectance response. Note that these hypothetical mechanisms were defined specifically for the stimulus and response space in this experiment design. They only abstract the association between stimulus and response with no assumptions of underlying sensory or cognitive mechanisms. In contrast to studies on operational color constancy where the choice is between a change in reflectance or illumination (Craven & Foster, 1992) or between congruent and incongruent changes (Foster, Nascimento, et al., 2001; Linnell & Foster, 1996), simultaneous identification (also in Ennis & Doerschner, 2019) introduces different requirements for the visual system. A uniform chromaticity shift in the stimulus can still signal that there was an illumination change without reflectance change. However, a nonuniform chromaticity shift can signal a reflectance change, but it can occur with or without an accompanying illumination change. Furthermore, the target reflectance can change similarly or inversely compared to illumination, producing either a large shift or no shift in the target, respectively. Neither is uniquely defined by a global or local (target) chromaticity shift alone. Responses that fail to separate changes in illumination and reflectance from this ambiguity lead to cross-talk between the model mechanisms. 
Based on model estimates (Figure 11), in all scene blocks, the illumination response was well aligned with illumination changes in the stimulus, indicating that reflectance changes were rarely misidentified as illumination changes. In contrast, the reflectance response was much more mixed between reflectance and illumination changes, indicating that changes in illumination were often misidentified as changes in reflectance. The independence of the reflectance mechanism in the modeled vectors mirrors the difference in reflectance identification we found between the scene blocks. However, they also show a difference for the illumination mechanism between the scene blocks while difference in illumination identification d′ did not reach statistical significance. Thus, even if local contrast and specular highlights do not produce a notable advantage for judging illumination changes in operational color constancy, they seem to help the visual system resolve at least some ambiguity. This further supports the notion that the visual cues tested here support the disentangling of illumination and reflectance, not absolute illumination chromaticity judgments. 
On average, there was no difference in discriminability for changes toward yellow and changes toward blue. However, this only held when averaged over all reference chromaticities. When different illumination reference chromaticities were analyzed separately, discriminability was higher and the decision criteria lower for illumination changes away from neutral compared to changes toward neutral (Figure 10). This bias in discriminability is in agreement with the “neutral bias” coined by Hurlbert and colleagues (Pastilha, Gupta, Gross, & Hurlbert, 2020; see also Aston, Radonjic, Brainard, & Hurlbert, 2019), but our results also demonstrated a strong complementary bias in the decision criteria, which was not reported in previous studies. Furthermore, the bias in decision criteria was not limited to illumination changes but was also found for reflectance changes toward and away from neutral depending on reflectance reference. Thus, there seems to be a significant component to the neutral bias driven by decision criteria, which is not limited to percepts of illumination chromaticity but influences surface color percepts as well. 
As discussed by previous studies that identified a neutral bias in illumination discrimination (Aston et al., 2019; Pastilha et al., 2020), the bias is not easily explained by low-level asymmetries of chromatic discrimination. For example, while detection thresholds for changes along the L − M cone-opponent axis are independent of the adaptation point on this axis, detection thresholds on the S − (L + M) axis increase with increasing S-cone excitation of the adapting stimulus (Krauskopf & Gegenfurtner, 1992). However, this means that changes toward both yellow and blue are easier to detect against a yellow background, whereas in the neutral bias, changes toward yellow are easier to detect from a yellow reference, and changes toward blue are easier to detect from a blue reference. Another plausible explanation is a strong prior for illumination chromaticity in a Bayesian framework of color constancy (Brainard & Freeman, 1997; Brainard et al., 2006; Gehler, Rother, Blake, Minka, & Sharp, 2008), as was suggested by Hurlbert (2019). The range of illumination chromaticity in our experiment was along the common axis of natural illumination variation of the daylight locus (Judd et al., 1964) and by extension natural scene color variation (McDermott & Webster, 2012; Webster, Mizokami, & Webster, 2007; Webster & Mollon, 1997). With respect to natural variation, however, our “neutral” references are still much more common than the saturated “yellower” and “bluer” references. As memory retention can increase the weighting of priors in color estimates (Olkkonen & Allred, 2014; Olkkonen, McCarthy, & Allred, 2014), this would produce a stronger influence of a neutral prior in the reference stimulus than the comparison stimulus. This would then lead to both a lower criterion and a higher sensitivity for identifying changes away from neutral. But if this prior is specific to illumination chromaticity, why do we also see a bias in the reflectance criterion? A speculative answer is that the bias for illumination judgments is partially transferred to reflectance judgments. We found that there is a strong interference of illumination changes in the stimulus on reflectance responses (Figure 11), but it is unclear if this could also influence the bias that depends on reflectance reference color (Figure 10f). Regardless, to confirm if an illumination prior could underlie the pronounced neutral bias in illumination estimation requires further study. 
Conclusions
Color constancy does not fully depend on correctly identifying the illumination chromaticity. Furthermore, cues from local contrast and specular highlights aid color constancy in accurate disentanglement of reflectance from illumination but not through improved illumination chromaticity judgments. Finally, changes away from the chromatic neutral point are not only better discriminated for illuminations but also more readily reported for both illuminations and surface colors. 
Acknowledgments
Funded by Emil Aaltonen Foundation (Grants #200240 N1 and #210249 N1). Open access funded by Helsinki University Library. 
Commercial relationships: none. 
Corresponding author: Lari S. Virtanen. 
Address: Department of Psychology, Faculty of Medicine, University of Helsinki, Haartmaninkatu 3, 00014, University of Helsinki, Helsinki, Finland. 
References
Arend, L., & Reeves, A. (1986). Simultaneous color constancy. Journal of the Optical Society of America A, 3(10), 1743–1751, https://doi.org/10.1364/JOSAA.3.001743. [CrossRef]
Ashby, F. G., & Townsend, J. T. (1986). Varieties of perceptual independence. Psychological Review, 93(2), 154–179. [CrossRef] [PubMed]
Aston, S., Radonjic, A., Brainard, D. H., & Hurlbert, A. C. (2019). Illumination discrimination for chromatically biased illuminations: Implications for color constancy. Journal of Vision, 19(3), 15, https://doi.org/10.1167/19.3.15. [CrossRef] [PubMed]
Boyaci, H., Doerschner, K., & Maloney, L. T. (2006). Cues to an equivalent lighting model. Journal of Vision, 6(2), 2, https://doi.org/10.1167/6.2.2. [CrossRef]
Brainard, D. H. (1997). The psychophysics toolbox. Spatial Vision, 10, 433–436, https://doi.org/10.1163/156856897X00357. [CrossRef] [PubMed]
Brainard, D. H. (1998). Color constancy in the nearly natural image. 2. Achromatic loci. Journal of the Optical Society of America A, 15(2), 307–325, https://doi.org/10.1364/JOSAA.15.000307. [CrossRef]
Brainard, D. H., Brunt, W. A., & Speigle, J. M. (1997). Color constancy in the nearly natural image. 1. Asymmetric matches. Journal of the Optical Society of America A, 14(9), 2091–2110, https://doi.org/10.1364/JOSAA.14.002091. [CrossRef]
Brainard, D. H., Cottaris, N. P., & Radonjic, A. (2018). The perception of colour and material in naturalistic tasks. Interface Focus, 8(4), 20180012, https://doi.org/10.1098/rsfs.2018.0012. [CrossRef] [PubMed]
Brainard, D. H., & Freeman, W. T. (1997). Bayesian color constancy. Journal of the Optical Society of America A, 14(7), 1393–1411, https://doi.org/10.1364/JOSAA.14.001393. [CrossRef]
Brainard, D. H., Longère, P., Delahunt, P. B., Freeman, W. T., Kraft, J. M., & Xiao, B. (2006). Bayesian model of human color constancy. Journal of Vision, 6(11), 10, https://doi.org/10.1167/6.11.10. [CrossRef]
Brainard, D. H., & Radonjić, A. (2014). Color constancy. In Werner, J. S. & Chalupa, L. M. (Eds.), The new visual neurosciences (pp. 545–556). Cambridge, MA: MIT Press.
Buchsbaum, G. (1980). A spatial processor model for object colour perception. Journal of the Franklin Institute, 310(1), 1–26, https://doi.org/10.1016/0016-0032(80)90058-7. [CrossRef]
Burnham, R. W., Evans, R. M., & Newhall, S. M. (1957). Prediction of color appearance with different adaptation illuminations. Journal of the Optical Society of America, 47(1), 35–42, https://doi.org/10.1364/JOSA.47.000035. [CrossRef]
CIE. (2018). Components of the relative spectral distribution of daylight, https://doi.org/10.25039/CIE.DS.w7zunnny.
Craven, B. J., & Foster, D. H. (1992). An operational approach to colour constancy. Vision Research, 32(7), 1359–1366, https://doi.org/10.1016/0042-6989(92)90228-B. [CrossRef] [PubMed]
Delahunt, P. B., & Brainard, D. H. (2004). Does human color constancy incorporate the statistical regularity of natural daylight? Journal of Vision, 4(2), 1, https://doi.org/10.1167/4.2.1. [CrossRef] [PubMed]
D'Zmura, M., & Lennie, P. (1986). Mechanisms of color constancy. Journal of the Optical Society of America A, 3(10), 1662–1672, https://doi.org/10.1364/JOSAA.3.001662. [CrossRef]
Ennis, R., & Doerschner, K. (2019). Disentangling simultaneous changes of surface and illumination. Vision Research, 158, 173–188, https://doi.org/10.1016/j.visres.2019.02.004. [CrossRef] [PubMed]
Fairchild, M. D., & Reniff, L. (1995). Time course of chromatic adaptation for colorappearance judgments. Journal of the Optical Society of America A, 12(5), 824–833, https://doi.org/10.1364/JOSAA.12.000824. [CrossRef]
Foster, D. H. (2011). Color constancy. Vision Research, 51(7), 674–700, https://doi.org/10.1016/j.visres.2010.09.006. [CrossRef] [PubMed]
Foster, D. H., Amano, K., & Nascimento, S. M. (2001). How temporal cues can aid colour constancy. Color Research and Application, 26(S1), S180–S185. [CrossRef]
Foster, D. H., Amano, K., & Nascimento, S. M. (2006). Color constancy in natural scenes explained by global image statistics. Visual Neuroscience, 23(3–4), 341–349, https://doi.org/10.1017/S0952523806233455. [PubMed]
Foster, D. H., Amano, K., Nascimento, S. M., & Foster, M. J. (2006). Frequency of metamerism in natural scenes. Journal of the Optical Society of America A, 23(10), 2359–2372, https://doi.org/10.1364/JOSAA.23.002359. [CrossRef]
Foster, D. H., & Nascimento, S. M. (1994). Relational colour constancy from invariant cone-excitation ratios. Proceedings of the Royal Society of London. Series B: Biological Sciences, 257(1349), 115–121, https://doi.org/10.1098/rspb.1994.0103.
Foster, D. H., Nascimento, S. M., Amano, K., Arend, L., Linnell, K. J., Nieves, J. L., ... Foster, J. S. (2001). Parallel detection of violations of color constancy. Proceedings of the National Academy of Sciences, 98(14), 8151–8156, https://doi.org/10.1073/pnas.141505198. [CrossRef]
Gegenfurtner, K. R., & Kiper, D. C. (2003). Color vision. Annual Review of Neuroscience, 26(1), 181–206, https://doi.org/10.1146/annurev.neuro.26.041002.13111. [CrossRef] [PubMed]
Gehler, P. V., Rother, C., Blake, A., Minka, T., & Sharp, T. (2008). Bayesian color constancy revisited. In 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK (pp. 1–8), https://doi.org/10.1109/CVPR.2008.4587765.
Golz, J., & MacLeod, D. I. (2002). Influence of scene statistics on colour constancy. Nature, 415(6872), 637–640, https://doi.org/10.1038/415637a. [CrossRef] [PubMed]
Gourevitch, V., & Galanter, E. (1967). A significance test for one parameter isosensitivity functions. Psychometrika, 32(1), 25–33, https://doi.org/10.1007/BF02289402. [CrossRef] [PubMed]
Granzier, J. J., Brenner, E., & Smeets, J. B. (2009). Can illumination estimates provide the basis for color constancy? Journal of Vision, 9(3), 18, https://doi.org/10.1167/9.3.18. [CrossRef] [PubMed]
Granzier, J. J., & Valsecchi, M. (2014). Variations in daylight as a contextual cue for estimating season, time of day, and weather conditions. Journal of Vision, 14(1), 22, https://doi.org/10.1167/14.1.22. [CrossRef] [PubMed]
Granzier, J. J., Vergne, R., & Gegenfurtner, K. R. (2014). The effects of surface gloss and roughness on color constancy for real 3D objects. Journal of Vision, 14(2), 16, https://doi.org/10.1167/14.2.16. [CrossRef] [PubMed]
Green, D. M., & Swets, J. A. (1966). Signal detection theory and psychophysics (Vol. 1). New York: Wiley.
Heasly, B. S., Cottaris, N. P., Lichtman, D. P., Xiao, B., & Brainard, D. H. (2014). Rendertoolbox3: Matlab tools that facilitate physically based stimulus rendering for vision research. Journal of Vision, 14(2), 6, https://doi.org/10.1167/14.2.6. [CrossRef] [PubMed]
Helson, H. (1943). Some factors and implications of color constancy. Journal of the Optical Society of America, 33(10), 555–567, https://doi.org/10.1364/JOSA.33.000555. [CrossRef]
Hering, E. (1920). Grundzuge der lehre vom lichtsinn. Berlin: Springer.
Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scandinavian Journal of Statistics, 6(2), 65–70, https://www.jstor.org/stable/4615733.
Hurlbert, A. (2019). Challenges to color constancy in a contemporary light. Current Opinion in Behavioral Sciences, 30, 186–193, https://doi.org/10.1016/j.cobeha.2019.10.004. [CrossRef]
Hurlbert, A., & Wolf, K. (2004). Color contrast: A contributory mechanism to color constancy. Progress in Brain Research, 144, 145–160, https://doi.org/10.1016/S0079-6123(03)14410-X. [CrossRef]
Ishihara, S. (1987). Test for colour-blindness. Tokyo, Japan: Kanehara.
Jakob, W. (2010). Mitsuba renderer, https://www.mitsuba-renderer.org.
Jameson, D., & Hurvich, L. M. (1959). Perceived color and its dependence on focal, surrounding, and preceding stimulus variables. Journal of the Optical Society of America, 49(9), 890–898, https://doi.org/10.1364/JOSA.49.000890. [CrossRef] [PubMed]
Jameson, D., & Hurvich, L. M. (1961). Opponent chromatic induction: Experimental evaluation and theoretical account. Journal of the Optical Society of America, 51(1), 46–53, https://doi.org/10.1364/JOSA.51.000046. [CrossRef] [PubMed]
Jameson, D., & Hurvich, L. M. (1964). Theory of brightness and color contrast in human vision. Vision Research, 4(1–2), 135–154, https://doi.org/10.1016/0042-6989(64)90037-9. [PubMed]
Judd, D. B., MacAdam, D. L., Wyszecki, G., Budde, H. W., Condit, H. R., Henderson, S. T., ... Simonds, J. L. (1964). Spectral distribution of typical daylight as a function of correlated color temperature. Journal of the Optical Society of America, 54(8), 1031–1040, https://doi.org/10.1364/JOSA.54.001031. [CrossRef]
Khang, B. G., & Zaidi, Q. (2004). Illuminant color perception of spectrally filtered spotlights. Journal of Vision, 4(9), 2, https://doi.org/10.1167/4.9.2. [CrossRef]
Kleiner, M., Brainard, D., & Pelli, D. (2007). What's new in psychtoolbox-3? Perception, 36(ECVP Abstract Supplement), 14.
Kraft, J. M., & Brainard, D. H. (1999). Mechanisms of color constancy under nearly natural viewing. Proceedings of the National Academy of Sciences, 96(1), 307–312, https://doi.org/10.1073/pnas.96.1.307. [CrossRef]
Kraft, J. M., Maloney, S. I., & Brainard, D. H. (2002). Surface-illuminant ambiguity and color constancy: Effects of scene complexity and depth cues. Perception, 31(2), 247–263, https://doi.org/10.1068/p08sp. [CrossRef] [PubMed]
Krauskopf, J., & Gegenfurtner, K. (1992). Color discrimination and adaptation. Vision Research, 32(11), 2165–2175, https://doi.org/10.1016/0042-6989(92)90077-V. [CrossRef] [PubMed]
Land, E. H., & McCann, J. J. (1971). Lightness and retinex theory. Journal of the Optical Society of America, 61(1), 1–11, https://doi.org/10.1364/JOSA.61.000001. [CrossRef] [PubMed]
Lee, H. C. (1986). Method for computing the scene-illuminant chromaticity from specular highlights. Journal of the Optical Society of America A, 3(10), 1694–1699, https://doi.org/10.1364/JOSAA.3.001694. [CrossRef]
Lee, R. J., & Smithson, H. E. (2016). Low levels of specularity support operational color constancy, particularly when surface and illumination geometry can be inferred. Journal of the Optical Society of America A, 33(3), A306–A318, https://doi.org/10.1364/JOSAA.33.00A306.
Lehmann, T. M., & Palm, C. (2001). Color line search for illuminant estimation in realworld scenes. Journal of the Optical Society of America A, 18(11), 2679–2691, https://doi.org/10.1364/JOSAA.18.002679.
Linnell, K. J., & Foster, D. H. (1996). Dependence of relational colour constancy on the extraction of a transient signal. Perception, 25(2), 221–228, https://doi.org/10.1068/p250221. [PubMed]
Mahy, M., Van Eycken, L., & Oosterlinck, A. (1994). Evaluation of uniform color spaces developed after the adoption of cielab and cieluv. Color Research & Application, 19(2), 105–121, https://doi.org/10.1111/j.1520-6378.1994.tb00070.x.
Maloney, L. T. (2002). Illuminant estimation as cue combination. Journal of Vision, 2(6), 6, https://doi.org/10.1167/2.6.6.
Maloney, L. T., & Wandell, B. A. (1987). Color constancy: A method for recovering surface spectral reflectance. In Fischler, M. A. & Firschein, O. (Eds.), Readings in computer vision: Issues, problems, principles, and paradigms (pp. 293–297). San Francisco: Morgan Kaufmann.
Maloney, L. T., & Yang, J. N. (2003). The illuminant estimation hypothesis and surface colour perception. In Mausfeld, R. & Heyer, D. (Eds.), Colour perception: Mind and the physical world (pp. 335–358). Oxford, UK: Oxford University Press.
Marascuilo, L. A. (1970). Extensions of the significance test for one-parameter signal detection hypotheses. Psychometrika, 35(2), 237–243, https://doi.org/10.1007/BF02291265.
McDermott, K. C., & Webster, M. A. (2012). Uniform color spaces and natural image statistics. Journal of the Optical Society of America A, 29(2), A182–A187, https://doi.org/10.1364/JOSAA.29.00A182.
Mizokami, Y. (2019). Three-dimensional stimuli and environment for studies of color constancy. Current Opinion in Behavioral Sciences, 30, 217–222, https://doi.org/10.1016/j.cobeha.2019.10.008.
Morimoto, T., Fukuda, K., & Uchikawa, K. (2016). Effects of surrounding stimulus properties on color constancy based on luminance balance. Journal of the Optical Society of America A, 33(3), A214–A227, https://doi.org/10.1364/JOSAA.33.00A214.
Morimoto, T., Kusuyama, T., Fukuda, K., & Uchikawa, K. (2021). Human color constancy based on the geometry of color distributions. Journal of Vision, 21(3), 7, https://doi.org/10.1167/jov.21.3.7. [PubMed]
Olkkonen, M., & Allred, S. R. (2014). Shortterm memory affects color perception in context. PLoS ONE, 9(1), e86488, https://doi.org/10.1371/journal.pone.0086488. [PubMed]
Olkkonen, M., & Ekroll, V. (2016). Color constancy and contextual effects on color appearance. In Kremers, J., Baraas, R. C., & Marshall, N. J. (Eds.), Human color vision (pp. 159–188). Cham: Springer International Publishing, https://doi.org/10.1007/978-3-319-44978-4_6.
Olkkonen, M., McCarthy, P. F., & Allred, S. R. (2014). The central tendency bias in color perception: Effects of internal and external noise. Journal of Vision, 14(11), 5, https://doi.org/10.1167/14.11.5. [PubMed]
Parkkinen, J. P., Hallikainen, J., & Jaaskelainen, T. (1989). Characteristic spectra of munsell colors. Journal of the Optical Society of America A, 6(2), 318–322, https://doi.org/10.1364/JOSAA.6.000318.
Pastilha, R., Gupta, G., Gross, N., & Hurlbert, A. (2020). Temporal dynamics of daylight perception: Detection thresholds. Journal of Vision, 20(13), 18, https://doi.org/10.1167/jov.20.13.18. [PubMed]
Pelli, D. G. (1997). The videotoolbox software for visual psychophysics: Transforming numbers into movies. Spatial Vision, 10, 437–442, https://doi.org/10.1163/156856897X00366. [PubMed]
Pointer, M. R. (1981). A comparison of the CIE 1976 colour spaces. Color Research & Application, 6(2), 108–118, https://doi.org/10.1002/col.5080060212.
Rinner, O., & Gegenfurtner, K. R. (2000). Time course of chromatic adaptation for color appearance and discrimination. Vision Research, 40(14), 1813–1826, https://doi.org/10.1016/S0042-6989(00)00050-X. [PubMed]
Rutherford, M. D., & Brainard, D. H. (2002). Lightness constancy: A direct test of the illumination-estimation hypothesis. Psychological Science, 13(2), 142–149, https://doi.org/10.1111/1467-9280.00426. [PubMed]
Saarela, T. (2018). Shapetoolbox: Creating 3D models for vision research. Journal of Vision, 18(10), 229, https://doi.org/10.1167/18.10.229.
Shevell, S. K. (1978). The dual role of chromatic backgrounds in color perception. Vision Research, 18(12), 1649–1661, https://doi.org/10.1016/0042-6989(78)90257-2. [PubMed]
Smithson, H., & Zaidi, Q. (2004). Colour constancy in context: Roles for local adaptation and levels of reference. Journal of Vision, 4(9), 3, https://doi.org/10.1167/4.9.3.
von Helmholtz, H. (1924). Handbuch der physiologischen optik, vol. II. In J. P. C. Southall (Ed. & Trans.), Helmholtz's treatise on physiological optics (Southall, J. P. C., Ed. & Trans.; 3rd ed.). Rochester: The Optical Society of America. (Original work published 1911), https://hdl.handle.net/2027/mdp.39015067341381.
Wallach, H. (1948). Brightness constancy and the nature of achromatic colors. Journal of Experimental Psychology, 38(3), 310, https://doi.org/10.1037/h0053804. [PubMed]
Webster, M. A., Mizokami, Y., & Webster, S. M. (2007). Seasonal variations in the color statistics of natural images. Network: Computation in Neural Systems, 18(3), 213–233, https://doi.org/10.1080/09548980701654405.
Webster, M. A., & Mollon, J. D. (1997). Adaptation and the color statistics of natural images. Vision Research, 37(23), 3283–3298, https://doi.org/10.1016/S0042-6989(97)00125-9. [PubMed]
Wedge-Roberts, R., Aston, S., Beierholm, U., Kentridge, R., Hurlbert, A., Nardini, M., ... Olkkonen, M. (2020). Specular highlights improve color constancy when other cues are weakened. Journal of Vision, 20(12), 4, https://doi.org/10.1167/jov.20.12.4. [PubMed]
Weiss, D., Witzel, C., & Gegenfurtner, K. (2017). Determinants of colour constancy and the blue bias. i-Perception, 8(6), 2041669517739635, https://doi.org/10.1177/2041669517739635. [PubMed]
Werner, A. (2014). Spatial and temporal aspects of chromatic adaptation and their functional significance for colour constancy. Vision Research, 104, 80–89, https://doi.org/10.1016/j.visres.2014.10.005. [PubMed]
Witzel, C., & Gegenfurtner, K. R. (2018). Color perception: Objects, constancy, and categories. Annual Review of Vision Science, 4, 475–499, https://doi.org/10.1146/annurev-vision-091517-034231. [PubMed]
Xiao, B., Hurst, B., MacIntyre, L., & Brainard, D. H. (2012). The color constancy of threedimensional objects. Journal of Vision, 12(4), 6, https://doi.org/10.1167/12.4.6. [PubMed]
Yang, J. N., & Maloney, L. T. (2001). Illuminant cues in surface color perception: Tests of three candidate cues. Vision Research, 41(20), 2581–2600, https://doi.org/10.1016/S0042-6989(01)00143-2. [PubMed]
Figure 1.
 
Illustrations of the five different types of stimuli used in the experiment. (a) Object stimuli only include a target with an otherwise black screen. (b) Scene stimuli include a target and a background affected by changing scene illumination, either as a gray flat surface or additional, colored blobs.
Figure 1.
 
Illustrations of the five different types of stimuli used in the experiment. (a) Object stimuli only include a target with an otherwise black screen. (b) Scene stimuli include a target and a background affected by changing scene illumination, either as a gray flat surface or additional, colored blobs.
Figure 2.
 
Chromaticities of one of the seven stimulus sets used in stimulus rendering displayed on CIE uv′ chromaticity diagram. The u′-dimension is on the x-axis and v′-dimension on the y-axis. Red diamonds, blue crosses, and black circles show the 13 different illuminants, 13 different reflectances, and reflected light from all their combinations, respectively (note that multiple circles overlap). Green crosses show the reflectances of the four “context” blobs and green circles their reflected lights with the middle illuminant (D65).
Figure 2.
 
Chromaticities of one of the seven stimulus sets used in stimulus rendering displayed on CIE uv′ chromaticity diagram. The u′-dimension is on the x-axis and v′-dimension on the y-axis. Red diamonds, blue crosses, and black circles show the 13 different illuminants, 13 different reflectances, and reflected light from all their combinations, respectively (note that multiple circles overlap). Green crosses show the reflectances of the four “context” blobs and green circles their reflected lights with the middle illuminant (D65).
Figure 3.
 
The course of a single trial. The reference stimulus was displayed for 1,000 ms, followed by a 250-ms ISI, and the comparison stimulus for 1,000 ms. The observer was then shown one or two indicators with written responses alternatives. When the observer gave their response, the experiment moved to the next trial after a 500-ms ITI.
Figure 3.
 
The course of a single trial. The reference stimulus was displayed for 1,000 ms, followed by a 250-ms ISI, and the comparison stimulus for 1,000 ms. The observer was then shown one or two indicators with written responses alternatives. When the observer gave their response, the experiment moved to the next trial after a 500-ms ITI.
Figure 4.
 
All possible changes in stimulus colors between the reference and the comparison stimulus in scene blocks. In this example, the reference stimulus would be identical to the stimulus in the middle. The comparison stimulus could be one of the nine possibilities in the figure. Changes in illumination are on different columns, and changes in target reflectance are on different rows. The red lines illustrate that similar changes in the illumination or reflectance result in identical changes in the reflected light from the target.
Figure 4.
 
All possible changes in stimulus colors between the reference and the comparison stimulus in scene blocks. In this example, the reference stimulus would be identical to the stimulus in the middle. The comparison stimulus could be one of the nine possibilities in the figure. Changes in illumination are on different columns, and changes in target reflectance are on different rows. The red lines illustrate that similar changes in the illumination or reflectance result in identical changes in the reflected light from the target.
Figure 5.
 
Experiment protocol. (a) Measurement Session 1 included a preliminary measurement used to choose an initial stimulus set. The following practice trials were analyzed to adjust the stimulus set toward a target level of performance. In Sessions 2 and 3, all main experiment trial blocks were repeated twice. (b) In demo and practice blocks, the practice trials were preceded by demo trials and dummy trials. (c) In main experiment blocks, sets of main experiment trials (without feedback) were interleaved with sets of dummy trials (with feedback).
Figure 5.
 
Experiment protocol. (a) Measurement Session 1 included a preliminary measurement used to choose an initial stimulus set. The following practice trials were analyzed to adjust the stimulus set toward a target level of performance. In Sessions 2 and 3, all main experiment trial blocks were repeated twice. (b) In demo and practice blocks, the practice trials were preceded by demo trials and dummy trials. (c) In main experiment blocks, sets of main experiment trials (without feedback) were interleaved with sets of dummy trials (with feedback).
Figure 6.
 
Change identification results from object blocks. Heatmaps of response counts across all observers for (a) flat object and (b) 3D object. Each stimulus change is shown on a separate row. Left-hand icons indicate the correct color change response, and bottom icons indicate observers’ color change responses. (c) Plot discriminability measured by average d between the neutral and the “large” color differences. Different object blocks are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bar indicates ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average.
Figure 6.
 
Change identification results from object blocks. Heatmaps of response counts across all observers for (a) flat object and (b) 3D object. Each stimulus change is shown on a separate row. Left-hand icons indicate the correct color change response, and bottom icons indicate observers’ color change responses. (c) Plot discriminability measured by average d between the neutral and the “large” color differences. Different object blocks are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bar indicates ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average.
Figure 7.
 
Change identification results from scene blocks. Heatmaps of response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene. Each stimulus change is shown in a separate heatmap. Top icons indicate the correct reflectance change response and right icons the correct illumination change response. Left icons indicate observers’ illumination responses and bottom icons indicate observers’ reflectance responses. (d) Plot discriminability measured by average d′ across stimulus reflectance and illumination changes, respectively. Different scene blocks with reflectance and illumination plotted separately are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bars indicate ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average. **p < 0.01, *p < 0.05.
Figure 7.
 
Change identification results from scene blocks. Heatmaps of response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene. Each stimulus change is shown in a separate heatmap. Top icons indicate the correct reflectance change response and right icons the correct illumination change response. Left icons indicate observers’ illumination responses and bottom icons indicate observers’ reflectance responses. (d) Plot discriminability measured by average d′ across stimulus reflectance and illumination changes, respectively. Different scene blocks with reflectance and illumination plotted separately are on the x-axis with d on the y-axis. Each dot is a single observer, and the error bars indicate ±1 SEM. Yellow dots show averages across changes toward yellow, blue dots toward blue, and gray the total average. **p < 0.01, *p < 0.05.
Figure 8.
 
Identification of reflectance changes in scene blocks depending on identification of illumination changes. Response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene, cross-tabulated based on the correctness of the illumination response (row) and reflectance response (column).
Figure 8.
 
Identification of reflectance changes in scene blocks depending on identification of illumination changes. Response counts across all observers for (a) flat scene, (b) 3D matte scene, and (c) 3D glossy scene, cross-tabulated based on the correctness of the illumination response (row) and reflectance response (column).
Figure 9.
 
Discriminability of reflectance changes in scene blocks measured by average d in trials with correct and incorrect illumination change identification, respectively. Correct and incorrect illumination change identification for yellow, no, and blue illumination changes within flat, 3D matte, and 3D glossy scene blocks are on the x-axis with d on the y-axis. Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. ***p < 0.001.
Figure 9.
 
Discriminability of reflectance changes in scene blocks measured by average d in trials with correct and incorrect illumination change identification, respectively. Correct and incorrect illumination change identification for yellow, no, and blue illumination changes within flat, 3D matte, and 3D glossy scene blocks are on the x-axis with d on the y-axis. Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. ***p < 0.001.
Figure 10.
 
Marginal decision criteria and discrimination indices calculated across all observers for data binned by reference stimulus colors and reference-comparison color changes. Trials were binned according to whether the reference had illumination (a–d), or reflectance (e–h), from the yellowest, bluest, or neutral third in the stimulus pool (x-axis). Panels (a–b) and (e–f) show criterion (c, y-axis), and panels (c–d) and (g–h) show discriminability (d′, y-axis) for identification of illumination and reflectance changes, respectively. d′ and c were calculated between identification of no change and yellow or blue change (yellow and blue lines). Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. Note that criteria are displayed in absolute values, and a criterion below 1 means a bias to respond with the change indicated by the line color. ***p < 0.001, **p < 0.01.
Figure 10.
 
Marginal decision criteria and discrimination indices calculated across all observers for data binned by reference stimulus colors and reference-comparison color changes. Trials were binned according to whether the reference had illumination (a–d), or reflectance (e–h), from the yellowest, bluest, or neutral third in the stimulus pool (x-axis). Panels (a–b) and (e–f) show criterion (c, y-axis), and panels (c–d) and (g–h) show discriminability (d′, y-axis) for identification of illumination and reflectance changes, respectively. d′ and c were calculated between identification of no change and yellow or blue change (yellow and blue lines). Error bars display one standard deviation of the mean, calculated by bootstrapping with 10,000 iterations and taking the range of the middle 68th percentile. Note that criteria are displayed in absolute values, and a criterion below 1 means a bias to respond with the change indicated by the line color. ***p < 0.001, **p < 0.01.
Figure 11.
 
(a) The vectors depicting the reflectance and illumination mechanisms in stimulus space for the three scene conditions: flat, 3D matte, and 3D glossy. The more aligned a given vector is with the corresponding stimulus axis, the better the model observer is at identifying the changes, and the fewer confusions between the two types of stimulus change occur. Dots indicate the nine stimulus changes used in the experiment. (b) The angle of the mechanism vectors with respect to the reflectance axis. (c) Average d′-values of the model observer for identifying reflectance and illumination changes. (d–f) Response distributions for the nine types of stimulus change plotted in the decision space of the model (the axes correspond to the mechanism responses) in the flat (d), 3D matte (e), and 3D glossy (f) conditions. The vertical and horizontal lines show the response criteria. Ellipses signify half a standard deviation of the response distribution around the mean. First letter in each distribution indicates stimulus illumination change; second letter reflectance change. Color of the circle indicates absolute color change of the target.
Figure 11.
 
(a) The vectors depicting the reflectance and illumination mechanisms in stimulus space for the three scene conditions: flat, 3D matte, and 3D glossy. The more aligned a given vector is with the corresponding stimulus axis, the better the model observer is at identifying the changes, and the fewer confusions between the two types of stimulus change occur. Dots indicate the nine stimulus changes used in the experiment. (b) The angle of the mechanism vectors with respect to the reflectance axis. (c) Average d′-values of the model observer for identifying reflectance and illumination changes. (d–f) Response distributions for the nine types of stimulus change plotted in the decision space of the model (the axes correspond to the mechanism responses) in the flat (d), 3D matte (e), and 3D glossy (f) conditions. The vertical and horizontal lines show the response criteria. Ellipses signify half a standard deviation of the response distribution around the mean. First letter in each distribution indicates stimulus illumination change; second letter reflectance change. Color of the circle indicates absolute color change of the target.
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×