We used GANalyze with AestheticsNet (
Kong et al., 2016) as the assessor and BigGAN-256 (
Brock, Donahue, & Simonyan, 2019) pretrained on ImageNet (
Russakovsky et al., 2015) as the generator to train the GANalyze model for 400,000 iterations. To do this, we used Python 3.6 with the TensorFlow libraries (
Abadi et al., 2016). The training resulted in GANalyze being able to produce image sequences based on 1,000 ImageNet categories. For each ImageNet category, we initially generated three different seeds (e.g., three different images of Siamese cats) to test for an effect of idiosyncratic effects of images on top of semantic category. Within each seed, GANalyze produced 21 images with a supposedly increasing amount of aesthetic value (
Figure 3). We picked a broad range of
α-values with increasing density close to zero to obtain stimuli with obvious changes (e.g.,
α = 0.25) but also subtle changes (e.g.,
α = 0.0025). We chose these values because we hypothesized there to be a ceiling effect where there would be no behavioral differences for extremely high
α-values due to oversaturation. We also hypothesized that for very low
α-values, the perceptual difference between stimuli would be too small to detect. The specific
α-values used in the final experiment were chosen based on the results of a pilot study with a wider variety of
αs. This pilot followed the same general structure as the full experiment, which is explained below. The main difference in the pilot is the use of the unfiltered data set with 1,000 ImageNet categories, 21
α-values, and 3 seeds. This pilot showed us that it would be justified to only use one of the three seeds for each category. More details concerning this pilot study are reported in the Results section.