For the PA rating, our analyses aimed at three questions: 1. Do the PA ratings follow the same or a similar function of SOA as the objective measures that we obtained in the two alternative forced-choice task? 2. Do PA ratings increase as a result of training? 3. Do the PA ratings differ depending on accuracy? We reasoned that the mean PA rating for correct responses should be higher than the mean PA rating for errors before as well as after the training if there was a genuine increase in PA, whereas no difference in the PA rating for correct and incorrect responses would indicate biasing effects such as over- or underconfidence. These three questions where first addressed in a repeated measures ANOVA with factors session (first, second), accuracy (correct responses, errors), and SOA (20, 40, 60, 80, 100, 120, 140, 160 ms). We found a significant main effect of session (
F(1, 7) = 7.869,
p = 0.0263,
η 2 = 0.529), a significant main effect of accuracy (
F(1, 7) = 18.730,
p = 0.0034,
η 2 = 0.728), and an effect of SOA that only approached significance (
F(2.770, 19.391) = 3.021,
p = 0.0580,
η 2 = 0.301). Furthermore, we found a significant interaction of session and accuracy (
F(1, 7) = 21.890,
p = 0.0023,
η 2 = 0.758) as well as a significant interaction of accuracy and SOA (
F(3.524, 24.669) = 3.509,
p = 0.0249,
η 2 = 0.334), but no interaction of session and SOA (
F(3.662, 25.635) = 1.122,
p = 0.3652,
η 2 = 0.138), and no interaction of session, accuracy, and SOA (
F(3.686, 25.805) = 0.729,
p = 0.5702,
η 2 = 0.094). These results indicate that the mean PA ratings do not differ between SOAs, but that they do differ before and after training, and that this difference is modulated by accuracy. To further elucidate these results, we first split up our data set by session in order to investigate whether the mean PA rating was higher for correct than for incorrect responses both before as well as after the training. This was confirmed for the data from the first threshold by means of a repeated measures ANOVA with factors SOA (20, 40, 60, 80, 100, 120, 140, 160 ms) and accuracy (correct responses, errors). We found a significant effect of accuracy (
F(1, 7) = 9.321,
p = 0.0185,
η 2 = 0.571), but no effect of SOA (
F(3.574, 25.015) = 2.107,
p = 0.1158,
η 2 = 0.231), and no interaction (
F(3.551, 24.858) = 0.771,
p = 0.5406,
η 2 = 0.099). Thus, the PA rating was sensitive to variations in accuracy at each SOA before the training, showing higher PA for correctly identified stimuli than for incorrectly identified stimuli (mean difference 0.1056,
SE 0.35). The same analysis for the PA scores of the second threshold yielded identical results: We found a significant effect of accuracy (
F(1, 7) = 23.6,
p = 0.0018,
η 2 = 0.771), but no effect of SOA (
F(3.891, 27.237) = 1.858,
p = 0.1480,
η 2 = 0.210), and no interaction (
F(3.177, 22.236) = 2.352,
p = 0.0967,
η 2 = 0.252). Thus, the PA scores retained their sensitivity for accuracy even after the training, with higher scores for correct responses than for errors (mean difference 0.2482,
SE 0.51). The results from the first and second threshold are illustrated in
Figures 5B and
5C.