Our primary hypothesis was that emotion recognition accuracy would be affected by both kinematic and spatial manipulation and that these effects would not interact with group. To test this hypothesis we conducted a mixed 2 x 3 x 3 x 3 ANOVA with the between-subjects factor group (ASD, control) and the within-subjects factors emotion (happy, angry, sad), stimulus spatial level (S1, S2, S3), and stimulus kinematic level (K1, K2, K3). This analysis revealed a significant main effect of emotion [F(2,116) = 17.79, p < .001, ηP2= .24, BF10 = 1.03e15; see Supplementary Materials B], a main effect of spatial level [F(2,116) = 259.57, p < .001, ηP2= .82, BF10 = 9.05e57; see Supplementary Materials B] which was qualified by an emotion x spatial interaction [F(4,232) = 88.42, p < .001, ηP2= .60, BF10 = 7.53e58], and an emotion x kinematic interaction [F(4,232) = 53.90, p < .001, ηP2= .48, BF10 = 1.90e20]. Furthermore, this analysis revealed a significant four-way emotion x spatial x kinematic x group interaction [F(8,464) = 2.438, p < .05, ηP2= .04, BF10 = 0.07]. Note that no kinematic x group interaction was found [p = .538, BF10 = 0.02], suggesting that autistic and control participants exhibit similar patterns of accuracy across the kinematic levels. Below, in order to shed light on the effects of the spatial and kinematic manipulations, we first unpack the emotion x kinematic and emotion x spatial interactions. Subsequently we fully unpack the emotion x spatial x kinematic x group interaction.
In line with Sowden et al., (23), we observed an emotion x spatial interaction [F(4,232) = 88.42, p< .001, ηP2= .60, BF10 = 7.53e58]. Post-hoc repeated measures ANOVAs revealed that whilst the effect of the spatial manipulation was present for all three emotions (all F > 7.00, all p < .01), the direction of the effect varied between high and low arousal emotions: recognition scores for angry and happy videos were highest for 150% spatial extent (S3) [angry mean (Standard Error of the Mean; SEM) = 5.21(.21); happy mean(SEM) = 5.70(.24)], followed by 100% spatial extent (S2) [angry mean(SEM) = 3.15(.22); happy mean(SEM) = 4.75(.23)], and finally 50% spatial extent (S1) [angry mean SEM) = 0.53(.22); happy mean(SEM) = 2.10(.25)]. In contrast, for sad videos, recognition scores were highest for S1 [sad mean(SEM) = 3.50(.22)], lowest for S3 [sad mean(SEM) = 2.78(.22)] and intermediate for S2 [sad mean(SEM) = 3.15(.20); Figure 2.]. This pattern matches the results reported by Sowden et al., (23) for neurotypical participants.
In addition, our analysis identified an emotion x kinematic interaction [F(4,232) = 53.90, p < .001, ηP2= .48, BF10 = 1.90e20]. Whilst there was a main effect of the kinematic manipulation for all three emotions (all F > 20, all p < .001), the direction of the effect differed between high and low arousal emotions. For angry and happy videos, emotion recognition improved with increasing speed [angry: K1 mean(SEM) = 2.28(.19); K2 mean(SEM) = 2.87(.19); K3 mean(SEM) = 3.73(.23); happy: K1 mean(SEM) = 3.50(.23); K2 mean(SEM) = 4.50(.22); K3 mean(SEM) = 4.55(.21)]. For sad videos, emotion recognition improved as speed decreased [K3 mean(SEM) = 2.03(.19); K2 mean(SEM) = 3.21(.22); K1 mean(SEM) = 4.18(.23); Figure 2. Bottom panel]. This pattern of results also matches the findings from Sowden et al. (23).[1]
In order to unpack the significant four-way interaction, we conducted post-hoc 2 x 3 x 3 (group, emotion, kinematic) ANOVAs for each spatial level. This analysis revealed a significant
emotion x kinematic x group interaction at the S2 [F(4,232) = 4.53, p < 0.01, ηP2= .07, BF10 = 5.92] but not S1 [p = .265, BF10 = 0.09] or S3 [p = .208, BF10 = 0.09] level. To unpack this emotion x kinematic x group interaction at the S2 level, we conducted separate post-hoc ANOVAs for each kinematic level at the 100% (S2) spatial level. This analysis revealed a significant emotion x group interaction at the K2 [F(2,116) = 6.48, p < .01, ηP2 = .10, BF10 = 17.09] but not K1 [p = .244, BF10 = 0.32] or K3 [p = .082, BF10 = 0.82] level. Bonferroni-corrected post-hoc independent sample t-tests revealed that control, relative to ASD, participants had higher accuracy for angry videos at the 100% spatial (S2) and speed (K2) level [t(58) = 2.78, pbonf. < .05, mean difference = 1.48, BF10 = 6.09; Figure 3.]. There were no significant group differences in emotion recognition accuracy for happy [p = .757, BF10 = 0.27] or sad [p = .085, BF10 = 0.93] videos at the S2K2 level. Thus, the groups significantly differed in accuracy for angry PLFs that were not spatially or kinematically manipulated.
To further unpack the emotion x kinematic x group interaction at the S2 level, we conducted separate post-hoc ANOVAs for each emotion at the S2 level. This analysis identified a significant kinematic x group interaction for angry [F(2,116) = 4.59, p < .05, ηP2= .07, BF10 = 3.49] but not happy [p = .070, BF10 = 0.95] or sad [p = .123, BF10 = 0.53] PLFs. Therefore, for angry videos at the normal spatial level, the effect of the kinematic manipulation varied as a function of group. Bonferroni-corrected paired sample t-tests demonstrated that whilst the control group exhibited increasing accuracy across all kinematic levels [K1-K2: t(28) = -4.31, pbonf < .001, mean difference = -1.62, BF10 = 153.77; K2-K3: t(28) = -2.86, pbonf < .05, mean difference = -0.95, BF10 = 5.52], the ASD group only showed improvement from the K2 to K3 [t(30) = -3.46, pbonf < .01, mean difference = -1.16, BF10 = 21.10] and not K1 to K2 [p = .865, BF10 = 0.19; Figure 4.]. Furthermore, the groups did not significantly differ at K1 (F(1,58) = .18, p > .05) or K3 (F(1,58) = 3.53 p > .05) but at K2, controls out-performed autistic participants (F(1,58) = 7.75, p < 0.01, ηP2= .12). These results suggest that, whilst controls improved in their accuracy for angry PLF stimuli across each level of increasing kinematic manipulation, for autistic participants, only the most extreme (K3) level of the kinematic manipulation resulted in an accuracy boost.
Multiple Regression Analyses
Our second hypothesis was that variation in emotion recognition accuracy would covary, not with ASD symptomatology but with scores on the self-report alexithymia scale (TAS-20). To test whether autistic or alexithymic traits were predictive of the effect of the spatial and kinematic manipulations, we conducted two multiple regression analyses. For the first analysis, we used the effect of spatial manipulation (defined as the difference in accuracy between S3 and S1) as the dependent variable (DV) and AQ and TAS-20 as predictor variables. This analysis resulted in a non-significant model overall [F(2,57) = .87, p= .425], neither AQ [standardized β = -.17, t(57) = -1.10, p = .274] nor TAS-20 [standardized β = .19, t(57) = 1.20, p = .236] were significant predictors of the effect of the spatial manipulation. In the second analysis, we used the effect of the kinematic manipulation (defined as the difference in accuracy between K3 and K1) as the DV and AQ and TAS-20 as predictors. Again, this analysis resulted in a non-significant model [F(2,57) = 1.63, p = .206], neither AQ [standardized β = .20, t(57) = 1.33, p = .189] nor TAS-20 [standardized β = .05, t(57) = .32 p = .752] were significant predictors of the effect of the kinematic manipulation. We then conducted a third multiple regression with mean emotion recognition accuracy (across all trials) as the DV. Once again, neither AQ [standardized β = -.19, t(57) = -1.24, p = .220] nor TAS-20 [standardized β = .12, t(57) = .81, p = .424] were significant predictors of mean recognition accuracy and the overall model did not explain a significant amount of variance in the data [F(2,57) = .78, p = .461]. To explore the possibility that only extreme scores on the TAS-20 predict performance, we compared mean accuracy for alexithymic (i.e. TAS-20 ≥ 61) and non-alexithymic (i.e. TAS-20 ≤ 51) participants (according to the cut-off scores outlined by Bagby, Taylor and Parker; 31), excluding ‘possibly alexithymic’ individuals. An independent samples t-test confirmed that there was no significant difference in mean accuracy between these groups [t(48) = -.18, p = .861, mean difference = -.05, BF10= 0.29].
Finally, building on our previous observation that the ASD and control groups differed in accuracy for angry videos at the normal (100%) spatial and speed level we conducted a multiple regression analysis to identify the extent to which autistic and alexithymic traits were predictive of accuracy for angry videos at the S2 and K2 levels. This analysis revealed that autistic [standardized β = -.44, t(57) = -3.05, p < .01], but not alexithymic [standardized β = .22, t(57) = 1.54, p = .130], traits were predictive of accuracy for angry videos at the normal spatial and speed level [overall model statistics: F(2, 57) = 4.67, p < .05, R2 = .141]. Bayesian analyses revealed that AQ [BFinclusion = 4.230] was over 16 times more likely to be included in a model to predict accuracy for angry videos at the normal spatial and speed level than alexithymic traits [BFinclusion = 0.263].
In order to ensure that AQ is not just a significant predictor of accuracy for angry expressions at the normal spatial and speed level due to variation across other co-variables (e.g. age, gender, and non-verbal reasoning), we completed an additional three-step forced entry hierarchical regression analysis following the procedures of Cook et al., (18). In the first step, the demographic variables (gender, age and NVR) were entered into the model, which overall accounted for 16% of the variance in accuracy at the S2K2 level [F(3,56) = 3.56, p < .05, R2 = .160]. Importantly, of the three demographic variables, only NVR was a significant predictor of accuracy for angry videos at the normal spatial and speed level [standardized β = .35, t(56) = 2.79, p < .01] (and not gender [standardized β = .15, t(56) = 1.20, p = .233] or age [standardized β = - .01, t(56) = -.06, p = .950]). In the second step, AQ was added [standardized β = -.36, t(55) = -3,13, p < .01], producing a statistically significant R2 change [F change (1, 55) = 9.80, p < .01, R2 change = .127]. Finally, when TAS-20 was entered into the model, the analysis revealed it was not a significant predictor of accuracy for angry videos at the normal level [standardized β = .17, t(54) = 1.26, p = .214] and resulted in a non-significant R2 change [F change (1, 54) = 1.58, p = .214, R2 change = .020; see Table 2.]. Hence, this analysis demonstrated that AQ (and not TAS-20) was a significant predictor of accuracy for angry videos at the normal level (S2, K2) even after age, gender and NVR have been accounted for.
These analyses suggest that alexithymia accounts for very little variance in accuracy for angry videos at the normal (S2K2) level once autistic traits have been accounted for. However, since our autism and alexithymia measures were correlated [R = .53, p < .001], when alexithymia is entered into a multiple regression after autistic traits, it may not be a significant predictor due to multicollinearity. Consequently, we ran one further hierarchical regression, with the demographic variables entered in Step 1, alexithymia in Step 2 and autistic traits in Step 3. Alexithymia failed to significantly improve the model [F change (1, 55) = .31, p = .581, R2 change = .005], explaining only 0.5% more variance than that explained by the demographic variables alone. Despite being highly correlated with alexithymia, autistic traits were again a significant predictor of accuracy for angry videos at the normal level [standardized β = -.45, t(54) = -3.33, p < .01] when added to the model in Step 3. Adding autistic traits at this step produced a statistically significant R2 change [F change (1, 54) = 11.12, p < .01, R2 change = .143], explaining an additional 14.3% of the variance in accuracy.
Table 2. Results of the forced entry hierarchical regression for accuracy for angry videos at the normal spatial and speed level. 1. predictors: age, gender, non-verbal reasoning; 2. predictors: age, gender, non-verbal reasoning, AQ; 3. predictors: age, gender, non-verbal reasoning, AQ, TAS-20
Model
|
R
|
R2
|
Adjusted R2
|
SEE
|
R2 change
|
F change
|
Sig. F change
|
1
|
.400
|
.160
|
.115
|
1.82
|
.160
|
3.556
|
.020
|
2
|
.536
|
.287
|
.235
|
1.69
|
.127
|
9.798
|
.003
|
3
|
.554
|
.307
|
.243
|
1.68
|
.020
|
1.581
|
.214
|