PreprintPDF Available

Modality-specific dysfunctional neural processing of social and non-social information in schizophrenia

Authors:

Abstract and Figures

Background: Schizophrenia (SZ) is characterized by marked social dysfunctions encompassing potential deficits in the processing of social and non-social information, especially in everyday settings where multiple modalities are present. To date, the neurobiological basis of these deficits remains elusive. Methods: In a functional magnetic resonance imaging (fMRI) study, 17 patients with schizophrenia or schizoaffective disorder, and 18 matched controls watched videos of an actor speaking, gesturing (unimodal), and both speaking and gesturing (bimodal) about social or non-social events in a naturalistic way. Participants had to judge whether each video contains person-related (social) or object-related (non-social) information. Results: When processing social content, controls activated the medial prefrontal cortex (mPFC) for both speech and gesture conditions; patients, in comparison to controls, showed no different activation in the speech condition but reduced activation in the mPFC in the gesture condition. For non-social content, across modalities, controls recruited the bilateral pre/postcentral gyrus, superior temporal gyrus, and insula, as well as the left occipitotemporal cortex; patients showed reduced activation of the left postcentral gyrus and the right insula only in the speech condition. Moreover, in the bimodal conditions, patients displayed improved task performance and comparable activation to controls in both social and non-social content. Conclusions: Patients with SZ displayed modality-specific aberrant neural processing of social and non-social information, which is not present for the bimodal conditions. This finding provides novel insights into dysfunctional social cognition in SZ, and may have potential therapeutic implications.
Content may be subject to copyright.
Modality-specific dysfunctional neural processing of social and non-social
information in schizophrenia
Yifei He1,4*, Miriam Steines1,4, Gebhard Sammer2, Arne Nagels3, Tilo Kircher1,4, Benjamin
Straube1,4
1. Department of Psychiatry and Psychotherapy, Philipps-University Marburg, Marburg,
Germany
2. Cognitive Neuroscience at Centre for Psychiatry, Justus-Liebig University Giessen,
Giessen, Germany
3. Department of General Linguistics, Johannes-Gutenberg University Mainz, Mainz,
Germany
4. Center for Mind, Brain and Behavior - CMBB, Hans-Meerwein-Straße 6, 35032 Marburg,
Germany
*Corresponding address: Yifei He, Department of Psychiatry and Psychotherapy, Philipps
University Marburg, Rudolf-Bultmann-Str. 8, 35039, Marburg, Germany
* Corresponding email: yifei.he@staff.uni-marburg.de, ORCID: 0000-0002-2826-4230
1
Abstract
Background: Schizophrenia (SZ) is characterized by marked social dysfunctions
encompassing potential deficits in the processing of social and non-social information,
especially in everyday settings where multiple modalities are present. To date, the
neurobiological basis of these deficits remains elusive.
Methods: In a functional magnetic resonance imaging (fMRI) study, 17 patients with
schizophrenia or schizoaffective disorder, and 18 matched controls watched videos of an
actor speaking, gesturing (unimodal), and both speaking and gesturing (bimodal) about
social or non-social events in a naturalistic way. Participants had to judge whether each
video contains person-related (social) or object-related (non-social) information.
Results: When processing social content, controls activated the medial prefrontal cortex
(mPFC) for both speech and gesture conditions; patients, in comparison to controls,
showed no different activation in the speech condition but reduced activation in the mPFC
in the gesture condition. For non-social content, across modalities, controls recruited the
bilateral pre/postcentral gyrus, superior temporal gyrus, and insula, as well as the left
occipitotemporal cortex; patients showed reduced activation of the left postcentral gyrus
and the right insula only in the speech condition. Moreover, in the bimodal conditions,
patients displayed improved task performance and comparable activation to controls in both
social and non-social content.
Conclusions: Patients with SZ displayed modality-specific aberrant neural processing of
social and non-social information, which is not present for the bimodal conditions. This
finding provides novel insights into dysfunctional social cognition in SZ, and may have
potential therapeutic implications.
Key words: social, multimodal processing, mPFC, gesture, speech, mentalizing,
schizophrenia
2
Introduction
Dysfunctional social cognition is a hallmark feature of schizophrenia (SZ) (1, 2). Social
cognition entails subdomains ranging from perception of social information to mentalizing
and social interaction (3). Patients with SZ, however, commonly suffer from deficits in these
processes. These deficits, correlated with negative symptoms (4), contribute directly to
impairments in social functioning (5). Among these processes, an important field in SZ
research is the processing of social-relevant stimuli and its underlying neurobiological
mechanisms.
In daily social communication, individuals encounter a diverse spectrum of information
from multiple modalities (6). These include non-linguistic inputs from others’ facial
expressions, body movements, e.g., postures and gestures, as well as linguistic stimuli
from auditory speech and written texts. Importantly, the multimodal information comprises
social and non-social information to individuals, and the processing of this information forms
the basis to further mentalize social intentions, and to perform appropriate social interaction
with others (7). To date, despite substantial reports of deficits in social processes in SZ (2),
neuroscientific studies on SZ’s social dysfunction have most extensively investigated
aberrant emotional perception of faces (8), and to a lesser degree, voices (9), while limited
research has directly examined potentially aberrant perception of social (person-related)
and non-social (object-related) information. In basic research, a seminal fMRI (functional
magnetic resonance imaging) study on social information processing, using linguistic
stimuli, has identified distinct brain regions for processing socially and non-socially relevant
information (10): when healthy participants were asked to judge whether visual word pairs
are person- or object-related, person-related social stimuli activated the medial prefrontal
cortex (mPFC), a crucial region forming the mentalizing network (7, 11-13). This functional
relevance of the mPFC for processing social information has been replicated in later studies
exploiting comparable tasks (14, 15). In addition, non-social vs. social content comparison
elicited a network including the bilateral insula and the left parietal lobe, these are regions
typically reported for processing concrete objects or tools (16, 17). Importantly, this line of
research has primarily examined linguistic stimuli such as single written words, largely
3
neglecting the multimodal nature of social and non-social information in everyday life: for
instance, marked social or non-social features can be delivered by either a “be silent”
emblem or a “hammeringpantomime via hand gestures, without providing any linguistic
information. Moreover, it is common to use both gesture and speech together. For example,
to ask someone else to stop, we often use a “stop” gesture (e.g., a raised hand) together
with its verbal counterpart.
Interestingly, irrespective of encoding modality, processing of social information is
shown to consistently activate a mentalizing network including the mPFC, at least in healthy
individuals (13, 18). For example, It is reported that a left-lateralized network, including the
mPFC, is activated when processing social-abstract information encoded in both auditory
speech and visual gestures (18). This ‘supramodal’ nature of social information processing
further concurs with the role of the mPFC (and the mentalizing network) for a wide range of
social tasks based on linguistic and non-linguistic stimuli (19). For non-social concrete
information, literature also suggests that humans may recruit brain networks that is modality
independent (20, 21). These characteristics of social and non-social information processing
may have profound implications in SZ research: given the well-documented deficits of SZ in
social processes (2, 22), it remains unknown whether social information processing is
impaired in SZ, and if so, whether this potential deficit is modality dependent. Similarly, for
processing of concrete, object-related non-social information, previous literature suggests
that patient’s processing may be impaired at least in visually presented linguistic form (23).
In the form of hand action, however, the reports on SZ’s potential neural deficits are mixed
(24, 25). Importantly, to date, no prior study has directly compared social/non-social
information processing in speech and gesture in SZ.
The current study directly addresses these gaps. We presented videos of an actor
communicating in a spontaneous and naturalistic manner. The actor performs either social
(person-related) or non-social (object-related) content in different modalities, where social
and non-social features are perceivable in gesture- and speech-only modalities. Similar to
approaches from previous research (10, 18), we directly compared social vs. non-social
videos, so as to identify neural perception of social and non-social information in both
speech and gesture modalities. Besides, we also showed to participants videos with
4
bimodal inputs (actor both speaking and gesturing). Based on previous basic research, we
hypothesized that the mentalizing network supports the processing of social (person)
information (10), and that non-social (object) information processing will activate left-
lateralized regions including the lateral occipitotemporal cortex (LOTC), the superior
temporal gyrus/sulcus (STG/STS), as well as pre/postcentral gyri forming the putative
mirror neuron system (26-29). We focused on group differences between a group of
patients suffering from schizophrenia or schizoaffective disorder, and their age- and
education-matched controls: for processing social content, we expected patients to show
reduced activation in the mentalizing regions, irrespective of encoding modality (22); for
non-social content, despite mixed findings from previous neuroimaging research on hand
action observation on SZ (24, 25), following previous report on dysfunctional processing of
non-social linguistic stimuli in SZ (23), we hypothesized neural modulation of the object-
related regions for patients with SZ for both gesture and speech modalities. Additionally, we
hypothesized that the presence of bimodal content might compensate for those unimodal
deficits, leading to improved performance and similar neural processing to the control
group.
5
Methods
Participants
We summarized participants’ demographic and clinical characteristics in Ta b le 1 .
Seventeen patients were recruited at the Department of Psychiatry and Psychotherapy at
the University of Marburg, and were diagnosed according to ICD-10 with schizophrenia
(F20.0, n=13, and F20.3, n=1) or schizoaffective disorder (F25.0, n=2, and F25.3, n=1).
Participants in both groups are native speakers of German, and have no knowledge of
Russian language. All except one of the patients received antipsychotic treatment; six were
additionally treated with antidepressive medication. Positive and negative symptoms were
assessed with the Scale for the Assessment of Positive Symptoms (SAPS) (30), and the
Scale for the Assessment of Negative Symptoms (SANS) (31). Eighteen age- and
education-matched healthy participants with no history of any mental disorders were
recruited from the same area. Exclusion criteria for both groups were brain injury and
neurological or other medical diseases affected by brain physiology. In both groups, we
conducted neuropsychological tests to assess working memory function, digital span, trail
making (TMT), verbal IQ (MWT-B) (32), and metaphoric language processing (concretism,
evaluated with the Proverb Interpretation Task) (33). These measures are reported in Ta b le
1. We report, additionally, scores from the subscales of SAPS and SANS, word fluency test,
as well as gesture production and perception (BAG, Brief Assessment of Gesture (34)) in
the supplement (Table S1). All participants had normal or corrected-to-normal vision and
hearing. Except for one control and one patient, all other participants are right-handed. All
participants gave written informed consent prior to participation in the experiment and were
compensated monetarily. The study was approved by the local ethics committee.
Table 1. Demographic, medication, symptom, and neuropsychological measures.
Patients (n=17)
Controls (n=18)
Age (years)
33.12 (12.35)
31.94 (10.21)
Gender male/female
13/4
13/5
Education (years)
11.82 (1.77)
12.72 (1.36)
TMT A (seconds)
31.49 (10.73)
26.17 (9.89)
TMT B (seconds)
68.56 (37.8)
52.93 (19.58)
6
Digit Span forward
7.94 (1.75)
8.05 (2.43)
Digit Span backward
6.35 (1.93)
6.61 (2.50)
Verbal IQ
28.8 (5.25)
28.5 (3.79)
*Concretism
1.38 (0.45)
1.14 (0.19)
SAPS (global)
15 (6.89)
SANS (global)
9 (6.02)
CPZ Equivalent
562.52 (372.63)
Values are presented as mean (SD). TMT: trail making test; CPZ: chlorpromazine.
Asterisk * indicates significant difference between controls and patients (p<0.05, two-
tailed t-test).
Materials and procedure
We employed a content judgement paradigm from a previous study (18), to investigate
multimodal (speech and gesture) processing of social and non-social information. We
showed to participants five-second videos of an actor spontaneously communicating both
social (S) and non-social (N) events in the following modalities: 1) incomprehensible
Russian sentences with gestures. This is considered as a gesture-only (G) condition
because social feature is only available to participants in the gesture form. 2)
comprehensible German sentences (S) without any gestures. Additionally, we also showed
to participants 3) German sentences with accompanying gestures as a bimodal input
condition (B). A filler condition is also included with videos of incomprehensible Russian
sentences with meaningless gestures. An example of both a social (S) and non-social (N)
bimodal videos is illustrated in Figure 1A. For a complete list of all videos, please refer to
Appendix in (35).
7
Figure 1. Panel A: Picture illustration for social (S) and non-social (N) videos in the bimodal condition (B). The
same stimuli were also presented in two additional modalities: gestures with foreign Russian sentences (G)
and German sentences without any gestures (S). For illustrative purposes, the spoken German sentences
were translated into English, and all spoken sentences were written into speech bubbles. Panel B: Illustration
of a sample trial. Participants performed a content judgment task for each video, indicating via button press
whether a stimulus was either person- or object-related. The face of the actor is masked in the manuscript to
avoid inclusion of identifying information of people. It is displayed to the participants during the experiment.
Experimental procedure
Altogether, 312 experimental video slips (26 videos per condition × 6 conditions × 2
sets) were included in the study. For each participant, an experimental session comprised
8
182 videos from one set of videos (156 critical videos and 26 filler videos), and consisted of
two 14-min runs. Each run contained 91 trials with a matched number of items from each
condition. The stimuli were presented in an event-related design in pseudo-randomized
order, and were counterbalanced across participants. Within each trial, each video-clip was
followed by a gray background with a variable duration of 2154-5846 ms (jitter average:
4000 ms), as illustrated in Figure 1B. Participants performed a content judgement task for
each video (10, 18), indicating via button press (with their left hand) whether a stimulus was
either person- or object-related. Participants were instructed to respond to the task as soon
as they had decided on an answer.
fMRI acquisition and preprocessing
All images were acquired using a 3T MRI scanner (Siemens MRT Trio series). The
functional images were obtained using a T2*-weighted echo-planar image sequence (TR =
2 s, TE = 30 ms, flip angle = 90°, slice thickness = 4 mm, interslice gap= 0.36 mm, field of
view= 230 mm, matrix = 64 × 64, voxel size = 3.6 x 3.6 x 4.0 mm, 30 axial slices orientated
parallel to the AC-PC line, ascending order). Two runs of 425 volumes each were acquired
during the experiment. Additionally, simultaneous EEG data from the participants were also
collected for other analyses not relevant for the current study, and are therefore not further
discussed here. MR images were preprocessed using the SPM12 software package
(Statistical Parametric Mapping, Welcome Trust Center for Neuroimaging, London, UK)
based on Matlab R2017a (version 9.2.0; MathWorks): after discarding the first five volumes
to minimize T1-saturation effects, all images were spatially and temporally realigned, and
normalized into the MNI space using the MNI template (resulting voxel size 2 × 2 × 2 mm),
smoothed (8 mm isotropic Gaussian filter), and high-pass filtered (cut-off period 128 s).
fMRI data analysis
We performed statistical whole-brain analysis in a two-level, mixed-effects procedure.
On the first level, single-participant BOLD responses were modeled by a design matrix
comprising the onset time points of each event (critical word of each sentence as used in
the previous event-related fMRI and EEG studies, e.g., (18, 35-38)), with a duration of 5
9
seconds for all experimental conditions. The micro-time onset was set to the average time
bin (8 of 16) to align the onset vector to the slice in the middle of the brain. For all
conditions, the duration of speech or gesture was used as parameters of no interests on a
single trial level. Six movement regressors (three rotations and three translations) were
entered in the single participant’s model to account for movement-induced effects on fMRI
results. HRF was defined as the canonical HRF. Contrasts images against implicit baseline
for all experimental conditions were used as summary measures and were included in the
between-group analysis. We applied a flexible factorial analysis of variance using condition
as main effect. To de te rm in e th e cluster extent threshold to correct for multiple
comparisons, we applied a Monte-Carlo simulation following Slotnick et al., (39, 40). For all
statistical comparisons, the whole-brain activation was simulated assuming a voxel type-I
error voxel activation of p<.05, this revealed a cluster extent of 2268 contiguous resampled
voxels as sufficient to correct for multiple comparisons at p<.0167 (Bonferroni-corrected for
three modalities). The reported voxel coordinates of activation peaks are located in MNI
space. For the anatomical localization, functional data were referenced to the AAL toolbox
(41).
For both groups, we firstly reported contrast images comparing the processing of social
vs. non-social conditions (S>N and N>S) within each modality for each group. Secondly, for
each modality, we performed interaction analyses to investigate group differences in the
processing of social or non-social conditions. Lastly, we tested the three-way interaction of
group*modality*content, and performed conjunction analyses between this contrast and
contrasts from the last step. This step revealed modality-specific group differences for the
processing of social and non-social content, which is reported in the results section.
Additionally, for patients, we tested the interaction between modalities on social vs. non-
social content processing, so as to reveal how bimodal stimuli might compensate potential
neural processing deficits for patients with SZ. This is reported in Supplement S2.
Based on the literature showing a potential relationship between symptom severity
(especially negative symptoms) and social/non-social cognition (4, 42), for patients with SZ,
we explored the relationship between clinical measures and brain activation in areas that
are relevant to social/non-social information processing. We conducted explorative
10
correlation (spearman) analyses between 1) parameter estimates from clusters showing
significant group difference for either social or non-social conditions, 2) behavioral
measures (reaction times and accuracy) for each experimental condition, and 3) scores
from sum/general and subscales of SAPS and SANS.
11
Results
Behavioral results
Descriptive and inferential statistics from the content judgement task are reported in
detail in the supplement (S1 and Table S2). In general, in the content judgement task,
patients responded slower and were less accurate than controls. Additionally, task accuracy
for the bimodal condition was higher than for other modalities in both patient and controls.
However, we observed no group interaction with either content or modality manipulations.
fMRI results
Social>Non-social: we report whole-brain fMRI results for S>N comparisons in Figure 2
and Tab l e S3 in the supplement. For the speech conditions (SS>NS), healthy controls
activated an extensive fronto-temporal-parietal network including the bilateral inferior frontal
gyrus (IFG) and the temporal lobe, the dorsolateral prefrontal cortex (dlPFC) and mPFC,
and the left supramarginal gyrus; patients revealed similar regions for this comparison, and
we observed no group difference for social > non-social speech. For the gesture conditions
(SG>NG), controls activated the bilateral PFC and IFG; patients activated the bilateral
prefrontal cortex. Group interaction (Control (SG>NG)>Patient (SG>NG)) suggests that
patients showed reduced activation in the mPFC and the anterior cingulate cortex for the
social gesture condition when compared to controls (Figure 2B). In the bimodal condition
(SB>NB), both controls and patients activated regions similar to that of the bimodal
condition. For patients, we additionally reported modality*content interaction in the
Supplement S2, which shows that patients’ aberrant processing of social gestures is
enhanced in the bimodal modality.
12
Figure 2. Activation maps for social>non-social videos (S>N). Panel A: S>N contrasts within each modality (S:
Speech, G: Gesture, B: bimodal) for controls and patients. Panel B: interaction analysis (Control>Patient) in
the gesture modality (SG>NG) together with box- and swarm-plots of eigenvariates for selected clusters.
Non-social>Social: we report whole-brain fMRI results for N>S comparisons in Figure 3
and Tab l e S4 in the supplement. For the speech conditions (NS>SS), healthy controls
activated the left pre/postcentral gyrus, supramarginal gyrus, and the left insula, whereas
patients did not reveal any significant activations for this comparison. The group interaction
(Control (NS>SS)>Patient (NS>SS)) suggests that, when compared to controls, patients
showed reduced activation in the left postcentral gyrus and the right insula for the
processing of non-social content in the speech-only modality (see Figure 3B). For gesture
conditions (NG>SG), controls showed increased activation for the non-social content in the
bilateral posterior temporal gyrus, supramarginal gyrus, and occipital cortices, as well as
the left pre/postcentral gyrus and the left insula. Patients also activated the bilateral
posterior temporal gyrus and occipital lobe, as well as the left pre/postcentral gyrus. The
group interaction revealed no significant clusters. For bimodal conditions (NB>SB), both
controls and patients activated regions that are comparable to that of the gesture
conditions. Additionally, for patients, bimodal input seems to enhance their aberrant
13
processing of non-social speech, as reported in Supplement S2.
Figure 3. Activation maps for non-social> social videos (N>S). Panel A: N>S contrasts within each modality
(S: Speech, G: Gesture, B: Bimodal) for controls and patients. Panel B: interaction analysis (Control>Patient)
in the speech modality (NS>SS) together with box- and swarm-plots of eigenvariates for selected clusters.
Correlation analyses
In patients, for the NS condition, we found that the accuracy for the NS condition
correlate negatively with the SANS composite scores of the patients (r = -0.52, p = 0.03,
power = 0.63; Figure 4A). Additionally, SANS 1 (flat affect) and SANS 2 (alogia) scores
correlate negatively with the accuracy for the NS condition (SANS 1: r = -0.62, p = 0.008,
power = 0.79, Figure 4B; SANS 2: r = -0.63, p = 0.006, power = 0.82, Figure 4C).
14
Figure 4. Significant negative correlations between patients’ accuracy for the NS condition and A) patients’
SANS composite scores, B) their SANS 1 (flat affect) scores, and C) their SANS 2 (alogia) scores.
AB C
15
Discussion
Social information processing in schizophrenia
In the current study, patients showed dissociable neural modulation during the
processing of social content in speech and gesture modalities. In the speech modality, both
controls and patients activated a left-lateralized set of brain regions, including the dlPFC,
mPFC, the IFG, the temporal lobe, and the angular/supramarginal gyrus, without any group
difference. This finding replicates results from our previous study showing supramodal
social-abstract processing of healthy individuals (18), and is consistent with earlier studies
in basic research on the role of the mPFC in both perceiving social-related stimuli and
mentalizing social intentions (10, 13, 14). The observed left IFG and temporal lobe
activation is also in line with the literature on the neural substrates of abstract vs. concrete
semantics (17, 43), as the videos in the social condition, irrespective of modality, are more
abstract than the non-social, object-related condition. The fact that we did not find any
group differences in social speech processing suggests that patients with SZ exhibit intact
neural processing of social content presented in this modality. This finding complies with a
previous language study in SZ, in which patients also activated a comparable left fronto-
temporal network to controls when they processed abstract vs. concrete visual sentences
(23). Tog et h er, these data might suggest that, although SZ patients often show marked
social cognition deficits (2, 3), they exhibit intact neural processing of social features
encoded with linguistic stimuli (either in the written or verbal form). In the gesture modality,
however, although patients activated the mPFC for the social vs. non-social stimuli, this
activation was reduced when compared to controls. Notably, such modality-specific neural
modulation is, for the first time, reported for social information processing in SZ. This finding
is further discussed below.
Non-social information processing in schizophrenia
For the processing of non-social (object-related) information, again, the neural
modulation in patients showed an apparent dissociation. In the gesture modality, both
controls and patients activated the bilateral occipital-parietal cortices, STG, LOTC, insula,
16
as well as the pre/postcentral gyrus. In the speech modality, although these regions were
similarly activated in controls, their brain activation was significantly reduced for patients.
The group comparison suggests reduced activation in SZ patients in the left insula and the
left postcentral gyrus for non-social speech. Of note, the observed regions for non-social
and object-related information processing overlap with part of the mirror neuron network,
which is not only important for action observation and imitation, but also for the
understanding of object- and motor-related features in verbal form (44-47). This process
would require mental simulation of sensorimotor experience (20, 21, 48). Additionally, the
LOTC is also crucially involved in the perception of biological motion, object, as well as tool-
use (26, 49, 50). Our data from the control group suggest that these regions directly support
the processing of non-social object-related features, irrespective of encoding modality. This
finding is in line with the embodiment view of action and language processing (21, 51). With
regard to the patients, we observed normal neural processing of non-social content in the
gesture modality, supporting a previous study (24), which reported intact mirror neuron
activity in SZ (but see (25)). However, as we also observed reduced bilateral postcentral
gyrus and right insula activity for patients vs. controls for non-social speech this would, in
turn, imply that motor simulation, as required for processing object-related features from
auditory speech, might still be impaired in SZ (23, 25). This impairment concurs with the
reported deficits of SZ in action imitation (25, 42, 52, 53), where certain degrees of motor
simulation is required. Moreover, in the NS condition, we also observed negative correlation
between patients’ SANS composite and subscores and their task accuracy. This evidence
converges with previous research, corroborating the potential role of the mirror neuron
system during embodiment of non-social information (e.g., action imitation and
observation), as well as its relation to the development and persistence of negative
symptoms (42, 54)
Enhancing modality-specific social and non-social information processing deficits with
bimodal input
The novelty of our findings lies in the dissociable modality-specificity concerning
dysfunctional neural processing of social and non-social features. Social and non-social
17
features are functionally and neurally dissociable at the representational level (10, 14).
Besides, they might be differentially processed through either linguistic (speech) or non-
linguistic (gesture) channels. It has been proposed that social-abstract concepts may be
preferentially represented in speech, and that non-social concrete concepts are
preferentially delivered in hand action and gesture (55, 56). Despite this theoretical
proposal, however, during comprehension, healthy participants seem to be able to process
both types of information in a supramodal manner (e.g., semantic processing with unitary
core systems, irrespective of encoding modality, as in (18, 20)). For patients with SZ, as
they exhibit similar neural activations when processing social speech and non-social
gestures to controls, this might be an indication that they are at least intact in processing
these contents at representational level. But, they might show activation reduction in
relevant regions when these features are conveyed in a ‘non-preferred’ modality, as the
processing of these features would require some form of mental simulation: In the case
non-social information, patients are impaired in the simulation of motor-related experience
from action to language (20); In contrast when patients are presented with social
information, they might be impaired when simulating social features encoded by hand
gestures (but not with speech), as shown in their reduced mPFC activation. This observed
modality-specific processing deficit might also suggest that patients, unlike controls, are not
capable of processing social/non-social information in a supramodal manner like healthy
participants, as reported in previous studies (18, 57). More importantly, extending previous
studies on aberrant processing of social/non-social content in SZ, our results indicate that
this neural deficit is not universally present for either a specific modality or content, but
rather appears only in specific combinations of these two factors.
Despite reduced neural processing of both social and non-social content in gesture and
speech modalities, patients displayed intact neural processing of these features, as well as
improved task accuracy in the bimodal conditions. This enhancement effect concurs with a
line of proposals (6), who argue for a bi-directional facilitative relation between speech and
gesture (for empirical evidence, see (58-62)). More importantly, our finding extends
previous basic research, suggesting the translational implication of this mechanism. In SZ
research, the past decade has witnessed substantial progress in the development of social
18
cognitive training in SZ (63, 64), with recent innovation regarding the incorporation of social
stimuli from a broader range of modalities (65). Our findings extend these approaches,
proposing potential therapeutic implications of deploying naturalistic and multimodal stimuli
during social cognitive training, as they might be able to normalize processing of both social
and non-social information, at least at a neural level. Future research is expected to further
explore whether the neural enhancements can be linked to functional outcome after social
cognitive training in a multimodal setting.
19
Acknowledgements
This research project is supported by a grant from the ‘Von-Behring-Röntgen-Stiftung’
(project no. 59-0002 and 64-0001) and by the ‘Deutsche Forschungsgemeinschaft’ (project
no. DFG: STR1146/11-2 & KI588/6-2 and CRC/TRR 135/2 project A3). Y.H. is supported by
the ‘Von-Behring- Röntgen-Stiftung’ (project no. 64-0001). B.S. is supported by the DFG
(project no. STR 1146/15-1). The study was also supported by the Core Facility Brain
Imaging, Faculty of Medicine, University of Marburg, Rudolf-Bultmann-Str. 9, 35039,
Marburg, Germany.
Conflict of interests
All authors declare no financial conflict of interests.
20
References
1. Savla GN, Vella L, Armstrong CC, Penn DL, Twamley EW (2013): Deficits in domains of social cognition
in schizophrenia: a meta-analysis of the empirical evidence. Schizophr Bull. 39:979-992.
2. Green MF, Horan WP, Lee J (2015): Social cognition in schizophrenia. Nature Reviews Neuroscience.
16:620.
3. Ochsner KN (2008): The social-emotional processing stream: five core constructs and their translational
potential for schizophrenia and beyond. Biological Psychiatry. 64:48-61.
4. Sergi MJ, Rassovsky Y, Widmark C, Reist C, Erhart S, Braff DL, et al. (2007): Social cognition in
schizophrenia: relationships with neurocognition and negative symptoms. Schizophrenia research. 90:316-
324.
5. Couture SM, Penn DL, Roberts DL (2006): The functional significance of social cognition in
schizophrenia: a review. Schizophrenia bulletin. 32:S44-S63.
6. Holler J, Levinson SC (2019): Multimodal Language Processing in Human Communication. Trends in
Cognitive Sciences.
7. Amodio DM, Frith CD (2006): Meeting of minds: the medial frontal cortex and social cognition. Nature
reviews neuroscience. 7:268.
8. Anticevic A, Van Snellenberg JX, Cohen RE, Repovs G, Dowd EC, Barch DM (2012): Amygdala
recruitment in schizophrenia in response to aversive emotional material: a meta-analysis of neuroimaging
studies. Schizophrenia bulletin. 38:608-621.
9. Witteman J, Van Heuven VJ, Schiller NO (2012): Hearing feelings: a quantitative meta-analysis on the
neuroimaging literature of emotional prosody perception. Neuropsychologia. 50:2752-2763.
10. Mitchell JP, Heatherton TF, Macrae CN (2002): Distinct neural systems subserve person and object
knowledge. Proceedings of the National Academy of Sciences. 99:15238-15243.
11. Saxe R, Powell LJ (2006): It's the thought that counts: specific brain regions for one component of theory
of mind. Psychological science. 17:692-699.
12. Frith CD, Frith U (2006): The neural basis of mentalizing. Neuron. 50:531-534.
13. Van Overwalle F (2009): Social cognition and the brain: a meta-analysis. Human brain mapping. 30:829-
858.
14. Mitchell JP, Banaji MR, Macrae CN (2005): General and specific contributions of the medial prefrontal
cortex to knowledge about mental states. Neuroimage. 28:757-762.
15. Ma N, Vandekerckhove M, Van Overwalle F, Seurinck R, Fias W (2011): Spontaneous and intentional trait
inferences recruit a common mentalizing network to a different degree: spontaneous inferences activate only
its core areas. Social neuroscience. 6:123-138.
16. Chao LL, Haxby JV, Martin A (1999): Attribute-based neural substrates in temporal cortex for perceiving
and knowing about objects. Nature neuroscience. 2:913-919.
17. Binder JR, Westbury CF, McKiernan KA, Possing ET, Medler DA (2005): Distinct Brain Systems for
Processing Concrete and Abstract Concepts. Journal of Cognitive Neuroscience. 17:905-917.
18. Straube B, He Y, Steines M, Gebhardt H, Kircher T, Sammer G, et al. (2013): Supramodal neural
processing of abstract information conveyed by speech and gesture. Front Behav Neurosci. 7:120.
19. Mar RA (2011): The neural bases of social cognition and story comprehension. Annual review of
psychology. 62:103-134.
20. Pulvermüller F, Shtyrov Y, Ilmoniemi R (2005): Brain signatures of meaning access in action word
21
recognition. Journal of Cognitive Neuroscience. 17:884-892.
21. Barsalou LW (2008): Grounded cognition. Annu Rev Psychol. 59:617-645.
22. Frith CD (2004): Schizophrenia and theory of mind. Psychological medicine. 34:385-389.
23. Kuperberg GR, West WC, Lakshmanan BM, Goff D (2008): Functional Magnetic Resonance Imaging
Reveals Neuroanatomical Dissociations During Semantic Integration in Schizophrenia. Biological Psychiatry.
64:407-418.
24. Horan WP, Iacoboni M, Cross KA, Korb A, Lee J, Nori P, et al. (2014): Self-reported empathy and neural
activity during action imitation and observation in schizophrenia. NeuroImage: Clinical. 5:100-108.
25. Thakkar KN, Peterman JS, Park S (2014): Altered Brain Activation During Action Imitation and
Observation in Schizophrenia: A Translational Approach to Investigating Social Dysfunction in Schizophrenia.
American Journal of Psychiatry. 171:539-548.
26. Lingnau A, Downing PE (2015): The lateral occipitotemporal cortex in action. Trends in Cognitive
Sciences. 19:268-277.
27. Johnson-Frey SH (2004): The neural bases of complex tool use in humans. Trends in Cognitive
Sciences. 8:71-78.
28. Järveläinen J, Schuermann M, Hari R (2004): Activation of the human primary motor cortex during
observation of tool use. Neuroimage. 23:187-192.
29. Johnson-Frey SH, Maloof FR, Newman-Norlund R, Farrer C, Inati S, Grafton ST (2003): Actions or hand-
object interactions? Human inferior frontal cortex and action observation. Neuron. 39:1053-1058.
30. Andreasen NC (1984): Scale for the Assessment of Positive Symptons:(SAPS). University of Iowa.
31. Andreasen NC (1981): Scale for the Assessment of Negative Symptoms (SANS).
32. Lehrl S (1999): Mehrfachwahl-Wortschatz-Intelligenztest: MWT-B. Spitta.
33. Barth A, Küfferle B (2001): Development of a proverb test for assessment of concrete thinking problems
in schizophrenic patients. Der Nervenarzt. 72:853-858.
34. Nagels A, Kircher T, Steines M, Grosvald M, Straube B (2015): A brief self-rating scale for the
assessment of individual differences in gesture perception and production. Learning and Individual
Differences. 39:73-80.
35. He Y, Gebhardt H, Steines M, Sammer G, Kircher T, Nagels A, et al. (2015): The EEG and fMRI
signatures of neural integration: An investigation of meaningful gestures and corresponding speech.
Neuropsychologia. 72:27-42.
36. He Y, Steines M, Sommer J, Gebhardt H, Nagels A, Sammer G, et al. (2018): Spatial-temporal dynamics
of gesture-speech integration: a simultaneous EEG-fMRI study. Brain Struct Funct. 223:3073-3089.
37. He Y, Nagels A, Schlesewsky M, Straube B (2018): The Role of Gamma Oscillations During Integration of
Metaphoric Gestures and Abstract Speech. Front Psychol. 9:1348.
38. Straube B, Wroblewski A, Jansen A, He Y (2018): The connectivity signature of co-speech gesture
integration: The superior temporal sulcus modulates connectivity between areas related to visual gesture and
auditory speech processing. NeuroImage. 181:539-549.
39. Slotnick SD, Moo LR, Segal JB, Hart Jr J (2003): Distinct prefrontal cortex activity associated with item
memory and source memory for visual shapes. Cognitive Brain Research. 17:75-82.
40. Slotnick SD (2017): Resting-state fMRI data reflects default network activity rather than null data: A
defense of commonly employed methods to correct for multiple comparisons. Cognitive neuroscience. 8:141-
143.
41. Tzourio-Mazoyer N, Landeau B, Papathanassiou D, Crivello F, Etard O, Delcroix N, et al. (2002):
Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI
22
MRI single-subject brain. Neuroimage. 15:273-289.
42. Mehta UM, Thirthalli J, Aneelraj D, Jadhav P, Gangadhar BN, Keshavan MS (2014): Mirror neuron
dysfunction in schizophrenia and its functional implications: a systematic review. Schizophrenia research.
160:9-19.
43. Binder JR, Desai RH, Graves WW, Conant LL (2009): Where is the semantic system? A critical review
and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex. 19:2767-2796.
44. Buccino G, Binkofski F, Fink GR, Fadiga L, Fogassi L, Gallese V, et al. (2001): Action observation
activates premotor and parietal areas in a somatotopic manner: an fMRI study. European Journal of
Neuroscience. 13:400-404.
45. Di Pellegrino G, Fadiga L, Fogassi L, Gallese V, Rizzolatti G (1992): Understanding motor events: a
neurophysiological study. Experimental Brain Research. 91:176-180.
46. Iacoboni M, Woods RP, Brass M, Bekkering H, Mazziotta JC, Rizzolatti G (1999): Cortical mechanisms of
human imitation. Science. 286:2526-2528.
47. Rizzolatti G, Craighero L (2004): The mirror-neuron system. Annual Review of Neuroscience. 27:169-
192.
48. Pulvermüller F (2005): Brain mechanisms linking language and action. Nature Reviews Neuroscience.
6:576-582.
49. Bi Y, Wang X, Caramazza A (2016): Object Domain and Modality in the Ventral Visual Pathway. Trends in
Cognitive Sciences. 20:282-290.
50. Higuchi S, Imamizu H, Kawato M (2007): Cerebellar Activity Evoked By Common Tool-Use Execution
And Imagery Tasks: An Fmri Study. Cortex. 43:350-358.
51. De Stefani E, De Marco D (2019): Language, Gesture, and Emotional Communication: An Embodied
View of Social Interaction. Frontiers in Psychology. 10.
52. Park S, Matthews N, Gibson C (2008): Imitation, simulation, and schizophrenia. Schizophrenia Bulletin.
34:698-707.
53. Matthews N, Gold BJ, Sekuler R, Park S (2013): Gesture imitation in schizophrenia. Schizophrenia
Bulletin. 39:94-101.
54. Buccino G, Amore M (2008): Mirror neurons and the understanding of behavioural symptoms in
psychiatric disorders. Current opinion in psychiatry. 21:281-285.
55. Paivio A (2010): Dual coding theory and the mental lexicon. The Mental Lexicon. 5:205-230.
56. Perlovsky LI, Ilin R (2013): Mirror neurons, language, and embodied cognition. Neural networks. 41:15-
22.
57. Straube B, Green A, Weis S, Kircher T (2012): A Supramodal Neural Network for Speech and Gesture
Semantics: An fMRI Study. Plos One. 7:e51207.
58. Wang L, Chu M (2013): The role of beat gesture and pitch accent in semantic processing: an ERP study.
Neuropsychologia. 51:2847-2855.
59. Drijvers L, Özyürek A, Jensen O (2018): Hearing and seeing meaning in noise: Alpha, beta, and gamma
oscillations predict gestural enhancement of degraded speech comprehension. Human Brain Mapping.
59:617-613.
60. Krahmer E, Swerts M (2007): The effects of visual beats on prosodic prominence: Acoustic analyses,
auditory perception and visual perception. Journal of Memory and Language. 57:396-414.
61. He Y, Steines M, Sammer G, Nagels A, Kircher T, Straube B (2018): Action-Related Speech Modulates
Beta Oscillations During Observation of Tool-Use Gestures. Brain Topogr. 31:838-847.
62. Cuevas P, Steines M, He Y, Nagels A, Culham J, Straube B (2019): The facilitative effect of gestures on
23
the neural processing of semantic complexity in a continuous narrative. NeuroImage. 195:38-47.
63. Kurtz MM, Gagen E, Rocha NB, Machado S, Penn DL (2016): Comprehensive treatments for social
cognitive deficits in schizophrenia: A critical review and effect-size analysis of controlled studies. Clinical
psychology review. 43:80-89.
64. Kurtz MM, Richardson CL (2011): Social cognitive training for schizophrenia: a meta-analytic
investigation of controlled research. Schizophrenia bulletin. 38:1092-110 4.
65. Nahum M, Fisher M, Loewy R, Poelke G, Ventura J, Nuechterlein KH, et al. (2014): A novel, online social
cognitive training program for young adults with schizophrenia: A pilot study. Schizophrenia Research:
Cognition. 1:e11-e19.
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
Spoken language is an innate ability of the human being and represents the most widespread mode of social communication. The ability to share concepts, intentions and feelings, and also to respond to what others are feeling/saying is crucial during social interactions. A growing body of evidence suggests that language evolved from manual gestures, gradually incorporating motor acts with vocal elements. In this evolutionary context, the human mirror mechanism (MM) would permit the passage from “doing something” to “communicating it to someone else.” In this perspective, the MM would mediate semantic processes being involved in both the execution and in the understanding of messages expressed by words or gestures. Thus, the recognition of action related words would activate somatosensory regions, reflecting the semantic grounding of these symbols in action information. Here, the role of the sensorimotor cortex and in general of the human MM on both language perception and understanding is addressed, focusing on recent studies on the integration between symbolic gestures and speech. We conclude documenting some evidence about MM in coding also the emotional aspects conveyed by manual, facial and body signals during communication, and how they act in concert with language to modulate other’s message comprehension and behavior, in line with an “embodied” and integrated view of social interaction.
Article
Full-text available
Metaphoric (MP) co-speech gestures are commonly used during daily communication. They communicate about abstract information by referring to gestures that are clearly concrete (e.g., raising a hand for “the level of the football game is high”). To understand MP co-speech gestures, a multisensory integration at semantic level is necessary between abstract speech and concrete gestures. While semantic gesture-speech integration has been extensively investigated using functional magnetic resonance imaging, evidence from electroencephalography (EEG) is rare. In the current study, we set out an EEG experiment, investigating the processing of MP vs. iconic (IC) co-speech gestures in different contexts, to reveal the oscillatory signature of MP gesture integration. German participants (n = 20) viewed video clips with an actor performing both types of gestures, accompanied by either comprehensible German or incomprehensible Russian (R) speech, or speaking German sentences without any gestures. Time-frequency analysis of the EEG data showed that, when gestures were accompanied by comprehensible German speech, MP gestures elicited decreased gamma band power (50–70 Hz) between 500 and 700 ms in the parietal electrodes when compared to IC gestures, and the source of this effect was localized to the right middle temporal gyrus. This difference is likely to reflect integration processes, as it was reduced in the R language and no-gesture conditions. Our findings provide the first empirical evidence suggesting the functional relationship between gamma band oscillations and higher-level semantic processes in a multisensory setting.
Article
Full-text available
The semantic integration between gesture and speech (GSI) is mediated by the left posterior temporal sulcus/middle temporal gyrus (pSTS/MTG) and the left inferior frontal gyrus (IFG). Evidence from electroencephalography (EEG) suggests that oscillations in the alpha and beta bands may support processes at different stages of GSI. In the present study, we investigated the relationship between electrophysiological oscillations and blood-oxygen-level-dependent (BOLD) activity during GSI. In a simultaneous EEG-fMRI study, German participants (n = 19) were presented with videos of an actor either performing meaningful gestures in the context of a comprehensible German (GG) or incomprehensible Russian sentence (GR), or just speaking a German sentence (SG). EEG results revealed reduced alpha and beta power for the GG vs. SG conditions, while fMRI analyses showed BOLD increase in the left pSTS/MTG for GG > GR ∩ GG > SG. In time-window-based EEG-informed fMRI analyses, we further found a positive correlation between single-trial alpha power and BOLD signal in the left pSTS/MTG, the left IFG, and several sub-cortical regions. Moreover, the alpha-pSTS/MTG correlation was observed in an earlier time window in comparison to the alpha-IFG correlation, thus supporting a two-stage processing model of GSI. Our study shows that EEG-informed fMRI implies multiple roles of alpha oscillations during GSI, and that the method is a best candidate for multidimensional investigations on complex cognitive functions such as GSI.
Article
Full-text available
Language and action have been thought of as closely related. Comprehending words or phrases that are related to actions commonly activates motor and premotor areas, and this comprehension process interacts with action preparation and/or execution. However, it remains unclear whether comprehending action-related language interacts with action observation. In the current study, we examined whether the observation of tool-use gesture subjects to interaction with language. In an electroencephalography (EEG) study (n = 20), participants were presented with video clips of an actor performing tool-use (TU, e.g., hammering with a fist) and emblematic (EM, e.g., the thumb up sign for ‘good job’) gestures accompanied by either comprehensible German (G) or incomprehensible Russian sentences (R). Participants performed a semantic judging task, evaluating whether the co-speech gestures were object- or socially-related. Behavioral results from the semantic task showed faster response for the TU versus EM gestures only in the German condition. For EEG, we found that TU elicited beta power decrease (~ 20 Hz) when compared to EM gestures, however this effect was reduced when gestures were accompanied by German instead of Russian sentences. We concluded that the processing of action-related sentences might facilitate gesture observation, in the sense that motor simulation required for TU gestures, as indexed by reduced beta power, was modulated when accompanied by comprehensible German speech. Our results corroborate the functional role of the beta oscillations during perception of hand gestures, and provide novel evidence concerning language–motor interaction.
Article
Full-text available
During face-to-face communication, listeners integrate speech with gestures. The semantic information conveyed by iconic gestures (e.g., a drinking gesture) can aid speech comprehension in adverse listening conditions. In this magnetoencephalography (MEG) study, we investigated the spatiotemporal neural oscillatory activity associated with gestural enhancement of degraded speech comprehension. Participants watched videos of an actress uttering clear or degraded speech, accompanied by a gesture or not and completed a cued-recall task after watching every video. When gestures semantically disambiguated degraded speech comprehension, an alpha and beta power suppression and a gamma power increase revealed engagement and active processing in the hand-area of the motor cortex, the extended language network (LIFG/pSTS/STG/MTG), medial temporal lobe, and occipital regions. These observed low- and high-frequency oscillatory modulations in these areas support general unification, integration and lexical access processes during online language comprehension, and simulation of and increased visual attention to manual gestures over time. All individual oscillatory power modulations associated with gestural enhancement of degraded speech comprehension predicted a listener's correct disambiguation of the degraded verb after watching the videos. Our results thus go beyond the previously proposed role of oscillatory dynamics in unimodal degraded speech comprehension and provide first evidence for the role of low- and high-frequency oscillations in predicting the integration of auditory and visual information at a semantic level.
Article
Full-text available
Analysis of functional magnetic resonance imaging (fMRI) data typically involves over one hundred thousand independent statistical tests; therefore, it is necessary to correct for multiple comparisons to control familywise error. Eklund, Nichols, and Knutsson (2016, Proceedings of the National Academy of Sciences of the United States of America, 113, 7900-7905) used resting-state fMRI data to evaluate commonly employed methods to correct for multiple comparison and reported unacceptable rates of familywise error. Eklund et al.'s analysis was based on the assumption that resting-state fMRI data reflects null data; however, their "null data" actually reflected default network activity that inflated familywise error. As such, Eklund et al.'s results provide no basis to question the validity of the tens of thousands of published fMRI studies that have corrected for multiple comparisons or the commonly employed methods to correct for multiple comparisons.
Article
The natural ecology of human language is face-to-face interaction comprising the exchange of a plethora of multimodal signals. Trying to understand the psycholinguistic processing of language in its natural niche raises new issues, first and foremost the binding of multiple, temporally offset signals under tight time constraints posed by a turn-taking system. This might be expected to overload and slow our cognitive system, but the reverse is in fact the case. We propose cognitive mechanisms that may explain this phenomenon and call for a multimodal, situated psycholinguistic framework to unravel the full complexities of human language processing.
Article
Gestures are elemental components of social communication and aid comprehension of verbal messages; however, little is known about the potential role of gestures in facilitating processing of semantic complexity in an ecologically valid setting. The goal of this study was to investigate whether cognitive load, as indexed by semantic complexity, is modulated by the presentation of gestures accompanying speech. Twenty healthy participants watched 16 video clips of a short narrative while instructed to carefully listen to and watch the narrator while functional magnetic resonance imaging (fMRI) data were acquired. The videos contained passages with and without various co-speech gestures, as well as passages where the semantic complexity was either low or high, as measured by the metric of idea density. Increasing semantic complexity led to reduced activation within the default mode network (DMN); whereas, presents of gestures decreased activation in language-related regions (left middle temporal gyrus and left inferior frontal gyrus) and increased activation in high-level visual and multimodal regions of occipitotemporal cortex. Most interestingly, an interaction between semantic complexity and gestures was observed in a language-related area in left anterior temporal cortex; specifically, increasing gestures led to a greater drop in activation with high vs. Low semantic complexity. These results provide evidence that the facilitation of gestures on semantic processing, particularly for complex narratives, is reflected in the neural substrates of language processing.
Article
Humans integrate information communicated by speech and gestures. Functional magnetic resonance imaging (fMRI) studies suggest that the posterior superior temporal sulcus (STS) and adjacent gyri are relevant for multisensory integration. However, a connectivity model representing this essential combinatory process is still missing. Here, we used dynamic causal modeling for fMRI to analyze the effective connectivity pattern between middle temporal gyrus (MTG), occipital cortex (OC) and STS associated with auditory verbal, visual gesture-related, and integrative processing, respectively, to unveil the neural mechanisms underlying integration of intrinsically meaningful gestures (e.g., "Thumbs-up gesture") and corresponding speech. 20 participants were presented videos of an actor either performing intrinsic meaningful gestures in the context of German or Russian sentences, or speaking a German sentence without gesture, while performing a content judgment task. The connectivity analyses resulted in a winning model that included bidirectional intrinsic connectivity between all areas. Furthermore, the model included modulations of both connections to the STS (OC→STS; MTG→STS), and non-linear modulatory effects of the STS on bidirectional connections between MTG and OC. Coupling strength in the occipital pathway (OC→STS) correlated with gesture related advantages in task performance, whereas the temporal pathway (MTG→STS) correlated with performance in the speech only condition. Coupling between MTG and OC correlated negatively with subsequent memory performance for sentences of the Gesture-German condition. Our model provides a first step towards a better understanding of speech-gesture integration on network level. It corroborates the importance of the STS during audio-visual integration by showing that this region inhibits direct auditory-visual coupling.
Article
The nature of domain-specific organization in higher-order visual cortex (ventral occipital temporal cortex, VOTC) has been investigated both in the case of visual experience deprivation and of modality of stimulation in sighted individuals. Object domain interacts in an intriguing and revelatory way with visual experience and modality of stimulation: selectivity for artifacts and scene domains is largely immune to visual deprivation and is multi-modal, whereas selectivity for animate items in lateral posterior fusiform gyrus is present only with visual stimulation. This domain-by-modality interaction is not readily accommodated by existing theories of VOTC representation. We conjecture that these effects reflect a distinction between the visual features that characterize different object domains and their interaction with different types of downstream computational systems.