Play all audios:
ABSTRACT Human nonverbal social signals are transmitted to a large extent by vocal and facial cues. The prominent importance of these cues is reflected in specialized cerebral regions which
preferentially respond to these stimuli, e.g. the temporal voice area (TVA) for human voices and the fusiform face area (FFA) for human faces. But it remained up to date unknown whether
there are respective specializations during resting state, i.e. in the absence of any cues, and if so, whether these representations share neural substrates across sensory modalities. In the
present study, resting state functional connectivity (RSFC) as well as voice- and face-preferential activations were analysed from functional magnetic resonance imaging (fMRI) data sets of
60 healthy individuals. Data analysis comprised seed-based analyses using the TVA and FFA as regions of interest (ROIs) as well as multi voxel pattern analyses (MVPA). Using the face- and
voice-preferential responses of the FFA and TVA as regressors, we identified several correlating clusters during resting state spread across frontal, temporal, parietal and occipital
regions. Using these regions as seeds, characteristic and distinct network patterns were apparent with a predominantly convergent pattern for the bilateral TVAs whereas a largely divergent
pattern was observed for the bilateral FFAs. One region in the anterior medial frontal cortex displayed a maximum of supramodal convergence of informative connectivity patterns reflecting
voice- and face-preferential responses of both TVAs and the right FFA, pointing to shared neural resources in supramodal voice and face processing. The association of individual voice- and
face-preferential neural activity with resting state connectivity patterns may support the perspective of a network function of the brain beyond an activation of specialized regions. SIMILAR
CONTENT BEING VIEWED BY OTHERS BASAL GANGLIA AND CEREBELLUM CONTRIBUTIONS TO VOCAL EMOTION PROCESSING AS REVEALED BY HIGH-RESOLUTION FMRI Article Open access 20 May 2021 THE
REPRESENTATIONAL DYNAMICS OF PERCEIVED VOICE EMOTIONS EVOLVE FROM CATEGORIES TO DIMENSIONS Article 11 March 2021 MAPPING OF FACIAL AND VOCAL PROCESSING IN COMMON MARMOSETS WITH ULTRA-HIGH
FIELD FMRI Article Open access 13 March 2024 INTRODUCTION Voices and faces are among the most salient cues in human life. This is reflected in the existence of specialized cerebral modules
which are hierarchically organized and specifically tuned to respond to these cues. Core components for the primary identification of human voices and faces are the temporal voice area (TVA)
for voices1,2,3,4 and the fusiform face area (FFA) for faces5,6,7,8. While not exclusively activated by these signals, they exhibit clearly voice- and face-preferential responses,
respectively. The FFA together with the occipital face area (OFA) respond mainly to invariant facial features (e.g. gender)5,8. Further processing of dynamic face aspects, and integration of
signals from voices and faces involves the posterior superior temporal sulcus (pSTS) and the thalamus9,10,11,12. The emotional information often present in faces and voices (e.g. in facial
expressions and emotional prosody) additionally converges in the amygdala9,13. Further processing of such emotional information involves further regions such as the inferior frontal cortex
(IFC) and orbitofrontal cortex (OFC)14,15. Convergent with the particular importance of voices and faces in human social communication, recent studies indicated that the responsivity to the
preferred cues of the basic modules for identification of human voices and faces is moderated by interindividual differences in social signal processing, e.g. social anxiety16, and emotional
intelligence17, even in the absence of emotional information. In some cases, as described above e.g. for the pSTS and thalamus, the hemodynamic correlates of cerebral processing of signals
from different sensory modalities overlap. This phenomenon will be termed supramodal throughout this manuscript. While a plethora of neuroimaging studies delineated the neural networks that
are active when we see faces or hear voices, it remains a completely open question if the brain’s activity patterns also reflect the individual cerebral responsivity to voices and faces in
the absence of these cues and if these representations may share neural substrates across sensory modalities. During the past three decades, the resting brain has become a major research
focus as it became clear that spontaneous physiological low-frequency fluctuations in brain activity occur non-randomly but simultaneously in various, partially overlapping neural networks
in the absence of any cues or stimulation or cognitive/emotional task18. Nevertheless, these fluctuation patterns are not independent from individual traits or diseases, as they have been
shown to correlate with various aspects of behavioural tendencies19,20,21, personality22,23, psychopathology21,24,25, and psychiatric disease (e.g. dementia and schizophrenia26) also
demonstrating that resting state data can be used to expand the neuroimaging perspective on their cerebral representation in a complementary manner with the potential to detect links between
the neural networks underlying various perceptual, cognitive or emotional functions not apparent in stimulation-based designs. In the area of face and voice processing, correlations of
resting state functional connectivity (RSFC) with behavioural outcomes, e.g. performance in various face- and voice-processing tasks have been observed27,28,29,30. One study compared
functional connectivity patterns during resting state and a passive viewing task and found for both conditions similar networks including posterior fusiform gyrus, inferior occipital gyrus
and superior temporal sulcus27. In this work the informative RSFC patterns were found exclusively within the network of modality-specific preferential processing areas27. Two studies
combined RSFC in the face processing network with behavioural performance in a face identification task and an emotional face matching task, respectively28,29 and found RSFC patterns between
modality-specific preferential processing areas but also with other parts of the brain28,29. One study in children revealed that performance in an auditory emotional prosody recognition
task was predicted by stronger connectivity between the inferior frontal gyrus and motor regions. Here, informative RSFC patterns were found exclusively outside the modality-specific
preferential processing networks30. In the present study, we intended to determine the neural correlates of voice- and face-preferential responses in the absence of voices and faces in the
resting state. Furthermore, we aimed to identify brain areas with RSFC patterns supramodally reflecting preferential responses to both, voices and faces. To this end, 60 healthy individuals
underwent functional magnetic resonance imaging (fMRI) at rest and during stimulation with voices, faces and various other classes of acoustic and visual stimuli. Individual voice- and
face-preferential responses were correlated with RSFC employing multi voxel pattern analyses (MVPA) and seed-based analyses focused on TVA and FFA. MATERIALS AND METHODS PARTICIPANTS 60
healthy individuals (mean age 25.8 years, s.d. = 4.5 years, 30 female) participated at the University of Tübingen. All of the participants were native German speakers and right-handed, as
assessed with the Edinburgh Inventory31. None of the participants was taking any regular medication, or had a history of substance abuse, or psychiatric or neurological illness. Hearing was
normal, vision normal or corrected to normal in all participants. The study was performed according to the Code of Ethics of the World Medical Association (Declaration of Helsinki) and the
protocol of human investigation was approved by the local ethics committees where the study was performed (i.e., the medical faculties of the Universities of Tübingen and Greifswald). All
individuals gave their written informed consent prior to their participation in the study. STIMULI AND EXPERIMENTAL DESIGN Two fMRI experiments were performed to localize face-sensitive5 and
voice-sensitive1 brain areas as described in previous publications9,10,14,16,17,32,33: For the face-sensitivity experiment, pictures from four different categories (faces, houses, objects,
and natural scenes) were employed within a block design. All stimuli used in the experiment were black-and-white photographs unknown to the participants17. The shown face stimuli had no
obvious emotional connotation, but rather showed neutral facial expressions. The house stimuli were multilevel apartment houses from different materials (brick, wooden, concrete). As object
stimuli different everyday life items were used (e.g. flat iron, spoon, T-shirt). The fourth category of natural scenes represented different countryside pictures (e.g. mountain, coastline,
waterfall). Each block and category contained 20 stimuli17. Within blocks, the stimuli were presented in random order for 300 ms. Stimuli were separated by 500 ms periods of fixation [1
block = 20 stimuli × (300 ms picture + 500 ms fixation) = 16 s]. Eight blocks of each category pseudorandomized within the experiment were shown separated by short ~ 1.5 s rest periods17. A
one-back task was employed, in which the participants had to press a button on a fibre optic system (LumiTouch, Photon Control, Burnaby, Canada) with their right index finger when they saw a
picture twice in a row, to ascertain constant attention17. The appearance of repeated stimuli was pseudorandomized ensuring a distribution across the entire experiment. Visual stimuli were
back-projected onto a screen placed in the magnet bore behind the participant’s head and viewed by the participant through a mirror system mounted onto the head coil. The voice-sensitivity
experiment was developed based on the study by Belin et al.1 in form of a block design experiment with 24 stimulation blocks and 12 silent periods (each 8 s) in a passive-listening design
without an explicit task. Between the blocks were short periods without sound (2 s). Participants were instructed to listen attentively with their eyes closed. The stimulus material
comprised 12 blocks of human vocal sounds (speech, sighs, laughs, cries), 6 blocks of animal sounds (e.g., gallops, various cries) and 6 blocks of environmental sounds (e.g., cars, planes,
doors, telephones). Stimuli were normalized with respect to mean acoustic energy17. Sound and silence blocks were pseudorandomized across the experiment with the restriction that with the
restriction that no two blocks of silence directly followed each other. Both experimental designs have been validated in previous studies9,10,14,17,32,33. Further details on the stimulus
material and experimental designs have been reported elsewhere9. For the resting state measurements (duration about 7 min and 15 s), the participants were instructed to keep their eyes
closed with no further task. IMAGE ACQUISITION MRI data were acquired with a TRIO 3T and a PRISMA scanner (Siemens, Erlangen, Germany). Structural T1-weighted images (176 slices, TR = 2300
ms, TE = 2.96 ms, TI = 1100 ms, voxel size: 1 × 1 × 1 mm3) and functional images (30 axial slices captured in sequential descending order, 3 mm thickness + 1 mm gap, TR = 1.7 s, TE = 30 ms,
voxel size: 3 × 3 × 4 mm3, field of view 192 × 192 mm2, 64 × 64 matrix, flip angle 90°) were recorded. For the resting state measurements, 245 images were recorded. The activation tasks were
performed after completion of the resting state measurements to avoid carry-over effects. The time series comprised 368 images for the face experiment and 232 images for the voice
experiment and 250 images for the resting state measurement. A field map with 36 slices (slice thickness 3 mm, TR = 400 ms, TE(1) = 5.19 ms, TE(2) = 7.65 ms) was recorded. ANALYSIS OF FMRI
DATA Statistical parametric mapping software (SPM8, Wellcome Department of Imaging Neuroscience, London, http://www.fil.ion.ucl.ac.uk/spm) was used to analyse the imaging data.
Pre-processing generally included the removal of the first five EPI images from each run to exclude measurements preceding T1 equilibrium. FACE- AND VOICE-SENSITIVITY EXPERIMENTS The
preprocessing procedure consisted of realignment, unwarping using a static field map, coregistration of anatomical and functional images, segmentation of the anatomical images, normalization
into MNI space (Montreal Neurological Institute34) with a resampled voxel size of 3 × 3 × 3 mm3, temporal smoothing with a high-pass filter (cut-off frequency of 1/128 Hz) and spatial
smoothing employing a Gaussian kernel (8 mm full width at half maximum, FWHM). The response to the single categories (faces (F), houses (H), objects (O), and natural scenes (S) in the face
localizer as well as vocal sounds (V), animal sounds (A), and environmental sounds (E) in the voice localizer were independently modelled with a box-car function corresponding to the
duration of the stimulation blocks (16 s in the face localizer and 8 s in the voice localizer) convolved with the hemodynamic response function (HRF). The error term was calculated as a
first order autoregressive process with a coefficient of 0.2 and a white noise component accounting for serial autocorrelations35. To minimize motion-associated error variance, the six
motion parameters (i.e. translation and rotation on the x-, y-, and z-axes) were included in the single subject models as covariates. Contrast images were constructed using data from the
first-level general linear models [face-sensitivity: F > (H, O, S); voice-sensitivity: V > (A, E)] for each subject. Taking these contrast images as sources, a second-level
random-effect analysis was performed with one-sample t-tests to define the face-sensitive fusiform face area (FFA) and the voice-sensitive temporal voice area (TVA) as functional regions of
interest (ROI) for further analyses. Statistical significance of activations was assessed at p < 0.001, uncorrected at voxel level and with FWE correction for multiple comparisons at
cluster level with p < 0.05. For the definition of the FFA, the fusiform gyrus was taken as a priori anatomical ROI; for definition of the TVA, the temporal gyri and the temporal pole
were selected. For definition of the functional ROIs (i.e. FFA and TVA), FWE-cluster level correction was performed across these a priori anatomical ROIs using small volume correction
(SVC36). We picked the maximum activation in the fusiform gyrus for the FFA and in the temporal lobe for the TVA respectively, and defined the surrounding 100 most sensitive voxels as masks
for the functional ROIs. Within these ROIs individual voice- and face-preferential responses were assessed using minimum difference criteria (for voices V > max[A, E], for faces F >
max[H, O, S])37. Intercorrelations of the four regressors were evaluated. Differences in the face- and voice-sensitive and -preferential responses between both hemispheres and
interhemispheric differences in cue-sensitivity and -preferentiality between TVA and FFA were post hoc tested using two-sided paired t-tests with Bonferroni correction. RESTING STATE
FUNCTIONAL CONNECTIVITY ANALYSIS For RSFC analyses we used the CONN toolbox (v 16b38) implemented in SPM8. The spatial preprocessing was performed analogously to the procedure described for
the face- and voice-sensitivity experiments. Denoising included linear regression of the following confounding effects: White matter and CSF components (6P each), effect of rest (2P,
temporal component and first order derivates) and motion regression (12 regressors: 6 motion parameters and 6 first-order temporal derivates) and band-pass filtering (0.008–0.09 Hz). Linear
detrending was added to remove linear trends. The participants’ movement parameters, their first order derivatives and the BOLD signal from white matter, cerebrospinal fluid and effect of
rest (each with five temporal components) were included in the analysis as covariates to reduce their confounding influences. In the individual first-level analyses, bivariate correlation
coefficients were calculated as linear measures of functional connectivity for the ensuing analyses. Coefficients were Z transformed to achieve comparability for group-level analyses, and
gender, age and scanner were included as regressors of no interest. The Automated Anatomic Labelling (AAL) toolbox39 was used for the definition of anatomical regions in MNI space. The
analysis targeted the correlation of individual resting state functional connectivity (RSFC) with face-/voice-preferential responses both with defined regions of interest (ROIs) and at whole
brain level. To this end, analyses were done on two different levels: ROI-to-voxel analyses should detect associations between individual voice- and face-preferential responses of the ROIs
and their RSFC with other brain regions. Here, the significance of observed connectivity patterns was assessed using a threshold of p < 0.001 at voxel level, two-tailed with FWE
correction (p < 0.05) for multiple comparisons at cluster level. Results were Bonferroni-corrected for the numbers of regressors (4) and ROIs (4), so that the effective cluster threshold
amounted to p < 0.00315. Second, a spatial hypothesis-free strategy was implemented using voxel-to-voxel multivariate multi voxel pattern analyses (MVPA). Here, for each voxel separately,
a low-dimensional multivariate representation of the connectivity pattern between this voxel and all other voxels in the brain was calculated. This representation was based on a principal
component analysis of the inter-subject variability of each separate voxel’s connectivity pattern enabling the investigation of differences across subjects using second-level multivariate
analyses. The number of principal components was set to three and number of dimensions was set to 64 (dimensionality reduction)40. The goal of the group-MVPA approach was to detect whole
brain resting state functional connectivity patterns correlating with individual voice-preferential responses of the TVA (i.e., V > max[A,E]) and face-preferential responses (i.e., F >
max[H,O,S]) of the FFA. These individual estimates were used as group level regressors in the RSFC analyses (four regressors: two for the FFAs, two for the TVAs). Results were evaluated at
a voxel-wise threshold of p < 0.001 and whole brain FWE-corrected at cluster level with additional Bonferroni-correction for the number of tested regressors (4) resulting in an effective
cluster threshold of p < 0.0125. Findings of the MVPA were further analysed using the significant clusters as seeds for ensuing seed-to-voxel analyses. Convergence of RSFC patterns
between different seeds was tested using conjunction analyses with a minimum statistic41. Results were assessed at a voxel-wise threshold of p < 0.001 and whole brain FWE-corrected at
cluster level with a cluster threshold of p < 0.05. RESULTS ROI CHARACTERISTICS The activation pattern of the right and left FFA showed a significant sensitivity for faces (rFFA t =
9.321, p < 0.001 and lFFA t = 7.585, p < 0.001), whereas significant face-preferential responses were observed in the right FFA (t = 4.344, p < 0.0001), but not the left FFA (t =
0.624, p = 0.535). The bilateral TVAs were highly sensitive to and preferential for voices (sensitivity: rTVA t = 18.265, p < 0.0001 and lTVA t = 17.457, p < 0.001; preferentiality:
rTVA t = 14.456, p < 0.001 and lTVA t = 14.023, p < 0.001). ROI characteristics are graphically displayed in Fig. 1. The ROIs’ preferential responses to their preferred cues were
significantly correlated within modality (voices: r(58) = 0.74, p < 0.001; faces: r(58) = 0.60, p < 0.001) but not across modalities (all abs(r(58)) < 0.12, all p > 0.05).
Comparison between the right and left hemisphere revealed no significant difference for voice-sensitivity or -preferentiality (all t < 2.03), all p > 0.187), but significant
differences for face-sensitivity and -preferentiality in favour of the right hemisphere (all t > 3.75, all p < 0.004). Comparison of hemispheric differences in modality-specific
differences in cue sensitivity and preferentiality between TVA and FFA corroborated the difference between the sensory modalities, both for sensitivity and for preferentiality (all t >
3.93, p < 0.002), i.e. a greater hemispheric difference in face-sensitivity and -preferentiality than in voice-sensitivity and -preferentiality. ROI-TO-VOXEL ANALYSIS In this analysis,
only individual voice-preferential responses of the lTVA were significantly associated with RSFC between the lTVA and a cluster in the right supramarginal gyrus extending into the inferior
parietal gyrus (peak: − 57x − 66y 27z; 143 voxels; p(FWE-corr.) = 0.0018). MULTI-VOXEL PATTERN ANALYSIS (MVPA) Using rFFA face-preferential responses as regressor, we identified one
informative cluster in the right middle frontal gyrus extending into the precentral gyrus. For the lFFA two clusters in the left caudate nucleus/olfactory gyrus and left superior temporal
pole were evident. For the rTVA and lTVA voice-preferential responses four overlapping clusters emerged: in the left superior occipital gyrus, the right inferior parietal gyrus, the right
superior temporal gyrus and the right frontal inferior orbital gyrus. For rTVA voice-preferentiality two additional clusters were detected in the left middle occipital gyrus and the right
thalamus, for the left TVA two additional clusters were located in the left frontal superior gyrus and the right parietal superior gyrus. A detailed description of the clusters can be found
in Table 1. A graphical representation is displayed in Fig. 2. For the four overlapping clusters informative of both the rTVA and lTVA voice-preferential responses common regions were
calculated and further on used as seeds. The characteristics of the resulting clusters are described in Table 2. Significant clusters were used as seeds for subsequent post-hoc explanatory
seed-to-voxel analyses. For the TVAs the convergence of informative MVPA clusters was accompanied by a relatively strong convergence of their RSFC patterns in contrast to the FFAs’ RSFC
patterns. Tables 3, 4 and 5 give an overview of convergent RSFC clusters across all informative regions observed in the MVPA analysis. Convergent clusters for the bilateral TVAs are listed
in Table 3, exemplary graphical representations are given in Fig. 3. In contrast to these results, for the FFAs, in addition to the lower number of informative clusters in the MVPA analysis,
the RSFC pattern was largely divergent as exemplarily shown for two MVPA clusters informative of FFA face-selective responses (rFFA: right middle frontal gyrus extending into the precentral
gyrus, lFFA: left caudate nucleus and olfactory gyrus). Only one significant common cluster was observed in the right supramarginal gyrus extending into the inferior parietal gyrus (peak:
57x − 27 < 45z; 81 voxels, p(FWE-corr.) = 0.010) using the of the right R middle frontal gyrus/precentral gyrus (rFFA) and the left caudate nucleus and olfactory gyri (− 6 6 − 15, lFFA)
as seeds. The results are illustrated in Fig. 4. SUPRAMODAL CONVERGENCE OF INFORMATIVE RSFC PATTERNS The combination of RSFC correlates of individual face-preferential responses in the right
and left FFA with RSFC correlates of individual voice-preferential responses in the right and left TVA can decipher supramodal convergence of RSFC patterns, i.e. combining voice- and
face-preferentiality. In our case, this was evident in eight clusters (Table 5). The convergence was more prominent using right-hemispheric voice- and face-preferentiality regressors with
five common clusters, whereas for the left-hemispheric regressors only one supramodal cluster was found. Two clusters derived from regressors of contralateral hemispheres. Only one region in
the anterior region of the rostral mediofrontal cortex (arMFC) exhibited supramodal convergence of informative RSFC patterns for more than two regressors: Convergence of the RSFC of the
rlTVA cluster in the left superior occipital gyrus with the lTVA cluster in the left frontal superior gyrus and the rFFA cluster in the right middle frontal gyrus delineated one common
region in the medial frontal gyrus (including the left orbital gyrus and the anterior cingulum as well as the right and left medial frontal gyrus, peak: 0 × 54y 9z; 83 voxels, p(FWE-corr.) =
0.011) indicative of right and left TVA voice-preferentiality as well as rFFA face-preferentiality (see also Fig. 5). DISCUSSION Combining seminal experiments used to localize voice- as
well as face-preferential areas in the human brain and resting state fMRI, this study provides the first description of hemodynamic functional connectivity patterns in the resting state that
are associated with voice- and face-preferential cerebral responses at the primary level of the TVA and FFA. Using functional connectivity in the resting state, we identified several
clusters correlating with voice- and face-preferentiality of the TVA and FFA. For the rFFA one right frontal/precentral cluster was evident, for the lFFA two clusters, one in the left
caudate/olfactory gyrus and one in the left superior temporal pole. Using the voice-preferentiality of the rTVA and lTVA as regressors, four common clusters emerged. These were widely
distributed the occipital, parietal, frontal and temporal cortex. For the rTVA two additional clusters in the left occipital cortex and the right thalamus, and for the lTVA in the left
frontal and right parietal cortex areas emerged. In explanatory seed-to-voxel analyses, the underlying connectivity patterns diverged markedly between the voice and face processing systems.
Whereas for the TVAs a largely convergent pattern of clusters was observed, among others in the occipital gyrus and bilateral insulae, the patterns for the FFAs were mainly divergent and
yielded only one common region in the right supramarginal gyrus extending into the inferior parietal gyrus. Moreover, we identified brain areas with RSFC patterns supramodally reflecting
preferential responses to both, voices and faces. One area in the anterior rostral mediofrontal cortex (arMFC) displayed a maximum of convergent RSFC patterns: its RSFC was indicative of
individual voice-preferential responses of both TVAs and face-preferential responses of the right FFA. Our results strengthen the view that cerebral voice and face processing is an
evolutionary important and therefore highly preserved mechanism, which is not only evidenced by several stages of very specialized processing in the brain, starting with the regions of
TVA1,2,3,4 and FFA5,6,7,8, but is also reflected in other networks, i.e. the resting state network that—per se—work independent from the aforementioned voice and face processing system.
Because during resting state participants were asked to lie quiescent without specific thought. But the independence could be impaired, in case the participants would have thought of human
voices and faces during the resting state measurement. To minimize this risk, we designed the experimental sequence with the resting state block first followed by the task-related parts. The
finding of a correlation of voice- and face-activation patterns with resting state parameters fits in quite well with the currently still limited literature applying both resting state and
voice/face processing measurements. Previous studies found diverging regions either exclusively in the modality-specific processing areas27,42, both in modality specific areas and other
parts of the brain28,29, or networks in the inferior frontal gyrus and motor regions which are not directly connected to modality specific processing30. It needs to be acknowledged however
that a broad range of diverse data analysis techniques were used in those studies27,28,29,30,42 which may account for the disparities to some extent. Our comprehensive analysis on RSFC
networks associated with voice- and face-preferentiality revealed large networks across whole brain, underpinning the notion that response patterns generated in basic voice and face
processing modules during the perception of these cues find a reflection in the coactivation of widespread cerebral networks at rest potentially indicating processes connected to voice and
face perception or a neural preparedness to respond to these stimuli. Speaking figuratively, the direct responses to stimulation with voices and faces can be imagined as the top of the
iceberg, the underlying resting state network structure as the part below the surface of the sea. It is known from the literature that resting state patterns reflect individual traits. In
fact, resting state functional connectivity has been shown to be associated with behavioural tendencies, personality or states of psychiatric disease, e.g. personality traits22,23, moral
behaviour43, violence proneness25, or the diagnosis of dementia or schizophrenia26. These results support the view that resting state patterns may reflect an adaptive system indicative of
different brain states and function. One could speculate about the connection between basal voice and face processing systems, as assessed in our work, and higher order social functioning
(e.g., emotional communication, empathy, theory of mind or moral behaviour), as effective voice and face perception appears as a prerequisite of the former to a certain degree. Certainly,
however, this link remains speculative presently. The novel and distinctive feature of this study is the combination of resting state and stimulation-based fMRI measurements for the visual
and the auditory system. The resting state pattern, i.e. a stimulation-free measurement, correlates with the propensity to respond to certain stimuli. Up to now, this form of association has
only scarcely been addressed. A similar approach revealed non-state-dependent cerebral markers of biased perception in social anxiety37. Another meta-analytic study focused on similarities
in resting state functional connectivity patterns and coactivation network configurations. Using an online database activation patterns of several different tasks were pooled together. A
high correlation between coactivation during task and resting-state correlation was detected44. In patients with first episode schizophrenia overlapping dysfunctions in the prefrontotemporal
pathway were evident45. Our study can serve as starting point for further combined analyses of resting state connectivity and activation patterns in stimulation-based designs from a network
perspective with a much more precise task design. Convergent with previous research which provided evidence for a greater functional similarity between the hemispheres in the cerebral voice
processing system than the face processing system46,47,48, in our study, both TVAs responded to voices in a sensitive (i.e. mixed contrast V > (A, E)) and preferential (i.e. minimum
contrast, V > max(A, E)) manner. In contrast, in the face processing system only the right FFA responds both in a sensitive and preferential way to faces, whereas the response of the left
FFA is only face-sensitive. We substantiated these results comparing voice- and face-sensitivity and -preferentiality of both hemispheres with lack of hemispheric differences in
voice-preferentiality, but significant hemispheric differences in the face processing system with greater face-preferentiality in the right hemisphere. This finding is in line with previous
results showing stronger and more consistent activation through faces in comparison to other stimulus categories in the right FFA compared to the left FFA46,49. The dominance of the right
hemisphere in face-related responses is not restricted to the FFA, but is also reflected in larger activation areas to faces in the right occipitotemporal cortex and the right amygdala and
an exclusive activation of the right inferior frontal gyrus46. Beyond this reliably replicated evidence, we found corresponding patterns in resting state measurements: The resting state
patterns predicting the face- and voice-sensitivity/-preferentiality, respectively, differed showing a convergent pattern for the voice processing system and a largely divergent pattern for
the face processing system as evidenced by the difference in significant overlaps of the informative connectivity patterns between the TVAs as compared to the FFAs. Thus, we conclude that
the different qualities of seeing faces and hearing voices do not work analogously, but that these two systems function in a unique and distinct way, with a higher hemispheric functional
similarity of the voice processing modules in comparison to the face processing system. In our supramodal approach combining voice and face processing networks with three regressors, one
common region in the medial frontal cortex correlated both voice- and face-preferentiality during resting state. The medial frontal cortex is known to be activated in higher order social
cognitive processing, the anterior rostral part especially in mentalizing tasks50. Additionally, it is involved in complex emotion processing13,51,52, independent of the presentation form,
e.g. visually via faces or bodies or acoustically via voices53. The activation of a region related to the processing of stimuli from different sensory modalities gives rise to the problem of
interpreting the results. Throughout this manuscript we use the term supramodal for the locally overlapping cerebral activation by signals from different sensory modalities which can be
identified using conjunction analyses, e.g. for mapping multisensory integration41,54. Limitations of the technique are that in our case the common region constitutes only a small part in
comparison to the complete connectivity pattern from each source, and that the local overlap not mandatorily represents a direct interaction or integration of signals from both sources, but
might indicate that the overlap region is simply linked to processing information from several sensory modalities. While the medial frontal cortex is not consistently activated in
stimulation experiments designed to localize voice- or face-specific brain areas, this notion would still appear quite plausible as effective processing of voices and faces might well be
required as basis for a variety of higher order social communication functions. In line with this, frontal areas were involved in the processing of incongruent but not congruent audiovisual
emotional stimuli55,56 and revealed emotion-specific activation regardless of the sensory modality of the emotional cue53. Whereas many studies assessing higher order social processing
employed emotional stimuli13,51,52, it is quite notable, that we found a convergence in this region even based on experimental designs without explicit emotional connotations. Limitations
concerning the assessment of neutral vs. emotional stimuli are discussed below. This seems to corroborate the notion that higher order social cognitive processes are linked to basic voice
and face perception irrespective of emotional information communicated via these stimuli. On the other hand, one might argue that there is no such thing as a voice or a face completely
devoid of emotional information in two ways: First, also stimuli not intended to carry emotional information by their sender may well contain subliminal emotional cues and, second, even a
putative completely neutral voice or face may automatically be scanned for emotional information and therefore become linked to emotion processing irrespective of its lack of emotional cues.
Previous results hint for a variability in the emotional perception of voices and faces depending on the previously experienced sensory input57,58. The posterior superior temporal sulcus
(pSTS), which has been shown to integrate simultaneously presented auditory and visual stimuli10, did not show an overlap of connectivity patterns indicating both voice- and
face-preferential responses. So, the pSTS’s role in combined face and voice processing might be more closely linked to the sensory integration of these stimuli during their simultaneous
perception and thus not be detectable in the resting state. Our work builds on the manifold confirmed and pioneering findings of regions that are preferentially activated by human stimuli in
comparison to environmental cues, i.e. especially the voice-preferential activation of the TVA and the face-preferential activation of the FFA4,5,6. And it broadens the perspective from
specialized regions for different tasks to a network perspective of regions exhibiting preferential responses both during and in the absence of human nonverbal cues. One could speculate that
the relevance of this finding lies in the reflection of relevant social situations during resting state, possibly including imagination of nonverbal cues. But to corroborate these ideas,
further research is necessary. The unique quality of our data stems from the combination of these individual cerebral processing characteristics of social stimuli with resting state
functional connectivity maps in a relatively large cohort. And it adds to the growing number of findings which advocate a readjustment of our view from specialized regions in the brain
responding to certain stimuli to a larger network perspective involving a multitude of regions across the whole brain in the presence and absence of tasks or/and stimuli. The specificity of
the activation is mediated not by the activation of single specialized regions itself but by the combination of simultaneously activated networks and therefore strengthens the view of a
network perspective. As gender-specific connectivity patterns were observed e.g. in the correlation of RSFC with personality traits22, this aspect represents a limitation of the present
study which focused on gender- and age-independent connectivity patterns. Due to the limited sample size, we did not perform subgroup analyses. Moreover, we did not assess and therefore were
able to correct for personality trait measures, such as the five-factor model of personality, which was shown to be associated with RSFC patterns22 and might therefore also represent a
moderator of the RSFC patterns associated with the propensity to respond to human voices and faces. Although based on seminal standard experiments to assess voice- and face-sensitive and
-preferential responses enabling direct comparisons with many previous studies, certain design-specific factors may have influenced the outcome of our study and should therefore be addressed
in further research: For one, the task-set differed considerably between the voice and face processing experiments (passive listening vs. one-back task) with potential influence on the
attentional status. As a further limitation we would like to address the problem of the assessment of human stimuli as neutral vs. emotional. Though not included in the experiments as
explicit factor employing face pictures with predominantly neutral expression, low-level emotional information in the experimental stimuli may have impacted the RSFC patterns predictive of
cerebral voice- and face-preferentiality. As a conclusion, these results emphasize that the individual cerebral propensity to respond to human voices and faces is reflected in the brain’s
activation patterns also in the absence of these cues as a possible neural corelate of mental reflections on relevant social situations including imagination of nonverbal cues during
“resting” state. The stronger convergence of informative connectivity patterns for the TVAs’ cue selectivity in contrast to the FFAs’ may indicate a higher hemispheric functional similarity
of the voice processing modules. The supramodal convergence of such informative connectivity patterns, in turn, points to the anterior medial prefrontal cortex as shared neural resource in
supramodal voice and face processing or potentially nonverbal communication. Similar to the underwater perspective on an iceberg, this experimental approach may open up interesting avenues
to the investigation of voice and face processing. In this regard, the resting state connectivity patterns correlating with individual voice and face selectivity may aid the understanding of
cerebral voice and face preference from a network perspective. DATA AVAILABILITY The datasets generated during and/or analysed during the current study are available from the corresponding
author on reasonable request. REFERENCES * Belin, P., Zatorre, R. J., Lafaille, P., Ahad, P. & Pike, B. Voice-selective areas in human auditory cortex. _Nature_ 403, 309–312.
https://doi.org/10.1038/35002078 (2000). Article ADS CAS PubMed Google Scholar * von Kriegstein, K. & Giraud, A. L. Implicit multisensory associations influence voice recognition.
_PLoS Biol._ 4, e326. https://doi.org/10.1371/journal.pbio.0040326 (2006). Article CAS Google Scholar * Ethofer, T. _et al._ Differential influences of emotion, task, and novelty on brain
regions underlying the processing of speech melody. _J. Cogn. Neurosci._ 21, 1255–1268. https://doi.org/10.1162/jocn.2009.21099 (2009). Article PubMed Google Scholar * Pernet, C. R. _et
al._ The human voice areas: Spatial organization and inter-individual variability in temporal and extra-temporal cortices. _Neuroimage_ 119, 164–174.
https://doi.org/10.1016/j.neuroimage.2015.06.050 (2015). Article PubMed Google Scholar * Kanwisher, N., McDermott, J. & Chun, M. M. The fusiform face area: A module in human
extrastriate cortex specialized for face perception. _J. Neurosci._ 17, 4302–4311 (1997). Article CAS Google Scholar * Kanwisher, N. & Yovel, G. The fusiform face area: A cortical
region specialized for the perception of faces. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 361, 2109–2128. https://doi.org/10.1098/rstb.2006.1934 (2006). Article PubMed PubMed Central
Google Scholar * Posamentier, M. T. & Abdi, H. Processing faces and facial expressions. _Neuropsychol. Rev._ 13, 113–143. https://doi.org/10.1023/a:1025519712569 (2003). Article PubMed
Google Scholar * Halgren, E. _et al._ Location of human face-selective cortex with respect to retinotopic areas. _Hum. Brain Mapp._ 7, 29–37 (1999). Article CAS Google Scholar *
Kreifelts, B., Ethofer, T., Huberle, E., Grodd, W. & Wildgruber, D. Association of trait emotional intelligence and individual fMRI-activation patterns during the perception of social
signals from voice and face. _Hum. Brain Mapp._ 31, 979–991. https://doi.org/10.1002/hbm.20913 (2010). Article PubMed Google Scholar * Kreifelts, B., Ethofer, T., Shiozawa, T., Grodd, W.
& Wildgruber, D. Cerebral representation of non-verbal emotional perception: fMRI reveals audiovisual integration area between voice- and face-sensitive regions in the superior temporal
sulcus. _Neuropsychologia_ 47, 3059–3066. https://doi.org/10.1016/j.neuropsychologia.2009.07.001 (2009). Article PubMed Google Scholar * Robins, D. L., Hunyadi, E. & Schultz, R. T.
Superior temporal activation in response to dynamic audio-visual emotional cues. _Brain Cogn._ 69, 269–278. https://doi.org/10.1016/j.bandc.2008.08.007 (2009). Article PubMed Google
Scholar * Ethofer, T., Pourtois, G. & Wildgruber, D. Investigating audiovisual integration of emotional signals in the human brain. _Prog. Brain Res._ 156, 345–361.
https://doi.org/10.1016/S0079-6123(06)56019-4 (2006). Article PubMed Google Scholar * Klasen, M., Kenworthy, C. A., Mathiak, K. A., Kircher, T. T. & Mathiak, K. Supramodal
representation of emotions. _J. Neurosci._ 31, 13635–13643. https://doi.org/10.1523/JNEUROSCI.2833-11.2011 (2011). Article CAS PubMed PubMed Central Google Scholar * Ethofer, T. _et
al._ Functional responses and structural connections of cortical areas for processing faces and voices in the superior temporal sulcus. _Neuroimage_ 76, 45–56.
https://doi.org/10.1016/j.neuroimage.2013.02.064 (2013). Article PubMed Google Scholar * Ethofer, T. _et al._ Cerebral pathways in processing of affective prosody: A dynamic causal
modeling study. _Neuroimage_ 30, 580–587. https://doi.org/10.1016/j.neuroimage.2005.09.059 (2006). Article PubMed Google Scholar * Kreifelts, B. _et al._ Tuned to voices and faces:
Cerebral responses linked to social anxiety. _Neuroimage_ 197, 450–456. https://doi.org/10.1016/j.neuroimage.2019.05.018 (2019). Article PubMed Google Scholar * Karle, K. N. _et al._
Neurobiological correlates of emotional intelligence in voice and face perception networks. _Soc. Cogn. Affect. Neurosci._ 13, 233–244. https://doi.org/10.1093/scan/nsy001 (2018). Article
PubMed PubMed Central Google Scholar * Smitha, K. A. _et al._ Resting state fMRI: A review on methods in resting state connectivity analysis and resting state networks. _Neuroradiol. J._
30, 305–317. https://doi.org/10.1177/1971400917697342 (2017). Article CAS PubMed PubMed Central Google Scholar * Hahn, T. _et al._ Reliance on functional resting-state network for
stable task control predicts behavioral tendency for cooperation. _Neuroimage_ 118, 231–236. https://doi.org/10.1016/j.neuroimage.2015.05.093 (2015). Article PubMed Google Scholar *
Inagaki, T. K. & Meyer, M. L. Individual differences in resting-state connectivity and giving social support: Implications for health. _Soc. Cogn. Affect. Neurosci._ 15, 1076–1085.
https://doi.org/10.1093/scan/nsz052 (2020). Article PubMed Google Scholar * Serafini, G., Pardini, M., Pompili, M., Girardi, P. & Amore, M. Understanding suicidal behavior: The
contribution of recent resting-state fMRI techniques. _Front. Psychiatry_ 7, 69. https://doi.org/10.3389/fpsyt.2016.00069 (2016). Article PubMed PubMed Central Google Scholar * Nostro,
A. D. _et al._ Predicting personality from network-based resting-state functional connectivity. _Brain Struct. Funct._ 223, 2699–2719. https://doi.org/10.1007/s00429-018-1651-z (2018).
Article PubMed PubMed Central Google Scholar * Markett, S., Montag, C. & Reuter, M. Network neuroscience and personality. _Pers. Neurosci._ 1, e14.
https://doi.org/10.1017/pen.2018.12 (2018). Article Google Scholar * Parkes, L., Satterthwaite, T. D. & Bassett, D. S. Towards precise resting-state fMRI biomarkers in psychiatry:
Synthesizing developments in transdiagnostic research, dimensional models of psychopathology, and normative neurodevelopment. _Curr. Opin. Neurobiol._ 65, 120–128.
https://doi.org/10.1016/j.conb.2020.10.016 (2020). Article CAS PubMed PubMed Central Google Scholar * Romero-Martinez, A. _et al._ The brain resting-state functional connectivity
underlying violence proneness: Is it a reliable marker for neurocriminology? A systematic review. _Behav. Sci. (Basel)._ https://doi.org/10.3390/bs9010011 (2019). Article PubMed PubMed
Central Google Scholar * van den Heuvel, M. P. & Hulshoff Pol, H. E. Exploring the brain network: A review on resting-state fMRI functional connectivity. _Eur. Neuropsychopharmacol._
20, 519–534. https://doi.org/10.1016/j.euroneuro.2010.03.008 (2010). Article CAS PubMed Google Scholar * Zhang, H., Tian, J., Liu, J., Li, J. & Lee, K. Intrinsically organized
network for face perception during the resting state. _Neurosci. Lett._ 454, 1–5. https://doi.org/10.1016/j.neulet.2009.02.054 (2009). Article CAS PubMed PubMed Central Google Scholar *
O’Neil, E. B., Hutchison, R. M., McLean, D. A. & Kohler, S. Resting-state fMRI reveals functional connectivity between face-selective perirhinal cortex and the fusiform face area
related to face inversion. _Neuroimage_ 92, 349–355. https://doi.org/10.1016/j.neuroimage.2014.02.005 (2014). Article PubMed Google Scholar * Kruschwitz, J. D. _et al._ Segregation of
face sensitive areas within the fusiform gyrus using global signal regression? A study on amygdala resting-state functional connectivity. _Hum. Brain Mapp._ 36, 4089–4103.
https://doi.org/10.1002/hbm.22900 (2015). Article PubMed PubMed Central Google Scholar * Correia, A. I. _et al._ Resting-state connectivity reveals a role for sensorimotor systems in
vocal emotional processing in children. _Neuroimage_ 201, 116052. https://doi.org/10.1016/j.neuroimage.2019.116052 (2019). Article PubMed Google Scholar * Oldfield, R. C. The assessment
and analysis of handedness: The Edinburgh inventory. _Neuropsychologia_ 9, 97–113. https://doi.org/10.1016/0028-3932(71)90067-4 (1971). Article CAS PubMed Google Scholar * Kreifelts, B.
_et al._ Non-verbal emotion communication training induces specific changes in brain function and structure. _Front. Hum. Neurosci._ 7, 648. https://doi.org/10.3389/fnhum.2013.00648 (2013).
Article PubMed PubMed Central Google Scholar * Kreifelts, B. _et al._ The neural correlates of face-voice-integration in social anxiety disorder. _Front. Psychiatry_ 11, 657.
https://doi.org/10.3389/fpsyt.2020.00657 (2020). Article PubMed PubMed Central Google Scholar * Collins, D. L., Neelin, P., Peters, T. M. & Evans, A. C. Automatic 3D intersubject
registration of MR volumetric data in standardized Talairach space. _J. Comput. Assist. Tomogr_ 18, 192–205 (1994). Article CAS Google Scholar * Friston, K. J. _et al._ Classical and
Bayesian inference in neuroimaging: Applications. _Neuroimage_ 16, 484–512. https://doi.org/10.1006/nimg.2002.1091 (2002). Article CAS PubMed Google Scholar * Worsley, K. J. _et al._ A
unified statistical approach for determining significant signals in images of cerebral activation. _Hum. Brain Mapp._ 4, 58–73.
https://doi.org/10.1002/(SICI)1097-0193(1996)4:1%3c58::AID-HBM4%3e3.0.CO;2-O (1996). Article CAS PubMed Google Scholar * Kreifelts, B. _et al._ Cerebral resting state markers of biased
perception in social anxiety. _Brain Struct. Funct._ 224, 759–777. https://doi.org/10.1007/s00429-018-1803-1 (2019). Article PubMed Google Scholar * Whitfield-Gabrieli, S. &
Nieto-Castanon, A. Conn: A functional connectivity toolbox for correlated and anticorrelated brain networks. _Brain Connect_ 2, 125–141. https://doi.org/10.1089/brain.2012.0073 (2012).
Article PubMed Google Scholar * Tzourio-Mazoyer, N. _et al._ Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject
brain. _Neuroimage_ 15, 273–289. https://doi.org/10.1006/nimg.2001.0978 (2002). Article CAS PubMed Google Scholar * Whitfield-Gabrieli, S. N.-C. A. _CONN Toolbox Manual_.
https://web.conn-toolbox.org/resources/documentation/manual (2017). Accessed 23 Aug 2017. * Nichols, T., Brett, M., Andersson, J., Wager, T. & Poline, J. B. Valid conjunction inference
with the minimum statistic. _Neuroimage_ 25, 653–660. https://doi.org/10.1016/j.neuroimage.2004.12.005 (2005). Article PubMed Google Scholar * Zhu, Q., Zhang, J., Luo, Y. L., Dilks, D. D.
& Liu, J. Resting-state neural activity across face-selective cortical regions is behaviorally relevant. _J. Neurosci._ 31, 10323–10330. https://doi.org/10.1523/JNEUROSCI.0873-11.2011
(2011). Article CAS PubMed PubMed Central Google Scholar * Liu, J., Yuan, B., Luo, Y. J. & Cui, F. Intrinsic functional connectivity of medial prefrontal cortex predicts the
individual moral bias in economic valuation partially through the moral sensitivity trait. _Brain Imaging Behav._ 14, 2024–2036. https://doi.org/10.1007/s11682-019-00152-1 (2020). Article
PubMed Google Scholar * Di, X., Gohel, S., Kim, E. H. & Biswal, B. B. Task vs. rest-different network configurations between the coactivation and the resting-state brain networks.
_Front. Hum. Neurosci._ 7, 493. https://doi.org/10.3389/fnhum.2013.00493 (2013). Article PubMed PubMed Central Google Scholar * Mwansisya, T. E. _et al._ Task and resting-state fMRI
studies in first-episode schizophrenia: A systematic review. _Schizophr. Res._ 189, 9–18. https://doi.org/10.1016/j.schres.2017.02.026 (2017). Article PubMed Google Scholar * Engell, A.
D. & McCarthy, G. Probabilistic atlases for face and biological motion perception: An analysis of their reliability and overlap. _Neuroimage_ 74, 140–151.
https://doi.org/10.1016/j.neuroimage.2013.02.025 (2013). Article PubMed Google Scholar * Brancucci, A., Lucci, G., Mazzatenta, A. & Tommasi, L. Asymmetries of the human social brain
in the visual, auditory and chemical modalities. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 364, 895–914. https://doi.org/10.1098/rstb.2008.0279 (2009). Article CAS PubMed Google Scholar
* Bonte, M. _et al._ Development from childhood to adulthood increases morphological and functional inter-individual variability in the right superior temporal cortex. _Neuroimage_ 83,
739–750. https://doi.org/10.1016/j.neuroimage.2013.07.017 (2013). Article PubMed Google Scholar * Schwarz, L. _et al._ Properties of face localizer activations and their application in
functional magnetic resonance imaging (fMRI) fingerprinting. _PLoS ONE_ 14, e0214997. https://doi.org/10.1371/journal.pone.0214997 (2019). Article CAS PubMed PubMed Central Google
Scholar * Amodio, D. M. & Frith, C. D. Meeting of minds: The medial frontal cortex and social cognition. _Nat. Rev. Neurosci._ 7, 268–277. https://doi.org/10.1038/nrn1884 (2006).
Article CAS PubMed Google Scholar * Wildgruber, D. _et al._ Different types of laughter modulate connectivity within distinct parts of the laughter perception network. _PLoS ONE_ 8,
e63441. https://doi.org/10.1371/journal.pone.0063441 (2013). Article ADS CAS PubMed PubMed Central Google Scholar * Bruck, C., Kreifelts, B., Gossling-Arnold, C., Wertheimer, J. &
Wildgruber, D. “Inner voices”: The cerebral representation of emotional voice cues described in literary texts. _Soc. Cogn. Affect. Neurosci._ 9, 1819–1827.
https://doi.org/10.1093/scan/nst180 (2014). Article PubMed PubMed Central Google Scholar * Peelen, M. V., Atkinson, A. P. & Vuilleumier, P. Supramodal representations of perceived
emotions in the human brain. _J. Neurosci._ 30, 10127–10134. https://doi.org/10.1523/JNEUROSCI.2161-10.2010 (2010). Article CAS PubMed PubMed Central Google Scholar * Friston, K. J.,
Penny, W. D. & Glaser, D. E. Conjunction revisited. _Neuroimage_ 25, 661–667. https://doi.org/10.1016/j.neuroimage.2005.01.013 (2005). Article PubMed Google Scholar * Muller, V. I.
_et al._ Incongruence effects in crossmodal emotional integration. _Neuroimage_ 54, 2257–2266. https://doi.org/10.1016/j.neuroimage.2010.10.047 (2011). Article PubMed Google Scholar *
Davies-Thompson, J. _et al._ Hierarchical brain network for face and voice integration of emotion expression. _Cereb. Cortex_ 29, 3590–3605. https://doi.org/10.1093/cercor/bhy240 (2019).
Article PubMed Google Scholar * Webster, M. A. & MacLeod, D. I. Visual adaptation and face perception. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 366, 1702–1725.
https://doi.org/10.1098/rstb.2010.0360 (2011). Article PubMed PubMed Central Google Scholar * Bestelmeyer, P. E., Rouger, J., DeBruine, L. M. & Belin, P. Auditory adaptation in vocal
affect perception. _Cognition_ 117, 217–223. https://doi.org/10.1016/j.cognition.2010.08.008 (2010). Article PubMed Google Scholar Download references ACKNOWLEDGEMENTS This work was
supported by the Clinician Scientist program of the University of Tübingen to KNE (Grant Number 367-0-0). We acknowledge support by Open Access Publishing Fund of University of Tübingen. The
funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. FUNDING Open Access funding enabled and organized by Projekt DEAL.
AUTHOR INFORMATION AUTHORS AND AFFILIATIONS * Department of Psychiatry and Psychotherapy, Tübingen Center for Mental Health (TüCMH), University of Tübingen, Calwerstrasse 14, 72076,
Tübingen, Germany Kathrin N. Eckstein, Dirk Wildgruber, Thomas Ethofer, Carolin Brück, Heike Jacob & Benjamin Kreifelts * Department for Biomedical Magnetic Resonance, University of
Tübingen, Tübingen, Germany Thomas Ethofer & Michael Erb Authors * Kathrin N. Eckstein View author publications You can also search for this author inPubMed Google Scholar * Dirk
Wildgruber View author publications You can also search for this author inPubMed Google Scholar * Thomas Ethofer View author publications You can also search for this author inPubMed Google
Scholar * Carolin Brück View author publications You can also search for this author inPubMed Google Scholar * Heike Jacob View author publications You can also search for this author
inPubMed Google Scholar * Michael Erb View author publications You can also search for this author inPubMed Google Scholar * Benjamin Kreifelts View author publications You can also search
for this author inPubMed Google Scholar CONTRIBUTIONS The study was conceptualized by D.W., T.E. and B.K. Data acquisition and curation was done by C.B., H.J. and M.E. Analyses were
performed by K.N.E., B.K., D.W., T.E., C.B., H.J. and M.E. K.N.E. and B.K. wrote the main manuscript text and prepared the figures. All authors reviewed the manuscript. CORRESPONDING AUTHOR
Correspondence to Kathrin N. Eckstein. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests. ADDITIONAL INFORMATION PUBLISHER'S NOTE Springer Nature
remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. SUPPLEMENTARY INFORMATION SUPPLEMENTARY FIGURE 1. RIGHTS AND PERMISSIONS OPEN ACCESS
This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as
long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third
party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the
article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the
copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Eckstein, K.N., Wildgruber,
D., Ethofer, T. _et al._ Correlates of individual voice and face preferential responses during resting state. _Sci Rep_ 12, 7117 (2022). https://doi.org/10.1038/s41598-022-11367-6 Download
citation * Received: 29 December 2021 * Accepted: 15 April 2022 * Published: 03 May 2022 * DOI: https://doi.org/10.1038/s41598-022-11367-6 SHARE THIS ARTICLE Anyone you share the following
link with will be able to read this content: Get shareable link Sorry, a shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature
SharedIt content-sharing initiative