Correlates of individual voice and face preferential responses during resting state

Correlates of individual voice and face preferential responses during resting state

Play all audios:

Loading...

ABSTRACT Human nonverbal social signals are transmitted to a large extent by vocal and facial cues. The prominent importance of these cues is reflected in specialized cerebral regions which


preferentially respond to these stimuli, e.g. the temporal voice area (TVA) for human voices and the fusiform face area (FFA) for human faces. But it remained up to date unknown whether


there are respective specializations during resting state, i.e. in the absence of any cues, and if so, whether these representations share neural substrates across sensory modalities. In the


present study, resting state functional connectivity (RSFC) as well as voice- and face-preferential activations were analysed from functional magnetic resonance imaging (fMRI) data sets of


60 healthy individuals. Data analysis comprised seed-based analyses using the TVA and FFA as regions of interest (ROIs) as well as multi voxel pattern analyses (MVPA). Using the face- and


voice-preferential responses of the FFA and TVA as regressors, we identified several correlating clusters during resting state spread across frontal, temporal, parietal and occipital


regions. Using these regions as seeds, characteristic and distinct network patterns were apparent with a predominantly convergent pattern for the bilateral TVAs whereas a largely divergent


pattern was observed for the bilateral FFAs. One region in the anterior medial frontal cortex displayed a maximum of supramodal convergence of informative connectivity patterns reflecting


voice- and face-preferential responses of both TVAs and the right FFA, pointing to shared neural resources in supramodal voice and face processing. The association of individual voice- and


face-preferential neural activity with resting state connectivity patterns may support the perspective of a network function of the brain beyond an activation of specialized regions. SIMILAR


CONTENT BEING VIEWED BY OTHERS BASAL GANGLIA AND CEREBELLUM CONTRIBUTIONS TO VOCAL EMOTION PROCESSING AS REVEALED BY HIGH-RESOLUTION FMRI Article Open access 20 May 2021 THE


REPRESENTATIONAL DYNAMICS OF PERCEIVED VOICE EMOTIONS EVOLVE FROM CATEGORIES TO DIMENSIONS Article 11 March 2021 MAPPING OF FACIAL AND VOCAL PROCESSING IN COMMON MARMOSETS WITH ULTRA-HIGH


FIELD FMRI Article Open access 13 March 2024 INTRODUCTION Voices and faces are among the most salient cues in human life. This is reflected in the existence of specialized cerebral modules


which are hierarchically organized and specifically tuned to respond to these cues. Core components for the primary identification of human voices and faces are the temporal voice area (TVA)


for voices1,2,3,4 and the fusiform face area (FFA) for faces5,6,7,8. While not exclusively activated by these signals, they exhibit clearly voice- and face-preferential responses,


respectively. The FFA together with the occipital face area (OFA) respond mainly to invariant facial features (e.g. gender)5,8. Further processing of dynamic face aspects, and integration of


signals from voices and faces involves the posterior superior temporal sulcus (pSTS) and the thalamus9,10,11,12. The emotional information often present in faces and voices (e.g. in facial


expressions and emotional prosody) additionally converges in the amygdala9,13. Further processing of such emotional information involves further regions such as the inferior frontal cortex


(IFC) and orbitofrontal cortex (OFC)14,15. Convergent with the particular importance of voices and faces in human social communication, recent studies indicated that the responsivity to the


preferred cues of the basic modules for identification of human voices and faces is moderated by interindividual differences in social signal processing, e.g. social anxiety16, and emotional


intelligence17, even in the absence of emotional information. In some cases, as described above e.g. for the pSTS and thalamus, the hemodynamic correlates of cerebral processing of signals


from different sensory modalities overlap. This phenomenon will be termed supramodal throughout this manuscript. While a plethora of neuroimaging studies delineated the neural networks that


are active when we see faces or hear voices, it remains a completely open question if the brain’s activity patterns also reflect the individual cerebral responsivity to voices and faces in


the absence of these cues and if these representations may share neural substrates across sensory modalities. During the past three decades, the resting brain has become a major research


focus as it became clear that spontaneous physiological low-frequency fluctuations in brain activity occur non-randomly but simultaneously in various, partially overlapping neural networks


in the absence of any cues or stimulation or cognitive/emotional task18. Nevertheless, these fluctuation patterns are not independent from individual traits or diseases, as they have been


shown to correlate with various aspects of behavioural tendencies19,20,21, personality22,23, psychopathology21,24,25, and psychiatric disease (e.g. dementia and schizophrenia26) also


demonstrating that resting state data can be used to expand the neuroimaging perspective on their cerebral representation in a complementary manner with the potential to detect links between


the neural networks underlying various perceptual, cognitive or emotional functions not apparent in stimulation-based designs. In the area of face and voice processing, correlations of


resting state functional connectivity (RSFC) with behavioural outcomes, e.g. performance in various face- and voice-processing tasks have been observed27,28,29,30. One study compared


functional connectivity patterns during resting state and a passive viewing task and found for both conditions similar networks including posterior fusiform gyrus, inferior occipital gyrus


and superior temporal sulcus27. In this work the informative RSFC patterns were found exclusively within the network of modality-specific preferential processing areas27. Two studies


combined RSFC in the face processing network with behavioural performance in a face identification task and an emotional face matching task, respectively28,29 and found RSFC patterns between


modality-specific preferential processing areas but also with other parts of the brain28,29. One study in children revealed that performance in an auditory emotional prosody recognition


task was predicted by stronger connectivity between the inferior frontal gyrus and motor regions. Here, informative RSFC patterns were found exclusively outside the modality-specific


preferential processing networks30. In the present study, we intended to determine the neural correlates of voice- and face-preferential responses in the absence of voices and faces in the


resting state. Furthermore, we aimed to identify brain areas with RSFC patterns supramodally reflecting preferential responses to both, voices and faces. To this end, 60 healthy individuals


underwent functional magnetic resonance imaging (fMRI) at rest and during stimulation with voices, faces and various other classes of acoustic and visual stimuli. Individual voice- and


face-preferential responses were correlated with RSFC employing multi voxel pattern analyses (MVPA) and seed-based analyses focused on TVA and FFA. MATERIALS AND METHODS PARTICIPANTS 60


healthy individuals (mean age 25.8 years, s.d. = 4.5 years, 30 female) participated at the University of Tübingen. All of the participants were native German speakers and right-handed, as


assessed with the Edinburgh Inventory31. None of the participants was taking any regular medication, or had a history of substance abuse, or psychiatric or neurological illness. Hearing was


normal, vision normal or corrected to normal in all participants. The study was performed according to the Code of Ethics of the World Medical Association (Declaration of Helsinki) and the


protocol of human investigation was approved by the local ethics committees where the study was performed (i.e., the medical faculties of the Universities of Tübingen and Greifswald). All


individuals gave their written informed consent prior to their participation in the study. STIMULI AND EXPERIMENTAL DESIGN Two fMRI experiments were performed to localize face-sensitive5 and


voice-sensitive1 brain areas as described in previous publications9,10,14,16,17,32,33: For the face-sensitivity experiment, pictures from four different categories (faces, houses, objects,


and natural scenes) were employed within a block design. All stimuli used in the experiment were black-and-white photographs unknown to the participants17. The shown face stimuli had no


obvious emotional connotation, but rather showed neutral facial expressions. The house stimuli were multilevel apartment houses from different materials (brick, wooden, concrete). As object


stimuli different everyday life items were used (e.g. flat iron, spoon, T-shirt). The fourth category of natural scenes represented different countryside pictures (e.g. mountain, coastline,


waterfall). Each block and category contained 20 stimuli17. Within blocks, the stimuli were presented in random order for 300 ms. Stimuli were separated by 500 ms periods of fixation [1


block = 20 stimuli × (300 ms picture + 500 ms fixation) = 16 s]. Eight blocks of each category pseudorandomized within the experiment were shown separated by short ~ 1.5 s rest periods17. A


one-back task was employed, in which the participants had to press a button on a fibre optic system (LumiTouch, Photon Control, Burnaby, Canada) with their right index finger when they saw a


picture twice in a row, to ascertain constant attention17. The appearance of repeated stimuli was pseudorandomized ensuring a distribution across the entire experiment. Visual stimuli were


back-projected onto a screen placed in the magnet bore behind the participant’s head and viewed by the participant through a mirror system mounted onto the head coil. The voice-sensitivity


experiment was developed based on the study by Belin et al.1 in form of a block design experiment with 24 stimulation blocks and 12 silent periods (each 8 s) in a passive-listening design


without an explicit task. Between the blocks were short periods without sound (2 s). Participants were instructed to listen attentively with their eyes closed. The stimulus material


comprised 12 blocks of human vocal sounds (speech, sighs, laughs, cries), 6 blocks of animal sounds (e.g., gallops, various cries) and 6 blocks of environmental sounds (e.g., cars, planes,


doors, telephones). Stimuli were normalized with respect to mean acoustic energy17. Sound and silence blocks were pseudorandomized across the experiment with the restriction that with the


restriction that no two blocks of silence directly followed each other. Both experimental designs have been validated in previous studies9,10,14,17,32,33. Further details on the stimulus


material and experimental designs have been reported elsewhere9. For the resting state measurements (duration about 7 min and 15 s), the participants were instructed to keep their eyes


closed with no further task. IMAGE ACQUISITION MRI data were acquired with a TRIO 3T and a PRISMA scanner (Siemens, Erlangen, Germany). Structural T1-weighted images (176 slices, TR = 2300 


ms, TE = 2.96 ms, TI = 1100 ms, voxel size: 1 × 1 × 1 mm3) and functional images (30 axial slices captured in sequential descending order, 3 mm thickness + 1 mm gap, TR = 1.7 s, TE = 30 ms,


voxel size: 3 × 3 × 4 mm3, field of view 192 × 192 mm2, 64 × 64 matrix, flip angle 90°) were recorded. For the resting state measurements, 245 images were recorded. The activation tasks were


performed after completion of the resting state measurements to avoid carry-over effects. The time series comprised 368 images for the face experiment and 232 images for the voice


experiment and 250 images for the resting state measurement. A field map with 36 slices (slice thickness 3 mm, TR = 400 ms, TE(1) = 5.19 ms, TE(2) = 7.65 ms) was recorded. ANALYSIS OF FMRI


DATA Statistical parametric mapping software (SPM8, Wellcome Department of Imaging Neuroscience, London, http://www.fil.ion.ucl.ac.uk/spm) was used to analyse the imaging data.


Pre-processing generally included the removal of the first five EPI images from each run to exclude measurements preceding T1 equilibrium. FACE- AND VOICE-SENSITIVITY EXPERIMENTS The


preprocessing procedure consisted of realignment, unwarping using a static field map, coregistration of anatomical and functional images, segmentation of the anatomical images, normalization


into MNI space (Montreal Neurological Institute34) with a resampled voxel size of 3 × 3 × 3 mm3, temporal smoothing with a high-pass filter (cut-off frequency of 1/128 Hz) and spatial


smoothing employing a Gaussian kernel (8 mm full width at half maximum, FWHM). The response to the single categories (faces (F), houses (H), objects (O), and natural scenes (S) in the face


localizer as well as vocal sounds (V), animal sounds (A), and environmental sounds (E) in the voice localizer were independently modelled with a box-car function corresponding to the


duration of the stimulation blocks (16 s in the face localizer and 8 s in the voice localizer) convolved with the hemodynamic response function (HRF). The error term was calculated as a


first order autoregressive process with a coefficient of 0.2 and a white noise component accounting for serial autocorrelations35. To minimize motion-associated error variance, the six


motion parameters (i.e. translation and rotation on the x-, y-, and z-axes) were included in the single subject models as covariates. Contrast images were constructed using data from the


first-level general linear models [face-sensitivity: F > (H, O, S); voice-sensitivity: V > (A, E)] for each subject. Taking these contrast images as sources, a second-level


random-effect analysis was performed with one-sample t-tests to define the face-sensitive fusiform face area (FFA) and the voice-sensitive temporal voice area (TVA) as functional regions of


interest (ROI) for further analyses. Statistical significance of activations was assessed at p < 0.001, uncorrected at voxel level and with FWE correction for multiple comparisons at


cluster level with p < 0.05. For the definition of the FFA, the fusiform gyrus was taken as a priori anatomical ROI; for definition of the TVA, the temporal gyri and the temporal pole


were selected. For definition of the functional ROIs (i.e. FFA and TVA), FWE-cluster level correction was performed across these a priori anatomical ROIs using small volume correction


(SVC36). We picked the maximum activation in the fusiform gyrus for the FFA and in the temporal lobe for the TVA respectively, and defined the surrounding 100 most sensitive voxels as masks


for the functional ROIs. Within these ROIs individual voice- and face-preferential responses were assessed using minimum difference criteria (for voices V > max[A, E], for faces F > 


max[H, O, S])37. Intercorrelations of the four regressors were evaluated. Differences in the face- and voice-sensitive and -preferential responses between both hemispheres and


interhemispheric differences in cue-sensitivity and -preferentiality between TVA and FFA were post hoc tested using two-sided paired t-tests with Bonferroni correction. RESTING STATE


FUNCTIONAL CONNECTIVITY ANALYSIS For RSFC analyses we used the CONN toolbox (v 16b38) implemented in SPM8. The spatial preprocessing was performed analogously to the procedure described for


the face- and voice-sensitivity experiments. Denoising included linear regression of the following confounding effects: White matter and CSF components (6P each), effect of rest (2P,


temporal component and first order derivates) and motion regression (12 regressors: 6 motion parameters and 6 first-order temporal derivates) and band-pass filtering (0.008–0.09 Hz). Linear


detrending was added to remove linear trends. The participants’ movement parameters, their first order derivatives and the BOLD signal from white matter, cerebrospinal fluid and effect of


rest (each with five temporal components) were included in the analysis as covariates to reduce their confounding influences. In the individual first-level analyses, bivariate correlation


coefficients were calculated as linear measures of functional connectivity for the ensuing analyses. Coefficients were Z transformed to achieve comparability for group-level analyses, and


gender, age and scanner were included as regressors of no interest. The Automated Anatomic Labelling (AAL) toolbox39 was used for the definition of anatomical regions in MNI space. The


analysis targeted the correlation of individual resting state functional connectivity (RSFC) with face-/voice-preferential responses both with defined regions of interest (ROIs) and at whole


brain level. To this end, analyses were done on two different levels: ROI-to-voxel analyses should detect associations between individual voice- and face-preferential responses of the ROIs


and their RSFC with other brain regions. Here, the significance of observed connectivity patterns was assessed using a threshold of p < 0.001 at voxel level, two-tailed with FWE


correction (p < 0.05) for multiple comparisons at cluster level. Results were Bonferroni-corrected for the numbers of regressors (4) and ROIs (4), so that the effective cluster threshold


amounted to p < 0.00315. Second, a spatial hypothesis-free strategy was implemented using voxel-to-voxel multivariate multi voxel pattern analyses (MVPA). Here, for each voxel separately,


a low-dimensional multivariate representation of the connectivity pattern between this voxel and all other voxels in the brain was calculated. This representation was based on a principal


component analysis of the inter-subject variability of each separate voxel’s connectivity pattern enabling the investigation of differences across subjects using second-level multivariate


analyses. The number of principal components was set to three and number of dimensions was set to 64 (dimensionality reduction)40. The goal of the group-MVPA approach was to detect whole


brain resting state functional connectivity patterns correlating with individual voice-preferential responses of the TVA (i.e., V > max[A,E]) and face-preferential responses (i.e., F >


 max[H,O,S]) of the FFA. These individual estimates were used as group level regressors in the RSFC analyses (four regressors: two for the FFAs, two for the TVAs). Results were evaluated at


a voxel-wise threshold of p < 0.001 and whole brain FWE-corrected at cluster level with additional Bonferroni-correction for the number of tested regressors (4) resulting in an effective


cluster threshold of p < 0.0125. Findings of the MVPA were further analysed using the significant clusters as seeds for ensuing seed-to-voxel analyses. Convergence of RSFC patterns


between different seeds was tested using conjunction analyses with a minimum statistic41. Results were assessed at a voxel-wise threshold of p < 0.001 and whole brain FWE-corrected at


cluster level with a cluster threshold of p < 0.05. RESULTS ROI CHARACTERISTICS The activation pattern of the right and left FFA showed a significant sensitivity for faces (rFFA t = 


9.321, p < 0.001 and lFFA t = 7.585, p < 0.001), whereas significant face-preferential responses were observed in the right FFA (t = 4.344, p < 0.0001), but not the left FFA (t = 


0.624, p = 0.535). The bilateral TVAs were highly sensitive to and preferential for voices (sensitivity: rTVA t = 18.265, p < 0.0001 and lTVA t = 17.457, p < 0.001; preferentiality:


rTVA t = 14.456, p < 0.001 and lTVA t = 14.023, p < 0.001). ROI characteristics are graphically displayed in Fig. 1. The ROIs’ preferential responses to their preferred cues were


significantly correlated within modality (voices: r(58) = 0.74, p < 0.001; faces: r(58) = 0.60, p < 0.001) but not across modalities (all abs(r(58)) < 0.12, all p > 0.05).


Comparison between the right and left hemisphere revealed no significant difference for voice-sensitivity or -preferentiality (all t < 2.03), all p > 0.187), but significant


differences for face-sensitivity and -preferentiality in favour of the right hemisphere (all t > 3.75, all p < 0.004). Comparison of hemispheric differences in modality-specific


differences in cue sensitivity and preferentiality between TVA and FFA corroborated the difference between the sensory modalities, both for sensitivity and for preferentiality (all t > 


3.93, p < 0.002), i.e. a greater hemispheric difference in face-sensitivity and -preferentiality than in voice-sensitivity and -preferentiality. ROI-TO-VOXEL ANALYSIS In this analysis,


only individual voice-preferential responses of the lTVA were significantly associated with RSFC between the lTVA and a cluster in the right supramarginal gyrus extending into the inferior


parietal gyrus (peak: − 57x − 66y 27z; 143 voxels; p(FWE-corr.) = 0.0018). MULTI-VOXEL PATTERN ANALYSIS (MVPA) Using rFFA face-preferential responses as regressor, we identified one


informative cluster in the right middle frontal gyrus extending into the precentral gyrus. For the lFFA two clusters in the left caudate nucleus/olfactory gyrus and left superior temporal


pole were evident. For the rTVA and lTVA voice-preferential responses four overlapping clusters emerged: in the left superior occipital gyrus, the right inferior parietal gyrus, the right


superior temporal gyrus and the right frontal inferior orbital gyrus. For rTVA voice-preferentiality two additional clusters were detected in the left middle occipital gyrus and the right


thalamus, for the left TVA two additional clusters were located in the left frontal superior gyrus and the right parietal superior gyrus. A detailed description of the clusters can be found


in Table 1. A graphical representation is displayed in Fig. 2. For the four overlapping clusters informative of both the rTVA and lTVA voice-preferential responses common regions were


calculated and further on used as seeds. The characteristics of the resulting clusters are described in Table 2. Significant clusters were used as seeds for subsequent post-hoc explanatory


seed-to-voxel analyses. For the TVAs the convergence of informative MVPA clusters was accompanied by a relatively strong convergence of their RSFC patterns in contrast to the FFAs’ RSFC


patterns. Tables 3, 4 and 5 give an overview of convergent RSFC clusters across all informative regions observed in the MVPA analysis. Convergent clusters for the bilateral TVAs are listed


in Table 3, exemplary graphical representations are given in Fig. 3. In contrast to these results, for the FFAs, in addition to the lower number of informative clusters in the MVPA analysis,


the RSFC pattern was largely divergent as exemplarily shown for two MVPA clusters informative of FFA face-selective responses (rFFA: right middle frontal gyrus extending into the precentral


gyrus, lFFA: left caudate nucleus and olfactory gyrus). Only one significant common cluster was observed in the right supramarginal gyrus extending into the inferior parietal gyrus (peak:


57x − 27 < 45z; 81 voxels, p(FWE-corr.) = 0.010) using the of the right R middle frontal gyrus/precentral gyrus (rFFA) and the left caudate nucleus and olfactory gyri (− 6 6 − 15, lFFA)


as seeds. The results are illustrated in Fig. 4. SUPRAMODAL CONVERGENCE OF INFORMATIVE RSFC PATTERNS The combination of RSFC correlates of individual face-preferential responses in the right


and left FFA with RSFC correlates of individual voice-preferential responses in the right and left TVA can decipher supramodal convergence of RSFC patterns, i.e. combining voice- and


face-preferentiality. In our case, this was evident in eight clusters (Table 5). The convergence was more prominent using right-hemispheric voice- and face-preferentiality regressors with


five common clusters, whereas for the left-hemispheric regressors only one supramodal cluster was found. Two clusters derived from regressors of contralateral hemispheres. Only one region in


the anterior region of the rostral mediofrontal cortex (arMFC) exhibited supramodal convergence of informative RSFC patterns for more than two regressors: Convergence of the RSFC of the


rlTVA cluster in the left superior occipital gyrus with the lTVA cluster in the left frontal superior gyrus and the rFFA cluster in the right middle frontal gyrus delineated one common


region in the medial frontal gyrus (including the left orbital gyrus and the anterior cingulum as well as the right and left medial frontal gyrus, peak: 0 × 54y 9z; 83 voxels, p(FWE-corr.) =


 0.011) indicative of right and left TVA voice-preferentiality as well as rFFA face-preferentiality (see also Fig. 5). DISCUSSION Combining seminal experiments used to localize voice- as


well as face-preferential areas in the human brain and resting state fMRI, this study provides the first description of hemodynamic functional connectivity patterns in the resting state that


are associated with voice- and face-preferential cerebral responses at the primary level of the TVA and FFA. Using functional connectivity in the resting state, we identified several


clusters correlating with voice- and face-preferentiality of the TVA and FFA. For the rFFA one right frontal/precentral cluster was evident, for the lFFA two clusters, one in the left


caudate/olfactory gyrus and one in the left superior temporal pole. Using the voice-preferentiality of the rTVA and lTVA as regressors, four common clusters emerged. These were widely


distributed the occipital, parietal, frontal and temporal cortex. For the rTVA two additional clusters in the left occipital cortex and the right thalamus, and for the lTVA in the left


frontal and right parietal cortex areas emerged. In explanatory seed-to-voxel analyses, the underlying connectivity patterns diverged markedly between the voice and face processing systems.


Whereas for the TVAs a largely convergent pattern of clusters was observed, among others in the occipital gyrus and bilateral insulae, the patterns for the FFAs were mainly divergent and


yielded only one common region in the right supramarginal gyrus extending into the inferior parietal gyrus. Moreover, we identified brain areas with RSFC patterns supramodally reflecting


preferential responses to both, voices and faces. One area in the anterior rostral mediofrontal cortex (arMFC) displayed a maximum of convergent RSFC patterns: its RSFC was indicative of


individual voice-preferential responses of both TVAs and face-preferential responses of the right FFA. Our results strengthen the view that cerebral voice and face processing is an


evolutionary important and therefore highly preserved mechanism, which is not only evidenced by several stages of very specialized processing in the brain, starting with the regions of


TVA1,2,3,4 and FFA5,6,7,8, but is also reflected in other networks, i.e. the resting state network that—per se—work independent from the aforementioned voice and face processing system.


Because during resting state participants were asked to lie quiescent without specific thought. But the independence could be impaired, in case the participants would have thought of human


voices and faces during the resting state measurement. To minimize this risk, we designed the experimental sequence with the resting state block first followed by the task-related parts. The


finding of a correlation of voice- and face-activation patterns with resting state parameters fits in quite well with the currently still limited literature applying both resting state and


voice/face processing measurements. Previous studies found diverging regions either exclusively in the modality-specific processing areas27,42, both in modality specific areas and other


parts of the brain28,29, or networks in the inferior frontal gyrus and motor regions which are not directly connected to modality specific processing30. It needs to be acknowledged however


that a broad range of diverse data analysis techniques were used in those studies27,28,29,30,42 which may account for the disparities to some extent. Our comprehensive analysis on RSFC


networks associated with voice- and face-preferentiality revealed large networks across whole brain, underpinning the notion that response patterns generated in basic voice and face


processing modules during the perception of these cues find a reflection in the coactivation of widespread cerebral networks at rest potentially indicating processes connected to voice and


face perception or a neural preparedness to respond to these stimuli. Speaking figuratively, the direct responses to stimulation with voices and faces can be imagined as the top of the


iceberg, the underlying resting state network structure as the part below the surface of the sea. It is known from the literature that resting state patterns reflect individual traits. In


fact, resting state functional connectivity has been shown to be associated with behavioural tendencies, personality or states of psychiatric disease, e.g. personality traits22,23, moral


behaviour43, violence proneness25, or the diagnosis of dementia or schizophrenia26. These results support the view that resting state patterns may reflect an adaptive system indicative of


different brain states and function. One could speculate about the connection between basal voice and face processing systems, as assessed in our work, and higher order social functioning


(e.g., emotional communication, empathy, theory of mind or moral behaviour), as effective voice and face perception appears as a prerequisite of the former to a certain degree. Certainly,


however, this link remains speculative presently. The novel and distinctive feature of this study is the combination of resting state and stimulation-based fMRI measurements for the visual


and the auditory system. The resting state pattern, i.e. a stimulation-free measurement, correlates with the propensity to respond to certain stimuli. Up to now, this form of association has


only scarcely been addressed. A similar approach revealed non-state-dependent cerebral markers of biased perception in social anxiety37. Another meta-analytic study focused on similarities


in resting state functional connectivity patterns and coactivation network configurations. Using an online database activation patterns of several different tasks were pooled together. A


high correlation between coactivation during task and resting-state correlation was detected44. In patients with first episode schizophrenia overlapping dysfunctions in the prefrontotemporal


pathway were evident45. Our study can serve as starting point for further combined analyses of resting state connectivity and activation patterns in stimulation-based designs from a network


perspective with a much more precise task design. Convergent with previous research which provided evidence for a greater functional similarity between the hemispheres in the cerebral voice


processing system than the face processing system46,47,48, in our study, both TVAs responded to voices in a sensitive (i.e. mixed contrast V > (A, E)) and preferential (i.e. minimum


contrast, V > max(A, E)) manner. In contrast, in the face processing system only the right FFA responds both in a sensitive and preferential way to faces, whereas the response of the left


FFA is only face-sensitive. We substantiated these results comparing voice- and face-sensitivity and -preferentiality of both hemispheres with lack of hemispheric differences in


voice-preferentiality, but significant hemispheric differences in the face processing system with greater face-preferentiality in the right hemisphere. This finding is in line with previous


results showing stronger and more consistent activation through faces in comparison to other stimulus categories in the right FFA compared to the left FFA46,49. The dominance of the right


hemisphere in face-related responses is not restricted to the FFA, but is also reflected in larger activation areas to faces in the right occipitotemporal cortex and the right amygdala and


an exclusive activation of the right inferior frontal gyrus46. Beyond this reliably replicated evidence, we found corresponding patterns in resting state measurements: The resting state


patterns predicting the face- and voice-sensitivity/-preferentiality, respectively, differed showing a convergent pattern for the voice processing system and a largely divergent pattern for


the face processing system as evidenced by the difference in significant overlaps of the informative connectivity patterns between the TVAs as compared to the FFAs. Thus, we conclude that


the different qualities of seeing faces and hearing voices do not work analogously, but that these two systems function in a unique and distinct way, with a higher hemispheric functional


similarity of the voice processing modules in comparison to the face processing system. In our supramodal approach combining voice and face processing networks with three regressors, one


common region in the medial frontal cortex correlated both voice- and face-preferentiality during resting state. The medial frontal cortex is known to be activated in higher order social


cognitive processing, the anterior rostral part especially in mentalizing tasks50. Additionally, it is involved in complex emotion processing13,51,52, independent of the presentation form,


e.g. visually via faces or bodies or acoustically via voices53. The activation of a region related to the processing of stimuli from different sensory modalities gives rise to the problem of


interpreting the results. Throughout this manuscript we use the term supramodal for the locally overlapping cerebral activation by signals from different sensory modalities which can be


identified using conjunction analyses, e.g. for mapping multisensory integration41,54. Limitations of the technique are that in our case the common region constitutes only a small part in


comparison to the complete connectivity pattern from each source, and that the local overlap not mandatorily represents a direct interaction or integration of signals from both sources, but


might indicate that the overlap region is simply linked to processing information from several sensory modalities. While the medial frontal cortex is not consistently activated in


stimulation experiments designed to localize voice- or face-specific brain areas, this notion would still appear quite plausible as effective processing of voices and faces might well be


required as basis for a variety of higher order social communication functions. In line with this, frontal areas were involved in the processing of incongruent but not congruent audiovisual


emotional stimuli55,56 and revealed emotion-specific activation regardless of the sensory modality of the emotional cue53. Whereas many studies assessing higher order social processing


employed emotional stimuli13,51,52, it is quite notable, that we found a convergence in this region even based on experimental designs without explicit emotional connotations. Limitations


concerning the assessment of neutral vs. emotional stimuli are discussed below. This seems to corroborate the notion that higher order social cognitive processes are linked to basic voice


and face perception irrespective of emotional information communicated via these stimuli. On the other hand, one might argue that there is no such thing as a voice or a face completely


devoid of emotional information in two ways: First, also stimuli not intended to carry emotional information by their sender may well contain subliminal emotional cues and, second, even a


putative completely neutral voice or face may automatically be scanned for emotional information and therefore become linked to emotion processing irrespective of its lack of emotional cues.


Previous results hint for a variability in the emotional perception of voices and faces depending on the previously experienced sensory input57,58. The posterior superior temporal sulcus


(pSTS), which has been shown to integrate simultaneously presented auditory and visual stimuli10, did not show an overlap of connectivity patterns indicating both voice- and


face-preferential responses. So, the pSTS’s role in combined face and voice processing might be more closely linked to the sensory integration of these stimuli during their simultaneous


perception and thus not be detectable in the resting state. Our work builds on the manifold confirmed and pioneering findings of regions that are preferentially activated by human stimuli in


comparison to environmental cues, i.e. especially the voice-preferential activation of the TVA and the face-preferential activation of the FFA4,5,6. And it broadens the perspective from


specialized regions for different tasks to a network perspective of regions exhibiting preferential responses both during and in the absence of human nonverbal cues. One could speculate that


the relevance of this finding lies in the reflection of relevant social situations during resting state, possibly including imagination of nonverbal cues. But to corroborate these ideas,


further research is necessary. The unique quality of our data stems from the combination of these individual cerebral processing characteristics of social stimuli with resting state


functional connectivity maps in a relatively large cohort. And it adds to the growing number of findings which advocate a readjustment of our view from specialized regions in the brain


responding to certain stimuli to a larger network perspective involving a multitude of regions across the whole brain in the presence and absence of tasks or/and stimuli. The specificity of


the activation is mediated not by the activation of single specialized regions itself but by the combination of simultaneously activated networks and therefore strengthens the view of a


network perspective. As gender-specific connectivity patterns were observed e.g. in the correlation of RSFC with personality traits22, this aspect represents a limitation of the present


study which focused on gender- and age-independent connectivity patterns. Due to the limited sample size, we did not perform subgroup analyses. Moreover, we did not assess and therefore were


able to correct for personality trait measures, such as the five-factor model of personality, which was shown to be associated with RSFC patterns22 and might therefore also represent a


moderator of the RSFC patterns associated with the propensity to respond to human voices and faces. Although based on seminal standard experiments to assess voice- and face-sensitive and


-preferential responses enabling direct comparisons with many previous studies, certain design-specific factors may have influenced the outcome of our study and should therefore be addressed


in further research: For one, the task-set differed considerably between the voice and face processing experiments (passive listening vs. one-back task) with potential influence on the


attentional status. As a further limitation we would like to address the problem of the assessment of human stimuli as neutral vs. emotional. Though not included in the experiments as


explicit factor employing face pictures with predominantly neutral expression, low-level emotional information in the experimental stimuli may have impacted the RSFC patterns predictive of


cerebral voice- and face-preferentiality. As a conclusion, these results emphasize that the individual cerebral propensity to respond to human voices and faces is reflected in the brain’s


activation patterns also in the absence of these cues as a possible neural corelate of mental reflections on relevant social situations including imagination of nonverbal cues during


“resting” state. The stronger convergence of informative connectivity patterns for the TVAs’ cue selectivity in contrast to the FFAs’ may indicate a higher hemispheric functional similarity


of the voice processing modules. The supramodal convergence of such informative connectivity patterns, in turn, points to the anterior medial prefrontal cortex as shared neural resource in


supramodal voice and face processing or potentially nonverbal communication. Similar to the underwater perspective on an iceberg, this experimental approach may open up interesting avenues


to the investigation of voice and face processing. In this regard, the resting state connectivity patterns correlating with individual voice and face selectivity may aid the understanding of


cerebral voice and face preference from a network perspective. DATA AVAILABILITY The datasets generated during and/or analysed during the current study are available from the corresponding


author on reasonable request. REFERENCES * Belin, P., Zatorre, R. J., Lafaille, P., Ahad, P. & Pike, B. Voice-selective areas in human auditory cortex. _Nature_ 403, 309–312.


https://doi.org/10.1038/35002078 (2000). Article  ADS  CAS  PubMed  Google Scholar  * von Kriegstein, K. & Giraud, A. L. Implicit multisensory associations influence voice recognition.


_PLoS Biol._ 4, e326. https://doi.org/10.1371/journal.pbio.0040326 (2006). Article  CAS  Google Scholar  * Ethofer, T. _et al._ Differential influences of emotion, task, and novelty on brain


regions underlying the processing of speech melody. _J. Cogn. Neurosci._ 21, 1255–1268. https://doi.org/10.1162/jocn.2009.21099 (2009). Article  PubMed  Google Scholar  * Pernet, C. R. _et


al._ The human voice areas: Spatial organization and inter-individual variability in temporal and extra-temporal cortices. _Neuroimage_ 119, 164–174.


https://doi.org/10.1016/j.neuroimage.2015.06.050 (2015). Article  PubMed  Google Scholar  * Kanwisher, N., McDermott, J. & Chun, M. M. The fusiform face area: A module in human


extrastriate cortex specialized for face perception. _J. Neurosci._ 17, 4302–4311 (1997). Article  CAS  Google Scholar  * Kanwisher, N. & Yovel, G. The fusiform face area: A cortical


region specialized for the perception of faces. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 361, 2109–2128. https://doi.org/10.1098/rstb.2006.1934 (2006). Article  PubMed  PubMed Central 


Google Scholar  * Posamentier, M. T. & Abdi, H. Processing faces and facial expressions. _Neuropsychol. Rev._ 13, 113–143. https://doi.org/10.1023/a:1025519712569 (2003). Article  PubMed


  Google Scholar  * Halgren, E. _et al._ Location of human face-selective cortex with respect to retinotopic areas. _Hum. Brain Mapp._ 7, 29–37 (1999). Article  CAS  Google Scholar  *


Kreifelts, B., Ethofer, T., Huberle, E., Grodd, W. & Wildgruber, D. Association of trait emotional intelligence and individual fMRI-activation patterns during the perception of social


signals from voice and face. _Hum. Brain Mapp._ 31, 979–991. https://doi.org/10.1002/hbm.20913 (2010). Article  PubMed  Google Scholar  * Kreifelts, B., Ethofer, T., Shiozawa, T., Grodd, W.


& Wildgruber, D. Cerebral representation of non-verbal emotional perception: fMRI reveals audiovisual integration area between voice- and face-sensitive regions in the superior temporal


sulcus. _Neuropsychologia_ 47, 3059–3066. https://doi.org/10.1016/j.neuropsychologia.2009.07.001 (2009). Article  PubMed  Google Scholar  * Robins, D. L., Hunyadi, E. & Schultz, R. T.


Superior temporal activation in response to dynamic audio-visual emotional cues. _Brain Cogn._ 69, 269–278. https://doi.org/10.1016/j.bandc.2008.08.007 (2009). Article  PubMed  Google


Scholar  * Ethofer, T., Pourtois, G. & Wildgruber, D. Investigating audiovisual integration of emotional signals in the human brain. _Prog. Brain Res._ 156, 345–361.


https://doi.org/10.1016/S0079-6123(06)56019-4 (2006). Article  PubMed  Google Scholar  * Klasen, M., Kenworthy, C. A., Mathiak, K. A., Kircher, T. T. & Mathiak, K. Supramodal


representation of emotions. _J. Neurosci._ 31, 13635–13643. https://doi.org/10.1523/JNEUROSCI.2833-11.2011 (2011). Article  CAS  PubMed  PubMed Central  Google Scholar  * Ethofer, T. _et


al._ Functional responses and structural connections of cortical areas for processing faces and voices in the superior temporal sulcus. _Neuroimage_ 76, 45–56.


https://doi.org/10.1016/j.neuroimage.2013.02.064 (2013). Article  PubMed  Google Scholar  * Ethofer, T. _et al._ Cerebral pathways in processing of affective prosody: A dynamic causal


modeling study. _Neuroimage_ 30, 580–587. https://doi.org/10.1016/j.neuroimage.2005.09.059 (2006). Article  PubMed  Google Scholar  * Kreifelts, B. _et al._ Tuned to voices and faces:


Cerebral responses linked to social anxiety. _Neuroimage_ 197, 450–456. https://doi.org/10.1016/j.neuroimage.2019.05.018 (2019). Article  PubMed  Google Scholar  * Karle, K. N. _et al._


Neurobiological correlates of emotional intelligence in voice and face perception networks. _Soc. Cogn. Affect. Neurosci._ 13, 233–244. https://doi.org/10.1093/scan/nsy001 (2018). Article 


PubMed  PubMed Central  Google Scholar  * Smitha, K. A. _et al._ Resting state fMRI: A review on methods in resting state connectivity analysis and resting state networks. _Neuroradiol. J._


30, 305–317. https://doi.org/10.1177/1971400917697342 (2017). Article  CAS  PubMed  PubMed Central  Google Scholar  * Hahn, T. _et al._ Reliance on functional resting-state network for


stable task control predicts behavioral tendency for cooperation. _Neuroimage_ 118, 231–236. https://doi.org/10.1016/j.neuroimage.2015.05.093 (2015). Article  PubMed  Google Scholar  *


Inagaki, T. K. & Meyer, M. L. Individual differences in resting-state connectivity and giving social support: Implications for health. _Soc. Cogn. Affect. Neurosci._ 15, 1076–1085.


https://doi.org/10.1093/scan/nsz052 (2020). Article  PubMed  Google Scholar  * Serafini, G., Pardini, M., Pompili, M., Girardi, P. & Amore, M. Understanding suicidal behavior: The


contribution of recent resting-state fMRI techniques. _Front. Psychiatry_ 7, 69. https://doi.org/10.3389/fpsyt.2016.00069 (2016). Article  PubMed  PubMed Central  Google Scholar  * Nostro,


A. D. _et al._ Predicting personality from network-based resting-state functional connectivity. _Brain Struct. Funct._ 223, 2699–2719. https://doi.org/10.1007/s00429-018-1651-z (2018).


Article  PubMed  PubMed Central  Google Scholar  * Markett, S., Montag, C. & Reuter, M. Network neuroscience and personality. _Pers. Neurosci._ 1, e14.


https://doi.org/10.1017/pen.2018.12 (2018). Article  Google Scholar  * Parkes, L., Satterthwaite, T. D. & Bassett, D. S. Towards precise resting-state fMRI biomarkers in psychiatry:


Synthesizing developments in transdiagnostic research, dimensional models of psychopathology, and normative neurodevelopment. _Curr. Opin. Neurobiol._ 65, 120–128.


https://doi.org/10.1016/j.conb.2020.10.016 (2020). Article  CAS  PubMed  PubMed Central  Google Scholar  * Romero-Martinez, A. _et al._ The brain resting-state functional connectivity


underlying violence proneness: Is it a reliable marker for neurocriminology? A systematic review. _Behav. Sci. (Basel)._ https://doi.org/10.3390/bs9010011 (2019). Article  PubMed  PubMed


Central  Google Scholar  * van den Heuvel, M. P. & Hulshoff Pol, H. E. Exploring the brain network: A review on resting-state fMRI functional connectivity. _Eur. Neuropsychopharmacol._


20, 519–534. https://doi.org/10.1016/j.euroneuro.2010.03.008 (2010). Article  CAS  PubMed  Google Scholar  * Zhang, H., Tian, J., Liu, J., Li, J. & Lee, K. Intrinsically organized


network for face perception during the resting state. _Neurosci. Lett._ 454, 1–5. https://doi.org/10.1016/j.neulet.2009.02.054 (2009). Article  CAS  PubMed  PubMed Central  Google Scholar  *


O’Neil, E. B., Hutchison, R. M., McLean, D. A. & Kohler, S. Resting-state fMRI reveals functional connectivity between face-selective perirhinal cortex and the fusiform face area


related to face inversion. _Neuroimage_ 92, 349–355. https://doi.org/10.1016/j.neuroimage.2014.02.005 (2014). Article  PubMed  Google Scholar  * Kruschwitz, J. D. _et al._ Segregation of


face sensitive areas within the fusiform gyrus using global signal regression? A study on amygdala resting-state functional connectivity. _Hum. Brain Mapp._ 36, 4089–4103.


https://doi.org/10.1002/hbm.22900 (2015). Article  PubMed  PubMed Central  Google Scholar  * Correia, A. I. _et al._ Resting-state connectivity reveals a role for sensorimotor systems in


vocal emotional processing in children. _Neuroimage_ 201, 116052. https://doi.org/10.1016/j.neuroimage.2019.116052 (2019). Article  PubMed  Google Scholar  * Oldfield, R. C. The assessment


and analysis of handedness: The Edinburgh inventory. _Neuropsychologia_ 9, 97–113. https://doi.org/10.1016/0028-3932(71)90067-4 (1971). Article  CAS  PubMed  Google Scholar  * Kreifelts, B.


_et al._ Non-verbal emotion communication training induces specific changes in brain function and structure. _Front. Hum. Neurosci._ 7, 648. https://doi.org/10.3389/fnhum.2013.00648 (2013).


Article  PubMed  PubMed Central  Google Scholar  * Kreifelts, B. _et al._ The neural correlates of face-voice-integration in social anxiety disorder. _Front. Psychiatry_ 11, 657.


https://doi.org/10.3389/fpsyt.2020.00657 (2020). Article  PubMed  PubMed Central  Google Scholar  * Collins, D. L., Neelin, P., Peters, T. M. & Evans, A. C. Automatic 3D intersubject


registration of MR volumetric data in standardized Talairach space. _J. Comput. Assist. Tomogr_ 18, 192–205 (1994). Article  CAS  Google Scholar  * Friston, K. J. _et al._ Classical and


Bayesian inference in neuroimaging: Applications. _Neuroimage_ 16, 484–512. https://doi.org/10.1006/nimg.2002.1091 (2002). Article  CAS  PubMed  Google Scholar  * Worsley, K. J. _et al._ A


unified statistical approach for determining significant signals in images of cerebral activation. _Hum. Brain Mapp._ 4, 58–73.


https://doi.org/10.1002/(SICI)1097-0193(1996)4:1%3c58::AID-HBM4%3e3.0.CO;2-O (1996). Article  CAS  PubMed  Google Scholar  * Kreifelts, B. _et al._ Cerebral resting state markers of biased


perception in social anxiety. _Brain Struct. Funct._ 224, 759–777. https://doi.org/10.1007/s00429-018-1803-1 (2019). Article  PubMed  Google Scholar  * Whitfield-Gabrieli, S. &


Nieto-Castanon, A. Conn: A functional connectivity toolbox for correlated and anticorrelated brain networks. _Brain Connect_ 2, 125–141. https://doi.org/10.1089/brain.2012.0073 (2012).


Article  PubMed  Google Scholar  * Tzourio-Mazoyer, N. _et al._ Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject


brain. _Neuroimage_ 15, 273–289. https://doi.org/10.1006/nimg.2001.0978 (2002). Article  CAS  PubMed  Google Scholar  * Whitfield-Gabrieli, S. N.-C. A. _CONN Toolbox Manual_.


https://web.conn-toolbox.org/resources/documentation/manual (2017). Accessed 23 Aug 2017. * Nichols, T., Brett, M., Andersson, J., Wager, T. & Poline, J. B. Valid conjunction inference


with the minimum statistic. _Neuroimage_ 25, 653–660. https://doi.org/10.1016/j.neuroimage.2004.12.005 (2005). Article  PubMed  Google Scholar  * Zhu, Q., Zhang, J., Luo, Y. L., Dilks, D. D.


& Liu, J. Resting-state neural activity across face-selective cortical regions is behaviorally relevant. _J. Neurosci._ 31, 10323–10330. https://doi.org/10.1523/JNEUROSCI.0873-11.2011


(2011). Article  CAS  PubMed  PubMed Central  Google Scholar  * Liu, J., Yuan, B., Luo, Y. J. & Cui, F. Intrinsic functional connectivity of medial prefrontal cortex predicts the


individual moral bias in economic valuation partially through the moral sensitivity trait. _Brain Imaging Behav._ 14, 2024–2036. https://doi.org/10.1007/s11682-019-00152-1 (2020). Article 


PubMed  Google Scholar  * Di, X., Gohel, S., Kim, E. H. & Biswal, B. B. Task vs. rest-different network configurations between the coactivation and the resting-state brain networks.


_Front. Hum. Neurosci._ 7, 493. https://doi.org/10.3389/fnhum.2013.00493 (2013). Article  PubMed  PubMed Central  Google Scholar  * Mwansisya, T. E. _et al._ Task and resting-state fMRI


studies in first-episode schizophrenia: A systematic review. _Schizophr. Res._ 189, 9–18. https://doi.org/10.1016/j.schres.2017.02.026 (2017). Article  PubMed  Google Scholar  * Engell, A.


D. & McCarthy, G. Probabilistic atlases for face and biological motion perception: An analysis of their reliability and overlap. _Neuroimage_ 74, 140–151.


https://doi.org/10.1016/j.neuroimage.2013.02.025 (2013). Article  PubMed  Google Scholar  * Brancucci, A., Lucci, G., Mazzatenta, A. & Tommasi, L. Asymmetries of the human social brain


in the visual, auditory and chemical modalities. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 364, 895–914. https://doi.org/10.1098/rstb.2008.0279 (2009). Article  CAS  PubMed  Google Scholar


  * Bonte, M. _et al._ Development from childhood to adulthood increases morphological and functional inter-individual variability in the right superior temporal cortex. _Neuroimage_ 83,


739–750. https://doi.org/10.1016/j.neuroimage.2013.07.017 (2013). Article  PubMed  Google Scholar  * Schwarz, L. _et al._ Properties of face localizer activations and their application in


functional magnetic resonance imaging (fMRI) fingerprinting. _PLoS ONE_ 14, e0214997. https://doi.org/10.1371/journal.pone.0214997 (2019). Article  CAS  PubMed  PubMed Central  Google


Scholar  * Amodio, D. M. & Frith, C. D. Meeting of minds: The medial frontal cortex and social cognition. _Nat. Rev. Neurosci._ 7, 268–277. https://doi.org/10.1038/nrn1884 (2006).


Article  CAS  PubMed  Google Scholar  * Wildgruber, D. _et al._ Different types of laughter modulate connectivity within distinct parts of the laughter perception network. _PLoS ONE_ 8,


e63441. https://doi.org/10.1371/journal.pone.0063441 (2013). Article  ADS  CAS  PubMed  PubMed Central  Google Scholar  * Bruck, C., Kreifelts, B., Gossling-Arnold, C., Wertheimer, J. &


Wildgruber, D. “Inner voices”: The cerebral representation of emotional voice cues described in literary texts. _Soc. Cogn. Affect. Neurosci._ 9, 1819–1827.


https://doi.org/10.1093/scan/nst180 (2014). Article  PubMed  PubMed Central  Google Scholar  * Peelen, M. V., Atkinson, A. P. & Vuilleumier, P. Supramodal representations of perceived


emotions in the human brain. _J. Neurosci._ 30, 10127–10134. https://doi.org/10.1523/JNEUROSCI.2161-10.2010 (2010). Article  CAS  PubMed  PubMed Central  Google Scholar  * Friston, K. J.,


Penny, W. D. & Glaser, D. E. Conjunction revisited. _Neuroimage_ 25, 661–667. https://doi.org/10.1016/j.neuroimage.2005.01.013 (2005). Article  PubMed  Google Scholar  * Muller, V. I.


_et al._ Incongruence effects in crossmodal emotional integration. _Neuroimage_ 54, 2257–2266. https://doi.org/10.1016/j.neuroimage.2010.10.047 (2011). Article  PubMed  Google Scholar  *


Davies-Thompson, J. _et al._ Hierarchical brain network for face and voice integration of emotion expression. _Cereb. Cortex_ 29, 3590–3605. https://doi.org/10.1093/cercor/bhy240 (2019).


Article  PubMed  Google Scholar  * Webster, M. A. & MacLeod, D. I. Visual adaptation and face perception. _Philos. Trans. R. Soc. Lond. B Biol. Sci._ 366, 1702–1725.


https://doi.org/10.1098/rstb.2010.0360 (2011). Article  PubMed  PubMed Central  Google Scholar  * Bestelmeyer, P. E., Rouger, J., DeBruine, L. M. & Belin, P. Auditory adaptation in vocal


affect perception. _Cognition_ 117, 217–223. https://doi.org/10.1016/j.cognition.2010.08.008 (2010). Article  PubMed  Google Scholar  Download references ACKNOWLEDGEMENTS This work was


supported by the Clinician Scientist program of the University of Tübingen to KNE (Grant Number 367-0-0). We acknowledge support by Open Access Publishing Fund of University of Tübingen. The


funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. FUNDING Open Access funding enabled and organized by Projekt DEAL.


AUTHOR INFORMATION AUTHORS AND AFFILIATIONS * Department of Psychiatry and Psychotherapy, Tübingen Center for Mental Health (TüCMH), University of Tübingen, Calwerstrasse 14, 72076,


Tübingen, Germany Kathrin N. Eckstein, Dirk Wildgruber, Thomas Ethofer, Carolin Brück, Heike Jacob & Benjamin Kreifelts * Department for Biomedical Magnetic Resonance, University of


Tübingen, Tübingen, Germany Thomas Ethofer & Michael Erb Authors * Kathrin N. Eckstein View author publications You can also search for this author inPubMed Google Scholar * Dirk


Wildgruber View author publications You can also search for this author inPubMed Google Scholar * Thomas Ethofer View author publications You can also search for this author inPubMed Google


Scholar * Carolin Brück View author publications You can also search for this author inPubMed Google Scholar * Heike Jacob View author publications You can also search for this author


inPubMed Google Scholar * Michael Erb View author publications You can also search for this author inPubMed Google Scholar * Benjamin Kreifelts View author publications You can also search


for this author inPubMed Google Scholar CONTRIBUTIONS The study was conceptualized by D.W., T.E. and B.K. Data acquisition and curation was done by C.B., H.J. and M.E. Analyses were


performed by K.N.E., B.K., D.W., T.E., C.B., H.J. and M.E. K.N.E. and B.K. wrote the main manuscript text and prepared the figures. All authors reviewed the manuscript. CORRESPONDING AUTHOR


Correspondence to Kathrin N. Eckstein. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests. ADDITIONAL INFORMATION PUBLISHER'S NOTE Springer Nature


remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. SUPPLEMENTARY INFORMATION SUPPLEMENTARY FIGURE 1. RIGHTS AND PERMISSIONS OPEN ACCESS


This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as


long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third


party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the


article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the


copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Eckstein, K.N., Wildgruber,


D., Ethofer, T. _et al._ Correlates of individual voice and face preferential responses during resting state. _Sci Rep_ 12, 7117 (2022). https://doi.org/10.1038/s41598-022-11367-6 Download


citation * Received: 29 December 2021 * Accepted: 15 April 2022 * Published: 03 May 2022 * DOI: https://doi.org/10.1038/s41598-022-11367-6 SHARE THIS ARTICLE Anyone you share the following


link with will be able to read this content: Get shareable link Sorry, a shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature


SharedIt content-sharing initiative