Correction: Global Genetic Variations Predict Brain Response to Faces

Face expressions are a rich source of social signals. Here we estimated the proportion of phenotypic variance in the brain response to facial expressions explained by common genetic variance captured by ~500,000 single nucleotide polymorphisms. Using genomic-relationship-matrix restricted maximum likelihood (GREML), we related this global genetic variance to that in the brain response to facial expressions, as assessed with functional magnetic resonance imaging (fMRI) in a community-based sample of adolescents (n = 1,620). Brain response to facial expressions was measured in 25 regions constituting a face network, as defined previously. In 9 out of these 25 regions, common genetic variance explained a significant proportion of phenotypic variance (40–50%) in their response to ambiguous facial expressions; this was not the case for angry facial expressions. Across the network, the strength of the genotype-phenotype relationship varied as a function of the inter-individual variability in the number of functional connections possessed by a given region (R2 = 0.38, p<0.001). Furthermore, this variability showed an inverted U relationship with both the number of observed connections (R2 = 0.48, p<0.001) and the magnitude of brain response (R2 = 0.32, p<0.001). Thus, a significant proportion of the brain response to facial expressions is predicted by common genetic variance in a subset of regions constituting the face network. These regions show the highest inter-individual variability in the number of connections with other network nodes, suggesting that the genetic model captures variations across the adolescent brains in co-opting these regions into the face network.


Introduction
Interactions with peers are of high relevance to our mental health. Patients with various psychological disorders show impairments in face perception and emotion recognition [1][2][3][4]. Similarly, differential neural responses to faces have been reported in various psychological disorders including depression [5], psychopathy and/or aggressive tendencies [6], autism [7], and schizophrenia [8].
Our ability to process faces is modulated by both environment and genes. Using a twin design, Zhu and colleagues observed that inter-individual variations in face perception are heritable, with the genetic component as high as 50% in adolescents performing various face tasks [9]. Although the key elements of the neural network underlying face processing are well known [10], whether or not brain response to facial expressions show comparable levels of heritability is unknown.
Here we address this question using genomic-relationshipmatrix restricted maximum likelihood (GREML), applied using Genome-wide Complex Trait Analysis (GCTA) software [11] to a dataset of functional magnetic resonance images (fMRI) obtained in over 1,600 typically developing adolescents while they were observing videoclips of ambiguous or angry facial expressions [The IMAGEN Study ; 12]. The GREML approach allows one to estimate how much phenotypic variance is attributable to the genetic variance captured by all common genetic variations (single nucleotide polymorphisms; SNPs) assayed in a typical Genome Wide Association Study (GWAS). We will ask here whether ''heritability'' of the response to facial expressions -as estimated using the GREML approach -varies across these regions. We will then examine possible reasons for such regional variations.
In fMRI, brain response to a stimulus is inferred from the variations in hemodynamics detected as the blood oxygenationlevel dependent (BOLD) signal on T 2 *-weighted MR images. This signal relies on the fact that brain activity is associated with an oversupply of oxygenated blood to the brain region engaged by the stimulus; consequently, small veins that drain this region contain some of the unused oxygenated blood. Thus, the BOLD signal reflects the proportion of oxygenated and de-oxygenated blood in a given brain region at a given moment; most likely, this hemodynamic signal is proportional to the local field potentials generated by (excitatory) inputs.
Here, summary measures for the BOLD response (%BSC) were calculated for two face viewing conditions (Ambiguous movements and Angry Expressions), each compared with a non-biological motion control condition; this was done for 25 brain regions of interest (ROIs), as defined previously using a probabilistic map of the brain response to facial expressions [13]. GREML-based estimates of ''heritability'' were calculated using the GCTA package [13] [11]. We observed significant estimates of ''heritability'' for the Ambiguous vs. Control contrast %BSC in 9 out of these 25 ROIs (Fig. 1, Supplementary Table S1). No significant estimates were observed for any ROI in the Angry vs. Control contrast (Supplementary Table S2). By chance we would expect 25*0.05 = 1.25 false positives when examining 25 ROIs (in each contrast), or 2.5 when examining 50 ROIs (both contrasts combined) at the a = 0.05 level. As we found 9 ROIs with P, 0.05 this appears to be evidence against the null hypothesis. To provide a P-value for this count-rate approach, we conducted a Monte Carlo simulation for each contrast using the observed correlation matrix (see Supplementary Figure S1 for the phenotypic correlation matrix between all 25 ROIs) [14]. Using 50,000 realizations, we simulated the null-hypothesis test statistics for each ROI (using correlated outcomes) and tabulated the distribution of the number of P-values significant at 0.05; this empirical distribution can be used to compute P-values for this count statistic. For the Ambiguous Facial Expressions contrast, this simulation confirmed that we could reject the null hypothesis of observing 9 significant tests by chance at P value of 0.03. For Angry Facial Expressions contrast, the P-value must be 1.0 for a count of 0 significant tests.
Given the high heritability of general intelligence [16], we have examined the possibility that intelligence correlates with the interindividual variations in the brain response to facial expressions in the Ambiguous condition. In a subset of 1,772 individuals with available scores on four subtests of the Wechsler Intelligence Scale for Children -IV (similarities, vocabulary, block design and matrix reasoning), we found no correlation between these scores and the mean %BSC across all ROIs (p.0.3) or between the scores and the mean %BCS in the Optional (p.0.2) and Obligatory (p.0.4) networks. The lack of the relationship between general intelligence and %BSC suggests that the former does not contribute to the above heritability estimates of the brain response to the ambiguous facial expressions.
Next, we asked whether the above GREML-based estimates of heritability reflect any properties of the brain response across the examined ROIs. For example, are the ROIs with stronger response to facial expressions more heritable? Table 1  The brain regions considered here may be viewed as nodes of a ''face'' network. The strength of each region's contribution to this network may differ across regions (ROIs) and across individuals. To quantify this phenomenon, we extracted mean time-courses in the BOLD signal from all ROIs in each participant and used these to calculate matrices of functional connectivity for all participants. From these matrices, we estimated the number of connections of a given region with the other members of the face network using the graph-theory metric of nodal ''degree'' [17]. Table 1 provides the population means and standard deviations for the nodal degree for all ROIs (Ambiguous Facial expressions).

Author Summary
We measured brain response to facial expressions in a large sample of typically developing adolescents (n = 1,620) and assessed ''heritability'' of the response using common genetic variations across the genome. In a subset of brain regions, we explained 40-50% of phenotypic variance by genetic variance. These brain regions appear to differ from the rest of the face network in the degree of inter-individual variations in their functional connectivity. We propose that these regions, including the prefrontal and premotor cortex, represent ''Optional'' part of the network co-opted by its ''Obligatory'' members, including the posterior part of the superior temporal sulcus, fusiform face area and the lateral occipital cortex, concerned with processing complex visual stimuli.
We then examined whether differences in this measure of functional connectivity across the 25 ROIs constituting the face network predict their GREML-based estimates of heritability (VG/Vp estimates). This is the case: the population (interindividual) variance in the nodal degree predicts strongly the ''heritability'' (R 2 = 0.38, p,0.001, Fig. 3a). Furthermore, the population variance in the nodal degree shows an inverse Ushaped relationship with the mean nodal degree (2 nd order polynomial fit: R 2 = 0.48, F(2,22) = 10.32, p,0.001, Fig. 3b) and the mean %BSC (2 nd order polynomial fit: R 2 = 0.32, F(2,22) = 5.07 p = 0.02, Fig. 3c). Similarity of this inverse Ushaped relationship for the mean nodal degree and the mean %BSC is not surprising given a strong correlation between these two measures (r = 0.84, p,0.001). Note, however, that the population means of nodal degree do not predict their GREMLbased estimates of heritability across the 25 brain regions (R 2 = 0.06, p = 0.23). This is likely related to the fact that the mean nodal degree shows an inverted-U relationship with the population variance of this measure across these regions (Fig. 3b). Finally, we have repeated these analyses for the Angry Facial expressions; the only significant relationship observed in this condition was that between the population variance and the population mean in nodal degree (Supplementary Figure S2). Population means and standard deviations for %BSC and nodal degree for all ROIs of the Angry condition are given in Supplemental Table S3.
To illustrate the relationship between population variance and mean in the number of connections (nodal degree) in the Ambiguous contrast, we selected two groups of ROIs that differ in the combination of these two measures of degree: (1) ROIs with the highest variance and an intermediate mean; and (2) ROIs with the highest mean and the lowest variance. As shown in Figure 4 (and Table 2), proportion of individuals with connections between a given pair of ROIs (i.e., pair-wise correlations with r .0.3) is intermediate within the first subnetwork (30 to 60% of participants) and very high within the second subnetwork (70 to 96% of participants). Importantly, the posterior STS (region #5 in Fig. 4) appears to act as a ''bridge'' between the two subnetworks: it is the only member of the second subnetwork with connections to all four nodes of the first subnetwork in 50% (or more) of participants. We then examined genetic covariance within and between the two subnetworks, which we term ''Optional'' and ''Obligatory'' (see Discussion for comparison with the ''Extended'' and ''Core'' systems of Haxby and colleagues [10]). Using a bivariate GCTA approach [18], we observed significant (p,0.05) genetic covariances in three pairs of ROIs: [R MVLFC -R Ant STS], [R MVLFC -R Post STS] and [R AntSTS -R Post STS]. Furthermore, we found marginal (p,0.1) covariances in four additional pairs of ROIs, three within the ''Optional'' network and one between the ''Optional'' and ''Obligatory'' network (L PMC -R Post STS). The full genetic-covariance matrix for the eight ROIs constituting the two subnetworks is provided in Supplemental Table S4.

Discussion
Using the GREML approach, we show that the aggregate of common genetic variations across the entire genome predicts brain response to facial expressions. This is the case, however, only for some brain regions when viewing ambiguous facial expressions. The fact that inter-individual variations in certain brain responses to certain faces can be predicted by genetic variations (across the entire genome) is in keeping with other evidence supporting heritability of many behavioral aspects of face perception. Indeed, prosopagnosia -an inability of recognize faces -is transmitted in some family pedigrees carrying an autosomal dominant pattern of transmission [19]. Using a twin design in adults, Wilmer et al. [20] reported that facial recognition has a strong heritable component. Greenwood et al. [21] found that memory for faces, as well as emotional recognition, were heritable phenotypes in a study of adult patients with schizophrenia and their non-affected relatives.
Most relevant to the current study, Zhu et al. [9] showed that several aspects of face perception, including facial recognition and the face-inversion effect, were heritable and that the heritability increased from childhood to adolescence.
Heritability of the brain response to faces has been estimated in a couple of twin studies. Thus, Anokhin et al. [22] used eventrelated potentials to changes in emotional expression in children and reported heritability values varying between 42 and 64% [22]. Moreover, Polk et al. [23] showed that fMRI responses to faces in the ventral visual stream were more similar within pairs of adult monozygotic (13 pairs) than dizygotic (11 pairs) twins; sample sizes were too small to calculate heritability in this study [23]. In the ambiguous viewing condition, we observed that brain regions with the most population variance in their contribution to the face-processing network, as indexed by the ''nodal degree'' metric, were the most heritable. In other words, the amount of population variance in functional connectivity of a given region was related to the probability of explaining this region's response (to faces) by global genetic variations. Furthermore, we also observed that -across the 25 ROIs -brain regions with highest estimates of heritability (and highest population variance in connectivity) appear to coincide with ROIs characterized by intermediate values of their response to facial expressions (and their connectivity). As illustrated in Figure 4, the combination of these two properties (i.e., population mean and variance in nodal degree) sets apart two ''subnetworks'' we term ''Obligatory'' and ''Optional'' to denote their hypothetical role in processing faces. The proposed model builds on the distinction between the ''Core'' and ''Extended'' neural systems for face perception, as outlined originally by Haxby and colleagues [10]. In their model, the ''Core'' system consists of the inferior occipital gyri, fusiform gyri and a set of cortical areas along the superior temporal sulcus; this system includes all ''Obligatory'' regions identified here. The ''Extended'' system encompasses the anterior temporal cortex (e.g., face identity) and amygdala (emotion), as well as the auditory cortex (speech-related mouth movements) and cortical areas in the intra-parietal sulcus (spatial attention); with the exception of the anterior temporal cortex, this system is different from the ''Optional'' system consisting primarily of fronto-cortical regions. Nonetheless, both the ''Extended'' system of Haxby and colleagues and the Optional one identified here are viewed as being ''…comprised of regions from neural systems for other cognitive functions that can be recruited to act in concert with the regions in the core system to extract meaning from faces'' [10].
We suggest that the ''Obligatory'' regions are brought online in most participants when viewing human facial expressions; hence the high mean number of connections but low inter-individual variance in connectivity. These regions coincide with visual areas located in the occipital and temporal cortex, known to respond robustly to complex visual stimuli in general and human faces in particular. On the other hand, the ''Optional'' regions are coopted to the face network by some but not all participants (hence the high population variance in their connectivity). Given the putative role of these regions in different aspects of working  memory (mid-dorsolateral and mid-ventrolateral prefrontal cortex) and motor resonance (premotor cortex), we speculate that individuals co-opting these regions while viewing faces do so in the context of their cognitive evaluation engaged spontaneously (no instructions were given to this effect). It is of interest to note that a number of ''Optional'' regions show shared genetic covariance with each other (albeit at marginal significance levels); this is not the case for any pair of ''Obligatory'' regions.
Finally, it appears that such a recruitment of the ''Optional'' regions may related to their connectivity with the posterior part of the STS (region #5 in Fig. 4), the only ''Obligatory'' region with a high number of connections to all four ''Optional'' regions. Of course, the posterior STS is an ideal candidate for bridging these two subnetworks, given its well-established role in extracting and processing social signals in non-biological motion [24]. This view is supported by our observation that the posterior STS is the only ''Obligatory'' region that shows genetic covariance with a number of ''Optional'' regions.
Overall, it is possible that regions with either low (floor effect) or high (ceiling effect) engagement by the stimulus (dynamic facial expressions) do not provide large enough range of (phenotypic) values across individuals and, in turn, are less likely to capture covariations with genetic variance. Note that given the modest sample size (1,620 unrelated individuals), we are powered here to detect only relatively high values of heritability [25]. Recent studies suggest that more than half of narrow-sense heritability (h 2 ), as estimated through studies of related individuals, can be explained by common genetic variations assessed with SNPs included in most DNA microarrays [26]. Furthermore, the degree of GREML-based estimates of heritability predicts strongly the number of SNPs that reach nominal significance in a GWAS across 47 quantitative traits [15]. We have observed similar relationship for the brain response to faces across the 25 brain regions.
The low GREML-based estimates of heritability in the Angry condition were unexpected but clear cut: VG/Vp values were close to 0 in 19/25 ROIs, with values in the remaining ROIs varying between 0.12 and 0.28 (Supplemental Table S2). As explained in the ''Limitations and Significance'' section, we have power to detect only relatively high values of ''heritability'', albeit with fairly large confidence intervals. If true heritability of the brain response to Angry Facial expressions is low (this is unknown), we are underpowered to estimate such low values using the GREML-approach in the current study. An alternative possibility is that, in general, neural processing of angry facial expressions is as heritable as that of any other facial expressions but our paradigm fails to elicit an ''adequate'' brain response. Given our observation of the strong relationship between GREML-based estimates of heritability (VG/Vp) for Ambiguous Facial expressions and the population variance in functional connectivity (Figure 2), it is important to note that the latter is lower in the Angry (vs. Ambiguous) contrast. Furthermore, unlike the Ambiguous contrast (Figure 3c), there is no relationship between the mean BOLD response and the population variance in functional Figure 3. Relationships between SNP-based estimates of heritability, population variance in functional connectivity and the BOLD response. A) Relationship between GREML-based heritability estimates (Genetic Variance/Phenotypic Variance) and the population variance (standard deviation, SD) of functional connectivity (node degree) across 25 ROIs for the Ambiguous face viewing contrast. B) Relationship between the population variance (SD) and the population mean of degree across the 25 ROIs. C) Relationship between population mean of the brain response (percent BOLD signal change, %BSC) and the population variance (SD) of degree across the 25 ROIs. For all three plots, colour is scaled according to the GREML results for %BSC (cyan for low values and red for high values). VG, Genetic Variance; Vp, Phenotypic Variance. doi:10.1371/journal.pgen.1004523.g003 connectivity across the 25 ROIs in the Angry contrast (Supplemental Figure S2c). Thus, it is possible that angry facial expressions did not engage the face network in the same manner as the ambiguous facial expressions, resulting in a suboptimal phenotype.

Limitations and Significance
The key limitation of the present report is sample size; with 1,620 unrelated individuals, we are at the lower limit of the GREML-based approach for estimating contributions of common SNPs to phenotypic variations. Therefore, we were able to detect only relatively high values of heritability; note that these estimates have fairly large confidence intervals (standard error [SE]; e.g., R-MDLFC: VG/Vp = 0.5260.22) and must be therefore interpreted cautiously. Simulations conditional on empirical GWAS data are consistent with these observations; sample size of 1,999 individuals is adequate for estimating correctly high (h 2 = 0.5) narrow-sense heritability [25]. The limited sample size also affected the significance values. Nonetheless, the uneven distribution of the nominally significant results between the Ambiguous (9/25) and Angry (0/25) speaks against a chance nature of these findings, as confirmed by the Monte-Carlo simulations.
The above sample-size limitation must be viewed in the context of the phenotype under study, however. The previous twin-based studies of fMRI-based phenotypes employed between 20 and 141 twin pairs, with heritability (a 2 ) estimates varying widely between 0 and 65 [27]. A total of 333 related individuals were included in a pedigree-based study of heritability of resting-state fMRI; h 2 for functional connectivity of the different components of so-called default-mode network varied between 10 (SE = 13) and 42 (SE = 17) [28]. Working with unrelated individuals, only a few other studies are acquiring functional brain phenotypes with a sample size comparable to the present report. For example, the Human Connectome Project plans to collect paradigm-based and resting-state fMRI datasets in 1,200 individuals [29]. In the Generation R cohort, scanning is under way to collected restingstate fMRI in up to 5,000 children (White, personal communication). Given the challenges related to test-retest reliability of fMRI data in general, and resting-state fMRI in particular, we suggest that the GREML approach provides an excellent test-bed for evaluating various approaches aimed at improving the fidelity of functional brain phenotypes. Such a GREML-based approach would be particularly powerful for fine-tuning functional phenotypes for meta-analyses of genome-wide association studies (similar to those carried out with structural brain phenotypes [30]), which require pooling of fMRI datasets (paradigm-based or resting) collected under varied conditions and on different scanners; GREML-based estimates of ''heritability'' would provide a useful metric for selecting appropriate post-processing steps and/or modifying inclusion criteria before launching the GWAS.
Overall, this report indicates that GREML-based estimates of heritability of the brain response to facial expressions vary across regions and paradigms, possibly as a function of inter-regional differences in the population variance of functional connectivity. As such, it demonstrates the usefulness of this approach in identifying functional phenotypes with properties suitable for genetic studies.

Participants
As part of the IMAGEN project [12], 2,000 adolescents (,14 years of age) were recruited through local high schools in eight European cities across four countries: France (Paris), Germany Table 2. Proportion of individuals with connections between a given pair of regions of interest.   During the fMRI session participants viewed short videoclips displaying ambiguous facial expressions (gestures such as nose twitching), angry facial expression or control stimuli (nonbiological motion). The control stimuli were adapted from a study of Beauchamp and colleagues [31]. The face stimuli were created as follows. Eight actors (four females) were filmed for the face movements. They were instructed to express different emotions starting from a neutral point. We also extracted short video-clips from the periods when the actors were not expressing the emotions but were nonetheless moving their face (e.g. twitching their nose, opening their mouth, blinking their eyes). Twenty video-clips were selected for the angry and ambiguous face movements respectively. Four raters judged the intensity of each of emotion from those clips. The average rating for the angry face movements, on a scale of 1 (not angry at all) to 9 (very angry) was 7.94 (Standard Deviation [SD] = 0.77). The average rating for the ambiguous facial expressions was 2.18 (SD = 0.84) for anger, 2.97 (SD = 1.07) for sadness and 3.49 (S = 1.03) for happiness; combined across the three scales, the rating of ambiguous facial expressions was 2.92 (SD = 1.18). The control stimuli consisted of black-and-white concentric circles of various contrasts, expanding and contracting at various speeds, roughly matching the contrast and motion characteristics of the faces and hands clips [32]. We presented dynamic video clips of faces because, compared with static faces, they elicit more robust responses in brain regions critical for face processing, such as the fusiform gyrus and amygdala, and engage a more elaborate network for face processing, including regions in the frontal cortex and along the superior temporal sulcus [33]. The three viewing conditions were organized into 18-second blocks (5 Ambiguous, 5 Angry, 9 control) for a total of 160 EPI volumes in a single six-min fMRI run.

MRI Acquisition and Initial Quality Control
From 2,000 participants, a total of 1,926 completed both the Face paradigm and T1-weighted scan. Data from 79 participants were excluded due to excessive head movement during functional MR scanning (more than 2 mm in translation or 2 degrees in rotation errors in either direction), 8 participants were excluded due to unknown age, 5 participants had poor quality of fMRI data, and 1 participant was excluded because of abnormal ventricles. Scans from 1,831 participants were preprocessed using SPM8 toolbox (Statistical Parameter Mapping: Wellcome Department of Cognitive Neurology, London, UK) in MATLAB 7.0 (www. mathworks.com). Functional (EPI) images were motion-corrected with respect to the first volume. Subsequently, the EPI images were aligned to the corresponding high-resolution T1-weighted images (co-registration). The co-registered EPI images were transformed to the ICBM152 template space using the deformation parameters from the nonlinear registration of the corresponding structural image to the ICBM152 template. The nonlinear registration was achieved using the Unified Segmentation tool in SPM package. Further details of the pre-processing pipeline is provided in Tahmasebi et al. [13].

Face Network: Definition and Analysis
Regions of interest (ROIs) relevant for face processing were defined from a probabilistic map computed in a subsample (n = 1,110) of the IMAGEN dataset, as reported in Tahmasebi et al [13]. From this map, 25 ROIs were defined that are consistently (population probability .0.5) engaged during the ambiguous and angry face processing, relative to control (non-biological motion) condition. For each ROI, mean percent BOLD signal change (%BSC) for each ROI was extracted for all participants, as in Tahmasebi et al. [13], and analyzed as phenotypes of interest in GREML analyses. Values of %BSC were standardized (Z-Scored) for each acquisition site to account for scanner effects. Sex was added as a covariate.
The connectivity matrix for each face condition was calculated as follows. Nuisance covariates including white matter (WM) signals, and cerebrospinal fluid (CSF) signals were regressed out from the BOLD signals; WM and CSF voxels were identified by thresholding (at 90%) the WM and CSF tissue probabilistic maps from the ICBM152 standard template. For each ROI, the mean BOLD signal time-series was calculated by averaging the BOLD signal from all voxels constituting the ROI at every time point (160 time points in total). The BOLD time-series for each face condition were then realized by concatenating the mean-centered signal from the corresponding blocks (5 blocks per face condition and 9 blocks for control; each block consists of 8 time points), shifted by 2 TRs (4.4 s) to accommodate for the rise in the hemodynamic response. The correlation matrix was calculated between the time-series from every pair of the 25 ROIs. This yielded a 25-by-25 symmetric functional-connectivity matrix for each participant and face condition. We reduce these matrices into undirected graphs by thresholding each pair-wise correlation at r. 0.3. This creates a graph (network) with ROI's as nodes and edges between them representing functional connections. Within each participant, we calculate node degree for each ROI (node) to summarize the graph. Node degree is simply the count of other ROI's in which the BOLD time series correlates (r.0.3) with the given ROI. This analysis was performed with the Brain Connectivity Toolbox [17].
Given the importance of measurement error in estimating heritability, we have evaluated reproducibility of the brain response to facial expressions by correlating -across the 25 ROIs -the %BSC values obtained in two randomly selected subsamples: Group A (434 males, 483 females) and Group B (448 males, 459 females). In the absence of test-retest reliability measurements, such a cross-group comparison provides an indirect index of measurement error. As shown in Supplemental Figure S3, variations of the %BSC across the 25 ROIs were highly predictable in Group B from measures obtained in Group A (R 2 = 0.96).

Genotyping, Genomic-Relationship-Matrix REstricted Maximum Likelihood (GREML) and Genome-Wide Association Study (GWAS) Analyses
Whole genome data were acquired from 2,089 participants using Illumina Human610-Quad Beadchip and Illumina Hu-man660-Quad Beadchip. Quality control of the genotypes was accomplished using Plink software [34]. Of the 588,875 SNPs overlapping present on both chips, a total of 42,506 single nucleotide polymorphisms (SNPs) were excluded for missingness of more than 5%, 15 individuals excluded for low genotyping rate (less than 97%), 16,385 SNPs were excluded for failing to reach Hardy-Weinberg equilibrium (p, = 0.0001), and 20,131 SNPs were excluded for low minor allele frequency (MAF,0.01).
In total, 511,089 SNPs were used to calculate genetic relationship matrices using GCTA (http://www. complextraitgenomics.com/software/gcta). We excluded adolescents with a genetic relationship .0.05 (i.e., more related than 2 nd degree cousins) to remove the influence of potential shared environment effects or familial causal variants not captured by SNPs. We included the top 10 principal components of the identity-by-state matrix as a covariate in all analyses to control for population stratification in our cohort. For each ROI, we have calculated GREML-based estimates of ''heritability'', defines as Genetic Variance [VG]/Phenotypic Variance [Vp], for the brain response to facial expressions (%BSC).
In order to examine the relationship between the VG/Vp estimates and the number of SNPs reaching a nominal level of significance [15] across the 25 ROIs, we have carried out Genome-wide Association Studies (GWAS) of %BSC using the same set of 1,620 unrelated adolescents. To do so, we used PLINK software [34]. Mean %BSC values where standardized (Z-Scores) in order to control for effects of Sex and Scanning Site. The top 10 principal components of the identity-by-state matrix as a covariate in all analyses to control for population stratification in our cohort.  S1 GREML results for percent BOLD Signal Change (%BSC) in response to Ambiguous Facial expressions (vs. Control Stimuli) in 1,620 adolescents. Regions in bold are those with GREML-based estimates of heritability of the brain response significant at an alpha 0.05 (uncorrected). The critical value for X 2 (1) in this context is 2.7055. Mid-ventrolateral frontal cortex (MVLFC); Mid-dorsolateral frontal cortex (MDLFC); premotor cortex (PMC), pre supplementary motor area (PreSMA); superior temporal sulcus (STS); fusiform face area (FFA); lateral occipital cortex (LOC); left (L); right (R). VG, Genetic Variance; Vp, Phenotypic Variance; df, degrees of freedom.