Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Spoken language processing activates the primary visual cortex

  • Anna Seydell-Greenwald ,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Writing – original draft, Writing – review & editing

    as2266@georgetown.edu (ASG); Ella.StriemAmit@georgetown.edu (ESA)

    Affiliation Center for Brain Plasticity and Recovery, Georgetown University Medical Center, Washington, DC, United States of America

  • Xiaoying Wang,

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Writing – review & editing

    Affiliation State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China

  • Elissa L. Newport,

    Roles Writing – review & editing

    Affiliation Center for Brain Plasticity and Recovery, Georgetown University Medical Center, Washington, DC, United States of America

  • Yanchao Bi,

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Writing – review & editing

    Affiliation State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China

  • Ella Striem-Amit

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing

    as2266@georgetown.edu (ASG); Ella.StriemAmit@georgetown.edu (ESA)

    Affiliations Center for Brain Plasticity and Recovery, Georgetown University Medical Center, Washington, DC, United States of America, Department of Neuroscience, Georgetown University Medical Center, Washington, DC, United States of America

Abstract

Primary visual cortex (V1) is generally thought of as a low-level sensory area that primarily processes basic visual features. Although there is evidence for multisensory effects on its activity, these are typically found for the processing of simple sounds and their properties, for example spatially or temporally-congruent simple sounds. However, in congenitally blind individuals, V1 is involved in language processing, with no evidence of major changes in anatomical connectivity that could explain this seemingly drastic functional change. This is at odds with current accounts of neural plasticity, which emphasize the role of connectivity and conserved function in determining a neural tissue’s role even after atypical early experiences. To reconcile what appears to be unprecedented functional reorganization with known accounts of plasticity limitations, we tested whether V1’s multisensory roles include responses to spoken language in sighted individuals. Using fMRI, we found that V1 in normally sighted individuals was indeed activated by comprehensible spoken sentences as compared to an incomprehensible reversed speech control condition, and more strongly so in the left compared to the right hemisphere. Activation in V1 for language was also significant and comparable for abstract and concrete words, suggesting it was not driven by visual imagery. Last, this activation did not stem from increased attention to the auditory onset of words, nor was it correlated with attentional arousal ratings, making general attention accounts an unlikely explanation. Together these findings suggest that V1 responds to spoken language even in sighted individuals, reflecting the binding of multisensory high-level signals, potentially to predict visual input. This capability might be the basis for the strong V1 language activation observed in people born blind, re-affirming the notion that plasticity is guided by pre-existing connectivity and abilities in the typically developed brain.

Introduction

What are the multisensory or cognitive inputs to primary sensory cortices? Despite the canonical view of the primary visual cortex (V1) as a unisensory, low-level processing station, recent decades have provided ample evidence for multisensory integration in the early visual cortex [13]. Studies have convincingly shown early visual cortex responses to, modulation of, or causal involvement in the processing of simple sound stimuli (e.g. [410]; see [1] for a review.) Usually, these V1 responses are based on low-level features of the auditory stimuli, such as spatial location or temporal synchrony with visual stimuli, allowing for efficient visual responses [1113]. It is less clear whether the early visual cortex also receives high-level information pertaining to object category, congruence, or language. While one study reported that V1 signal patterns allow discrimination based on the categorical content of sounds [14], several other studies did not find such high-level modulations [1522].

There is one case, however, where the early visual cortex is known without a doubt to respond to higher-level non-visual processing: the primary visual cortex (V1) of people born blind has been shown to be involved in language comprehension and production [2329], and its involvement in language is often left-lateralized [23, 24, 26, 3035] like the typical fronto-temporal language activation observed in many neuroimaging studies of sighted people (for reviews, see [36, 37]). This apparent role of V1 in language processing in people born blind appears to be a marked deviation from its function in sighted individuals, where V1 is involved in the processing of basic visual features [3841]. Although evidence for V1 language activation in congenitally blind people is compelling, persuasive evidence for a mechanism by which such extreme functional change from low-level visual to language processing might occur has not been provided to date. Beyond deterioration of the visual pathways [4245], no drastic differences in anatomical connectivity of the visual cortex have been found between congenitally blind people and sighted controls. Importantly, most recent research suggests that brain organization is strongly determined by anatomical connectivity present already at birth [4649]. This view implies that functional reorganization needs to build on, and is limited by, pre-existing capacities and connections of the available tissue, even in cases of sensory deprivation since birth [50, 51]. How can findings of language processing in primary visual cortex in the congenitally blind be reconciled with such a view? If the hypothesis of pre-existing anatomical connectivity and its constraints is correct, then to accord with language processing in primary visual cortex in the blind, there must also be language processing, or related inputs, in primary visual cortex in sighted people.

Here we tested whether language processing recruits V1 in sighted adults, with the dual goal of testing cognitive engagement of early sensory cortices, and addressing the roots of plasticity in blindness. We report a series of experiments that demonstrates V1 activation by spoken language while addressing potential confounds of visual imagery and attention. Experiment 1 employs a robust auditory sentence comprehension task, as compared with a low-level control (backward speech, which is not comprehensible), in 20 neurologically healthy sighted young adults, and Experiments 2 and 3 examine the potential influence of visual imagery in an independent second cohort by testing responses to auditorily presented abstract words, which are hard to visualize. If language indeed activates primary visual cortex during sentence processing in a typically-developed cohort, such activation may be the basis for the more extreme, previously unaccounted for, plasticity in blindness. Further, in the sighted brain, it would support the notion of cognitive responsivity in the primary visual cortex.

Materials and methods

Participants

Experiment 1.

Participants were 20 young adults (5 men, ages 18 to 38, mean 21.8 years) with normal or corrected-to-normal vision and no history of neurological disorder from the Georgetown University community. All were native speakers of English and had not been fluent in any other language by the age of 12. All experimental protocols were approved by the institutional review board of Georgetown University Medical Center, in accordance with the Declaration of Helsinki. Participants provided informed consent and were compensated for their time.

Experiments 2, 3.

Participants were 14 adults with normal or corrected-to-normal vision and no history of neurological disorder (8 men, ages 23 to 66, mean 43.85 years). All were native speakers of Mandarin Chinese. All experimental protocols were approved by the institutional review board of the Department of Psychology at Peking University, China, as well as by the institutional review board of Harvard University, in accordance with the Declaration of Helsinki. Participants provided informed consent and were compensated for their time.

Experimental design

Experiment 1.

The fMRI language task used here was a modified version of an Auditory Description Decision Task used to determine language dominance prior to epilepsy surgery [52, 53]. In the Forward Speech condition, participants heard short English sentences (e.g., “A big gray animal is an elephant”, “Birthday cake lights are candles”, “Something that reflects your image is a beaver”) and pushed a button if they considered the sentence a true statement. In the Reverse Speech condition, they heard the same sentences played in reverse (thus rendered incomprehensible) and pushed a button when they heard a soft beep inserted at the end of the utterance. Audio files with example stimuli can be found in the S1S4 Audios. The proportion of correct statements and reverse speech utterances with beeps was 50%. The task was designed to be easy; performance was nearly perfect (median performance at 100% for both tasks, mean 97.2±4.5% for sentence comprehension, mean 99.5±1.1% for beep detection). Each participant completed two fMRI runs of 5 min and 48 s duration, each containing four 30-second blocks of each of two experimental conditions (Forward and Reversed Speech, six utterances per block) in counterbalanced order, with 12-second silent rest periods at the beginning and end of the run, as well as in between each of the eight active blocks. Aside from a fixation cross that participants were asked to rest their eyes on throughout the scan, no visual stimulation was provided. The Forward>Reverse activation differences evoked by this task are highly robust and reproducible, making them suitable for localizing language-associated brain areas across development [54] and even in cases of atypical functional organization, such as participants with a history of chronic epilepsy [53] or perinatal stroke [55]. Importantly, the sentences were spoken with neutral prosody, so that potential modulation of V1 activation by emotional auditory stimuli [56] is unlikely.

Imaging data were acquired on Georgetown’s research-dedicated 3T Siemens Trio Tim scanner with a 12-channel birdcage head coil. Auditory stimuli were delivered via insert earphones (Sensimetrics S14) worn under ear defenders (Bilsom Thunder T1). Stimulus presentation and response collection (via a Cedrus fiber optics button box) were coordinated by E-Prime 2.0 software. Each of the two functional runs contained 100 functional (T2*-weighted) volumes covering the whole brain in 50 slices acquired in descending order and oriented parallel to the AC-PC plane (EPI parameters: TE = 30 ms, TR = 3000 ms, flip angle = 90°, matrix 64 × 64, slice thickness = 2.8 mm, distance factor = 7%, resulting in an effective voxel size of 3 × 3 × 3 mm). A high-resolution anatomical (T1-weighted) scan was acquired for co-registration (MPRAGE parameters: whole-brain coverage in 176 sagittal slices, TE = 3.5 ms, TR = 2530 ms, TI = 1100 ms, flip angle = 7°, matrix 256 × 256, voxel size = 1 × 1 × 1 mm).

Experiments 2, 3.

Stimuli in these experiments were spoken words, each a two-character word in Mandarin Chinese, belonging to eight categories: abstract concepts (e.g. “freedom”, “truth”, “wish”), concrete everyday object names (e.g. “cup”, “closet”, “computer”), and six additional categories which were not analyzed in the current manuscript (astral/weather phenomena—e.g. “rainbow”, “rain”; scenes—“island”, “beach”; and object features—colors and shapes, e.g. “red”, “square”; see full detail in [57]). Each category included ten words whose imaginability and attentional arousal (as well as other measures not used here) were rated on a 7-point scale [58] by an independent sample of 45 sighted Chinese participants with similar levels of education. The concrete objects and abstract concepts compared in Fig 2B differed significantly in imaginability (Welch t-test contrast, p < 0.001, significant after Bonferroni correction for multiple comparisons including other measures not used here; [57]), but not quite in arousal (p = 0.06 uncorrected; does not surpass the corrected threshold for multiple comparisons).) During both experiments, participants kept their eyes closed and responded with a button push to rare catch trials (occurrences of a fruit name among the other words.) Runs with more than one missed catch trial were excluded from the analysis, as were the imaging data associated with catch events.

During Experiment 2, the participants heard short lists of words in a block design paradigm (8-second blocks with eight words each, baseline between blocks 8 seconds). Each run began with a 12-second rest period. Each block contained words from one of the eight concept categories. Experiment 3 was an item-level slow event-related design and was conducted at a different scanning session on the same participants that participated in Exp. 2. The stimuli were eight of the ten words of each category from Experiment 2, except for the concrete object names (see detail above). During each of eight slow event-related runs, the participants heard each word once, in a random order, followed by a 5-second baseline period.

Imaging data for Experiments 2 and 3 were acquired using a Siemens Prisma 3-T scanner with a 20-channel phase-array head coil at the Imaging Center for MRI Research, Peking University. Functional imaging data for Experiment 2 were comprised of four functional runs, each containing 251 continuous whole-brain functional volumes. Functional imaging data for the single-item-level event-related Experiment 3 were comprised of eight functional runs, each containing 209 continuous whole-brain functional volumes. Data were acquired with a simultaneous multi-slice (SMS) sequence supplied by Siemens: slice planes scanned along the rectal gyrus, 64 slices, phase encoding direction from posterior to anterior; 2 mm thickness; 0.2mm gap; multi-band factor = 2; TR = 2000 ms; TE = 30 ms; FA = 90°; matrix size = 112 × 112; FOV = 224 × 224 mm; voxel size = 2 × 2 × 2 mm. T1-weighted anatomical images were acquired for coregistration using a 3D MPRAGE sequence: 192 sagittal slices; 1mm thickness; TR = 2530 ms; TE = 2.98 ms; inversion time = 1100 ms; FA = 7°; FOV = 256 × 224 mm; voxel size = 0.5 × 0.5 × 1 mm, interpolated; matrix size = 512 × 448.

Data analysis

Preprocessing.

Imaging data were analyzed using BrainVoyager (BVQX 3.6). Anatomical images were corrected for field inhomogeneities and transformed into Talairach space using 9-parameter affine transformation based on manually identified anatomical landmarks. Functional runs underwent slice time correction, removal of linear trends, and 3D motion correction to the first volume of each run using rigid-body transformation. The first two volumes of each run were discarded to allow for magnetization stabilization. Each run was coregistered with the native-space anatomical image of the same participant using 9-parameter gradient-based alignment, and subsequently warped into Talairach space using the same affine transformation used for warping the anatomical data.

Whole-brain group-level analysis

To create group-level activation maps (Figs 1A and 2A), we smoothed the Talairach-warped functional data with a 3D Gaussian kernel of 8 mm FWHM and conducted a hierarchical random effects analysis (RFX GLM; [59]). Each experimental condition’s predictor was modeled by convolving the boxcar predictor describing the condition’s time-course with a standard hemodynamic response function (two gamma, peak at 5 s, undershoot peak at 15 s). In addition, the model included nuisance predictors to capture participant- and run-specific effects as well as motion-related effects (using the z-transformed motion estimates generated during preprocessing). During modeling, voxel time courses were normalized using percent signal change transformation and corrected for serial autocorrelations (AR2).

thumbnail
Fig 1. Left primary visual cortex is engaged in spoken language comprehension.

A. A contrast of comprehensible vs. reversed spoken sentences is shown on brain slices and inflated cortical hemispheres. In addition to the left-lateralized fronto-parieto-temporal language network, significant activation is found in the ventral primary visual cortex. CaS–Calcarine Sulcus. B. GLM parameter estimates (betas) were sampled in the left retinotopically-defined primary visual cortex, showing significant activation for comprehensible speech and selectivity for comprehensible vs. reversed speech. Error bars denote standard error of the mean, *p < 0.05, **p < 0.005 FDR corrected. C. Selectivity for comprehensible speech (the beta difference between forward and reversed speech) is higher in the left V1 than in right V1, showing slight lateralization for language, and stronger in left V1 compared with left V2 (p < 0.05 FDR corrected for both comparisons). Error bars denote standard error of the mean.

https://doi.org/10.1371/journal.pone.0289671.g001

thumbnail
Fig 2. Left primary visual cortex spoken language activation is found for abstract, unimaginable words.

A. Activation for spoken abstract words as compared to the rest baseline is shown on brain slices and inflated cortical hemispheres. In addition to the auditory cortex and inferior frontal cortex, significant activation is found in the primary visual cortex, despite the inability to visually imagine abstract concepts. CaS–Calcarine Sulcus. B. GLM parameter estimates (betas) were sampled in the left retinotopically-defined primary visual cortex, showing significant activation for spoken words, which does not differ between abstract and concrete words. C. Activation for abstract words is significantly higher in left than right V1, showing slight lateralization for language. Error bars denote standard error of the mean, *p < 0.05, **p < 0.005 FDR-corrected.

https://doi.org/10.1371/journal.pone.0289671.g002

Activation maps contrasting the beta values (GLM parameter estimates) for the different conditions via voxel-wise t-tests were thresholded by applying an uncorrected single-voxel threshold of at least p < 0.001 and running BrainVoyager’s Cluster-Level Statistical Threshold Estimator Plugin to determine a cluster-size threshold corresponding to a corrected threshold k < 0.05.

To control for any attention effects elicited by the onset of sound after periods of rest, GLM analyses for Exp. 1 and Exp. 2 (Figs 1A and 2A) were replicated (S3 and S4 Figs respectively) with a brief (1TR) condition modelling auditory signal onset at the beginning of each block as a separate predictor.

Region-of-interest analyses

Regions-of-interest (ROIs) for the primary and secondary visual cortex (V1 and V2, respectively) were defined from an external group localizer [51]. The external retinotopy localizer was acquired in a separate group of 14 normally sighted participants using a standard phase-encoded retinotopic mapping protocol, with eccentricity and polar mapping of ring and wedge stimuli, respectively, to measure visual retinotopic mapping [41, 6062], delivered during two separate experiments. The stimuli were projected by an LCD projector onto a tangent screen positioned over the subject’s forehead and viewed through a tilted mirror. In the eccentricity experiment, an expanding annulus was presented, expanding from 0 to 34 degrees of the subject’s visual field in 30 s, repeated 10 times. The polar angle experiment presented a wedge with a polar angle of 22.5 degrees that rotated around the fixation point, completing a full cycle in 30 s, repeated 20 times. Both the annulus in the eccentricity experiment and the wedge in the polar angle experiment contained a flickering (6 Hz) radial checkerboard pattern according to standard retinotopic procedures (Engel et al., 1994) for field map mapping. In both cases there was a 30-second period of baseline (fixation) before and after the visual stimulus for baseline. Group phase analysis was conducted on the two experiments as done in other studies [63, 64] resulting in group maps depicting the eccentricity and angle mapping aligned to the Talairach-transformed Colin27 brain. Full experimental detail can be found at [51]. Angle (polar) mapping was used to define the borders of V1 and V2 in both hemispheres, used as a ROI to sample activation for the language conditions in the early visual cortices (Figs 1B, 1C and 2B, 2C). V1 was further divided into three portions largely representing foveal, middle, and peripheral visual fields based on the eccentricity mapping (S2B Fig).

Beta values (GLM parameter estimates) for each condition were sampled in individuals, and comparisons across conditions within the same ROI (Figs 1B and 2B) were computed with a two-tailed paired t-test. Comparisons across areas were computed based on the subtraction of beta values between forward and reversed speech for each individual (Fig 1C), and applying a one-tailed paired t-test between regions, under the prediction that language activation would be localized to the left V1, as seen in blindness [23, 24, 26, 35]. Comparably, we used a one-tailed paired t-test between regions to test the hypothesis that abstract word activation vs. baseline would be localized to the left V1 (Fig 2C). In addition, to investigate imaginability and arousal as potential contributors to any observed activation, we explored correlations between language activation (GLM parameter estimate, beta values) in the left V1 ROI and imaginability and arousal behavioral ratings of the words presented in Experiment 3, across all 56 words used in the experiment. Lastly, to ensure that the time-course of activation characteristics resemble a genuine neural response, the averaged percent signal change with relation to condition onset was sampled from the left V1 ROI and the standard errors were calculated for each condition and plotted for each time point (S2A and S2C Fig).

To statistically correct for these multiple comparisons conducted on our ROI analyses, we report p-values computed using the false-discovery-rate (FDR) approach [65]. Specifically, we corrected for 5 statistical comparisons for the ROI analyses in Experiment 1: (1) V1 forward vs. reverse speech (Fig 1B), (2) V1 forward speech vs. silent baseline, (3) left vs. right V1 and (4) Left V1 vs. V2 (Fig 1C), and (5) the V1 eccentricity effect (S2B Fig). For Experiment 2, we corrected for 4 statistical comparisons: (1) V1 abstract vs. concrete words (Fig 2B), (2) V1 abstract words vs. silent baseline, (3) V1 concrete words vs. silent baseline, and (4) left vs. right V1 (Fig 2C).

Results

Experiment 1

In contrasting activation by forward and reverse speech in a whole-brain analysis, a typical left-lateralized fronto-temporal “language” network emerged (Fig 1A), as identified by numerous neuroimaging studies (for reviews, see [36, 37]). The primary auditory cortex was not significantly activated because the contrasted conditions are matched in low-level auditory information (see similarly [66]), including the sound envelope change rate, which has been suggested to activate V1 [67]. Despite this, the primary visual cortex was significantly activated by forward speech (S1 Fig) and more strongly activated by forward than by reverse speech (Fig 1A).

This preference was confirmed in region-of-interest (ROI) analyses extracting percent signal change from left and right V1 and V2. Left V1, our primary ROI, showed a significantly stronger response to forward than to reverse speech (Fig 1B; paired t-test, t(19) = 4.02, FDR-corrected p = 0.002, d’ = 1.89, one-tailed and displayed a standard stimulus-evoked hemodynamic response for spoken sentences (S2A Fig). The response to forward speech was also significant compared to baseline (t(19) = 2.88, FDR-corrected p = 0.012).

A comparison of the forward>reverse speech effect in retinotopically defined V1 in the left and right hemisphere (Fig 1C) confirmed the impression from the whole-brain analysis that the activation seemed to be at least somewhat stronger in the left hemisphere (paired t-test, t(19) = 1.98, FDR-corrected p = 0.039, d’ = 0.91). Moreover, the forward>reverse speech effect was slightly weaker in left V2 than in left V1 (paired t-test, t(19) = 2.06, FDR-corrected p = 0.045, d’ = 0.95).

To test whether the V1 activations we observed here might reflect increased attention to the onset of auditory stimulation [68], we repeated the analyses while including a confound predictor modelling a short response to the onset of the conditions. This control analysis replicated the main effects (S3 Fig), making simple auditory attention effects an unlikely explanation for the V1 language activations. It is also notable that the whole-brain activations observed here did not include any areas of the fronto-parietal attention network [6972]) that one might expect to see activated if the forward speech condition elicited a significantly stronger attentional response than the control condition.

Experiment 2

Could the V1 language activations in Experiment 1 stem from visual imagery, due to the concrete content of the spoken sentences? To explore this possibility, we investigated, in a separate group of sighted adults, whether V1 would show differential activation for abstract (less imaginable) and concrete (easily imaginable) spoken words (e.g., “freedom”, “truth”, “wish” vs. “cup”, “closet”, “computer”). Just as for spoken sentence comprehension (Experiment 1 above), whole-brain activation for listening to blocks of abstract words as compared to inter-block rest interval baseline (Experiment 2; see also [57]) included, in addition to vast activation of the temporal lobe and inferior frontal cortex, also significant localized activation in the calcarine sulcus (Fig 2A). Again, activation was stronger in left than in right V1 (Fig 2C; paired t-test, t(13) = 2.77, FDR-corrected p = 0.048, d’ = 1.5). Activation time-courses extracted from left V1 resembled a typical hemodynamic response function for both abstract and concrete words (S2C Fig). Importantly, left V1 activation did not differ between abstract and concrete words (object names; Fig 2B; t(13) = 0.48, FDR-corrected p = 0.975, d’ = 0.27), even though the latter were significantly more imaginable according to behavioral ratings (t(9) = 1074, p < 0.001 uncorrected, significant with correction for multiple comparisons, d’ = 716). As in Exp. 1, modelling the potential attention-arousing effect of the auditory onset at the beginning of each block as a nuisance condition did not affect the main findings (S4 Fig).

Experiment 3

Last, a separate study, an event-related design of spoken words (performed on the same participants and using most of the abstract and easily imaginable words used in Exp. 2; see methods), allowed us to test whether left V1 activation correlated with imaginability and attentional arousal ratings for spoken words of a variety of imaginable and abstract concept types [57]. We computed the correlation between behavioral ratings collected for these words and beta values for each presented word within the left V1 region of interest. No correlation was found between left V1 activation and imaginability ratings (r2(54) = 0.003, p = 0.69 FDR corrected) or arousal ratings (r2(54) = 0.01, p = 0.90, FDR corrected), although left V1 still showed activation (above baseline) for abstract words (S5 Fig). Together, these findings suggest that the observed forward>reverse speech activation in V1 did not result from imagery or attention confounds.

Discussion

The primary visual cortex is widely thought to be a low-level sensory station devoted to the processing of simple visual features [3841]. However, there is also increasing evidence implicating the primary visual cortex in aspects of low-level multisensory integration [13], and some recent evidence, although still controversial, indicates that it may also receive signals related to higher level non-visual representations, specifically non-visual imagery [14] and working memory [73]. The present results further expand on the known multisensory information reaching V1. We observed activation for spoken sentences in V1 of typically developed individuals, which showed a preference for comprehensible over incomprehensible speech (Fig 1A, 1B). Moreover, this activation tended to be stronger in left than right V1 (Figs 1A, 1C and 2C), just like the V1 language activation in blindness [24, 26, 34, 35] and in line with the observation of left-lateralization for language activation in the vast majority of adults regardless of handedness [74, 75]. It also tended to be stronger in V1 than in V2, suggesting that it did not emerge from feedback cortico-cortical connectivity from visual language areas via higher retinotopic cortical stations (e.g., V2; Fig 1C). The same pattern was evident in response to spoken abstract words in a separate sighted cohort in two separate experiments (Fig 2 and S5 Fig). Together these findings suggest that left-lateralized primary visual cortex responds to spoken language information in typically developed sighted adults. These findings reflect on several key issues regarding the multisensory properties of primary visual cortex, the developmental origins of reorganization in the blind brain, and the nature of brain plasticity itself.

Before we discuss these implications, we must first address why our study is the first to highlight V1 language activation in sighted individuals despite the large number of functional neuroimaging studies that have investigated language activation in the typically-developed population. One likely reason is that V1 activation is small relative to that in other regions of the frontotemporal language network, both regarding peak signal change and regarding spatial extent. Thus, depending on the statistical power of the experiment and applied thresholds, V1 activation may not be apparent in all functional neuroimaging studies contrasting comprehensible speech with silence, non-speech stimuli, or incomprehensible speech stimuli. However, it is apparent in some [7684]. In studies directly looking for early visual cortex recruitment in blindness, which included sighted participants as a control group, there is at times mention of responses in the sighted [24, 85], but often, possibly due to statistical analysis and power or to the specific contrasts used, significant activation in the sighted is either not found [23, 28, 32, 86, 87], or is (accurately) reported as smaller than that of the blind group without being tested for significance in and of itself [26]. In summary, while our experiments are not the first to show V1 activation in language tasks in sighted adults, our focus on this activation and examination of its properties with relevant controls allows us to interpret it as a meaningful response to spoken language.

To do so, we must address whether our findings could be explained by confounds such as differences in visual stimuli, visual imagery or increased attention to speech which in themselves can generate responses in V1 [8891], as has been suggested for several other studies observing early visual cortex activation by language stimuli [76, 77, 79, 82]. Visual stimulation was matched and essentially absent throughout our experiments: In Experiment 1, participants were instructed to rest their eyes on a fixation cross throughout, matching visual stimulation during forward and reverse speech blocks; and during Experiments 2 and 3, participants listened with their eyes closed and blindfolded. This rules out unmatched visual stimulation between the conditions as a potential explanation for the observed V1 activation.

There are also several reasons to argue that global attention is an unlikely explanation for the V1 activation observed here. First, when attention to sounds activates V1, it likely serves a role in spatial attention orientation, found more strongly in peripheral V1 [13, 91, 92] and stems from direct and indirect anatomical connectivity between auditory cortices and mostly (though not only) peripheral retinotopic locations of V1 found in primates and humans [9399]. In contrast, the activation pattern in comparing forward and reversed speech in our study was not peripherally localized (S2B Fig; ANOVA for an eccentricity effect in left V1 F(2,143) = 0.01, p = 0.99). Second, no activation was observed in the fronto-parietal attention network (typically bilateral or right-lateralized; [6972]) in our whole-brain analysis, which would be expected if there were significant attention differences between the conditions (either due to larger top-down attention allocated to processing the comprehensible stimulus, or to larger effort to selectively attend to that stimulus over the scanner noise). Moreover, including an additional nuisance regressor to capture attention effects associated with sound onset did not abolish the V1 response in either block-design experiment (S3 and S4 Figs). Lastly, activation in left V1 was not correlated with the arousal ratings of the heard words in Experiment 3. All this makes it unlikely that attention is the primary driver for the V1 responses we observed.

Similarly, it does not appear to stem from visual imagery. Visual imagery may activate and its content can be decoded from the primary visual cortex [89, 100106]. However, imagery responses are stronger in association rather than primary cortex [107, 108] and are typically bilateral [104, 108, 109]. When V1 activation is reported for visual imagery, it is usually associated with explicit imagery of high-resolution detail of images [89, 110, 111]. In contrast, our sentence comprehension task did not require explicit imagery or attention to visual detail, activation was slightly stronger in V1 than V2 (Fig 1A, 1C), and elicited no activation or deactivation as compared to rest in other retinotopic and association visual areas (S1 Fig). Moreover, we observed the same localized V1 activation in a whole-brain analysis for abstract words (Fig 2A and 2C), and it was no weaker for abstract than concrete words (Fig 2B) even though the former were rated as significantly less imaginable. The V1 response also was not correlated with imaginability ratings (Experiment 3). This pattern of results argues against visual imagery as an explanation for the observed V1 activation.

Having ruled out these potential confounds to the best of our ability, we lean towards interpreting the V1 responses to forward>reverse speech (Experiment 1) and to spoken words of varying imaginability (Experiments 2 and 3) as activation driven by spoken language. However, our results do not reveal which aspect(s) of spoken language processing are contributing to this activation. Our reversed speech condition controls for several aspects, such as the overall spectral and temporal envelope of the forward speech stimuli, which may in itself drive V1 activation [67, 68]. Nonetheless, it does differ from the forward speech condition in several ways beyond the lack of linguistic information and meaning. One difference between the forward and reversed speech is that reverse speech contains sounds that are not usually (and in some cases cannot be) produced by the human vocal apparatus. If V1 response are driven by the visual or cross-modal associations of spoken language’s common speech sounds [112], this could also form a difference in Exp. 1. Further, although reversed speech is well matched to forward speech for long-term spectrotemporal characteristics, including the overall spectral and temporal envelopes, time-reversal disrupts local spectrotemporal patterns. Yet, it appears unlikely that V1 would be more sensitive to these features than A1, which does not show differences between reversed and forward speech (Fig 1A). The most obvious difference between forward and reverse speech remains the presence of and attention to linguistic information and meaning. This difference is what the present study shares with the other studies that also found V1 language activation in sighted adults [7684]. All these studies were conducted using different designs and stimuli, and while each individual one may have potential confounds, the most parsimonious explanation for the V1 activation common to all of them is what all studies’ contrasts have in common with each other: linguistic and semantic processing.

How might linguistic information reach V1 and what role could it play? Primary sensory cortices receive information from multiple cortical and subcortical stations. Specifically, beyond thalamic LGN and pulvinar projections, primary visual cortex receives input from auditory cortices, parietal cortex and other regions including frontal cortex in primates and other mammals [9496, 98, 113119]. These feedback pathways [119, 120] allow for multisensory integration even in V1 [13, 9, 121], along with integration of reward value information [122, 123]. Relevant non-visual information can affect early visual cortex excitability [4], and interact with visual EEG alpha wave phase [6, 8, 124]. Many studies have shown early visual cortex responses to, modulation of, or causal involvement in the processing of simple sound stimuli (e.g. [410]; see [1] for a review), often based on their spatial location or temporal synchrony [1113]. However, it is less clear whether early visual cortex receives information pertaining to high-level auditory information such as object category, congruence, or language. Vetter and colleagues [14, 125] showed that V1 signal patterns can discriminate the categorical content of sounds both in sighted and congenitally blind individuals, but others did not find such high-level modulations [1522]. Moreover, recent studies showed that the primary visual cortex may represent information pertaining to the temporal envelope for complex sounds [67, 68], suggesting that the successful discrimination of sound category from V1 activation patterns could have been based on mid-level properties of the sounds rather than high-level semantic category information. However, in our experiment, the sound temporal envelope modulation frequency was maintained in the reversed speech control condition, which demonstrates that the preference of V1 for comprehensible speech goes beyond sound envelope.

Although the representational content of language responses in V1 will need to be further addressed by multivariate analyses, our results provide additional evidence for semantic processing of auditory stimuli in V1. Theoretically, these cross-modal and higher-level inputs to V1 could play a role in predictive coding, whereby predictions of future states and inputs enables visual cortex to anticipate coming events [119, 126128] and allows for efficient coding and adapting to the everchanging environment [3, 117, 119]. How language comprehension fits into this framework is uncertain. Language input into V1 may allow integrating contextual information that enables visual cortex to anticipate coming events [119, 126, 127]. Alternatively, it may play a simpler role in alerting spatial or overall attention [13, 129], without conveying specific content. What we tentatively interpret as V1 language activation may even be epiphenomenal altogether; our data do not speak directly to these alternative explanations, which will need to be addressed in future work. Importantly, accounts of predictive use of speech information would have to reconcile the level of representation of incoming high-level inputs with the spatial and low-level nature of V1 (e.g., [129]), such that these types of information can be integrated in a meaningful way.

While our data cannot speak to the functional role of V1 language activation in the sighted, the resemblance of the V1 activation observed here to that seen in people born blind is intriguing. In people born blind early retinotopic visual cortices, including V1, are activated by high-level cognitive tasks such as language, verbal memory and executive function [2328, 31, 32, 34, 86, 130132]. Stimulating primary visual cortex affects Braille reading [133] and verb generation [33] suggesting that, at least in people born blind, this activation may indeed contribute to language processing. The fact that similar (albeit weaker) V1 language activation can be seen in the sighted brain suggests that such activation in the blind may not require massive changes in brain organization. This is particularly important because despite evidence suggesting increased functional connectivity between early visual cortex and the inferior frontal lobe in the blind [51, 134138], no anatomical evidence for such large changes has been identified to date. Rather, differences in anatomical connectivity between sighted and early blind individuals, although evident in some animal models [139142], in humans appear to be limited in scope, mostly to the deterioration of the visual pathways in the blind [4245]. Responses to language in V1 in the sighted indicate that even major functional reorganization (e.g. functional pluripotency [143]) may not be needed either. Rather, our data suggest a more conservative explanation of V1’s language recruitment in blindness: little reorganization of V1 structure or function (perhaps in the shape of unmasking [144], and additional local changes) is required to support language recruitment of deprived cortex because it also recruits non-deprived cortex, albeit to a lower extent. This is in line with similar explanations that have already been proposed for non-visual responses in the visual cortex in blindness for other perceptual [48, 145147] and cognitive [148, 149] domains. Importantly, no comparable explanation was possible for language recruitment in V1, given that V1 language responses were not reported in the sighted brain. Our evidence here closes this gap, and reconciles the seemingly inordinate plasticity for language in people born blind with current views of connectivity-driven functional brain organization [4649, 150]. Thus, we contribute to a unifying explanatory framework for findings in the primary and association cortices in the blind, based on extant non-visual functions of the visual cortex.

In summary, our findings expand on the known non-visual, cognitive inputs to the primary visual cortex and suggest its modulation also by cognitive inputs even in the sighted brain, possibly as part of a predictive coding mechanism guiding visual perception. These findings also provide evidence that language-driven visual cortex activation in the blind can be explained without proposing drastic changes to cortical tissue connectivity or function. This suggests that human cortical plasticity is still limited by innate anatomical structures and functional characteristics and is not unconstrained even following extreme changes in early experience.

Supporting information

S1 Fig. Left primary visual cortex is activated by spoken language.

A contrast of comprehensible spoken sentences as compared to baseline is shown on brain slices and inflated cortical hemispheres. In addition to the auditory cortex and left-lateralized fronto-parieto-temporal language network, significant activation is found in the primary visual cortex, accompanied by deactivation of other parts of early visual cortex. CaS–Calcarine Sulcus.

https://doi.org/10.1371/journal.pone.0289671.s001

(DOCX)

S2 Fig. Activation for speech comprehension in primary visual cortex.

(A) Time course of activation from Experiment 1 was sampled from retinotopic left V1, showing typical BOLD-shaped response in V1 for speech, which is higher for forward as compared to reversed speech. (B) GLM parameter estimates (betas) were sampled in the left retinotopically-defined primary visual cortex divided based on eccentricity, showing that the activation for forward speech does not differ between foveal, middle and peripheral-representing V1 sections. (C) Time course of activation from Experiment 2 was sampled from retinotopic left V1, showing typical BOLD-shaped response in V1 for abstract and concrete words.

https://doi.org/10.1371/journal.pone.0289671.s002

(DOCX)

S3 Fig. Left primary visual cortex is engaged in spoken language comprehension when including a “block onset” predictor to control for bottom-up attention effects.

This figure is provided for comparison with Fig 1. The underlying analyses are the same except for the inclusion of an additional nuisance predictor in the GLM to capture the bottom-up attention effects that might occur at the beginning of each block, at the onset of auditory stimulation. Even after including this additional predictor, a left-lateralized fronto-temporal language network is clearly evident (A), left V1 activation is significantly stronger for comprehensible forward than incomprehensible reverse speech (B), and the forward>reverse effect is slightly stronger in left V1 compared to right V1 and left V2 (C). **p < 0 .01, *p < 0.05 FDR corrected.

https://doi.org/10.1371/journal.pone.0289671.s003

(DOCX)

S4 Fig. Left primary visual cortex spoken language activation is found for abstract, unimaginable words when including a “block onset” predictor to control for bottom-up attention effects This figure is provided for comparison with Fig 2.

The underlying analyses are the same except for inclusion of an additional nuisance predictor in the GLM to capture the bottom-up attention effects that might occur at the beginning of each block, at the onset of auditory stimulation. Even after including this additional predictor, activation of the primary visual cortex for abstract words is clearly evident (A), with no difference in response between abstract and concrete (object name) words (B; p = 0.94), and the response to abstract words is stronger in left V1 compared to right V1 (C). *p < 0.05 FDR corrected.

https://doi.org/10.1371/journal.pone.0289671.s004

(DOCX)

S5 Fig. Left primary visual cortex spoken language activation is replicated for abstract words in Experiment 3.

Activation for spoken abstract words in Experiment 3 as compared to the rest baseline is shown on brain slices and inflated cortical hemispheres. In addition to the auditory cortex and inferior frontal cortex, significant activation is found in the left primary visual cortex, replicating the findings of Experiments 1 and 2. CaS–Calcarine Sulcus.

https://doi.org/10.1371/journal.pone.0289671.s005

(DOCX)

S1 Audio. Example of a correct spoken sentence as used in Experiment 1.

https://doi.org/10.1371/journal.pone.0289671.s006

(WAV)

S2 Audio. Example of an incorrect spoken sentence as used in Experiment 1.

https://doi.org/10.1371/journal.pone.0289671.s007

(WAV)

S3 Audio. S1 Audio played in reverse and thus rendered incomprehensible.

https://doi.org/10.1371/journal.pone.0289671.s008

(WAV)

S4 Audio. S2 Audio played in reverse and thus rendered incomprehensible.

https://doi.org/10.1371/journal.pone.0289671.s009

(WAV)

References

  1. 1. Murray MM, Thelen A, Thut G, Romei V, Martuzzi R, Matusz PJ. The multisensory function of the human primary visual cortex. Neuropsychologia. 2016;83:161–9. pmid:26275965
  2. 2. Ghazanfar AA, Schroeder CE. Is neocortex essentially multisensory? Trends in Cognitive Sciences. 2006;10(6):278–85. pmid:16713325
  3. 3. van Atteveldt N, Murray Micah M, Thut G, Schroeder Charles E. Multisensory Integration: Flexible Use of General Operations. Neuron. 2014;81(6):1240–53. pmid:24656248
  4. 4. Romei V, Murray MM, Cappe C, Thut G. Preperceptual and Stimulus-Selective Enhancement of Low-Level Human Visual Cortex Excitability by Sounds. Current Biology. 2009;19(21):1799–805. pmid:19836243
  5. 5. Martuzzi R, Murray MM, Michel CM, Thiran J-P, Maeder PP, Clarke S, et al. Multisensory Interactions within Human Primary Cortices Revealed by BOLD Dynamics. Cerebral Cortex. 2007;17(7):1672–9. pmid:16968869
  6. 6. Mercier MR, Foxe JJ, Fiebelkorn IC, Butler JS, Schwartz TH, Molholm S. Auditory-driven phase reset in visual cortex: Human electrocorticography reveals mechanisms of early multisensory integration. Neuroimage. 2013;79(0):19–29. pmid:23624493
  7. 7. Spierer L, Manuel AL, Bueti D, Murray MM. Contributions of pitch and bandwidth to sound-induced enhancement of visual cortex excitability in humans. Cortex. 2013;49(10):2728–34. Epub 2013/02/20. pmid:23419789.
  8. 8. Cecere R, Rees G, Romei V. Individual Differences in Alpha Frequency Drive Crossmodal Illusory Perception. Current Biology. 2015;25(2):231–5. pmid:25544613
  9. 9. Rohe T, Noppeney U. Distinct Computational Principles Govern Multisensory Integration in Primary Sensory and Association Cortices. Current Biology. 2016;26(4):509–14. pmid:26853368
  10. 10. Ferraro S, Van Ackeren MJ, Mai R, Tassi L, Cardinale F, Nigri A, et al. Stereotactic electroencephalography in humans reveals multisensory signal in early visual and auditory cortices. Cortex. 2020. pmid:32092494
  11. 11. Noppeney U, Lee HL. Causal inference and temporal predictions in audiovisual perception of speech and music. Annals of the New York Academy of Sciences. 2018;1423(1):102–16. pmid:29604082
  12. 12. Meienbrock A, Naumer MJ, Doehrmann O, Singer W, Muckli L. Retinotopic effects during spatial audio-visual integration. Neuropsychologia. 2007;45(3):531–9. Epub 2006/06/27. pmid:16797610.
  13. 13. Azevedo FA, Ortiz-Rios M, Li Q, Logothetis NK, Keliris GA. A Potential Role of Auditory Induced Modulations in Primary Visual Cortex. Multisens Res. 2015;28(3–4):331–49. Epub 2015/08/21. pmid:26288903.
  14. 14. Vetter P, Smith FW, Muckli L. Decoding Sound and Imagery Content in Early Visual Cortex. Curr Biol. 2014;24(11):1256–62. Epub 2014/05/27. pmid:24856208.
  15. 15. Fort A, Delpuech C, Pernier J, Giard M-H. Early auditory–visual interactions in human cortex during nonredundant target identification. Cognitive Brain Research. 2002;14(1):20–30. pmid:12063127
  16. 16. Molholm S, Ritter W, Javitt DC, Foxe JJ. Multisensory visual-auditory object recognition in humans: a high-density electrical mapping study. Cereb Cortex. 2004;14(4):452–65. pmid:15028649.
  17. 17. Yuval-Greenberg S, Deouell LY. What You See Is Not (Always) What You Hear: Induced Gamma Band Responses Reflect Cross-Modal Interactions in Familiar Object Recognition. The Journal of Neuroscience. 2007;27(5):1090–6. pmid:17267563
  18. 18. Brandman T, Avancini C, Leticevscaia O, Peelen MV. Auditory and Semantic Cues Facilitate Decoding of Visual Object Category in MEG. Cerebral Cortex. 2020;30(2):597–606. pmid:31216008
  19. 19. de Haas B, Schwarzkopf DS, Urner M, Rees G. Auditory modulation of visual stimulus encoding in human retinotopic cortex. NeuroImage. 2013;70:258–67. pmid:23296187
  20. 20. Lu L, Zhang G, Xu J, Liu B. Semantically Congruent Sounds Facilitate the Decoding of Degraded Images. Neuroscience. 2018;377:12–25. pmid:29408368
  21. 21. Gu J, Liu B, Li X, Wang P, Wang B. Cross-modal representations in early visual and auditory cortices revealed by multi-voxel pattern analysis. Brain Imaging Behav. 2020;14(5):1908–20. Epub 2019/06/12. pmid:31183774.
  22. 22. Csonka M, Mardmomen N, Webster PJ, Brefczynski-Lewis JA, Frum C, Lewis JW. Meta-Analyses Support a Taxonomic Model for Representations of Different Categories of Audio-Visual Interaction Events in the Human Brain. Cerebral Cortex Communications. 2021;2(1). pmid:33718874
  23. 23. Burton H, Diamond JB, McDermott KB. Dissociating cortical regions activated by semantic and phonological tasks: a FMRI study in blind and sighted people. J Neurophysiol. 2003;90(3):1965–82. Epub 2003 Jun 4. pmid:12789013
  24. 24. Burton H, Snyder AZ, Diamond JB, Raichle ME. Adaptive changes in early and late blind: a FMRI study of verb generation to heard nouns. J Neurophysiol. 2002;88(6):3359–71. pmid:12466452
  25. 25. Röder B, Stock O, Bien S, Neville H, Rösler F. Speech processing activates visual cortex in congenitally blind humans. Eur J Neurosci. 2002;16(5):930–6. pmid:12372029
  26. 26. Bedny M, Pascual-Leone A, Dodell-Feder D, Fedorenko E, Saxe R. Language processing in the occipital cortex of congenitally blind adults. Proc Natl Acad Sci U S A. 2011;108(11):4429–34. Epub 2011/03/04. pmid:21368161; PubMed Central PMCID: PMC3060248.
  27. 27. Dietrich S, Hertrich I, Ackermann H. Ultra-fast speech comprehension in blind subjects engages primary visual cortex, fusiform gyrus, and pulvinar—a functional magnetic resonance imaging (fMRI) study. BMC Neurosci. 2013;14(1):74. Epub 2013/07/25. pmid:23879896.
  28. 28. Lane C, Kanjlia S, Omaki A, Bedny M. “Visual” Cortex of Congenitally Blind Adults Responds to Syntactic Movement. The Journal of Neuroscience. 2015;35(37):12859–68. pmid:26377472
  29. 29. Abboud S, Engemann D-A, Cohen L. Semantic coding in the occipital cortex of early blind individuals. bioRxiv. 2019:539437.
  30. 30. Burton H, Snyder AZ, Conturo TE, Akbudak E, Ollinger JM, Raichle ME. Adaptive changes in early and late blind: a fMRI study of Braille reading. J Neurophysiol. 2002;87(1):589–607. pmid:11784773
  31. 31. Sadato N, Pascual-Leone A, Grafman J, Deiber MP, Ibanez V, Hallett M. Neural networks for Braille reading by the blind. Brain. 1998;121(Pt 7):1213–29. pmid:9679774
  32. 32. Amedi A, Raz N, Pianka P, Malach R, Zohary E. Early ’visual’ cortex activation correlates with superior verbal memory performance in the blind. Nat Neurosci. 2003;6(7):758–66. pmid:12808458
  33. 33. Amedi A, Floel A, Knecht S, Zohary E, Cohen LG. Transcranial magnetic stimulation of the occipital pole interferes with verbal processing in blind subjects. Nat Neurosci. 2004;7(11):1266–70. pmid:15467719.
  34. 34. Ofan RH, Zohary E. Visual Cortex Activation in Bilingual Blind Individuals during Use of Native and Second Language. Cereb Cortex. 2006;17(6):1249–59. pmid:16861335.
  35. 35. Burton H. Visual cortex activity in early and late blind people. J Neurosci. 2003;23(10):4005–11. pmid:12764085.
  36. 36. Vigneau M, Beaucousin V, Hervé PY, Duffau H, Crivello F, Houdé O, et al. Meta-analyzing left hemisphere language areas: phonology, semantics, and sentence processing. Neuroimage. 2006;30(4):1414–32. Epub 2006/01/18. pmid:16413796.
  37. 37. Price CJ. A review and synthesis of the first 20 years of PET and fMRI studies of heard speech, spoken language and reading. Neuroimage. 2012. Epub 2012/05/16. [pii] pmid:22584224.
  38. 38. Hubel DH, Wiesel TN. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. The Journal of physiology. 1962;160(1):106–54. pmid:14449617.
  39. 39. Hubel DH, Wiesel TN. Receptive fields and functional architecture of monkey striate cortex. The Journal of Physiology. 1968;195(1):215–43. pmid:4966457
  40. 40. Grill-Spector K, Malach R. The human visual cortex. Annu Rev Neurosci. 2004;27:649–77. pmid:15217346.
  41. 41. Wandell BA, Dumoulin SO, Brewer AA. Visual field maps in human cortex. Neuron. 2007;56(2):366–83. Epub 2007/10/30. pmid:17964252.
  42. 42. Noppeney U, Friston KJ, Ashburner J, Frackowiak R, Price CJ. Early visual deprivation induces structural plasticity in gray and white matter. Curr Biol. 2005;15(13):R488–90. pmid:16005276.
  43. 43. Shimony JS, Burton H, Epstein AA, McLaren DG, Sun SW, Snyder AZ. Diffusion tensor imaging reveals white matter reorganization in early blind humans. Cereb Cortex. 2006;16(11):1653–61. pmid:16400157.
  44. 44. Yu C, Shu N, Li J, Qin W, Jiang T, Li K. Plasticity of the corticospinal tract in early blindness revealed by quantitative analysis of fractional anisotropy based on diffusion tensor tractography. Neuroimage. 2007;36(2):411–7. pmid:17442594.
  45. 45. Shu N, Li J, Li K, Yu C, Jiang T. Abnormal diffusion of cerebral white matter in early blindness. Hum Brain Mapp. 2009;30(1):220–7. Epub 2007/12/12. pmid:18072278.
  46. 46. Mahon BZ, Caramazza A. What drives the organization of object knowledge in the brain? Trends Cogn Sci. 2011;15(3):97–103. Epub 2011/02/15. pmid:21317022.
  47. 47. Hannagan T, Amedi A, Cohen L, Dehaene-Lambertz G, Dehaene S. Origins of the specialization for letters and numbers in ventral occipitotemporal cortex. Trends in Cognitive Sciences. 2015;(0). pmid:26072689
  48. 48. Heimler B, Striem-Amit E, Amedi A. Origins of task-specific sensory-independent organization in the visual and auditory brain: neuroscience evidence, open questions and clinical implications. Curr Opin Neurobiol. 2015;35:169–77. Epub 2015/10/16. pmid:26469211.
  49. 49. Saygin ZM, Osher DE, Norton ES, Youssoufian DA, Beach SD, Feather J, et al. Connectivity precedes function in the development of the visual word form area. Nat Neurosci. 2016;advance online publication(9):1250–5. pmid:27500407; PubMed Central PMCID: PMC5003691.
  50. 50. Striem-Amit E, Vannuscorps G, Caramazza A. Plasticity based on compensatory effector use in the association but not primary sensorimotor cortex of people born without hands. Proceedings of the National Academy of Sciences. 2018;115(30):7801–6. pmid:29997174; PubMed Central PMCID: PMC6065047.
  51. 51. Striem-Amit E, Ovadia-Caro S, Caramazza A, Margulies DS, Villringer A, Amedi A. Functional connectivity of visual cortex in the blind follows retinotopic organization principles. Brain. 2015;138(6):1679–95. pmid:25869851; PubMed Central PMCID: PMC4614142.
  52. 52. Gaillard WD, Berl MM, Moore EN, Ritzl EK, Rosenberger LR, Weinstein SL, et al. Atypical language in lesional and nonlesional complex partial epilepsy. Neurology. 2007;69(18):1761–71. pmid:17967992
  53. 53. Berl MM, Zimmaro LA, Khan OI, Dustin I, Ritzl E, Duke ES, et al. Characterization of atypical language activation patterns in focal epilepsy. Ann Neurol. 2014;75(1):33–42. Epub 2013/09/17. pmid:24038442; PubMed Central PMCID: PMC4209919.
  54. 54. Olulade OA, Seydell-Greenwald A, Chambers CE, Turkeltaub PE, Dromerick AW, Berl MM, et al. The neural basis of language development: Changes in lateralization over age. Proceedings of the National Academy of Sciences. 2020:201905590. pmid:32900940
  55. 55. Newport EL, Landau B, Seydell-Greenwald A, Turkeltaub PE, Chambers CE, Dromerick AW, et al. Revisiting Lenneberg’s Hypotheses About Early Developmental Plasticity: Language Organization After Left-Hemisphere Perinatal Stroke. Biolinguistics (Nicos). 2017;11:407–22. Epub 12/31. pmid:30556058.
  56. 56. Skouras S, Gray M, Critchley H, Koelsch S. fMRI Scanner Noise Interaction with Affective Neural Processes. PLOS ONE. 2013;8(11):e80564. pmid:24260420
  57. 57. Striem-Amit E, Wang X, Bi Y, Caramazza A. Neural Representation of Visual Concepts in People Born Blind. Nature Communications. 2018;9(1):5250. pmid:30531889; PubMed Central PMCID: PMC6286313.
  58. 58. Barca L, Burani C, Arduino LS. Word naming times and psycholinguistic norms for Italian nouns. Behav Res Methods Instrum Comput. 2002;34(3):424–34. Epub 2002/10/25. pmid:12395559.
  59. 59. Friston KJ, Holmes AP, Worsley KJ. How many subjects constitute a study? Neuroimage. 1999;10(1):1–5. pmid:10385576
  60. 60. Engel SA, Rumelhart DE, Wandell BA, Lee AT, Glover GH, Chichilnisky EJ, et al. fMRI of human visual cortex. Nature. 1994;369(6481):525. Epub 1994/06/16. pmid:8031403.
  61. 61. Sereno MI, Dale AM, Reppas JB, Kwong KK, Belliveau JW, Brady TJ, et al. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science. 1995;268(5212):889–93. Epub 1995/05/12. pmid:7754376.
  62. 62. Wandell BA, Winawer J. Imaging retinotopic maps in the human brain. Vision Research. 2011;51(7):718–37. pmid:20692278
  63. 63. Hertz U, Amedi A. Disentangling unisensory and multisensory components in audiovisual integration using a novel multifrequency fMRI spectral analysis. Neuroimage. 2010;52(2):617–32. Epub 2010/04/24. pmid:20412861.
  64. 64. Striem-Amit E, Hertz U, Amedi A. Extensive Cochleotopic Mapping of Human Auditory Cortical Fields Obtained with Phase-Encoding fMRI. PLoS ONE. 2011;6(3):e17832. pmid:21448274
  65. 65. Benjamini Y, Yekutieli D. The control of the false discovery rate in multiple testing under dependency. Annals of statistics. 2001:1165–88.
  66. 66. Seydell-Greenwald A, Chambers CE, Ferrara K, Newport EL. What you say versus how you say it: Comparing sentence comprehension and emotional prosody processing using fMRI. NeuroImage. 2020;209:116509. pmid:31899288
  67. 67. Martinelli A, Handjaras G, Betta M, Leo A, Cecchetti L, Pietrini P, et al. Auditory features modelling demonstrates sound envelope representation in striate cortex. bioRxiv. 2020:2020.04.15.043174.
  68. 68. Deneux T, Harrell ER, Kempf A, Ceballo S, Filipchuk A, Bathellier B. Context-dependent signaling of coincident auditory and visual events in primary visual cortex. Elife. 2019;8. Epub 2019/05/23. pmid:31115334; PubMed Central PMCID: PMC6544434.
  69. 69. Sturm W, Willmes K. On the functional neuroanatomy of intrinsic and phasic alertness. Neuroimage. 2001;14(1 Pt 2):S76–84. Epub 2001/05/25. pmid:11373136.
  70. 70. Westerhausen R, Moosmann M, Alho K, Belsby SO, Hämäläinen H, Medvedev S, et al. Identification of attention and cognitive control networks in a parametric auditory fMRI study. Neuropsychologia. 2010;48(7):2075–81. Epub 2010/04/07. pmid:20363236.
  71. 71. Corbetta M, Shulman GL. Spatial neglect and attention networks. Annu Rev Neurosci. 2011;34:569–99. Epub 2011/06/23. pmid:21692662; PubMed Central PMCID: PMC3790661.
  72. 72. Petersen SE, Posner MI. The attention system of the human brain: 20 years after. Annu Rev Neurosci. 2012;35:73–89. Epub 2012/04/25. pmid:22524787; PubMed Central PMCID: PMC3413263.
  73. 73. Lawrence SJD, van Mourik T, Kok P, Koopmans PJ, Norris DG, de Lange FP. Laminar Organization of Working Memory Signals in Human Visual Cortex. Current Biology. 2018;28(21):3435–40.e4. pmid:30344121
  74. 74. Knecht S, Deppe M, Dräger B, Bobe L, Lohmann H, Ringelstein E, et al. Language lateralization in healthy right-handers. Brain. 2000;123 (Pt 1):74–81. Epub 1999/12/28. pmid:10611122.
  75. 75. Szaflarski JP, Binder JR, Possing ET, McKiernan KA, Ward BD, Hammeke TA. Language lateralization in left-handed and ambidextrous people: fMRI data. Neurology. 2002;59(2):238–44. Epub 2002/07/24. pmid:12136064.
  76. 76. Bookheimer SY, Zeffiro TA, Blaxton TA, Gaillard WD, Malow B, Theodore WH. Regional cerebral blood flow during auditory responsive naming: evidence for cross-modality neural activation. Neuroreport. 1998;9(10):2409–13. pmid:9694237.
  77. 77. Gaillard WD, Sachs BC, Whitnah JR, Ahmad Z, Balsamo LM, Petrella JR, et al. Developmental aspects of language processing: fMRI of verbal fluency in children and adults. Human Brain Mapping. 2003;18(3):176–85. pmid:12599275
  78. 78. Binder JR, Liebenthal E, Possing ET, Medler DA, Ward BD. Neural correlates of sensory and decision processes in auditory object identification. Nature Neuroscience. 2004;7(3):295–301. pmid:14966525
  79. 79. Zekveld AA, Heslenfeld DJ, Festen JM, Schoonhoven R. Top–down and bottom–up processes in speech comprehension. NeuroImage. 2006;32(4):1826–36. pmid:16781167
  80. 80. Wolmetz M, Poeppel D, Rapp B. What does the right hemisphere know about phoneme categories? J Cogn Neurosci. 2011;23(3):552–69. Epub 2010/03/31. pmid:20350179; PubMed Central PMCID: PMC4471482.
  81. 81. Kovelman I, Norton ES, Christodoulou JA, Gaab N, Lieberman DA, Triantafyllou C, et al. Brain basis of phonological awareness for spoken language in children and its disruption in dyslexia. Cereb Cortex. 2012;22(4):754–64. Epub 2011/06/23. pmid:21693783; PubMed Central PMCID: PMC4498147.
  82. 82. Rämä P, Relander-Syrjänen K, Carlson S, Salonen O, Kujala T. Attention and semantic processing during speech: an fMRI study. Brain Lang. 2012;122(2):114–9. Epub 2012/06/08. pmid:22672735.
  83. 83. Bedny M, Thompson-Schill SL. Neuroanatomically separable effects of imageability and grammatical class during single-word comprehension. Brain and Language. 2006;98(2):127–39. pmid:16716387
  84. 84. Venezia JH, Richards VM, Hickok G. Speech-Driven Spectrotemporal Receptive Fields Beyond the Auditory Cortex. Hearing Research. 2021;408:108307. pmid:34311190
  85. 85. Watkins KE, Cowey A, Alexander I, Filippini N, Kennedy JM, Smith SM, et al. Language networks in anophthalmia: maintained hierarchy of processing in ’visual’ cortex. Brain. 2012. Epub 2012/03/20. [pii] pmid:22427328.
  86. 86. Röder B, Rösler F, Neville HJ. Event-related potentials during auditory language processing in congenitally blind and sighted people. Neuropsychologia. 2000;38(11):1482–502. Epub 2000/07/25. [pii]. pmid:10906374.
  87. 87. Kanjlia S, Pant R, Bedny M. Sensitive Period for Cognitive Repurposing of Human Visual Cortex. Cerebral Cortex. 2018:bhy280–bhy. pmid:30418533
  88. 88. O’Leary DS, Andreasen NC, Hurtig RR, Torres IJ, Flashman LA, Kesler ML, et al. Auditory and visual attention assessed with PET. Human Brain Mapping. 1997;5(6):422–36. pmid:20408245
  89. 89. Klein I, Paradis AL, Poline JB, Kosslyn SM, Le Bihan D. Transient activity in the human calcarine cortex during visual-mental imagery: an event-related fMRI study. J Cogn Neurosci. 2000;12 Suppl 2:15–23. pmid:11506644.
  90. 90. Shaywitz BA, Shaywitz SE, Pugh KR, Fulbright RK, Skudlarski P, Mencl WE, et al. The Functional Neural Architecture of Components of Attention in Language-Processing Tasks. NeuroImage. 2001;13(4):601–12. pmid:11305889
  91. 91. Cate AD, Herron TJ, Yund EW, Stecker GC, Rinne T, Kang X, et al. Auditory attention activates peripheral visual cortex. PLoS ONE. 2009;4(2):e4645. Epub 2009/02/28. pmid:19247451.
  92. 92. Brang D, Towle VL, Suzuki S, Hillyard SA, Di Tusa S, Dai Z, et al. Peripheral sounds rapidly activate visual cortex: evidence from electrocorticography. Journal of Neurophysiology. 2015;114(5):3023–8. pmid:26334017
  93. 93. Rockland KS, Ojima H. Multisensory convergence in calcarine visual areas in macaque monkey. Int J Psychophysiol. 2003;50(1–2):19–26. Epub 2003/09/27. [pii]. pmid:14511833.
  94. 94. Rockland KS, Van Hoesen GW. Direct temporal-occipital feedback connections to striate cortex (V1) in the macaque monkey. Cereb Cortex. 1994;4(3):300–13. pmid:8075534.
  95. 95. Falchier A, Clavagnier S, Barone P, Kennedy H. Anatomical evidence of multimodal integration in primate striate cortex. J Neurosci. 2002;22(13):5749–59. pmid:12097528.
  96. 96. Falchier A, Renaud L, Barone P, Kennedy H. Extensive projections from the primary auditory cortex and polisensory area STP to peripheral area V1 in the macaque. Society for Neuroscience Abstracts. 2001;31:511–21.
  97. 97. Borra E, Rockland KS. Projections to early visual areas v1 and v2 in the calcarine fissure from parietal association areas in the macaque. Frontiers in neuroanatomy. 2011;5:35–. pmid:21734867.
  98. 98. Majka P, Rosa MGP, Bai S, Chan JM, Huo B-X, Jermakow N, et al. Unidirectional monosynaptic connections from auditory areas to the primary visual cortex in the marmoset monkey. Brain structure & function. 2019;224(1):111–31. Epub 10/04. pmid:30288557.
  99. 99. Eckert MA, Kamdar NV, Chang CE, Beckmann CF, Greicius MD, Menon V. A cross-modal system linking primary auditory and visual cortices: evidence from intrinsic fMRI connectivity analysis. Hum Brain Mapp. 2008;29(7):848–57. Epub 2008/04/17. pmid:18412133.
  100. 100. Kosslyn SM, Pascual-Leone A, Felician O, Camposano S, Keenan JP, Thompson WL, et al. The role of area 17 in visual imagery: convergent evidence from PET and rTMS. Science. 1999;284(5411):167–70. pmid:10102821.
  101. 101. Kosslyn SM, Thompson WL, Kim IJ, Alpert NM. Topographical representations of mental images in primary visual cortex. Nature. 1995;378(6556):496–8. pmid:7477406
  102. 102. Chen W, Kato T, Zhu XH, Ogawa S, Tank DW, Ugurbil K. Human primary visual cortex and lateral geniculate nucleus activation during visual imagery. Neuroreport. 1998;9(16):3669–74. pmid:9858377.
  103. 103. Slotnick SD, Thompson WL, Kosslyn SM. Visual Mental Imagery Induces Retinotopically Organized Activation of Early Visual Areas. Cerebral Cortex. 2005;15(10):1570–83. pmid:15689519
  104. 104. Cichy RM, Heinzle J, Haynes JD. Imagery and perception share cortical representations of content and location. Cereb Cortex. 2012;22(2):372–80. Epub 2011/06/15. pmid:21666128.
  105. 105. Senden M, Emmerling TC, van Hoof R, Frost MA, Goebel R. Reconstructing imagined letters from early visual cortex reveals tight topographic correspondence between visual mental imagery and perception. Brain Structure and Function. 2019. pmid:30637491
  106. 106. Ragni F, Tucciarelli R, Andersson P, Lingnau A. Decoding stimulus identity in occipital, parietal and inferotemporal cortices during visual mental imagery. Cortex. 2020. pmid:32289581
  107. 107. Reddy L, Tsuchiya N, Serre T. Reading the mind’s eye: decoding category information during mental imagery. Neuroimage. 2010;50(2):818–25. Epub 2009/12/17. pmid:20004247; PubMed Central PMCID: PMC2823980.
  108. 108. Lee SH, Kravitz DJ, Baker CI. Disentangling visual imagery and perception of real-world objects. Neuroimage. 2012;59(4):4064–73. Epub 2011/11/02. pmid:22040738; PubMed Central PMCID: PMC3288657.
  109. 109. O’Craven KM, Kanwisher N. Mental imagery of faces and places activates corresponding stiimulus-specific brain regions. J Cogn Neurosci. 2000;12(6):1013–23. pmid:11177421.
  110. 110. Kosslyn SM, Thompson WL. When is early visual cortex activated during visual mental imagery? Psychol Bull. 2003;129(5):723–46. Epub 2003/09/06. pmid:12956541.
  111. 111. Dijkstra N, Bosch S, Gerven M. Shared Neural Mechanisms of Visual Perception and Imagery. Trends in Cognitive Sciences. 2019;23. pmid:30876729
  112. 112. Chandrasekaran C, Trubanova A, Stillittano S, Caplier A, Ghazanfar AA. The Natural Statistics of Audiovisual Speech. PLOS Computational Biology. 2009;5(7):e1000436. pmid:19609344
  113. 113. Felleman DJ, Van Essen DC. Distributed hierarchical processing in the primate cerebral cortex. Cereb Cortex. 1991;1(1):1–47. pmid:1822724.
  114. 114. Batardiere A, Barone P, Dehay C, Kennedy H. Area-specific laminar distribution of cortical feedback neurons projecting to cat area 17: quantitative analysis in the adult and during ontogeny. J Comp Neurol. 1998;396(4):493–510. pmid:9651007.
  115. 115. Clavagnier S, Falchier A, Kennedy H. Long-distance feedback projections to area V1: implications for multisensory integration, spatial awareness, and visual consciousness. Cogn Affect Behav Neurosci. 2004;4(2):117–26. pmid:15460918.
  116. 116. Hall AJ, Lomber SG. Auditory cortex projections target the peripheral field representation of primary visual cortex. Exp Brain Res. 2008;190(4):413–30. Epub 2008/07/22. pmid:18641978.
  117. 117. Muckli L, Petro LS. Network interactions: non-geniculate input to V1. Current Opinion in Neurobiology. 2013;23(2):195–201. pmid:23402951
  118. 118. Henschke J, Noesselt T, Scheich H, Budinger E. Possible anatomical pathways for short-latency multisensory integration processes in primary sensory cortices. Brain Structure and Function. 2014:1–23. pmid:24384580
  119. 119. Pennartz CMA, Dora S, Muckli L, Lorteije JAM. Towards a Unified View on Pathways and Functions of Neural Recurrent Processing. Trends in Neurosciences. 2019;42(9):589–603. pmid:31399289
  120. 120. Muckli L, De Martino F, Vizioli L, Petro Lucy S, Smith Fraser W, Ugurbil K, et al. Contextual Feedback to Superficial Layers of V1. Current Biology. 2015;25(20):2690–5. pmid:26441356
  121. 121. Schroeder CE, Foxe J. Multisensory contributions to low-level, ’unisensory’ processing. Curr Opin Neurobiol. 2005;15(4):454–8. Epub 2005/07/16. pmid:16019202.
  122. 122. Stănişor L, van der Togt C, Pennartz CMA, Roelfsema PR. A unified selection signal for attention and reward in primary visual cortex. Proceedings of the National Academy of Sciences. 2013;110(22):9136–41. pmid:23676276
  123. 123. Roth ZN, Ryoo M, Merriam EP. Task-related activity in human visual cortex. PLOS Biology. 2020;18(11):e3000921. pmid:33156829
  124. 124. Romei V, Gross J, Thut G. Sounds Reset Rhythms of Visual Cortex and Corresponding Human Visual Perception. Current Biology. 2012;22(9):807–13. pmid:22503499
  125. 125. Vetter P, Bola Ł, Reich L, Bennett M, Muckli L, Amedi A. Decoding Natural Sounds in Early "Visual" Cortex of Congenitally Blind Individuals. Current Biology. 2020. pmid:32559449
  126. 126. Battistoni E, Stein T, Peelen MV. Preparatory attention in visual cortex. Annals of the New York Academy of Sciences. 2017. pmid:28253445
  127. 127. Petro LS, Paton AT, Muckli L. Contextual modulation of primary visual cortex by auditory signals. Philosophical Transactions of the Royal Society B: Biological Sciences. 2017;372(1714). pmid:28044015
  128. 128. Hogendoorn H. Perception in real-time: predicting the present, reconstructing the past. Trends Cogn Sci. 2022;26(2):128–41. Epub 2022/01/03. pmid:34973925.
  129. 129. Störmer VS. Orienting spatial attention to sounds enhances visual processing. Current Opinion in Psychology. 2019;29:193–8. pmid:31022562
  130. 130. Abboud S, Cohen L. Distinctive Interaction Between Cognitive Networks and the Visual Cortex in Early Blind Individuals. Cereb Cortex. 2019;29(11):4725–42. Epub 2019/02/05. pmid:30715236.
  131. 131. Hamilton R, Keenan JP, Catala M, Pascual-Leone A. Alexia for Braille following bilateral occipital stroke in an early blind woman. Neuroreport. 2000;11(2):237–40. pmid:10674462.
  132. 132. Hull T, Mason H. Performance of blind-children on digit-span tests. Journal of Visual Impairment and Blindness. 1995;89(2):166–9.
  133. 133. Cohen LG, Celnik P, Pascual-Leone A, Corwell B, Falz L, Dambrosia J, et al. Functional relevance of cross-modal plasticity in blind humans. Nature. 1997;389(6647):180–3. pmid:9296495
  134. 134. Liu Y, Yu C, Liang M, Li J, Tian L, Zhou Y, et al. Whole brain functional connectivity in the early blind. Brain. 2007;130:2085–96. pmid:17533167.
  135. 135. Yu C, Liu Y, Li J, Zhou Y, Wang K, Tian L, et al. Altered functional connectivity of primary visual cortex in early blindness. Hum Brain Mapp. 2008;29(5):533–43. pmid:17525980.
  136. 136. Wang D, Qin W, Liu Y, Zhang Y, Jiang T, Yu C. Altered resting-state network connectivity in congenital blind. Human Brain Mapping. 2013;35(6):2573–81. pmid:24038713; PubMed Central PMCID: PMC6869085.
  137. 137. Burton H, Snyder AZ, Raichle ME. Resting state functional connectivity in early blind humans. Front Syst Neurosci. 2014;8:51. Epub 2014/04/30. pmid:24778608; PubMed Central PMCID: PMC3985019.
  138. 138. Qin W, Xuan Y, Liu Y, Jiang T, Yu C. Functional Connectivity Density in Congenitally and Late Blind Subjects. Cerebral Cortex. 2014. pmid:24642421
  139. 139. Nicolelis MA, Chapin JK, Lin RC. Neonatal whisker removal in rats stabilizes a transient projection from the auditory thalamus to the primary somatosensory cortex. Brain Res. 1991;567(1):133–9. Epub 1991/12/13. [pii]. pmid:1726139.
  140. 140. Karlen SJ, Kahn DM, Krubitzer L. Early blindness results in abnormal corticocortical and thalamocortical connections. Neuroscience. 2006;142(3):843–58. pmid:16934941.
  141. 141. Henschke JU, Oelschlegel AM, Angenstein F, Ohl FW, Goldschmidt J, Kanold PO, et al. Early sensory experience influences the development of multisensory thalamocortical and intracortical connections of primary sensory cortices. Brain Struct Funct. 2017. Epub 2017/11/03. pmid:29094306.
  142. 142. Magrou L, Barone P, Markov NT, Killackey H, Giroud P, Berland M, et al. Cortical Connectivity In A Macaque Model Of Congenital Blindness. bioRxiv. 2017.
  143. 143. Bedny M. Evidence from Blindness for a Cognitively Pluripotent Cortex. Trends in Cognitive Sciences. 2017;21(9):637–48. pmid:28821345
  144. 144. Pascual-Leone A, Hamilton R. The metamodal organization of the brain. Prog Brain Res. 2001;134:427–45. pmid:11702559.
  145. 145. Kupers R, Ptito M. Compensatory plasticity and cross-modal reorganization following early visual deprivation. Neurosci Biobehav Rev. 2013. Epub 2013/08/21. pmid:23954750.
  146. 146. Ricciardi E, Bonino D, Pellegrini S, Pietrini P. Mind the blind brain to understand the sighted one! Is there a supramodal cortical functional architecture? Neuroscience & Biobehavioral Reviews. 2013;(0). pmid:24157726
  147. 147. Bi Y, Wang X, Caramazza A. Object Domain and Modality in the Ventral Visual Pathway. Trends in Cognitive Sciences. 2016. pmid:26944219
  148. 148. Kanjlia S, Lane C, Feigenson L, Bedny M. Absence of visual experience modifies the neural basis of numerical thinking. Proceedings of the National Academy of Sciences. 2016. pmid:27638209
  149. 149. Crollen V, Lazzouni L, Bellemare A, Rezk M, Lepore F, Noel M-P, et al. Recruitment of occipital cortex by arithmetic processing follows computational bias in early blind. Neuroimage. 2019;186:549–56.
  150. 150. Saygin ZM, Osher DE, Koldewyn K, Reynolds G, Gabrieli JD, Saxe RR. Anatomical connectivity patterns predict face selectivity in the fusiform gyrus. Nat Neurosci. 2011;15(2):321–7. Epub 2011/12/27. pmid:22197830; PubMed Central PMCID: PMC3267901.