Figures
Abstract
The ability to selectively attend to one talker in the presence of competing talkers is crucial to communication in noisy, real-world environments. In this study, we investigated the relationship between intellectual ability and speech perception under multitalker conditions. Since neurodivergent individuals show a wide range of intellectual ability, from above average IQ to intellectual disability, intellectual ability may be an important individual characteristic that impacts multitalker speech perception, but this is not currently well understood. We tested individuals with autism, fetal alcohol spectrum disorder, and an age- and sex-matched comparison group, all with typical hearing. We found a strong positive correlation between IQ and multitalker speech perception thresholds. This demonstrates that deficits in intellectual ability, despite intact peripheral encoding of sound, are associated with difficulty listening under complex conditions for individuals with autism and fetal alcohol spectrum disorder. Future research is needed to investigate specific cognitive control mechanisms that contribute to difficulty listening under complex conditions. These findings suggest that audiological services to improve communication in real-world environments for neurodivergent individuals should be considered during clinical evaluations.
Citation: Lau BK, Emmons K, Maddox RK, Estes A, Dager SR, (Astley) Hemingway SJ, et al. (2025) The relationship between intellectual ability and auditory multitalker speech perception in neurodivergent individuals. PLoS One 20(9): e0329581. https://doi.org/10.1371/journal.pone.0329581
Editor: Ramandeep Kaur, Father Muller Charitable Institutions, INDIA
Received: January 31, 2025; Accepted: July 18, 2025; Published: September 24, 2025
Copyright: © 2025 Lau et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data supporting the findings of this study are publicly available in the Dryad Digital Repository at https://doi.org/10.5061/dryad.5mkkwh7hp.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Multitalker speech perception, the ability to selectively attend to one talker in the presence of several competing talkers, is a crucial yet often underappreciated skill employed in everyday life. To purchase a drink in a coffee shop with loud background noise, one must selectively tune-in to the barista’s voice while filtering out the voices of other customers. To learn in a classroom, a student needs to ignore the voices of other students to focus on the teacher’s voice. In real-world settings involving multiple speakers, neurotypical listeners with typical hearing can effortlessly attend to the desired talker even when competing voices are significantly louder [1]. The ability to organize and extract meaning from competing sounds in these complex acoustic environments is often referred to as auditory scene analysis or the cocktail party problem [2].
Successful auditory scene analysis relies first on the accurate encoding of the incoming sound mixture by the ear and auditory nerve fibers, followed by the integration of sounds between the ears [see 3 for conceptual model]. These spectro-temporal patterns encoded by the auditory periphery are then segregated and grouped into separate auditory streams based on acoustic features, such as pitch or spatial cues, that indicate where the sound is coming from [4–7]. From there, more complex processing of the lexical, syntactic, and semantic information in the speech signal occurs [8,9]. Importantly, cognitive processes, including long and short-term memory, as well as attentional control, are thought to be required throughout both auditory grouping and complex processing stages [10]. The interplay between these cognitive and auditory perceptual processes are often modelled as complex functions, with involvement of both feedback and feedforward mechanisms [3]. Moreover, there is evidence that the selective enhancement of relevant speech is influenced by an attention-mediated feedback loop [3,11–14]. In the presence of multiple simultaneous talkers, listeners must also selectively direct their attention to the speaker of interest [6,15]. Many aspects of the relationship between cognitive and auditory processes, including how they function during multitalker speech perception, are not fully understood. However, attention and memory play a critical role in all stages of speech perception under complex conditions [3].
Auditory processing differences, including hyper- and hypo-sensitivity to sound, aversions to sound, difficulty listening in noisy environments, are commonly reported by individuals with autism and fetal alcohol spectrum disorder [FASD; 16–19]. Autism is a developmental condition characterized by difficulties with social communication and interactions, as well as restricted interests and repetitive behaviors, according to the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition [20]. Autism is estimated to affect 1 in a 100 people globally [21] and 1 in 36 children in the United States [22]. FASD describes a range of physical, functional, and neurological outcomes associated with prenatal alcohol exposure. It is estimated to affect 8 in 1000 children globally [23]. On the most severe end of the spectrum, Fetal Alcohol Syndrome (FAS) is a permanent birth defect characterized by significant central nervous system (CNS) abnormalities, growth deficiency, and a unique cluster of three minor facial anomalies—small eyes, thin upper lip, and a flat philtrum [24]. The majority of children will not receive a diagnosis of FAS [25], but a diagnosis of FASD, which can be characterized by physical differences such as facial dysmorphias as well as cognitive challenges, such as deficits in working memory, learning, communication, and attention [24,26].
Difficulty understanding speech under noisy conditions is commonly reported by autistic individuals [e.g., 27,28]. While there are no prior studies measuring speech perception abilities in individuals with FASD, the studies conducted in autistic individuals report variable findings of whether speech perception deficits are indeed observed and if so, under what conditions [29–32]. Variability in the results of these studies may result from heterogeneity in the participants themselves. Since individuals with autism and FASD show a wide range of intellectual ability, from above average IQ to intellectual disability, intellectual ability may be an important individual characteristic that impacts multitalker speech perception, but this is not currently well understood for neurodivergent individuals.
Given that cognitive processes are known to be active in speech perception [33–35] and the results of prior studies show a relationship between intellectual ability and speech perception under complex conditions [e.g., 36–39], the goal of this study is to investigate whether deficits in intellectual ability is related to speech perception difficulty for neurodivergent individuals. We tested neurodivergent individuals with varying intellectual quotient (IQ) and with clinically normal hearing thresholds who belonged to one of three diagnostic groups: autism, FASD, and an age- and sex-matched comparison group. We use the term neurodivergent to refer to individuals who process, learn, or behave in ways that differ, but are not necessarily inferior to neurotypical individuals. The rationale for testing these two groups is that individuals with autism and FASD have a wide range of IQ, including individual with intellectual disability, and they commonly report difficulty perceiving speech under noisy conditions.
Methods
Participants
Forty-nine subjects ( Autism;
FASD,
Comparison) participated in the study (see Table 1 for demographic information). Twelve adults diagnosed with autism spectrum disorder (ASD) were recruited from a larger longitudinal study conducted at the University of Washington [40]. The original cohort consisted of 72 children diagnosed with autism between the ages of 3 and 4 years. Diagnoses of ASD, according to criteria from the fourth edition of the Diagnostic and Statistical Manual of Mental Disorders [41] were made by a licensed clinical psychologist or supervised graduate student using: 1) the Autism Diagnostic Interview-Revised [42], 2) the Autism Diagnostic Observation Schedule [ADOS; 43], 3) medical and family history, 4) cognitive test scores, and 5) clinical observation and judgment [see 40 for further details]. These participants were tested again at ages 6, 9, and 13–15 years. For this current study, forty-six participants from the original cohort were re-contacted and invited to participate in this current study. The remaining 26 were not contacted because they were already recruited for another study. Of the 46 participants contacted, 4 had moved out of state, 2 were not interested in participating, 1 could not be scheduled, 25 did not respond to phone calls or emails, and 2 did not meet our eligibility criteria of being able to speak in 3-word phrases. Thus, twelve autistic participants from the original cohort were enrolled in the study and tested at age 21–23 years of age (see Table 1 for sample demographics). This response rate was consistent with our expectations as the participants in this study were first recruited between the ages of 3 and 4 years and many families have moved, changed contact information, or were no longer interested in participating in research.
Ten participants with FASD were recruited from a university based FASD diagnostic clinic. All participants had a diagnosis of FASD as determined by the 4-Digit Diagnostic Code [24,26], which is an interdisciplinary approach to diagnosis guided by the magnitude of expression of the four features of FAS: growth deficiency, FAS facial phenotype, CNS structural/functional abnormalities, and prenatal alcohol exposure. Two participants had a diagnosis of fetal alcohol syndrome, two with a diagnosis of static encephalography/alcohol exposed, and six with a diagnosis of neurodevelopmental disorder/alcohol exposed (see Table 2 for speech perception thresholds and intellectual ability by FASD diagnoses). In terms of degree of dysfunction, “static encephalopathy, alcohol exposed (SE/AE)” can be used to indicate a significant CNS dysfunction in the context of prenatal alcohol exposure, and the term “neurodevelopmental disorder, alcohol exposed (ND/AE)” can be used to indicate a mild-to-moderate CNS dysfunction in the context of prenatal alcohol exposure [24]. Nine of ten FASD participants were between the ages of 13 and 22; however, one adult aged 47 was also included in the study. The initial recruitment goal was for 12 participants with FASD to match the Autism group sample size, however, we were only able to recruit 10 participants.
The Comparison group consisted of 27 participants. For each participant in the Autism and FASD groups, one age- and biological sex-matched participant was recruited as a pair-matched comparison subject. Age matching in comparison participants was ± 1 year except the 47-year-old participant in the FASD group was matched to a 45-year-old comparison subject. A one-way analysis of variance confirmed that there was no significant difference in age between groups (F(2,46) = 0.611, p = 0.547). Participants were also matched on biological sex due to the higher incidence of males with a diagnosis of autism. Besides the pair-matched comparison subjects, there were also an additional five Comparison group participants tested. These five additional participants were recruited for the magnetoencephalography component of the larger study. Comparison group participants all reported no history of cognitive, developmental, or other health concerns except attention-deficit/hyperactivity disorder (ADHD). As there is a high incidence of ADHD in individuals with FASD and autism, ADHD was not included in the exclusion criteria for any group. In our participant sample, 25% had parent-reported ADHD (Autism = 17%, Comparison = 8%, FASD = 80%). We found no evidence for a relationship between report of ADHD and performance on our multitalker speech perception task (further details in Discussion and Supporting Information). Additional exclusion criteria for all three participant groups included: a history of sensory or motor impairment such as hearing loss, traumatic brain injury, major physical anomalies, genetic disorders associated with ASD such as Fragile X, or other neurological impairments such as seizures.
Stimuli
The multitalker listening environment was constructed from three sentences spoken simultaneously. The sentence stimuli were from the Coordinate Response Measure (CRM) corpus [44], which consists of 256 sentences of the form “Ready (CALLSIGN) go to (COLOR) (NUMBER) now.” There are eight possible call signs (Arrow, Baron, Charlie, Eagle, Hopper, Laker, Ringo, Tiger), four colors (red, green, white, blue) and the numbers 1–8. Two male and two female speakers speaking all combinations of the call signs, colors, and numbers were included in the stimulus set. Using custom Python software, head-related transfer functions (HRTFs) from the CIPIC HRTF database [45], were resampled to 24,414 Hz to match the stimulus sampling rate then convolved with the target and masker sentences to simulate the source locations tested in this experiment: the target stream to be attended at 0º azimuth and the two spatially separated masking speech streams at ± 45º azimuth. The target talker was always male, while the two maskers were either male/male or female/female. Sentences were randomly selected on each presentation.
Procedures
The following measures were obtained over the course of several visits and as part of a larger study that included additional neurophysiological and behavioral measures. The number of visits for the larger study ranged from two to five, based on the preference of participants. In the present study, the Autism Diagnostic Observation Scale, Second Edition [ADOS-2; 46], and The Weschler Abbreviated Scale of Intelligence – Second Edition [WASI-II; 47] was administered on the same visit. For most participants, the audiological screening and the multitalker adaptive speech perception threshold task were obtained on a separate visit, but within the span of 4 weeks. Written informed consent was obtained from all participants or their Legally Authorized Representative for participants under 16 years of age. All methods were performed in accordance with the protocols approved by the Institutional Review Board at the University of Washington where the research was conducted. Participants were provided with monetary compensation for their time. An autistic community member was consulted during the preparation of this manuscript.
Autism symptoms and categorization
The ADOS-2 [43], a measure of autism symptom severity, was administered to all participants at the time of testing to confirm group inclusion. All participants in the Autism group received a classification of autism or autism spectrum based on their performance on the ADOS-2 while all participants in the FASD or Comparison received a classification of non-spectrum.
Intellectual ability
The Weschler Abbreviated Scale of Intelligence – Second Edition [47] was used as a measure of intellectual ability. The Block Design, Vocabulary, Matrix Reasoning, and Similarities subtests were administered to each participant to estimate their Full Scale IQ (FSIQ-4) or general intellectual ability. The Block Design subtest measures the ability to analyze and synthesize abstract visual stimuli. The Vocabulary subtest is designed to measure word knowledge and verbal concept formation. On the Matrix Reasoning subtest, the participants view an incomplete matrix or series and selects a response option that completes the matrix or series. This subtest measures fluid intelligence, broad visual intelligence, classification, spatial ability, knowledge of part-whole relationships, simultaneous processing and perceptual organization. Finally, the Similarities subtest measures verbal concept formation and reasoning. As a measure of verbal IQ, the Verbal Composite Index (VCI) score was computed from the Vocabulary and Similarities subtests and as a measure of nonverbal IQ, the Perceptual Reasoning Index (PRI) scores were computed from the Block Design and Matrix Reasoning subtests.
Audiological screening
Typical hearing was an inclusion criterion to participate in this study. To ensure clinically normal hearing thresholds, all participants were required to pass an audiometric screen ( dB hearing level at octave frequencies between 500 and 8000 Hz) as well as a distortion product otoacoustic emission (DPOAE) screen. For the DPOAE screening, DPOAE growth functions were estimated at the distortion frequency of 2f1-f2. The frequency of the f2 primary tone was 4 kHz and the frequency and level of the f1 tone were varied according to formula in Johnson et al. (2006) [48]. One participant with ASD was excluded from this study because they failed the audiometric screening.
Multitalker adaptive threshold task
To assess multitalker speech perception, participants listened to three simultaneous sentences and were asked to identify the color and number spoken by the target talker, who was always situated in front. The two competing talkers, who we refer to as “maskers,” were presented off to the side (Fig 1). Participants only heard the sentence stimuli (auditory only) and the spatial location of the talkers were simulated using head-related transfer functions. At the start of the task, the target talker is louder than the competing maskers which are presented at a much lower sound level. To increase task difficulty, we increased the sound level of the competing maskers adaptively to obtain a speech perception threshold for each participant.
A) Participants heard three simultaneous sentences, with the target talker in front at 0° azimuth presented with two spatially separated simultaneous competing talkers, “maskers”, on each side at ±45° azimuth. B) Participants were asked to identify via mouse click the color and number spoken by the target talker on a computer screen.
These experimental procedures closely follow those outlined in Gallun et al. (2013) [49], except that stimuli were presented via insert ear tips in both ears as opposed to in sound field. Participants listened to three simultaneous sentences with the target stream to be attended was always positioned in front at 0º azimuth and two spatially separated masking speech streams at ± 45º azimuth. The goal was to attend to one of the three CRM sentences, identified by the callsign “Charlie,” and correctly identify the color and number spoken by the target talker. Speech perception thresholds, in terms of target-to-masker ratios (TMRs), were estimated using a one-up-one-down procedure, to estimate 50% correct [50]; the target level was fixed at 40 dB SPL and the masker levels were adaptively varied. The starting level for the masker sentences was 40 dB and they were changed in level by 5 dB for the first three reversals and then 1 dB for six additional reversals. Thresholds were estimated to be the geometric mean of the last six reversals. TMR is expressed as the level difference (dB) between the target and the two competing talkers, where a positive TMR indicated that the target had to be louder than the competing talkers, and a negative TMR indicated that the target talker would be quieter than the competing talkers. Thus, TMR scores in the positive range reflect poor speech perception performance, meaning that the listener needed the target talker to be louder than the maskers to accurately perceive the target talker’s speech. Each participant completed five adaptive tracks; the worst run (i.e., highest threshold value) was dropped and the best four runs were averaged.
To ensure audibility of the target sentence and to demonstrate the ability to perform the task, participants had to obtain 8 out of 10 trials correct on a training run where the target sentences were presented in quiet before testing began. If participants were not able to obtain 80% correct in quiet after a maximum of three attempts, they were excluded from the study; no participants were excluded for failing training. Feedback was provided for correct and incorrect trials only during the training run(s).
All auditory stimuli were presented via Etymotic ER-2 insert earphones in a sound-attenuated booth. Participants initiated each trial and indicated the color and number keywords associated with the target callsign “Charlie” by clicking a virtual button on a grid of color and number targets. All participants completed the five runs in a single session of about 30 minutes.
Statistical analyses
To test the main hypothesis that deficits in intellectual ability are associated with poor speech perception thresholds, a simple linear regression was used to investigate whether WASI-II FSIQ-4 scores was significantly correlated with TMR. Three further analyses were conducted to assess the stability of this relationship. A simple linear regression was used to investigate whether intellectual ability was significantly correlated with speech perception thresholds within each group. Next, to investigate the relationship between verbal and non-verbal abilities, paired sample t-tests were first used to test for differences in average VCI and PRI scores within each diagnostic group. A simple linear regression was then used to investigate the relationship between verbal (VCI) or non-verbal ability (PCI) and speech perception. Finally, a simple linear regression was used to investigate the relationship between speech perception and performance on each individual WASI-II subtest. All tests were evaluated against at two-tailed p < 0.05 level of significance. Pair matching was conducted during the recruitment of the Comparison group participants to achieve a balance in age and sex between the study groups. This pair matching was not maintained in the statistical analyses because preliminary analyses conducted using pair matched linear regression with Autism and FASD groups and their respective Comparison groups revealed no effect of age; thus, all 27 Comparison participants were combined into one group. The total sample size of 49 was sufficient to detect the relationship between speech perception and intellectual ability using a one-sided test with a significance level of 0.025 to test whether the observed slope of −0.68 is greater than zero with a power (1-β) of > 0.99.
Results
Multitalker speech perception as a function of IQ
Multitalker speech perception thresholds across groups are shown in in Fig 2 and speech perception thresholds as a function of intellectual ability are shown in Fig 3. Lower WASI-II FSIQ-4 scores were related to higher speech perception thresholds in the overall sample (Fig 3A; linear regression, As intellectual ability decreased, speech perception performance (the ability of the listener to selectively attend to one talker in the presence of maskers) also decreased. We then investigated whether the relationship between intellectual ability and speech perception thresholds held within each diagnostic group. Indeed, intellectual ability (FSIQ-4 score) was significantly correlated with speech perception thresholds in all three groups (Fig 3B-D; linear regression, Autism, n = 12:
Comparison, n = 27:
; FASD, n = 10:
). However, the strength of the correlation is lower for the Comparison group than for the Autism and FASD groups.
The Autism group is in blue (left), the Comparison group is in red (middle), and the FASD group is in orange (right).
A) Individual speech perception thresholds (TMR at 50% correct) as a function of intellectual ability (WASI-II FSIQ-4), with Autism group in blue, Comparison group in red, and FASD group in orange. Negative TMRs reflect better speech perception performance. Intellectual ability was correlated with TMR in the overall sample. B-D) Speech perception thresholds as a function of intellectual ability for the Autism (B), Comparison (C), and FASD (D) groups. Individual data points shown with solid circles. FSIQ-4 was correlated with TMR in Autism, Comparison, and FASD groups.
During secondary analyses, the presence of influential outliers was assessed using Cook’s distance. Three influential outliers were identified and thus, a clean model with the outliers removed was conducted. Comparing the coefficients between the original versus clean model revealed that removing influential observations did not meaningfully alter the model. Both the slope (−0.182 to −0.181) and intercept (19.41 to 19.39) changed only slightly, suggesting that the findings of the original model are robust. An additional multivariate linear regression model was also conducted to control for biological sex, age, and participant group. While FSIQ-4 score remained a significant predictor of speech perception thresholds (multiple linear regression, , age, sex, and group were not statistically significant (p < 0.05 for all). These results suggest that in our study sample, age, sex, and participant group did not have a significant effect on the relationship between IQ and multitalker speech perception.
Verbal and non-verbal abilities
One possibility for these findings is that verbal ability is driving the relationship between the WASI-II FSIQ-4 scores and speech perception thresholds. To further investigate, we tested the relationship between verbal and non-verbal abilities with the Verbal Comprehension Index score (VCI) and the Perceptual Reasoning Index score (PRI) of the WASI-II. We conducted paired-samples t-tests to compare average VCI and PRI scores within each diagnostic group. We did not observe a difference between the two scores in any of the three groups (Autism paired Furthermore, both WASI-II VCI (Fig 4A, linear regression,
) and WASI-II PRI were correlated with speech perception performance in the overall sample (Fig 4B, linear regression,
, with comparable strength and effect size. The higher the VCI and PRI scores, the better the speech perception performance, showing that the relationship holds for both verbal and nonverbal abilities; thus, ruling out the possibility that verbal ability alone drove the relationship.
Individual data points shown with solid circles. A) Verbal comprehension as a function of multitalker speech perception performance. WASI-II VCI was significantly correlated with TMR. B) Perceptual reasoning ability as a function of multitalker speech perception performance. WASI-II PRI was significantly correlated with TMR.
Individual subtests
Each WASI-II subtest measures different intellectual abilities, so it also remains possible that multitalker speech perception is related to the specific intellectual abilities measured in some subtests but not others. To further investigate, we tested the relationship between speech perception thresholds and performance on each subtest of the WASI-II for the overall sample. Mean WASI-II subtest T scores for each participant group are shown in Table 3. On average, both Autism and FASD groups had lower T scores than the Comparison group for all four subtests. However, the FASD group showed more variability in performance in all four subtests as well, while the Autism group showed more variability than the Comparison group on the Similarities and Vocabulary subtests. We found a significant correlation between the speech perception thresholds and each of the four subtests: Block Design (Fig 5A, linear regression, ), Similarities (Fig 5B, linear regression,
), Matrix Design (Fig 5C, linear regression,
), and Vocabulary (Fig 5D, linear regression,
), with comparable strength and effect size for each subtest. In all cases, the higher the scores, the better the speech perception performance, showing that the relationship holds also for each individual subtest.
Individual data points shown with solid circles. A) Block Design subtest T Score as a function of TMR. B) Similarities subtest T Score as a function of TMR. C) Matrix reasoning subtest T Score as a function of TMR. D) Vocabulary subtest T Score as a function of TMR.
Discussion
We observed an association between intellectual ability and multitalker speech perception thresholds; as intellectual ability increased, multitalker speech perception thresholds improved. This relationship was not driven by verbal abilities alone as the correlation between speech perception performance held for both the VCI and the PRI. Moreover, the correlation between multitalker speech perception thresholds was also observed for each individual WASI-II subtest, suggesting that performance on the speech perception task is indeed related to general intelligence. Although we did not have a sufficient sample size to statistically test group differences on average, the Autism and FASD groups required the target speaker to be louder than the two competing talkers, as reflected by positive target-to-masker ratios. In contrast, on average, individuals in the Comparison group were able to selectively attend to the target talker even when competing voices were louder, although some individuals did have positive target-to-masker ratios. These results are consistent with prior studies showing worse speech perception in autistic adults and children [29,30,32]. They are also consistent with self/caregiver reports of auditory processing differences including difficulty listening to speech under noisy conditions in individuals with FASD and Autism [18,51–55]. To our knowledge, this is the first study to obtain multitalker speech perception thresholds in individuals with FASD.
A strength of this study is that we considered several characteristics of individuals with Autism and FASD in designing our study and analyses. First, we conducted an audiometric screen at octave frequencies between 250–8000 Hz (≤20 dB Hearing Level), as well as an otoacoustic emissions screen, to ensure clinically normal hearing thresholds in all participants, as there is an increased incidence of hearing loss in both populations [18,56–58]. Second, we chose a receptive speech perception task in which participants did not require verbal responses, as our intent was to assess speech perception rather than speech production. Third, we chose an adaptive threshold task, so that task difficulty would be varied systematically based on individual performance. Importantly, an adaptive threshold task design ensured that participants were able to perform the speech perception task when conditions were favorable (i.e., target talker is louder than the distractors).
The relationship between intellectual abilities and multitalker speech perception was observed in all participant groups, including the Comparison group. Listening in noisy environments is known to be difficult for individuals who are hard of hearing, with hearing assistive technology and auditory rehabilitation therapy often employed to address this challenge. However, this study suggests that individuals with typical hearing but lower intellectual ability may also have trouble listening under complex acoustic conditions. This suggests that those children who are intellectually challenged and the furthest behind their peers will have the most need to learn in noisy classrooms but, may also have the most trouble extracting information in complex listening environments. Consequently, future studies are needed to develop targeted assessment and intervention strategies for speech perception difficulties among neurodivergent individuals living in noisy real-world conditions.
The WASI-II is the gold-standard measure of general intellectual ability; however, it does not provide information on the specific cognitive mechanisms involved in multitalker speech perception. We ruled out the possibility that verbal abilities are driving the relationship and also showed that speech perception is related to each of the individual subtests, suggesting that it is in fact general intellectual ability that is related to the multitalker speech perception task employed in this study. While we did see the relationship to each individual subtest, future research is needed to determine why. For example, how are the abilities measured by matrix reasoning related to complex speech perception? Future research also needs to employ distinct cognitive tasks to study the mechanisms that play a role in complex listening as this current study was not designed to identify specific cognitive mechanisms.
In this study, we demonstrate that general intellectual ability is related to multitalker speech perception. However, we previously reported no effect of intellectual ability on stream segregation and selective attention in autism or a comparison group [59]. One possible explanation for these divergent findings is the difference in cognitive load between the tasks [60]. In Emmons et al. we presented a dual stream selective attention task with two talkers who each spoke two words. In this current paradigm, there are three simultaneous talkers speaking full sentences, that requires greater listening effort. For individuals with higher intellectual ability, this increased cognitive load may not have a discernable effect on complex listening. For individuals with lower ability, however, the effort associated with segregating and selectively listening in a multitalker situation may be sufficient to reduce their ability to remember a color and a number that was spoken by the target speaker. This may be especially true for individuals with FASD, who often experience deficits in working memory. Moreover, this conjecture is consistent with the fact that all participants in this study, regardless of ability, were able to perform the multitalker speech perception task with positive TMRs when the target talker was louder than the competing talkers.
There are several limitations to consider when interpretating findings from the present study. First, FASD and ASD are both heterogenous populations and the sample size in this study is small [61]. Future studies with larger cohorts will be required to further investigate multitalker speech perception differences between FASD, Autism, and Comparison groups, while controlling for IQ and to characterize individual variability. Second, the primary goal of this study was to investigate whether difficulty with multitalker speech perception is observed with deficits in intellectual ability, in the absence of hearing loss. The challenge with this study design was to determine participant groups that did not have hearing loss but did have variability in intellectual ability. Thus, we recruited from two groups that met these criteria: autism and FASD, along with a matched comparison group. While we had strict exclusion criteria to rule out most co-morbid health and neurological concerns, with the high incidence of attention deficit in FASD and autism, it was not possible for ADHD to be an exclusion criterion in this participant sample. Thus, we obtained a medical history from each participant or their parent/guardian prior to participation and conducted an additional analysis on whether ADHD status affected their performance on our multitalker speech perception task. We found no evidence for a relationship between report of ADHD and performance on our task (S1 Fig; see Supporting Information for details of the analysis). Nonetheless, an important limitation to note in this study is that for the participants that reported ADHD, we cannot determine the contributions of attention deficit to task performance, although at group level, no statistically significant difference was observed. Moreover, while an ADHD diagnosis in combination with autism or FASD did not affect performance on this particular speech perception task, it is unlikely the case for all complex speech perception scenarios. In fact, whether ADHD impairs complex speech perception is still in debate, with conflicting findings from different studies, with some reporting worse speech perception in individuals with ADHD but not others [62–64]. Future studies designed to assess both attention and complex speech perception are required to evaluate how attention deficits in addition to autism or FASD could impact multitalker speech perception. The critical point to note is that regardless of ADHD status, all participants were screened for clinically normal-hearing thresholds, and thus, our results demonstrate that the multitalker speech perception difficulty observed was not caused by peripheral deficits in sound encoding and had a strong association with intellectual ability. These results provide novel evidence that listening in complex situations may be impaired in individuals with lower IQ, with or without ADHD.
This study attempted to create a real-world listening scenario using full sentence stimuli and multiple speakers talking at the same time. While we simulated a realistic multitalker situation in the laboratory, there are many possible differences that may arise in situ, in the face of a multisensory environment. Future studies should consider how neurodiverse individuals listen in the context of non-human sounds alongside multiple talkers, as well as a range of multisensory input such as smells, lighting, and a visually crowded environment. Another aspect of multitalker speech perception is the communication demands on the listener. For example, a student in a lecture may be focused on listening; however, in a seminar, they would not only be expected to listen but also to participate by responding. Speech perception in situ might also be affected by both speaker behavior and expectations, guided by eye contact, nodding, and even verbal manipulation of attention. Besides these intricate dynamics of social interaction, speaker interest and motivation in the conversation topic as well as the mode of communication also contribute to speech perception in the real world.
In conclusion, in this investigation of multitalker speech perception in neurodivergent individuals with typical hearing, we found a highly significant relationship between directly assessed intellectual ability and multitalker speech perception. This relationship was observed within each diagnostic group separately, suggesting this effect may cross diagnostic boundaries. This is one of the first studies to investigate multitalker speech perception in individuals with FASD. From a clinical perspective, these findings suggest that assessment of complex listening ability and support for listening under complex, real-world environments may improve outcomes for neurodivergent individuals with lower cognitive ability. Finally, the findings raise the question of whether there could be shared etiology between multitalker speech perception and intellectual ability, genetic or neurologic, versus a causal relationship between the two.
Supporting information
S1 Fig. Multitalker speech perception thresholds as a function of ADHD status.
https://doi.org/10.1371/journal.pone.0329581.s001
(TIF)
Acknowledgments
We thank Dr. Erick Gallun for helpful discussions relating to this project and Dr. Lucas Harrington for consultation on using neurodiversity affirming language in this manuscript.
References
- 1. Brungart DS. Informational and energetic masking effects in the perception of two simultaneous talkers. J Acoust Soc Am. 2001;109(3):1101–9. pmid:11303924
- 2.
Bregman AS. Auditory scene analysis: the perceptual organization of sound. Cambridge, Mass: MIT Press. 1990.
- 3. Bronkhorst AW. The cocktail-party problem revisited: early processing and selection of multi-talker speech. Atten Percept Psychophys. 2015;77(5):1465–87. pmid:25828463
- 4. Best V, Gallun FJ, Ihlefeld A, Shinn-Cunningham BG. The influence of spatial separation on divided listening. J Acoust Soc Am. 2006;120(3):1506–16. pmid:17004472
- 5. Bizley JK, Cohen YE. The what, where and how of auditory-object perception. Nat Rev Neurosci. 2013;14(10):693–707. pmid:24052177
- 6. Shinn-Cunningham B, Best V, Lee AKC. Auditory object formation and selection. In: Middlebrooks JC, Simon JZ, Popper AN, Fay RR, editors. The auditory system at the Cocktail Party. Cham: Springer International Publishing; 2017. pp. 7–40.
- 7. Bronkhorst A. The cocktail party phenomenon: a review of research on speech intelligibility in multiple-talker conditions. Acta Acust United Acust. 2000;86:117–28.
- 8. Davis MH, Johnsrude IS. Hearing speech sounds: top-down influences on the interface between audition and speech perception. Hear Res. 2007;229(1–2):132–47. pmid:17317056
- 9. Mattys SL, Davis MH, Bradlow AR, Scott SK. Speech recognition in adverse conditions: a review. Lang Cogn Process. 2012;27(7–8):953–78.
- 10. Michalka SW, Kong L, Rosen ML, Shinn-Cunningham BG, Somers DC. Short-term memory for space and time flexibly recruit complementary sensory-biased frontal lobe attention networks. Neuron. 2015;87(4):882–92. pmid:26291168
- 11. Ahveninen J, Hämäläinen M, Jääskeläinen IP, Ahlfors SP, Huang S, Lin F-H, et al. Attention-driven auditory cortex short-term plasticity helps segregate relevant sounds from noise. Proc Natl Acad Sci U S A. 2011;108(10):4182–7. pmid:21368107
- 12. Ding N, Simon JZ. Emergence of neural encoding of auditory objects while listening to competing speakers. Proc Natl Acad Sci U S A. 2012;109(29):11854–9. pmid:22753470
- 13. O’Sullivan JA, Power AJ, Mesgarani N, Rajaram S, Foxe JJ, Shinn-Cunningham BG, et al. Attentional selection in a Cocktail Party environment can be decoded from single-trial EEG. Cereb Cortex. 2015;25(7):1697–706. pmid:24429136
- 14. Woldorff MG, Hillyard SA. Modulation of early auditory processing during selective listening to rapidly presented tones. Electroencephalogr Clin Neurophysiol. 1991;79(3):170–91. pmid:1714809
- 15. Lee AKC, Larson E, Maddox RK, Shinn-Cunningham BG. Using neuroimaging to understand the cortical mechanisms of auditory selective attention. Hear Res. 2014;307:111–20. pmid:23850664
- 16. Haesen B, Boets B, Wagemans J. A review of behavioural and electrophysiological studies on auditory processing and speech perception in autism spectrum disorders. Res Autism Spectr Disord. 2011;5(2):701–14.
- 17. Lau BK, Maddox RK, Estes A, Dager S, Lee AKC. Combining clinical, behavioral, and neurophysiological measures to investigate auditory processing abnormalities in individuals with autism spectrum disorder. Proc Meet Acoust. 2018;35(1):050004. pmid:35949247
- 18. McLaughlin SA, Thorne JC, Jirikowic T, Waddington T, Lee AKC, Astley Hemingway SJ. Listening difficulties in children with fetal alcohol spectrum disorders: more than a problem of audibility. J Speech Lang Hear Res. 2019;62(5):1532–48. pmid:31039324
- 19. O’Connor K. Auditory processing in autism spectrum disorder: a review. Neurosci Biobehav Rev. 2012;36(2):836–54. pmid:22155284
- 20.
American Psychiatric Association, editor. Diagnostic and statistical manual of mental disorders: DSM-5. Washington, D.C: American Psychiatric Association. 2013.
- 21. Zeidan J, Fombonne E, Scorah J, Ibrahim A, Durkin MS, Saxena S, et al. Global prevalence of autism: a systematic review update. Autism Res. 2022;15(5):778–90. pmid:35238171
- 22. Maenner MJ, Warren Z, Williams AR, Amoakohene E, Bakian AV, Bilder DA, et al. Prevalence and characteristics of autism spectrum disorder among children aged 8 years - autism and developmental disabilities monitoring network, 11 Sites, United States, 2020. MMWR Surveill Summ. 2023;72(2):1–14. pmid:36952288
- 23. Lange S, Probst C, Gmel G, Rehm J, Burd L, Popova S. Global prevalence of fetal alcohol spectrum disorder among children and youth: a systematic review and meta-analysis. JAMA Pediatr. 2017;171(10):948–56. pmid:28828483
- 24.
(Astley) Hemingway SJ. Diagnostic guide for fetal alcohol spectrum disorders. 4th Edition, v.1.01. Seattle, WA: University of Washington. 2024.
- 25. Astley SJ. Profile of the first 1,400 patients receiving diagnostic evaluations for fetal alcohol spectrum disorder at the Washington State Fetal Alcohol Syndrome Diagnostic & Prevention Network. Can J Clin Pharmacol. 2010;17(1):e132-e164.
- 26. Astley S. Validation of the fetal alcohol spectrum disorder (FASD) 4-digit diagnostic code. J Popul Ther Clin Pharmacol. 2013;20.
- 27.
Birch J. Congratulations! It’s asperger syndrome. Jessica Kingsley Publishers. 2003.
- 28.
Grandin T. Thinking in pictures: And other reports from my life with autism. Vintage. 2006.
- 29. Alcántara JI, Weisblatt EJL, Moore BCJ, Bolton PF. Speech-in-noise perception in high-functioning individuals with autism or Asperger’s syndrome. J Child Psychol Psychiatry. 2004;45(6):1107–14. pmid:15257667
- 30. DePape A-MR, Hall GBC, Tillmann B, Trainor LJ. Auditory processing in high-functioning adolescents with Autism Spectrum Disorder. PLoS One. 2012;7(9):e44084. pmid:22984462
- 31. Groen WB, van Orsouw L, Huurne N ter, Swinkels S, van der Gaag R-J, Buitelaar JK, et al. Intact spectral but abnormal temporal processing of auditory stimuli in autism. J Autism Dev Disord. 2009;39(5):742–50. pmid:19148738
- 32. Schelinski S, von Kriegstein K. Brief report: speech-in-noise recognition and the relation to vocal pitch perception in adults with autism spectrum disorder and typical development. J Autism Dev Disord. 2020;50(1):356–63. pmid:31583624
- 33. Heald SLM, Nusbaum HC. Speech perception as an active cognitive process. Front Syst Neurosci. 2014;8:35. pmid:24672438
- 34. Holt LL, Lotto AJ. Speech perception within an auditory cognitive science framework. Curr Dir Psychol Sci. 2008;17(1):42–6. pmid:19060961
- 35.
McQueen JM, Cutler A. Cognitive processes in speech perception. 2010.
- 36. Gallun FJ, Best V. Age-related changes in segregation of sound sources. In: Helfer KS, Bartlett EL, Popper AN, Fay RR, editors. Aging and hearing: causes and consequences. Cham: Springer International Publishing; 2020. pp. 143–171.
- 37. Humes LE. The contributions of audibility and cognitive factors to the benefit provided by amplified speech to older adults. J Am Acad Audiol. 2007;18(7):590–603. pmid:18236646
- 38. McCreery RW, Walker EA, Spratford M, Lewis D, Brennan M. Auditory, cognitive, and linguistic factors predict speech recognition in adverse listening conditions for children with hearing loss. Front Neurosci. 2019;13:1093. pmid:31680828
- 39. Taljaard DS, Olaithe M, Brennan-Jones CG, Eikelboom RH, Bucks RS. The relationship between hearing impairment and cognitive function: a meta-analysis in adults. Clin Otolaryngol. 2016;41(6):718–29. pmid:26670203
- 40. Dawson G, Toth K, Abbott R, Osterling J, Munson J, Estes A, et al. Early social attention impairments in autism: social orienting, joint attention, and attention to distress. Dev Psychol. 2004;40(2):271–83. pmid:14979766
- 41. American Psychiatric Association. Diagnostic and statistical manual of mental disorders, Fourth Edition, Text Revision (DSM-IV-TR). 4th ed. Arlington, VA: American Psychiatric Association. 2000.
- 42. Lord C, Rutter M, Le Couteur A. Autism Diagnostic Interview-Revised: a revised version of a diagnostic interview for caregivers of individuals with possible pervasive developmental disorders. J Autism Dev Disord. 1994;24(5):659–85. pmid:7814313
- 43. Lord C, Rutter M, Goode S, Heemsbergen J, Jordan H, Mawhood L, et al. Autism diagnostic observation schedule: a standardized observation of communicative and social behavior. J Autism Dev Disord. 1989;19(2):185–212. pmid:2745388
- 44. Bolia RS, Nelson WT, Ericson MA, Simpson BD. A speech corpus for multitalker communications research. J Acoust Soc Am. 2000;107(2):1065–6. pmid:10687719
- 45. Algazi VR, Duda RO, Thompson DM, Avendano C. The CIPIC HRTF database. Proceedings of the 2001 IEEE Workshop on the Applications of Signal Processing to Audio and Acoustics (Cat No01TH8575). New Platz, NY, USA: IEEE; 2001. pp. 99–102.
- 46.
Lord C, Rutter M, DiLavore PC, Risi S, Gotham K, Bishop SL. Autism diagnostic observation schedule, 2nd. Torrance, CA: Western Psychological Services. 2012.
- 47.
Weschler D. Weschler abbreviated scale of intelligence, second edition (WASI-II). San Antonio, TX: NCS Pearson. 2011.
- 48. Johnson TA, Neely ST, Kopun JG. Reducing reflected contributions to ear-canal distortion product otoacoustic emissions in humans. J Acoust Soc Am. 2006;119(6):3896–07. https://doi.org/10.1121/1.2200048
- 49. Gallun FJ, Diedesch AC, Kampel SD, Jakien KM. Independent impacts of age and hearing loss on spatial release in a complex auditory environment. Front Neurosci. 2013;7:252. https://doi.org/10.3389/fnins.2013.00252
- 50. Levitt, H. Transformed up‐down methods in psychoacoustics. J Acoust Soc Am. 1971;49(2B):467–77.
- 51. Ben-Sasson A, Hen L, Fluss R, Cermak SA, Engel-Yeger B, Gal E. A meta-analysis of sensory modulation symptoms in individuals with autism spectrum disorders. J Autism Dev Disord. 2009;39(1):1–11. pmid:18512135
- 52. Church MW, Gerkin KP. Hearing disorders in children with fetal alcohol syndrome: findings from case reports. Pediatrics. 1988;82(2):147–54.
- 53. Franklin L, Deitz J, Jirikowic T, Astley S. Children with fetal alcohol spectrum disorders: problem behaviors and sensory processing. Am J Occup Ther. 2008;62(3):265–73. pmid:18557002
- 54. Gabriels RL, Agnew JA, Miller LJ, Gralla J, Pan Z, Goldson E, et al. Is there a relationship between restricted, repetitive, stereotyped behaviors and interests and abnormal sensory response in children with autism spectrum disorders? Res Autism Spectr Disord. 2008;2(4):660–70.
- 55. Green D, Chandler S, Charman T, Simonoff E, Baird G. Brief report: DSM-5 sensory behaviours in children with and without an autism spectrum disorder. J Autism Dev Disord. 2016;46(11):3597–606. pmid:27475418
- 56. Klin A. Auditory brainstem responses in autism: brainstem dysfunction or peripheral hearing loss?. J Autism Dev Disord. 1993;23(1):15–35. pmid:8463195
- 57. Rosenhall U, Nordin V, Sandström M, Ahlsén G, Gillberg C. Autism and hearing loss. J Autism Dev Disord. 1999;29(5):349–57. pmid:10587881
- 58. Szymanski CA, Brice PJ, Lam KH, Hotto SA. Deaf children with autism spectrum disorders. J Autism Dev Disord. 2012;42(10):2027–37. pmid:22290585
- 59. Emmons KA, Kc Lee A, Estes A, Dager S, Larson E, McCloy DR, et al. Auditory attention deployment in young adults with autism spectrum disorder. J Autism Dev Disord. 2022;52(4):1752–61. pmid:34013478
- 60. Peelle JE. Listening effort: how the cognitive consequences of acoustic challenge are reflected in brain and behavior. Ear Hear. 2018;39(2):204–14. pmid:28938250
- 61. Schönbrodt FD, Perugini M. At what sample size do correlations stabilize? J Res Personal. 2013;47(5):609–12.
- 62. Michalek AMP, Watson SM, Ash I, Ringleb S, Raymer A. Effects of noise and audiovisual cues on speech processing in adults with and without ADHD. Int J Audiol. 2014;53(3):145–52. pmid:24456181
- 63. Pillsbury HC, Grose JH, Coleman WL, Conners CK, Hall JW. Binaural function in children with attention-deficit hyperactivity disorder. Arch Otolaryngol Head Neck Surg. 1995;121(12):1345–50. pmid:7488361
- 64. Söderlund GBW, Jobs EN. Differences in speech recognition between children with attention deficits and typically developed children disappear when exposed to 65 dB of auditory noise. Front Psychol. 2016;7:34. pmid:26858679