Figures
Abstract
Building on research demonstrating the benefits of music training for emotional prosody recognition in nontonal languages, this study delves into its unexplored influence on tonal languages. In tonal languages, the acoustic similarity between lexical tones and music, along with the dual role of pitch in conveying lexical and affective meanings, create a unique interplay. We evaluated 72 participants, half of whom had extensive instrumental music training, with the other half serving as demographically matched controls. All participants completed an online test consisting of 210 Chinese pseudosentences, each designed to express one of five emotions: happiness, sadness, fear, anger, or neutrality. Our robust statistical analyses, which included effect size estimates and Bayesian factors, revealed that music and nonmusic groups exhibit similar abilities in identifying the emotional prosody of various emotions. However, the music group attributed higher intensity ratings to emotional prosodies of happiness, fear, and anger compared to the nonmusic group. These findings suggest that while instrumental music training is not related to emotional prosody recognition, it does appear to be related to perceived emotional intensity. This dissociation between emotion recognition and intensity evaluation adds a new piece to the puzzle of the complex relationship between music training and emotion perception in tonal languages.
Citation: Liu M, Teng X, Jiang J (2024) Instrumental music training relates to intensity assessment but not emotional prosody recognition in Mandarin. PLoS ONE 19(8): e0309432. https://doi.org/10.1371/journal.pone.0309432
Editor: Junchen Shang, Southeast University - Jiulonghu Campus, CHINA
Received: March 26, 2024; Accepted: August 12, 2024; Published: August 30, 2024
Copyright: © 2024 Liu et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The raw stimuli, procedure, data, and code are available from the Harvard Dataverse Repository at https://doi.org/10.7910/DVN/EEI1QR.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Accurate recognition of emotional prosody is crucial for effective social interactions [1, 2]. Individuals adept at identifying others’ emotional tones are better prepared for nuanced social engagement [3–6]. In contrast, those with challenges in understanding emotional prosody often face social complications [e.g., 7–12]. It is crucial to examine how individuals from varied linguistic backgrounds perceive emotional prosody. Exploring this, particularly the impact of specific training on social well-being through improved prosody recognition, unveils insights into cross-cultural variations in the utilization of acoustic cues, especially among individuals with and without tonal language experience [13, 14].
Recognizing emotional prosody hinges on acoustic features like pitch, intensity, tempo, and voice quality/timbre [15–17], which also serve as essential elements in music, such as Western tonal music. Different emotional prosodies are characterized by distinct features: for instance, happy prosody involves moderately high average pitch, high average voice intensity, and a rapid speech rate, whereas sad prosody is marked by low average pitch, low average voice intensity, and a slow speech rate [18–21]. Effective recognition of emotional prosody requires the efficient extraction of these acoustic markers and appropriate perceptual analysis. Given the shared acoustic elements between music and speech in conveying emotion [18, 22–24], instrumental music training could enhance speech perception [25], as posited by Patel’s OPERA hypothesis [26, 27]. The OPERA hypothesis posits that shared sensory and cognitive mechanisms underlie both music and speech perception [28, 29], with music imposing more demanding conditions on these systems. Therefore, training in one domain (e.g., music listening) could have cross-modal benefits for the other (e.g., prosody recognition).
Furthermore, individuals with musical training and tonal languages listeners may exhibit advantages in perceiving lexical tones and musical pitch by paying greater attention to the relevant sound cues acquired in speech or music [13, 29]. Tonal languages listeners and musicians have an advantage in recognizing and learning lexical tones, exhibiting cross-domain transfer in pitch perception [13, 30]. Pitch plays a crucial role in both speech and music, serving as a predictive factor in the perception of emotional aspects of vocals. In nontonal languages, it is primarily utilized for emotional expression in speech. In contrast, tonal languages use pitch variations not only as emotional markers but also to convey semantic information [31–36]. As a result, the task of recognizing emotional prosody in tonal languages becomes more intricate, potentially requiring additional cognitive resources to mitigate interference from semantic aspects during emotion recognition. While music training is associated with enhanced perception of lexical tones in native speakers of tonal languages [37], it remains uncertain whether it is associated with improved perception of emotional prosody in their native language.
Many studies have examined the relationship between music training and emotional prosody recognition in native speakers of nontonal languages, such as English, German, Portuguese, Dutch, and Hebrew [38–41] (see [42–44], for reviews). Most of these studies utilize a cross-sectional design comparing musically trained and untrained individuals, while a few use a randomized experimental design, often randomly assigning musically untrained individuals to experiment and control groups (see S1 Table for an overview). Cross-sectional studies have yielded mixed findings on this relationship. Some studies demonstrate that adults with at least 6 years of instrumental music training [38–41, 45–48] and children with 3 years of extracurricular instruction [49] exhibit greater accuracy in identifying emotional prosodies. These benefits, however, do not extend to measures such as reaction times or emotional intensity ratings [39, 40, 50]. Contrarily, other research finds no significant benefit of music training in this domain, even with 5–14 years of instrumental practice [50–53]. These divergences may be attributable to varying sample sizes, impacting statistical significance, as well as the quantity of music training [54]. Specifically, studies which observed an effect of music training had larger sample sizes (20 ≤ n ≤ 58), making it easier to detect group differences. Meanwhile, studies which did not observe an effect of music training [50, 51, 53] had smaller sample sizes (10 ≤ n ≤ 13) and thus failed to obtain statistically significant results. This is because sample size is related to statistical significance (p-value); the smaller the sample size, the greater the p-value [55, 56]. However, the discrepancy between Mualem and Lavido’s [52] study and other studies [45, 46] cannot be explained by sample size, and may be due to differences in the amount of music training. It is believed that at least 10 years of mixed music training is necessary to have an effect on auditory emotion recognition [54]. The average years of music training of the participants in the study of Mualem and Lavido was approximately 9 years, which may explain why their performance was not altered. Notably, the above research predominantly concentrates on nontonal languages, leaving a significant gap in understanding the implications for tonal languages like Mandarin, Thai, or Vietnamese.
Recognizing that lexical tones are a common element in both language and music [57], existing studies highlight how language and cultural backgrounds influence the processing of emotional cues. Notable differences exist in how listeners of tonal and nontonal languages employ acoustic cues to interpret these tones [58, 59]. Furthermore, East Asians are more sensitive to verbal cues of an emotional utterance, whereas Westerners rely more on semantic cues for understanding and interpreting emotional information [60–62]. While longitudinal experimental studies have demonstrated that musical training enhances the recognition of emotional prosody in children [46, 63] and young adults [52, 64] but not in older adults [65, 66] who speak nontonal languages, the link between musical training and the perception of emotional prosody in speakers of tonal languages remains insufficiently established. To fill this gap in current research, the present study aims to explore the connection between music training and the recognition of emotional prosody in tonal languages. For this purpose, undergraduate students majoring in music performance or musicology, all of whom had received instrumental training since childhood, were compared with students without professional instrumental training. The focus was on their ability to recognize emotional prosody—specifically happiness, sadness, fear, anger, and neutrality—in Mandarin pseudosentences. Building on insights from existing studies [38–41], we hypothesized that music majors would excel in accurately identifying Mandarin emotional prosody compared to nonmusic majors.
Method
Participants
This study was approved by the Ethics Committee of the Harbin Conservatory of Music on 04/06/2023 and all participants provided verbal informed consent form and received monetary compensation for their participation. To achieve 90% power to detect a large main effect (d = 0.8) of group (music vs. nonmusic) at the 0.05 level, a minimum sample size of 34 for each group was required using G*Power (Version 3.1.9.7) [67]. Consequently, 36 music students and 36 nonmusic students who were native Mandarin speakers were recruited for the study from 04/08/2023 to 05/29/2023. The music students had been playing their chosen instruments for a minimum of 12 years, with an average of 13.9 hr of practice per week, or at least 3 hr of practice per day [54, 68]. The instruments covered a wide range, including piano (n = 22), violin (n = 4), double bass (n = 1), viola (n = 1), flute (n = 1), clarinet (n = 1), oboe (n = 1), harp (n = 1), yangqin (n = 1), trumpet (n = 1), suona (n = 1), and erhu (n = 1). In contrast, the nonmusic students had only received compulsory music education in school. To ensure differences in self-reported musical skills and behaviors between music and nonmusic students, the Goldsmiths Musical Sophistication Index (Gold-MSI) [69] was administered. As shown in Table 1, there is moderate evidence that the two groups were matched for age, sex (Fisher’s exact test), years of education, nonverbal IQ as assessed by the Matrix Reasoning Item Bank (MaRs-IB) [70], and handedness as assessed by the Edinburgh Handedness Inventory [71]. However, there is very strong evidence that the music group had more years of training than the nonmusic group and scored higher on all Gold-MSI subtests as well as on the general musical sophistication factor. All participants had normal hearing and no history of neurological or psychiatric disorders.
Stimuli
This study utilized emotional prosody materials drawn from the Chinese vocal emotional stimuli database [72]. The database features 874 pseudosentences, produced by four native Mandarin-Chinese speakers (two male), each conveying one of seven emotions (happiness, sadness, fear, anger, disgust, pleasant surprise, and neutrality) through prosodic variation in 35 unique utterances. Each utterance, ranging from 1 to 2 s in duration, adhered to Chinese grammatical rules while remaining semantically meaningless to eliminate the influence of semantics on emotional prosody perception [72–74]. The stimuli were validated with a group of native Mandarin speakers who participated in a seven-option forced choice task to identify the expressed emotion and a 5-point Likert item (1 = very weak, 5 = very intense) to rate its intensity, excluding neutrality. Stimuli meeting satisfactory identification accuracy criteria were included in the database; specifically, those exceeding a 42.86% average recognition rate (3 times the chance level) for the target emotion and remaining below 42.86% for any nontarget emotion [72, 75, 76].
Twenty-one utterances each expressing happiness, sadness, fear, anger and neutrality were selected from recordings of a male and a female speaker, yielded 42 stimuli per emotion. Table 2 presents the recognition rate and intensity rating for each emotion except neutrality. A one-way analysis of variance (ANOVA) was conducted on the average accuracy rates. Although normality assumptions were met (Ws ≥ 0.94, ps ≥ .200), sphericity was violated (W = 0.26, p = .003). Therefore, the degrees of freedom were corrected using the Greenhouse-Geisser method ( = 0.709). The analysis revealed moderate evidence for a main effect of emotion with a large effect size, F(2.83, 56.70) = 5.87, ωp2 = 0.185, 95% CI [0.018, 0.343], BF = 6.09. Pairwise comparisons (see S2 Table) showed that happiness was recognized less accurately than sadness, anger, and neutrality. No differences were found between any other emotion pairs. Similar analyses were conducted for the average intensity ratings. The data met normality assumptions (Ws ≥ 0.94, ps ≥ .223) but violated sphericity (W = 0.55, p = .048). ANOVA with the Greenhouse-Geisser correction (
= 0.703) revealed very strong evidence for a main effect of emotion with a large effect size, F(2.11, 42.20) = 14.50, ωp2 = 0.386, 95% CI [0.150, 0.557], BF = 1.52 × 103. Pairwise comparisons (see S2 Table) indicated that happiness was rated as less intense than sadness, fear, and anger, while sadness was rated as more intense than fear. No differences were found between anger and sadness or fear.
To assess whether the main acoustic features (mean fundamental frequency [f0 mean], f0 range, mean amplitude [Amp mean], mean harmonics-to-noise ratio [HNR mean], HNR variation [HNR SD], and speech rate) of the stimuli (see Table 2) could predict the intended emotion categories, we conducted a discriminant analysis following the procedures of Liu and Pell [72] and Paulmann and Uskul [77]. Using IBM SPSS Statistics (Version 29.0.2), we treated the main acoustic features as predictor variables and the intended emotions as the outcome variable for the pseudosentences. The analysis revealed that 84.76% (178/210) of the pseudosentences were correctly classified in terms of their intended emotion: happiness 78.57% (33/42), sadness 97.62% (41/42), fear 76.19% (32/42), anger 80.95% (34/42), and neutrality 90.48% (38/42). These results suggest that listeners can reliably identify the intended emotions based on the analyzed acoustic features.
Measures
MaRs-IB.
The MaRs-IB was employed to measure participants’ nonverbal IQ [70]. This open-access online assessment has demonstrated good reliability and validity and is similar to Raven’s Advanced Progressive Matrices [78]. The test consists of 80 items, each featuring a 3×3 matrix in which eight out of nine cells contain abstract shapes that vary along four dimensions: color, size, shape and position. Participants must select the missing shape from an array of four options. Each item answered correctly is worth one point, while incorrect items are worth zero points. A higher score or percentage of accuracy is indicative of a stronger spatial reasoning ability and a higher nonverbal IQ [79].
Gold-MSI.
The Gold-MSI [69] is a questionnaire designed to assess individual differences in self-reported musical skills and behaviors among both musicians and the general population. It consists of 38 items divided into five subscales: Musical Training (7 items), Perceptual Abilities (9 items), Singing Abilities (7 items), Active Engagement (9 items), and Emotions (6 items). Respondents are asked to rate their agreement with each statement on a 7-point Likert scale (from 1 strongly disagree to 7 strongly agree) for the first 31 items or select one of seven options for the remaining 7 items. The scores for each dimension are then summed, with higher scores indicating a higher level of musical skills and greater musical engagement or emotional experiences. Additionally, the General Sophistication consists of 18 items, with higher scores indicating a greater degree of musical sophistication. The Chinese version of the Gold-MSI has been found to demonstrate high reliability and validity [80, 81] and was used in the present study. In the current study, the reliability of the Musical Training, Perceptual Abilities, Singing Abilities, Active Engagement, Emotions, and General Sophistication factors, as evaluated by McDonald’s ω, was 0.98 (95% CI [0.97, 0.99]), 0.95 (95% CI [0.93, 0.96]), 0.94 (95% CI [0.91, 0.96]), 0.93 (95% CI [0.91, 0.96]), 0.84 (95% CI [0.78, 0.90]), and 0.97 (95% CI [0.96, 0.98]) respectively.
Procedure
Participants completed the online experiment through the NAODAO platform (www.naodao.com) between 04/15/2023 and 06/06/2023. Before the experiment began, participants were required to read and fully understand the informed consent form. The experiment only commenced automatically upon their explicit confirmation of participation. To ensure participants were familiar with the procedure, a practice session preceded the formal experiment. This practice session included four practice stimuli and feedback on accuracy. Additionally, participants could adjust the headphone volume to a comfortable level.
Fig 1 displays the design of the trial in the experiment. At the start of each trial, a red fixation cross appeared in the center of the screen for 0.5 s to indicate that a stimulus was about to come. The fixation cross then turned green as the stimulus began to play for approximately 1–2 s. After listening to the stimulus, a response interface appeared, and participants were asked to click on one of the five emotional prosody labels and icons to match the emotion expressed in the speech as quickly and accurately as possible. Subsequently, a 7-point Likert item of emotion intensity (1 = very low, 7 = very high) was presented on the screen, and participants were asked to rate the intensity of the emotion in the stimulus without a time limit. To control for potential sex-related differences in vocal acoustic features [82, 83] and reduce participants’ cognitive load from sudden change in speaker sex, we presented sentences spoken by male and female speakers in two separate blocks, with 105 randomly presented sentences per block. The order of the blocks was counterbalanced across participants, with a 1–2 min break between the two blocks.
Data analysis
Before data analysis, trials with response times (RTs) falling below 0.2 s (anticipatory responses or fast guesses) or exceeding 2 SD above the mean for each participant (distraction) were excluded as outliers. Trials with errors were also omitted for RT and intensity ratings. However, in speeded decision tasks, a pervasive speed-accuracy tradeoff (SAT) [84, 85] exists. Analyzing mean RT or accuracy rates separately could, therefore, lead to contradictory or invalid conclusions [86–88]. To control for this potential SAT, a composite metric incorporating speed and accuracy should be used for performance evaluation. Composite metrics offer more sensitive and reliable assessments than relying solely on accuracy or RT measures [89, 90].
Several such metrics exist, including inverse efficiency score, rate correct score, linear integrated speed-accuracy score, and balanced integration score (BIS). Among these, only the BIS exhibits relatively insensitivity to SAT [91, 92]. To minimize the SAT effect and preserve “true” effects, we opted for the BIS, assigning equal weights to speed and accuracy. The BIS is calculated by subtracting the standardized (i.e., z scores) mean correct RT from the standardized mean proportions of correct responses across all conditions and participants [91–93]. Previous studies have widely adopted this metric [e.g., 94–102]. Using this approach, participants achieving high accuracy at the expense of RT or fast RTs at the expense of accuracy will have BISs near zero compared to the sample average. Participants with both high accuracy and fast RTs will have higher positive BISs, while those with low accuracy and slow RTs will have lower negative BISs. Therefore, higher BIS represents faster and more accurate performance. To investigate the SAT in this study, we first analyzed accuracy and RT data. If an SAT is identified, we will subsequently analyze the data using the BIS framework. Both the frequentist and Bayesian approaches were applied to the data.
Frequentist analysis.
In frequentist analysis, effect sizes and confidence intervals are used to evaluate the magnitude and precision of estimation of an effect and to draw conclusions, instead of relying on p-values from traditional null hypothesis significance tests [103–108]. This method of interpretation is endorsed by the American Psychological Association [109, 110] and several journals [111–113] and researchers [114–117]. If the confidence interval for an effect size does not include zero, the effect is considered to be present; while if the confidence interval (CI) does include zero, the effect is considered to be nonexistent [118–121]. Additionally, effect sizes can be categorized as small, medium, or large based on established thresholds: g or d = 0.20, 0.50, and 0.80 [122], r = 0.10, 0.30, and 0.50 [123], and ω2 = 0.01, 0.06, and 0.14 [124]. This study employed effect sizes and their confidence intervals to interpret the results.
We analyzed the demographic variables of age, years of education, MaRs-IB, handedness, and Gold-MSI using Welch’s t-tests with JASP (Version 0.19.0). Hedges’ g* and their 95% CIs [125] were calculated using the deffectsize package (Version 0.0.0.9000) in R (Version 4.4.1) and RStudio (Version 2024.04.2+764). For sex, we used Fisher’s exact test in JASP and calculated log OR and its 95% CI. For years of music training, we performed a one-sample t-test using the statsExpressions package (Version 1.5.5) and calculated g and its 95% CI.
For each perceptual feature, we conducted an ANOVA with emotion (happiness, sadness, fear, anger, and/or neutrality) as a within-items variable using jamovi (Version 2.5.6). Prior to the ANOVAs, we tested for normality with the Shapiro-Wilk test and sphericity with Mauchly’s test. If assumptions were met, parametric tests were conducted. If any assumption was violated, the Greenhouse-Geisser correction () was applied to adjust the degrees of freedom.
BIS was used to assess overall task performance, but we also conducted a correlation analysis between accuracy and RT to test the SAT. We used repeated measures correlation [126–128] implemented in the rmcorr package (Version 0.7.0). The effect size (correlation coefficient) and its 95% CI were estimated using the bootstrapping with the suggested number of 10000 bootstrap resamples [129–132].
For accuracy, RT, BIS and emotional intensity, we conducted a two-way mixed design ANOVA with group (music and nonmusic) as the between-subjects variable and emotion (happiness, sadness, fear, anger, and neutrality) as the within-subjects variable. Data were tested for normality, homogeneity of variance with Levene’s test, and sphericity using JASP. Parametric tests were conducted if assumptions were met; otherwise, nonparametric tests were conducted with ARTool (Version 0.11.1). Post hoc comparisons with Bonferroni correction were performed using the ART-C program for significant interaction effects. We used the effectsize package (Version 0.8.9) to calculate ωp2 and its 95% CI for each ANOVA effect, and d and its 95% CI for each pairwise or post hoc comparison.
Bayesian analysis.
To assess the strength of the evidence and obtain more conservative results, we conducted Bayesian hypothesis testing. The Bayes factor (BF) was calculated to quantify the relative plausibility of the null hypothesis (H0) and alternative hypothesis (H1) given the observed data [133–136]. Conventionally, a BF of 1–3 indicates weak evidence for H1, 3–10 indicates moderate evidence, 10–30 indicates strong evidence, and above 30 indicates very strong evidence for H1. Conversely, BFs between 0.33–1, 0.10–0.33, 0.03–0.10, and below 0.03 indicate weak, moderate, strong, and very strong evidence for H0, respectively [137, 138]. While acknowledging the inherent subjectivity in interpreting evidential strength, this study adopted the common practice of considering moderate evidence as a sufficient threshold for drawing meaningful conclusions. For example, a BF of 5 indicates that the data are 5 times more likely under H1 than H0, implying substantial support for the presence of an effect.
For age, years of education, MaRs-IB, handedness and Gold-MSI, we employed the Bain module in JASP to perform Bayesian Welch’s t-tests [139] and calculate BFs. For sex, we employed JASP to perform a Bayesian contingency table test based on an independent multinomial sampling scheme [140] and obtain the BF. For years of music training, we used JASP to execute a Bayesian one-sample t-test with the default prior (Cauchy r = 0.707) and determine the BF.
For perceptual features, accuracy, RT, BIS, or emotional intensity, we calculated BFs using anovaBFcalc (Version 0.1.0) with default α = −0.5 (MWS method) based on F-values and degrees of freedom obtained from frequentist analysis. For pairwise or post hoc comparisons, we computed BFs using the BayesFactor package (Version 0.9.12–4.7) with a prior scale factor of r = 0.707, based on t values and sample sizes obtained from frequentist analysis.
Results
We removed 24 music group trials (0.16%) and 86 nonmusic group trials (0.57%) with RT < 0.2 s. Subsequently, we discarded 312 music group trials (2.08%) and 368 nonmusic group trials (2.45%) with RT > (M + 2 SD) s. Finally, we eliminated 753 music group trials (5.25%) and 1742 nonmusic group trials (12.16%) due to error responses.
Accuracy and response time
Regarding the accuracy data, all dataset except for happiness in the nonmusic group (W = 0.95, p = .146) deviated from normality (Ws ≤ 0.93, ps ≤ .025). Additionally, all datasets except for happiness (F(1, 70) = 2.40, p = .126) violated the assumptions of homogeneity of variance (Fs(1, 70) ≥ 5.01, ps ≤ .028) and sphericity (W = 0.54, p < .001). As for the RT data, all datasets except for happiness (W = 0.95, p = .138) and sadness (W = 0.95, p = .121) in the nonmusic group violated the normality assumption (Ws ≤ 0.93, ps ≤ .024). Furthermore, all RT datasets violated the assumptions of homogeneity of variance (Fs(1, 70) ≥ 4.47, ps ≤ .038) and sphericity (W = 0.65, p < .001).
Fig 2 illustrates the accuracy and RTs of the music and nonmusic groups across all emotions. It is evident that the music group exhibited higher accuracy but slower RTs compared to the nonmusic group for all emotions. To validate these visual observations, separate two-way ANOVAs were conducted on the accuracy and RT data. There was very strong evidence for a large main effect of group on accuracy (F(1,70) = 62.10, ωp2 = 0.459, 95% CI = [0.290, 0.589], BF = 3.13 × 108), indicating that the music group (M = 89.47, SD = 11.67) performed better than the nonmusic group (M = 75.39, SD = 20.34). Similarly, there was very strong evidence for a large main effect of group on RTs (F(1,70) = 14.33, ωp2 = 0.156, 95% CI = [0.033, 0.312], BF = 59.08), with the music group exhibiting slower RTs (M = 1.28, SD = 0.60) than the nonmusic group (M = 0.89, SD = 0.37). These results suggest the presence of an SAT for both music and nonmusic subjects.
Accuracy (A) and response time (B) as a function of group and emotion. These raincloud plots, combining a halved violin plot, a boxplot, and jittered individual data points, depict the data distribution. They were generated using ggplot2 (Version 3.5.1), PupillometryR (Version 0.0.5), ggpp (Version 0.5.8–1), and cowplot (Version 1.1.3) in R.
Fig 3 depicts the speed–accuracy correlations for the combined group and each individual group across all emotions. A moderate negative correlation between RT and accuracy was observed for the combined group, r(287) = −0.397, 95% CI [−0.485, −0.306], consistent with the SAT phenomenon. This correlation remained similar for the music group, r(143) = −0.366, 95% CI [−0.504, −0.204], but was larger for the nonmusic group, r(143) = −0.510, 95% CI [−0.616, −0.392]. These findings suggest that faster RT generally come at the expense of reduced accuracy within individuals, highlighting the suitability of BIS as a more comprehensive measure of task performance.
Each dot represents the average response time and accuracy for a specific emotion. Participant identity is indicated by color, and colored lines show the repeated measures correlation fits for each participant’s data. These scatterplots were created using the following R packages: rmcorr (Version 0.7.0), ggplot2 (Version 3.5.1), pals (Version 1.9), and cowplot (Version 1.1.3).
Balanced integration score
All but one dataset (happiness for the nonmusic group: W = 0.96, p = .281) deviated from normality (Ws ≤ 0.94, ps ≤ .037). Homogeneity of variance was confirmed for all datasets (Fs(1, 70) ≤ 0.76, ps ≥ .388), but the data violated the assumption of sphericity (W = 0.62, p < .001).
Fig 4A illustrates the BIS of the music and nonmusic groups across all emotions, suggesting comparable performance. A nonparametric ANOVA corroborated these observations. There was substantial evidence for the absence of a main effect of group, F(1,70) = 0.14, ωp2 = 0.000, 95% CI = [0.000, 0.000], BF = 0.103, indicating that the music group (M = 0.03, SD = 1.40) did not outperform the nonmusic group (M = −0.03, SD = 1.19). In contrast, there was very strong evidence for a large main effect of emotion, F(4, 280) = 34.62, ωp2 = 0.321, 95% CI = [0.230, 0.396], BF = 8.41 × 1019. Pairwise comparisons (see S3 Table) showed that participants performed best in detecting neutral emotional prosody (M = 0.74, SD = 1.19), followed by sadness (M = 0.19, SD = 1.20), happiness (M = 0.17, SD = 1.08), or anger (M = −0.08, SD = 1.40), and worst in detecting fear (M = −1.03, SD = 1.70). Notably, there was very strong evidence for the absence of an interaction effect between group and emotion, F(4, 280) = 2.07, ωp2 = 0.015, 95% CI = [0.000, 0.041], BF = 2.21 × 10−3. These findings suggest that instrumental training is not associated with the ability to recognize emotional prosody in tonal languages.
Balanced integration score (A) and intensity ratings (B) as a function of group and emotion. These raincloud plots, combining a halved violin plot, a boxplot, and jittered individual data points, depict the data distribution. They were generated using ggplot2 (Version 3.5.1), PupillometryR (Version 0.0.5), ggpp (Version 0.5.8–1), and cowplot (Version 1.1.3) in R.
Intensity ratings
All but four datasets (neutrality for the music group, and sadness, anger, and neutrality for the nonmusic group: Ws ≤ 0.90, ps ≤ .004) exhibited normality (Ws ≥ 0.94, ps ≥ .060). Homogeneity of variance was met for all datasets (Fs(1, 70) ≤ 3.02, ps ≥ .087), but the assumption of sphericity was violated (W = 0.35, p < .001).
Fig 4B demonstrates the intensity ratings of the music and nonmusic groups across all emotions, suggesting higher ratings for the music group. A nonparametric ANOVA revealed moderate evidence for a medium main effect of group, F(1,70) = 8.47, ωp2 = 0.094, 95% CI = [0.006, 0.240], BF = 4.92. This indicates that the music group (M = 5.07, SD = 0.85) assigned higher intensity ratings than the nonmusic group (M = 4.68, SD = 0.89). Additionally, a very strong evidence for a large main effect of emotion emerged, F(4, 280) = 74.55, ωp2 = 0.508, 95% CI = [0.428, 0.571], BF = 3.24 × 1039. Pairwise comparisons (see S3 Table) showed that angry emotional prosody received the highest rating (M = 5.47, SD = 0.79), followed by happy (M = 5.05, SD = 0.71) and sad (M = 4.97, SD = 0.84) or fearful (M = 4.81, SD = 0.75), with neutral receiving the lowest rating (M = 4.09, SD = 0.77). Furthermore, moderate evidence was found for a medium interaction effect between group and emotion, F(4, 280) = 6.15, ωp2 = 0.067, 95% CI = [0.014, 0.121], BF = 4.80. Post hoc comparisons (Table 3) revealed very strong evidence for higher ratings by the music group for happy emotion, moderate evidence for higher ratings for fearful and angry emotions, similar ratings for neutral emotion, and only weak evidence for higher ratings for sad emotion. These findings suggest that instrumental training is associated with heightened sensitivity to the intensity of emotional prosodies in tonal languages.
Discussion
This study uniquely explores the link between instrumental music training and emotional prosody identification in tonal languages, focusing on Mandarin. Utilizing robust analysis with effect size estimates and Bayesian factors, we draw strong conclusions based on our sizable sample of 72 participants. Interestingly, the results revealed no difference in emotional prosody identification between musically trained and untrained participants. However, music majors assigned higher intensity ratings to happy, fearful, and angry prosodies. These findings suggest that music training is not positively correlated with an improved ability to recognize emotional prosody in tonal languages but is instead associated with perceived emotional intensity. This adds a new dimension to our understanding of an intricate relationship between music training and emotion perception in tonal languages, distinct from research focused on nontonal languages.
Utilizing the BIS to mitigate speed-accuracy tradeoff effects in our data, we discovered that music and nonmusic majors exhibited similar proficiency in recognizing emotional prosodies. This outcome contrasts with previous studies suggesting that instrumentalists [38–41, 48] and mixed musicians [45, 46, 49] exhibit superior performance on emotional prosody recognition in nontonal languages. This inconsistency may be explained by the intricate interplay between music and language experience in shaping emotional prosody recognition. Indeed, prior research has demonstrated that musicians with a nontonal language background (i.e., English) outperform their nonmusician counterparts in tasks involving linguistic pitch perception [141, 142] and lexical tone perception [14, 143–145]. However, this advantage is not as pronounced among musicians with a tonal language background such as Mandarin [146], Cantonese [147], and Thai [144, 148]. The null result observed in this study may be related to the characteristics of the music subjects. Specifically, the music subjects in our study are more likely to possess absolute pitch—the ability to name a tone without reference—due to their experience with a tonal language. Although we did not directly assess absolute pitch ability, previous research indicates that 53%-72% of Chinese music students possess absolute pitch [149–152]. Given that emotional prosody perception primarily relies on relative pitch processing [153, 154], which focuses on the pitch changes and patterns rather than absolute pitch, the presence of absolute pitch ability may not necessarily confer an advantage in emotional prosody recognition tasks. In contrast, among musical subjects who spoke a nontonal language, only 3%-10% have absolute pitch [149, 151, 155, 156], while the majority possess relative pitch. This heightened capacity for relative pitch processing [157, 158] likely contributes to their enhanced recognition of emotional prosody.
We also discovered that music majors gave higher intensity ratings to happy, fearful, and angry emotional prosody compared to nonmusic majors. This observation aligns with previous evidence that musicians provided higher intensity ratings for happy emotional prosody in English than did nonmusicians [159]. However, it is contradictory to the studies reporting no differences in intensity or arousal ratings for emotional prosody in Portuguese or German between musicians and nonmusicians [40, 160]. The discrepancy may be attributed to the difference in sample sizes; our study included 36 music and 36 nonmusic subjects, whereas Dibben et al.’s and Lima and Castro’s studies involved 17 music and 15 nonmusic subjects, and 20 music and 20 nonmusic subjects, respectively. The larger sample size in our study, being at least 1.8 times larger than those of the other studies, likely facilitated a more sensitive detection the effect of music training. Our findings might be explained by enhanced auditory processing and selective attention abilities in music subjects. Through extensive instrument training and practice, music subjects develop advanced auditory skills that make them more sensitive to subtle changes in acoustic features (e.g., pitch, intensity, and tempo) shared by speech and music. Studies show musicians excel in pitch [161–165], intensity [166], and tempo [167, 168] discrimination compared to nonmusicians. High-arousal emotional speech (e.g., happiness, fear, and anger), often characterized by high pitch and/or large pitch variability, rapid speech rate, and/or increased intensity [18, 21], exhibits more pronounced shifts in these features. As a result, even slight increases in these features might be more salient to music subjects, potentially leading to higher intensity ratings. Furthermore, high-arousal emotional prosodies tend to capture more attention than low-arousal ones. Musicians typically demonstrate superior auditory attentional capacities relative to nonmusicians [169–171]. This heightened selective attention allows them to allocate more cognitive resources to processing emotional information in high-arousal speech prosody. This deeper processing amplifies the perceived intensity of the emotion. This dual effect of enhanced perception and attentional capacity may result in a heightened emotional experience, making high-arousal emotions seem more intense than those with low arousal.
Music majors in our study displayed similar levels of emotion recognition as nonmusic majors but provided higher intensity ratings for emotional prosodies in Mandarin. This dissociation pattern may arise from differences in task demands between emotion recognition and intensity ratings. Emotion recognition tasks, such as categorizing emotions, are generally regarded as explicit tasks; intensity rating tasks are seen as more implicit [172–174]. This distinction stems from the fact that categorizing emotions require direct access to a stable internal representation of a stimulus and a comparison of the presented stimulus with mental prototypes of each emotion. This process necessitates higher activation levels and a conscious representation of the stimulus. On the other hand, rating the intensity of an emotion may not require verbal categorization of a given emotion and relies on a weaker internal representation of the stimulus, with less precise knowledge about this stimulus being necessary for a more global judgement [173–175]. The dissociation between emotion categorization and emotion intensity judgment has been documented in the processing of musical [173, 175–177] and vocal [174, 178] emotion. Thus, our findings suggest that instrumental music training may primarily relate to the implicit, rather than explicit, processing of emotional prosody in tonal languages. This cross-domain transfer likely stems from a statistical learning mechanism employed in music and language acquisition. Statistical learning refers to a domain-general perceptual mechanism where learners unconsciously extract statistical patterns or regularities from sensory input over time, without explicit instruction [29, 179]. This type of learning is particularly relevant for implicit knowledge acquisition that relies on specific acoustic cues [180]. Therefore, even passive exposure to music allows both music and nonmusic subjects to acquire implicit knowledge about the statistical regularities between acoustic cues and musical emotions [181, 182]. On the basis of this implicit knowledge, all subjects are likely to show heightened attention to these cues in emotional prosody and make intensity judgments. Nonetheless, music training enhances music subjects’ sensitivity to the acoustic cues [183, 184], leading to a superior ability to attend to them compared to nonmusic subjects. This heightened sensitivity arises from the focused and intensive nature of music training, which makes music subjects particularly adept at detecting the underlying statistical regularities. Indeed, research demonstrates that musicians outperform nonmusicians in auditory statistical learning tasks involving pitch [161, 185–187] and rhythm [188–190]. Considering the implicit nature of statistical learning [191], its impact is more likely to be observed in implicit tasks rather than explicit ones.
This study has several limitations that warrant careful consideration. First, the quasi-experimental design precludes establishing a causal link between instrumental music training and emotional prosody processing. Future studies should employ randomized experimental designs to address this issue. Second, limited sample size likely contributed to the weakness of evidence for some findings. Consequently, subsequent future studies should use larger sample sizes to obtain more reliable results. Third, this investigation utilized only one tonal language, raising questions about the generalizability of the findings to other tonal languages. Future research is needed to explore this broader applicability. Finally, the mechanisms by which instrumental music training is related to the implicit processing of emotional prosody remain unclear and merit further exploration.
Conclusion
Despite the limitations outlined above, our findings do not support an association between instrumental music training and enhanced recognition of emotional prosody in tonal languages. However, they do suggest a link between music training and participants’ tendency to assign higher intensity ratings to emotional prosody. This dissociation between emotion recognition and emotion intensity evaluation underscores the complexity of the relationship between music training and emotion perception in tonal languages.
The present study is the first to investigate the relationship between music training and emotional prosody perception in Mandarin, a representative of tonal language. Future research should examine this question using different tonal languages to test the reliability of our results and deepen the understanding of the relationship. Such research has important implications for educators and clinicians, as it may provide a new method to treat impaired implicit processing of emotional prosody.
Supporting information
S1 Table. Overview of selected study characteristics for musically trained subjects.
https://doi.org/10.1371/journal.pone.0309432.s001
(XLSX)
S2 Table. Pairwise comparisons of emotional prosodies.
https://doi.org/10.1371/journal.pone.0309432.s002
(DOCX)
S3 Table. Pairwise comparisons of emotional prosodies.
All statistics were obtained from the aligned-and-ranked data.
https://doi.org/10.1371/journal.pone.0309432.s003
(DOCX)
Acknowledgments
We are grateful to Dr. Pan Liu from the Department of Psychology at University of Alberta for providing us with the necessary experimental materials.
References
- 1. Neves L, Martins M, Correia AI, Castro SL, Lima CF. Associations between vocal emotion recognition and socio-emotional adjustment in children. Royal Society Open Science. 2021;8(11). pmid:34804582
- 2. Chronaki G, Wigelsworth M, Pell MD, Kotz SA. The development of cross-cultural recognition of vocal emotion during childhood and adolescence. Scientific Reports. 2018;8(1). pmid:29904120
- 3. Alba-Ferrara L, Hausmann M, Mitchell RL, Weis S. The neural correlates of emotional prosody comprehension: Disentangling simple from complex emotion. PLOS ONE. 2011;6(12). pmid:22174872
- 4. Grandjean D. Brain networks of emotional prosody processing. Emotion Review. 2021;13(1):34–43.
- 5. Scherer KR, Banse R, Wallbott HG. Emotion inferences from vocal expression correlate across languages and cultures. Journal of Cross-Cultural Psychology. 2001;32(1):76–92.
- 6. Young AW, Frühholz S, Schweinberger SR. Face and voice perception: Understanding commonalities and differences. Trends in Cognitive Sciences. 2020;24(5):398–410. pmid:32298625
- 7. Fridenson-Hayo S, Berggren S, Lassalle A, Tal S, Pigat D, Bölte S, et al. Basic and complex emotion recognition in children with autism: Cross-cultural findings. Molecular Autism. 2016;7(1). pmid:28018573
- 8. Schelinski S, von Kriegstein K. The relation between vocal pitch and vocal emotion recognition abilities in people with autism spectrum disorder and typical development. Journal of Autism and Developmental Disorders. 2019;49(1):68–82. pmid:30022285
- 9. Gong B, Li Q, Zhao Y, Wu C. Auditory emotion recognition deficits in schizophrenia: A systematic review and meta-analysis. Asian Journal of Psychiatry. 2021;65. pmid:34482183
- 10. Hoekert M, Kahn RS, Pijnenborg M, Aleman A. Impaired recognition and expression of emotional prosody in schizophrenia: Review and meta-analysis. Schizophrenia Research. 2007;96(1):135–45. pmid:17766089
- 11. Icht M, Zukerman G, Ben-Itzchak E, Ben-David BM. Keep it simple: Identification of basic versus complex emotions in spoken language in individuals with autism spectrum disorder without intellectual disability: A meta-analysis study. Autism Research. 2021;14(9):1948–64. pmid:34101373
- 12. Pawełczyk A, Łojek E, Radek M, Pawełczyk T. Prosodic deficits and interpersonal difficulties in patients with schizophrenia. Psychiatry Research. 2021;306. pmid:34673310
- 13. Ong JH. Distributional learning of lexical tone and musical pitch by naïve and experienced adult learners [Doctoral dissertation]: Western Sydney University; 2016.
- 14. Lee C-Y, Tao L, Bond ZS. Identification of acoustically modified Mandarin tones by native listeners. Journal of Phonetics. 2008;36(4):537–63.
- 15. Banse R, Scherer KR. Acoustic profiles in vocal emotion expression. Journal of Personality and Social Psychology. 1996;70(3):614–36. pmid:8851745
- 16.
Johnstone T, Scherer KR. Vocal communication of emotion. In: Lewis M, Haviland-Jones JM, editors. Handbook of emotions. 2nd ed. New York, NY, US: Guilford Press; 2000. p. 220–35.
- 17. Goudbeek M, Scherer K. Beyond arousal: Valence and potency/control cues in the vocal expression of emotion. The Journal of the Acoustical Society of America. 2010;128(3):1322–36. pmid:20815467
- 18. Juslin PN, Laukka P. Communication of emotions in vocal expression and music performance: Different channels, same code? Psychological Bulletin. 2003;129(5):770–814. pmid:12956543
- 19. Juslin PN, Laukka P. Impact of intended emotion intensity on cue utilization and decoding accuracy in vocal expression of emotion. Emotion. 2001;1(4):381–412. pmid:12901399
- 20. Leitman DI, Laukka P, Juslin PN, Saccente E, Butler P, Javitt DC. Getting the cue: Sensory contributions to auditory emotion recognition impairments in schizophrenia. Schizophrenia Bulletin. 2008;36(3):545–56. pmid:18791077
- 21. Scherer KR. Vocal affect expression: A review and a model for future research. Psychological Bulletin. 1986;99(2):143–65. pmid:3515381
- 22. Curtis ME, Bharucha JJ. The minor third communicates sadness in speech, mirroring its use in music. Emotion. 2010;10(3):335–48. pmid:20515223
- 23. Ilie G, Thompson WF. A comparison of acoustic cues in music and speech for three dimensions of affect. Music Perception. 2006;23(4):319–30.
- 24. Coutinho E, Dibben N. Psychoacoustic cues to emotion in speech prosody and music. Cognition and Emotion. 2013;27(4):658–84. pmid:23057507
- 25. Thorndike EL, Woodworth RS. The influence of improvement in one mental function upon the efficiency of other functions. (I). Psychological Review. 1901;8(3):247–61.
- 26. Patel AD. Why would musical training benefit the neural encoding of speech? The OPERA hypothesis. Frontiers in Psychology. 2011;2. pmid:21747773
- 27. Patel AD. Can nonlinguistic musical training change the way the brain processes speech? The expanded OPERA hypothesis. Hearing Research. 2014;308:98–108. pmid:24055761
- 28. Besson M, Chobert J, Marie C. Transfer of training between music and speech: Common processing, attention, and memory. Frontiers in Psychology. 2011;2. pmid:21738519
- 29.
Patel AD. Music, Language, and the Brain: Oxford University Press; 2008 22 Mar 2012.
- 30. Ong JH, Burnham D, Escudero P. Distributional learning of lexical tones: A comparison of attended vs. unattended listening. PLOS ONE. 2015;10(7). pmid:26214002
- 31. Asaridou SS, Hagoort P, McQueen JM. Effects of early bilingual experience with a tone and a non-tone language on speech-music integration. PLOS ONE. 2015;10(12). pmid:26659377
- 32. Chen A, Peter V, Wijnen F, Schnack H, Burnham D. Are lexical tones musical? Native language’s influence on neural response to pitch in different domains. Brain and Language. 2018;180–182:31–41. pmid:29689493
- 33. Creel SC. The familiar-melody advantage in auditory perceptual development: Parallels between spoken language acquisition and general auditory perception. Attention, Perception, & Psychophysics. 2019;81(4):948–57. pmid:30635834
- 34. Creel SC, Obiri-Yeboah M, Rose S. Language-to-music transfer effects depend on the tone language: Akan vs. East Asian tone languages. Memory & Cognition. 2023;51(7):1624–39. pmid:37052771
- 35. Ip MHK, Cutler A. Universals of listening: Equivalent prosodic entrainment in tone and non-tone languages. Cognition. 2020;202. pmid:32502869
- 36. Peng G, Zheng H-Y, Gong T, Yang R-X, Kong J-P, Wang WSY. The influence of language experience on categorical perception of pitch contours. Journal of Phonetics. 2010;38(4):616–24.
- 37. Ong JH, Wong PCM, Liu F. Musicians show enhanced perception, but not production, of native lexical tones. The Journal of the Acoustical Society of America. 2020;148(6):3443–54. pmid:33379922
- 38. Correia AI, Castro SL, MacGregor C, Müllensiefen D, Schellenberg EG, Lima CF. Enhanced recognition of vocal emotions in individuals with naturally good musical abilities. Emotion. 2022;22(5):894–906. pmid:32718172
- 39. Farmer E, Jicol C, Petrini K. Musicianship enhances perception but not feeling of emotion from others’ social interaction through speech prosody. Music Perception. 2020;37(4):323–38.
- 40. Lima CF, Castro SL. Speaking to the trained ear: Musical expertise enhances the recognition of emotions in speech prosody. Emotion. 2011;11(5):1021–31. pmid:21942696
- 41. Twaite JT. Examining relationships between basic emotion perception and musical training in the prosodic, facial, and lexical channels of communication and in music [Doctoral dissertation]: City University of New York; 2016.
- 42. Martins M, Pinheiro AP, Lima CF. Does music training improve emotion recognition abilities? A critical review. Emotion Review. 2021;13(3):199–210.
- 43. Schellenberg EG, Lima CF. Music training and nonmusical abilities. Annual Review of Psychology. 2024;75:87–128. pmid:37738514
- 44. Nussbaum C, Schweinberger SR. Links between musicality and vocal emotion perception. Emotion Review. 2021;13(3):211–24.
- 45. Thompson WF, Schellenberg EG, Husain G. Perceiving prosody in speech. Annals of the New York Academy of Sciences. 2003;999(1):530–2. pmid:14681180
- 46. Thompson WF, Schellenberg EG, Husain G. Decoding speech prosody: Do music lessons help? Emotion. 2004;4(1):46–64. pmid:15053726
- 47. Fuller CD, Galvin JJ, Maat B, Free RH, Başkent D. The musician effect: does it persist under degraded pitch conditions of cochlear implant simulations? Frontiers in Neuroscience. 2014;8. pmid:25071428
- 48. Nussbaum C, Schirmer A, Schweinberger SR. Musicality–Tuned to the melody of vocal emotions. British Journal of Psychology. 2023. pmid:37851369
- 49. Schellenberg EG, Mankarious M. Music training and emotion comprehension in childhood. Emotion. 2012;12(5):887–91. pmid:22642351
- 50. Francisco RMC. Pianos and microphones: Does the type of musical training affect emotion recognition? [Master’s thesis]: Universidade de Lisboa; 2021.
- 51. Başkent D, Fuller CD, Galvin JJ III, Schepel L, Gaudrain E, Free RH. Musician effect on perception of spectro-temporally degraded speech, vocal emotion, and music in young adolescents. The Journal of the Acoustical Society of America. 2018;143(5):EL311–EL6. pmid:29857757
- 52. Mualem O, Lavidor M. Music education intervention improves vocal emotion recognition. International Journal of Music Education. 2015;33(4):413–25.
- 53. Park M, Gutyrchik E, Welker L, Carl P, Pöppel E, Zaytseva Y, et al. Sadness is unique: Neural processing of emotions in speech prosody in musicians and non-musicians. Frontiers in Human Neuroscience. 2015;8. pmid:25688196
- 54. Zhang JD, Susino M, McPherson GE, Schubert E. The definition of a musician in music psychology: A literature review and the six-year rule. Psychology of Music. 2020;48(3):389–409.
- 55. Gómez-de-Mariscal E, Guerrero V, Sneider A, Jayatilaka H, Phillip JM, Wirtz D, et al. Use of the p-values as a size-dependent function to address practical differences when analyzing large datasets. Scientific Reports. 2021;11(1). pmid:34686696
- 56. Whitley E, Ball J. Statistics review 4: Sample size calculations. Critical Care. 2002;6(4):335–41. pmid:12225610
- 57.
Brown S. The "musilanguage" model of music evolution. In: Wallin NL, Merker B, Brown S, editors. The origins of music: MIT Press; 1999. p. 271–300.
- 58. Burnham D, Jones C. Categorical perception of lexical tone by tonal and non-tonal language speakers. 9th International Conference on Speech Science and Technology; Melbourne, VIC, Australia 2002, December 2–5. p. 515–20.
- 59. Hallé PA, Chang Y-C, Best CT. Identification and discrimination of Mandarin Chinese tones by Mandarin Chinese vs. French listeners. Journal of Phonetics. 2004;32(3):395–421.
- 60. Kitayama S, Ishii K. Word and voice: Spontaneous attention to emotional utterances in two languages. Cognition and Emotion. 2002;16(1):29–59.
- 61. Ishii K, Reyes JA, Kitayama S. Spontaneous attention to word content versus emotional tone: Differences among three cultures. Psychological Science. 2003;14(1):39–46. pmid:12564752.
- 62. Yang Y, Wang L, Wang Q. Take your word or tone for it? European American and Chinese children’s attention to emotional cues in speech. Child Development. 2021;92(3):844–52. pmid:33881172
- 63. Good A, Gordon KA, Papsin BC, Nespoli G, Hopyan T, Peretz I, et al. Benefits of music training for perception of emotional speech prosody in deaf children with cochlear implants. Ear and Hearing. 2017;38(4). pmid:28085739
- 64. Bodner E, Aharoni R, Iancu I. The effect of training with music on happiness recognition in social anxiety disorder. Journal of Psychopathology and Behavioral Assessment. 2012;34(4):458–66.
- 65. Chari DA, Barrett KC, Patel AD, Colgrove TR, Jiradejvong P, Jacobs LY, et al. Impact of auditory-motor musical training on melodic pattern recognition in cochlear implant users. Otology & Neurotology. 2020;41(4):e422–e31. pmid:32176126
- 66. Fuller CD, Galvin JJ, Maat B, Başkent D, Free RH. Comparison of two music training approaches on music and speech perception in cochlear implant users. Trends in Hearing. 2018;22. pmid:29621947.
- 67. Faul F, Erdfelder E, Lang A-G, Buchner A. G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods. 2007;39(2):175–91. pmid:17695343
- 68. Zhang JD, Schubert E. A single item measure for identifying musician and nonmusician categories based on measures of musical sophistication. Music Perception. 2019;36(5):457–67.
- 69. Müllensiefen D, Gingras B, Musil J, Stewart L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. PLOS ONE. 2014;9(2). pmid:24586929
- 70. Chierchia G, Fuhrmann D, Knoll LJ, Pi-Sunyer BP, Sakhardande AL, Blakemore S-J. The matrix reasoning item bank (MaRs-IB): Novel, open-access abstract reasoning items for adolescents and adults. Royal Society Open Science. 2019;6(10). pmid:31824684
- 71. Oldfield RC. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia. 1971;9(1):97–113. pmid:5146491
- 72. Liu P, Pell MD. Recognizing vocal emotions in Mandarin Chinese: A validated database of Chinese vocal emotional stimuli. Behavior Research Methods. 2012;44(4):1042–51. pmid:22539230
- 73. Pell MD, Jaywant A, Monetta L, Kotz SA. Emotional speech processing: Disentangling the effects of prosody and semantic cues. Cognition and Emotion. 2011;25(5):834–53. pmid:21824024
- 74. Larrouy-Maestri P, Poeppel D, Pell MD. The sound of emotional prosody: Nearly 3 decades of research and future directions. Perspectives on Psychological Science. 2024. pmid:38232303.
- 75. Castro SL, Lima CF. Recognizing emotions in spoken language: A validated set of Portuguese sentences and pseudosentences for research on emotional prosody. Behavior Research Methods. 2010;42(1):74–81. pmid:20160287
- 76. Liu P, Rigoulot S, Jiang X, Zhang S, Pell MD. Unattended emotional prosody affects visual processing of facial expressions in Mandarin-speaking Chinese: A comparison with English-speaking Canadians. Journal of Cross-Cultural Psychology. 2021;52(3):275–94. pmid:33958813
- 77. Paulmann S, Uskul AK. Cross-cultural emotional prosody recognition: Evidence from Chinese and British listeners. Cognition and Emotion. 2014;28(2):230–44. pmid:23862740
- 78. Raven J. The Raven Progressive Matrices and measuring aptitude constructs. The International Journal of Educational and Psychological Assessment. 2009;2:2–38. 2010-11467-001.
- 79. Zorowitz S, Chierchia G, Blakemore S-J, Daw ND. An item response theory analysis of the matrix reasoning item bank (MaRs-IB). Behavior Research Methods. 2023. pmid:37020082
- 80. Li J, Lin H-R, Wolf A, Lothwesen K. Measuring musical sophistication in the Chinese general population: Validation and replication of the Simplified Chinese Gold-MSI. Musicae Scientiae. 2023.
- 81. Lin H-R, Kopiez R, Müllensiefen D, Wolf A. The Chinese version of the Gold-MSI: Adaptation and validation of an inventory for the measurement of musical sophistication in a Taiwanese sample. Musicae Scientiae. 2019;25(2):226–51.
- 82. Santos AO, Godoy J, Silverio K, Brasolotto A. Vocal changes of men and women from different age decades: An analysis from 30 years of age. Journal of Voice. 2023;37(6):840–50. pmid:34284927
- 83. Hillenbrand JM, Clark MJ. The role of f0 and formant frequencies in distinguishing the voices of men and women. Attention, Perception, & Psychophysics. 2009;71(5):1150–66. pmid:19525544
- 84. Heitz RP. The speed-accuracy tradeoff: History, physiology, methodology, and behavior. Frontiers in Neuroscience. 2014;8. pmid:24966810
- 85. Wickelgren WA. Speed-accuracy tradeoff and information processing dynamics. Acta Psychologica. 1977;41(1):67–85.
- 86. Alhowaide A, Alsmadi I, Tang J. Ensemble detection model for IoT IDS. Internet of Things. 2021;16.
- 87. Lerche V, Voss A. Speed–accuracy manipulations and diffusion modeling: Lack of discriminant validity of the manipulation or of the parameter estimates? Behavior Research Methods. 2018;50(6):2568–85. pmid:29542062
- 88. Lerche V, Voss A. When accuracy rates and mean response times lead to false conclusions: A simulation study based on the diffusion model. TQMP. 2020;16(2):107–19.
- 89. Mueller ST, Alam L, Funke GJ, Linja A, Ibne Mamun T, Smith SL. Examining methods for combining speed and accuracy in a go/no-go vigilance task. Proceedings of the Human Factors and Ergonomics Society Annual Meeting. 2020;64(1):1202–6.
- 90. Bakun Emesh T, Garbi D, Kaplan A, Zelicha H, Yaskolka Meir A, Tsaban G, et al. Retest reliability of integrated speed–accuracy measures. Assessment. 2022;29(4):717–30. pmid:33522278.
- 91. Liesefeld HR, Janczyk M. Combining speed and accuracy to control for speed-accuracy trade-offs(?). Behavior Research Methods. 2019;51(1):40–60. pmid:30022459
- 92. Liesefeld HR, Janczyk M. Same same but different: Subtle but consequential differences between two measures to linearly integrate speed and accuracy (LISAS vs. BIS). Behavior Research Methods. 2023;55(3):1175–92. pmid:35595937
- 93. Liesefeld HR, Fu X, Zimmer HD. Fast and careless or careful and slow? Apparent holistic processing in mental rotation is explained by speed-accuracy trade-offs. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2015;41(4):1140–51. pmid:25528084
- 94. Bouvet L, Simard-Meilleur A-A, Paignon A, Mottron L, Donnadieu S. Auditory local bias and reduced global interference in autism. Cognition. 2014;131(3):367–72. pmid:24637103
- 95. Hassan EK, Jones AM, Buckingham G. A novel protocol to induce mental fatigue. Behavior Research Methods. 2023. pmid:37537491
- 96. Paas Oliveros LK, Cieslik EC, Pieczykolan A, Pläschke RN, Eickhoff SB, Langner R. Brain functional characterization of response-code conflict in dual-tasking and its modulation by age. Cerebral Cortex. 2023;33(18):10155–80. pmid:37540164
- 97. English MCW, Maybery MT, Visser TAW. Magnitude of sex differences in visual search varies with target eccentricity. Psychonomic Bulletin & Review. 2021;28(1):178–88. pmid:32875533
- 98. Weissbach A, Moyé J, Takacs A, Verrel J, Chwolka F, Friedrich J, et al. Perception–action integration is altered in functional movement disorders. Movement Disorders. 2023;38(8):1399–409. pmid:37315159
- 99. Robson SG, Tangen JM, Searston RA. The effect of expertise, target usefulness and image structure on visual search. Cognitive Research: Principles and Implications. 2021;6. pmid:33709197
- 100.
Yu Z, Dong Y. The emergence of a complex language skill: Evidence from the self-organization of interpreting competence in interpreting students. Bilingualism: Language and Cognition. 2022;25(2):269–82. Epub 2021/09/28. https://doi.org/10.1017/S1366728921000870
- 101. McGarrigle R, Knight S, Hornsby BWY, Mattys S. Predictors of listening-related fatigue across the adult life span. Psychological Science. 2021;32(12):1937–51. pmid:34751602.
- 102. McGarrigle R, Mattys S. Sensory-processing sensitivity predicts fatigue from listening, but not perceived effort, in young and older adults. Journal of Speech, Language, and Hearing Research. 2023;66(2):444–60. pmid:36657070
- 103. Calin-Jageman RJ, Cumming G. Estimation for better inference in neuroscience. eNeuro. 2019;6(4). pmid:31453316
- 104. Calin-Jageman RJ, Cumming G. The new statistics for better science: Ask how much, how uncertain, and what else is known. The American Statistician. 2019;73(sup1):271–80. pmid:31762475
- 105. Cohen J. Things I have learned (so far). American Psychologist. 1990;45(12):1304–12.
- 106. Nakagawa S, Cuthill IC. Effect size, confidence interval and statistical significance: A practical guide for biologists. Biological Reviews. 2007;82(4):591–605. pmid:17944619
- 107. Wasserstein RL, Lazar NA. The ASA statement on p-values: Context, process, and purpose. The American Statistician. 2016;70(2):129–33.
- 108. Wasserstein RL, Schirm AL, Lazar NA. Moving to a world beyond “p < 0.05”. The American Statistician. 2019;73(sup1):1–19.
- 109. American Psychological Association. Publication manual of the American Psychological Association. 6th ed 2010.
- 110. American Psychological Association. Publication manual of the American Psychological Association. 7th ed 2020.
- 111. Points of significance. Nature Human Behaviour. 2023;7(3):293–4. pmid:36964301
- 112. Harrington DD’Agostino RB, Gatsonis C, Hogan JW, Hunter DJ, Normand S-LT, et al. New guidelines for statistical reporting in the Journal. The New England Journal of Medicine. 2019;381(3):285–6. pmid:31314974.
- 113. Trafimow D, Marks M. Editorial. Basic and Applied Social Psychology. 2015;37(1):1–2.
- 114. Cumming G. The new statistics: A how‐to guide. Australian Psychologist. 2013;48(3):161–70.
- 115. Cumming G. The new statistics: Why and how. Psychological Science. 2014;25(1):7–29. pmid:24220629.
- 116. Griffiths P, Needleman J. Statistical significance testing and p-values: Defending the indefensible? A discussion paper and position statement. International Journal of Nursing Studies. 2019;99. pmid:31442781
- 117. Karadaghy OA, Hong H, Scott-Wittenborn N, Sinha P, Suko J, Tait S, et al. Reporting of effect size and confidence intervals in JAMA Otolaryngology–Head & Neck Surgery. JAMA Otolaryngology–Head & Neck Surgery. 2017;143(11):1075–80. pmid:28910425
- 118. Cheung SF, Pesigan IJA, Vong WN. DIY bootstrapping: Getting the nonparametric bootstrap confidence interval in SPSS for any statistics or function of statistics (when this bootstrapping is appropriate). Behavior Research Methods. 2023;55(2):474–90. pmid:35292932
- 119. Perdices M. Null hypothesis significance testing, p-values, effects sizes and confidence intervals. Brain Impairment. 2018;19(1):70–80. Epub 2017/12/07.
- 120. Sohn D. Sex differences in achievement self-attributions: An effect-size analysis. Sex Roles. 1982;8(4):345–57.
- 121. Steiger JH. Beyond the F test: Effect size confidence intervals and tests of close fit in the analysis of variance and contrast analysis. Psychological Methods. 2004;9(2):164–82. pmid:15137887
- 122.
Cohen J. Statistical power analysis for the behavioral sciences. 2nd ed. Hillsdale, NJ: Erlbaum; 1988.
- 123. Cohen J. A power primer. Psychological Bulletin. 1992;112(1):155–9. pmid:19565683
- 124. Kirk RE. Practical significance: A concept whose time has come. Educational and Psychological Measurement. 1996;56(5):746–59.
- 125. Delacre M, Lakens D, Ley C, Liu L, Leys C. Why Hedges’ g*s based on the non-pooled standard deviation should be reported with Welch’s t-test 2021.
- 126. Bakdash JZ, Marusich LR. Repeated measures correlation. Frontiers in Psychology. 2017;8. pmid:28439244
- 127. Bland JM, Altman DG. Calculating correlation coefficients with repeated observations: Part 1—correlation within subjects. BMJ. 1995;310(6977). pmid:7873953
- 128. Marusich L, Bakdash J. rmcorrShiny: A web and standalone application for repeated measures correlation. F1000Research. 2021;10. pmid:34621514
- 129.
Chihara LM, Hesterberg TC. Mathematical statistics with resampling and R. 3rd ed: John Wiley & Sons; 2022.
- 130.
Hair JF, Hult GTM, Ringle CM, Sarstedt M. A primer on partial least squares structural equation modeling (PLS-SEM). 3rd ed: SAGE Publications; 2022.
- 131. Streukens S, Leroi-Werelds S. Bootstrapping and PLS-SEM: A step-by-step guide to get more out of your bootstrap results. European Management Journal. 2016;34(6):618–32.
- 132. Hesterberg TC. What teachers should know about the bootstrap: Resampling in the undergraduate statistics curriculum. The American Statistician. 2015;69(4):371–86. pmid:27019512
- 133. Brydges CR, Bielak AAM. A Bayesian analysis of evidence in support of the null hypothesis in gerontological psychology (or lack thereof). The Journals of Gerontology: Series B. 2020;75(1):58–66. pmid:30877301
- 134. Heck DW, Boehm U, Böing-Messing F, Bürkner P-C, Derks K, Dienes Z, et al. A review of applications of the Bayes factor in psychological research. Psychological Methods. 2023;28(3):558–79. pmid:35298215
- 135. Schmalz X, Biurrun Manresa J, Zhang L. What is a Bayes factor? Psychological Methods. 2023;28(3):705–18. pmid:34780246
- 136. Wagenmakers E-J, Marsman M, Jamil T, Ly A, Verhagen J, Love J, et al. Bayesian inference for psychology. Part I: Theoretical advantages and practical ramifications. Psychonomic Bulletin & Review. 2018;25(1):35–57. pmid:28779455
- 137.
Jeffreys H. Theory of probability. 3rd ed. Oxford, UK: Oxford University Press; 1961.
- 138.
Lee MD, Wagenmakers E-J. Bayesian cognitive modeling: A practical course. New York, NY: Cambridge University Press; 2014. xiii, 264–xiii, p.
- 139. Fu Q, Hoijtink H, Moerbeek M. Sample-size determination for the Bayesian t test and Welch’s test using the approximate adjusted fractional Bayes factor. Behavior Research Methods. 2021;53(1):139–52. pmid:32632740
- 140. Jamil T, Ly A, Morey RD, Love J, Marsman M, Wagenmakers E-J. Default “Gunel and Dickey” Bayes factors for contingency tables. Behavior Research Methods. 2017;49(2):638–52. pmid:27325166
- 141. Wayland R, Herrera E, Kaan E. Effects of musical experience and training on pitch contour perception. Journal of Phonetics. 2010;38(4):654–62.
- 142. Wong PCM, Skoe E, Russo NM, Dees T, Kraus N. Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nature Neuroscience. 2007;10(4):420–2. pmid:17351633
- 143. Alexander J, Wong P, Bradlow A. Lexical tone perception in musicians and non-musicians. 6th Interspeech 2005 and 9th European Conference on Speech Communication and Technology; Lisbon, Portugal: International Speech Communication Association; 2005, September 4–8. p. 397–400.
- 144. Cooper A, Wang Y. The influence of linguistic and musical experience on Cantonese word learning. The Journal of the Acoustical Society of America. 2012;131(6):4756–69. pmid:22712948
- 145. Lee C-Y, Lekich A, Zhang Y. Perception of pitch height in lexical and musical tones by English-speaking musicians and nonmusiciansa). The Journal of the Acoustical Society of America. 2014;135(3):1607–15. pmid:24606295
- 146. Chen S, Zhu Y, Wayland R, Yang Y. How musical experience affects tone perception efficiency by musicians of tonal and non-tonal speakers? PLOS ONE. 2020;15(5). pmid:32384088
- 147. Mok PKP, Zuo D. The separation between music and speech: Evidence from the perception of Cantonese tonesa). The Journal of the Acoustical Society of America. 2012;132(4):2711–20. pmid:23039463
- 148. Maggu AR, Wong PCM, Liu H, Wong FCK. Experience-dependent influence of music and language on lexical pitch learning is not additive. 19th Annual Conference of the International Speech Communication Association (Interspeech 2018); Hyderabad, India. Hyderabad, India2018, September 2–6. p. 3791–4.
- 149. Deutsch D, Henthorn T, Marvin E, Xu H. Absolute pitch among American and Chinese conservatory students: Prevalence differences, and evidence for a speech-related critical perioda). The Journal of the Acoustical Society of America. 2006;119(2):719–22. pmid:16521731
- 150. Deutsch D, Li X, Shen J. Absolute pitch among students at the Shanghai Conservatory of Music: A large-scale direct-test study. The Journal of the Acoustical Society of America. 2013;134(5):3853–9. pmid:24180794
- 151. Ki Miyazaki, Rakowski A, Makomaska S, Jiang C, Tsuzaki M, Oxenham AJ, et al. Absolute pitch and relative pitch in music students in the East and the West: Implications for aural-skills education. Music Perception. 2018;36(2):135–55.
- 152. Lee C-Y, Lee Y-F. Perception of musical pitch and lexical tones by Mandarin-speaking musicians. The Journal of the Acoustical Society of America. 2010;127(1):481–90. pmid:20058993
- 153. Pell MD, Paulmann S, Dara C, Alasseri A, Kotz SA. Factors in the recognition of vocally expressed emotions: A comparison of four languages. Journal of Phonetics. 2009;37(4):417–35.
- 154. Breitenstein C, Lancker DV, Daum I. The contribution of speech rate and pitch variation to the perception of vocal emotions in a German and an American sample. Cognition and Emotion. 2001;15(1):57–79.
- 155. Ki Miyazaki, Makomaska S, Rakowski A. Prevalence of absolute pitch: A comparison between Japanese and Polish music students. The Journal of the Acoustical Society of America. 2012;132(5):3484–93. pmid:23145628
- 156. Leite RBC, Mota-Rolim SA, Queiroz CMT. Music proficiency and quantification of absolute pitch: A large-scale study among Brazilian musicians. Frontiers in Neuroscience. 2016;10. pmid:27790084
- 157. Ngo MK, Vu K-PL, Strybel TZ. Effects of music and tonal language experience on relative pitch performance. The American Journal of Psychology. 2016;129(2):125–34. pmid:27424415
- 158. Schellenberg EG, Moreno S. Music lessons, pitch processing, and g. Psychology of Music. 2010;38(2):209–21.
- 159. Faber S, Fiveash A. Emotion without words: A comparison study of music and speech prosody. Canadian Journal of Music Therapy. 2014;20(2):86–101.
- 160. Dibben N, Coutinho E, Vilar JA, Estévez-Pérez G. Do individual differences influence moment-by-moment reports of emotion perceived in music and speech prosody? Frontiers in Behavioral Neuroscience. 2018;12. pmid:30210316
- 161. Vasuki PRM, Sharma M, Demuth K, Arciuli J. Musicians’ edge: A comparison of auditory processing, cognitive abilities and statistical learning. Hearing Research. 2016;342:112–23. pmid:27770623
- 162. Micheyl C, Delhommeau K, Perrot X, Oxenham AJ. Influence of musical and psychoacoustical training on pitch discrimination. Hearing Research. 2006;219(1):36–47. pmid:16839723
- 163. Liang C, Earl B, Thompson I, Whitaker K, Cahn S, Xiang J, et al. Musicians are better than non-musicians in frequency change detection: behavioral and electrophysiological evidence. Frontiers in Neuroscience. 2016;10. pmid:27826221
- 164. Kishon-Rabin L, Amir O, Vexler Y, Zaltz Y. Pitch discrimination: Are professional musicians better than non-musicians? Journal of Basic and Clinical Physiology and Pharmacology. 2001;12(2):125–44. pmid:11605682
- 165. Sares AG, Foster NEV, Allen K, Hyde KL. Pitch and time processing in speech and tones: The effects of musical training and attention. Journal of Speech, Language, and Hearing Research. 2018;61(3):496–509. pmid:29466555
- 166. Slade T, Gascon A, Comeau G, Russell D. Just noticeable differences in sound intensity of piano tones in non-musicians and experienced pianists. Psychology of Music. 2023;51(3):924–37.
- 167. Madsen CK. Modulated beat discrimination among musicians and nonmusicians. Journal of Research in Music Education. 1979;27(2):57–67.
- 168. Nguyen T, Sidhu RK, Everling JC, Wickett MC, Gibbings A, Grahn JA. Beat perception and production in musicians and dancers. Music Perception. 2022;39(3):229–48.
- 169. Strait DL, Kraus N, Parbery-Clark A, Ashley R. Musical experience shapes top-down auditory mechanisms: Evidence from masking and auditory attention performance. Hearing Research. 2010;261(1):22–9. pmid:20018234
- 170. Weijkamp J, Sadakata M. Attention to affective audio-visual information: Comparison between musicians and non-musicians. Psychology of Music. 2017;45(2):204–15.
- 171. Bialystok E, DePape A-M. Musical expertise, bilingualism, and executive functioning. Journal of Experimental Psychology: Human Perception and Performance. 2009;35(2):565–74. pmid:19331508
- 172. Cunningham WA, Raye CL, Johnson MK. Implicit and explicit evaluation: fMRI correlates of valence, emotional intensity, and control in the processing of attitudes. Journal of Cognitive Neuroscience. 2004;16(10):1717–29. pmid:15701224
- 173. Lévêque Y, Teyssier P, Bouchet P, Bigand E, Caclin A, Tillmann B. Musical emotions in congenital amusia: Impaired recognition, but preserved emotional intensity. Neuropsychology. 2018;32(7):880–94. pmid:30047757
- 174. Pralus A, Fornoni L, Bouet R, Gomot M, Bhatara A, Tillmann B, et al. Emotional prosody in congenital amusia: Impaired and spared processes. Neuropsychologia. 2019;134. pmid:31647961
- 175. Pralus A, Belfi A, Hirel C, Lévêque Y, Fornoni L, Bigand E, et al. Recognition of musical emotions and their perceived intensity after unilateral brain damage. Cortex. 2020;130:78–93. pmid:32645502
- 176. Hirel C, Lévêque Y, Deiana G, Richard N, Cho TH, Mechtouff L, et al. Amusie acquise et anhédonie musicale. Revue Neurologique. 2014;170(8):536–40. pmid:24856610
- 177. Lévêque Y, Roulet-Perez E, Deonna T, Moulin A, Fornoni L, Mayor-Dubois C, et al. Music processing deficits in Landau-Kleffner syndrome: Four case studies in adulthood. Cortex. 2020;129:99–111. pmid:32442777
- 178. Bourgeois–Vionnet J, Moulin A, Hermier M, Pralus A, Nighoghossian N. A case of verbal and emotional prosody processing dissociation after a right temporal venous infarct. Neurological Sciences. 2020;41(6):1615–8. pmid:31902013
- 179. Frost R, Armstrong BC, Siegelman N, Christiansen MH. Domain generality versus modality specificity: the paradox of statistical learning. Trends in Cognitive Sciences. 2015;19(3):117–25. pmid:25631249
- 180. Ong JH, Burnham D, Stevens CJ, Escudero P. Naïve learners show cross-domain transfer after distributional learning: The case of lexical and musical pitch. Frontiers in Psychology. 2016;7. pmid:27551272
- 181. Pearce MT. Statistical learning and probabilistic prediction in music cognition: mechanisms of stylistic enculturation. Annals of the New York Academy of Sciences. 2018;1423(1):378–95. pmid:29749625
- 182. Bigand E, Poulin-Charronnat B. Are we “experienced listeners”? A review of the musical capacities that do not depend on formal musical training. Cognition. 2006;100(1):100–30. pmid:16412412
- 183. François C, Schön D. Neural sensitivity to statistical regularities as a fundamental biological process that underlies auditory learning: The role of musical practice. Hearing Research. 2014;308:122–8. pmid:24035820
- 184. Schön D, François C. Musical expertise and statistical learning of musical and linguistic structures. Frontiers in Psychology. 2011;2. pmid:21811482
- 185. Vasuki PRM, Sharma M, Ibrahim RK, Arciuli J. Musicians’ Online Performance during Auditory and Visual Statistical Learning Tasks. Frontiers in Human Neuroscience. 2017;11. pmid:28352223
- 186. Paraskevopoulos E, Kuchenbuch A, Herholz SC, Pantev C. Statistical learning effects in musicians and non-musicians: An MEG study. Neuropsychologia. 2012;50(2):341–9. pmid:22197571
- 187. François C, Schön D. Musical expertise boosts implicit learning of both musical and linguistic structures. Cerebral Cortex. 2011;21(10):2357–65. pmid:21383236
- 188. Daikoku T, Yumoto M. Musical expertise facilitates statistical learning of rhythm and the perceptive uncertainty: A cross-cultural study. Neuropsychologia. 2020;146. pmid:32649945
- 189. van Zuijen TL, Sussman E, Winkler I, Näätänen R, Tervaniemi M. Auditory organization of sound sequences by a temporal or numerical regularity—a mismatch negativity study comparing musicians and non-musicians. Cognitive Brain Research. 2005;23(2):270–6. pmid:15820634
- 190. Vasuki PRM, Sharma M, Ibrahim R, Arciuli J. Statistical learning and auditory processing in children with music training: An ERP study. Clinical Neurophysiology. 2017;128(7):1270–81. pmid:28545016
- 191. Perruchet P, Pacton S. Implicit learning and statistical learning: one phenomenon, two approaches. Trends in Cognitive Sciences. 2006;10(5):233–8. pmid:16616590