Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Instrumental music training relates to intensity assessment but not emotional prosody recognition in Mandarin

Abstract

Building on research demonstrating the benefits of music training for emotional prosody recognition in nontonal languages, this study delves into its unexplored influence on tonal languages. In tonal languages, the acoustic similarity between lexical tones and music, along with the dual role of pitch in conveying lexical and affective meanings, create a unique interplay. We evaluated 72 participants, half of whom had extensive instrumental music training, with the other half serving as demographically matched controls. All participants completed an online test consisting of 210 Chinese pseudosentences, each designed to express one of five emotions: happiness, sadness, fear, anger, or neutrality. Our robust statistical analyses, which included effect size estimates and Bayesian factors, revealed that music and nonmusic groups exhibit similar abilities in identifying the emotional prosody of various emotions. However, the music group attributed higher intensity ratings to emotional prosodies of happiness, fear, and anger compared to the nonmusic group. These findings suggest that while instrumental music training is not related to emotional prosody recognition, it does appear to be related to perceived emotional intensity. This dissociation between emotion recognition and intensity evaluation adds a new piece to the puzzle of the complex relationship between music training and emotion perception in tonal languages.

Introduction

Accurate recognition of emotional prosody is crucial for effective social interactions [1, 2]. Individuals adept at identifying others’ emotional tones are better prepared for nuanced social engagement [36]. In contrast, those with challenges in understanding emotional prosody often face social complications [e.g., 712]. It is crucial to examine how individuals from varied linguistic backgrounds perceive emotional prosody. Exploring this, particularly the impact of specific training on social well-being through improved prosody recognition, unveils insights into cross-cultural variations in the utilization of acoustic cues, especially among individuals with and without tonal language experience [13, 14].

Recognizing emotional prosody hinges on acoustic features like pitch, intensity, tempo, and voice quality/timbre [1517], which also serve as essential elements in music, such as Western tonal music. Different emotional prosodies are characterized by distinct features: for instance, happy prosody involves moderately high average pitch, high average voice intensity, and a rapid speech rate, whereas sad prosody is marked by low average pitch, low average voice intensity, and a slow speech rate [1821]. Effective recognition of emotional prosody requires the efficient extraction of these acoustic markers and appropriate perceptual analysis. Given the shared acoustic elements between music and speech in conveying emotion [18, 2224], instrumental music training could enhance speech perception [25], as posited by Patel’s OPERA hypothesis [26, 27]. The OPERA hypothesis posits that shared sensory and cognitive mechanisms underlie both music and speech perception [28, 29], with music imposing more demanding conditions on these systems. Therefore, training in one domain (e.g., music listening) could have cross-modal benefits for the other (e.g., prosody recognition).

Furthermore, individuals with musical training and tonal languages listeners may exhibit advantages in perceiving lexical tones and musical pitch by paying greater attention to the relevant sound cues acquired in speech or music [13, 29]. Tonal languages listeners and musicians have an advantage in recognizing and learning lexical tones, exhibiting cross-domain transfer in pitch perception [13, 30]. Pitch plays a crucial role in both speech and music, serving as a predictive factor in the perception of emotional aspects of vocals. In nontonal languages, it is primarily utilized for emotional expression in speech. In contrast, tonal languages use pitch variations not only as emotional markers but also to convey semantic information [3136]. As a result, the task of recognizing emotional prosody in tonal languages becomes more intricate, potentially requiring additional cognitive resources to mitigate interference from semantic aspects during emotion recognition. While music training is associated with enhanced perception of lexical tones in native speakers of tonal languages [37], it remains uncertain whether it is associated with improved perception of emotional prosody in their native language.

Many studies have examined the relationship between music training and emotional prosody recognition in native speakers of nontonal languages, such as English, German, Portuguese, Dutch, and Hebrew [3841] (see [4244], for reviews). Most of these studies utilize a cross-sectional design comparing musically trained and untrained individuals, while a few use a randomized experimental design, often randomly assigning musically untrained individuals to experiment and control groups (see S1 Table for an overview). Cross-sectional studies have yielded mixed findings on this relationship. Some studies demonstrate that adults with at least 6 years of instrumental music training [3841, 4548] and children with 3 years of extracurricular instruction [49] exhibit greater accuracy in identifying emotional prosodies. These benefits, however, do not extend to measures such as reaction times or emotional intensity ratings [39, 40, 50]. Contrarily, other research finds no significant benefit of music training in this domain, even with 5–14 years of instrumental practice [5053]. These divergences may be attributable to varying sample sizes, impacting statistical significance, as well as the quantity of music training [54]. Specifically, studies which observed an effect of music training had larger sample sizes (20 ≤ n ≤ 58), making it easier to detect group differences. Meanwhile, studies which did not observe an effect of music training [50, 51, 53] had smaller sample sizes (10 ≤ n ≤ 13) and thus failed to obtain statistically significant results. This is because sample size is related to statistical significance (p-value); the smaller the sample size, the greater the p-value [55, 56]. However, the discrepancy between Mualem and Lavido’s [52] study and other studies [45, 46] cannot be explained by sample size, and may be due to differences in the amount of music training. It is believed that at least 10 years of mixed music training is necessary to have an effect on auditory emotion recognition [54]. The average years of music training of the participants in the study of Mualem and Lavido was approximately 9 years, which may explain why their performance was not altered. Notably, the above research predominantly concentrates on nontonal languages, leaving a significant gap in understanding the implications for tonal languages like Mandarin, Thai, or Vietnamese.

Recognizing that lexical tones are a common element in both language and music [57], existing studies highlight how language and cultural backgrounds influence the processing of emotional cues. Notable differences exist in how listeners of tonal and nontonal languages employ acoustic cues to interpret these tones [58, 59]. Furthermore, East Asians are more sensitive to verbal cues of an emotional utterance, whereas Westerners rely more on semantic cues for understanding and interpreting emotional information [6062]. While longitudinal experimental studies have demonstrated that musical training enhances the recognition of emotional prosody in children [46, 63] and young adults [52, 64] but not in older adults [65, 66] who speak nontonal languages, the link between musical training and the perception of emotional prosody in speakers of tonal languages remains insufficiently established. To fill this gap in current research, the present study aims to explore the connection between music training and the recognition of emotional prosody in tonal languages. For this purpose, undergraduate students majoring in music performance or musicology, all of whom had received instrumental training since childhood, were compared with students without professional instrumental training. The focus was on their ability to recognize emotional prosody—specifically happiness, sadness, fear, anger, and neutrality—in Mandarin pseudosentences. Building on insights from existing studies [3841], we hypothesized that music majors would excel in accurately identifying Mandarin emotional prosody compared to nonmusic majors.

Method

Participants

This study was approved by the Ethics Committee of the Harbin Conservatory of Music on 04/06/2023 and all participants provided verbal informed consent form and received monetary compensation for their participation. To achieve 90% power to detect a large main effect (d = 0.8) of group (music vs. nonmusic) at the 0.05 level, a minimum sample size of 34 for each group was required using G*Power (Version 3.1.9.7) [67]. Consequently, 36 music students and 36 nonmusic students who were native Mandarin speakers were recruited for the study from 04/08/2023 to 05/29/2023. The music students had been playing their chosen instruments for a minimum of 12 years, with an average of 13.9 hr of practice per week, or at least 3 hr of practice per day [54, 68]. The instruments covered a wide range, including piano (n = 22), violin (n = 4), double bass (n = 1), viola (n = 1), flute (n = 1), clarinet (n = 1), oboe (n = 1), harp (n = 1), yangqin (n = 1), trumpet (n = 1), suona (n = 1), and erhu (n = 1). In contrast, the nonmusic students had only received compulsory music education in school. To ensure differences in self-reported musical skills and behaviors between music and nonmusic students, the Goldsmiths Musical Sophistication Index (Gold-MSI) [69] was administered. As shown in Table 1, there is moderate evidence that the two groups were matched for age, sex (Fisher’s exact test), years of education, nonverbal IQ as assessed by the Matrix Reasoning Item Bank (MaRs-IB) [70], and handedness as assessed by the Edinburgh Handedness Inventory [71]. However, there is very strong evidence that the music group had more years of training than the nonmusic group and scored higher on all Gold-MSI subtests as well as on the general musical sophistication factor. All participants had normal hearing and no history of neurological or psychiatric disorders.

Stimuli

This study utilized emotional prosody materials drawn from the Chinese vocal emotional stimuli database [72]. The database features 874 pseudosentences, produced by four native Mandarin-Chinese speakers (two male), each conveying one of seven emotions (happiness, sadness, fear, anger, disgust, pleasant surprise, and neutrality) through prosodic variation in 35 unique utterances. Each utterance, ranging from 1 to 2 s in duration, adhered to Chinese grammatical rules while remaining semantically meaningless to eliminate the influence of semantics on emotional prosody perception [7274]. The stimuli were validated with a group of native Mandarin speakers who participated in a seven-option forced choice task to identify the expressed emotion and a 5-point Likert item (1 = very weak, 5 = very intense) to rate its intensity, excluding neutrality. Stimuli meeting satisfactory identification accuracy criteria were included in the database; specifically, those exceeding a 42.86% average recognition rate (3 times the chance level) for the target emotion and remaining below 42.86% for any nontarget emotion [72, 75, 76].

Twenty-one utterances each expressing happiness, sadness, fear, anger and neutrality were selected from recordings of a male and a female speaker, yielded 42 stimuli per emotion. Table 2 presents the recognition rate and intensity rating for each emotion except neutrality. A one-way analysis of variance (ANOVA) was conducted on the average accuracy rates. Although normality assumptions were met (Ws ≥ 0.94, ps ≥ .200), sphericity was violated (W = 0.26, p = .003). Therefore, the degrees of freedom were corrected using the Greenhouse-Geisser method ( = 0.709). The analysis revealed moderate evidence for a main effect of emotion with a large effect size, F(2.83, 56.70) = 5.87, ωp2 = 0.185, 95% CI [0.018, 0.343], BF = 6.09. Pairwise comparisons (see S2 Table) showed that happiness was recognized less accurately than sadness, anger, and neutrality. No differences were found between any other emotion pairs. Similar analyses were conducted for the average intensity ratings. The data met normality assumptions (Ws ≥ 0.94, ps ≥ .223) but violated sphericity (W = 0.55, p = .048). ANOVA with the Greenhouse-Geisser correction ( = 0.703) revealed very strong evidence for a main effect of emotion with a large effect size, F(2.11, 42.20) = 14.50, ωp2 = 0.386, 95% CI [0.150, 0.557], BF = 1.52 × 103. Pairwise comparisons (see S2 Table) indicated that happiness was rated as less intense than sadness, fear, and anger, while sadness was rated as more intense than fear. No differences were found between anger and sadness or fear.

thumbnail
Table 2. Acoustic and perceptual measures for each emotion category averaged across two speakers (n = 21).

https://doi.org/10.1371/journal.pone.0309432.t002

To assess whether the main acoustic features (mean fundamental frequency [f0 mean], f0 range, mean amplitude [Amp mean], mean harmonics-to-noise ratio [HNR mean], HNR variation [HNR SD], and speech rate) of the stimuli (see Table 2) could predict the intended emotion categories, we conducted a discriminant analysis following the procedures of Liu and Pell [72] and Paulmann and Uskul [77]. Using IBM SPSS Statistics (Version 29.0.2), we treated the main acoustic features as predictor variables and the intended emotions as the outcome variable for the pseudosentences. The analysis revealed that 84.76% (178/210) of the pseudosentences were correctly classified in terms of their intended emotion: happiness 78.57% (33/42), sadness 97.62% (41/42), fear 76.19% (32/42), anger 80.95% (34/42), and neutrality 90.48% (38/42). These results suggest that listeners can reliably identify the intended emotions based on the analyzed acoustic features.

Measures

MaRs-IB.

The MaRs-IB was employed to measure participants’ nonverbal IQ [70]. This open-access online assessment has demonstrated good reliability and validity and is similar to Raven’s Advanced Progressive Matrices [78]. The test consists of 80 items, each featuring a 3×3 matrix in which eight out of nine cells contain abstract shapes that vary along four dimensions: color, size, shape and position. Participants must select the missing shape from an array of four options. Each item answered correctly is worth one point, while incorrect items are worth zero points. A higher score or percentage of accuracy is indicative of a stronger spatial reasoning ability and a higher nonverbal IQ [79].

Gold-MSI.

The Gold-MSI [69] is a questionnaire designed to assess individual differences in self-reported musical skills and behaviors among both musicians and the general population. It consists of 38 items divided into five subscales: Musical Training (7 items), Perceptual Abilities (9 items), Singing Abilities (7 items), Active Engagement (9 items), and Emotions (6 items). Respondents are asked to rate their agreement with each statement on a 7-point Likert scale (from 1 strongly disagree to 7 strongly agree) for the first 31 items or select one of seven options for the remaining 7 items. The scores for each dimension are then summed, with higher scores indicating a higher level of musical skills and greater musical engagement or emotional experiences. Additionally, the General Sophistication consists of 18 items, with higher scores indicating a greater degree of musical sophistication. The Chinese version of the Gold-MSI has been found to demonstrate high reliability and validity [80, 81] and was used in the present study. In the current study, the reliability of the Musical Training, Perceptual Abilities, Singing Abilities, Active Engagement, Emotions, and General Sophistication factors, as evaluated by McDonald’s ω, was 0.98 (95% CI [0.97, 0.99]), 0.95 (95% CI [0.93, 0.96]), 0.94 (95% CI [0.91, 0.96]), 0.93 (95% CI [0.91, 0.96]), 0.84 (95% CI [0.78, 0.90]), and 0.97 (95% CI [0.96, 0.98]) respectively.

Procedure

Participants completed the online experiment through the NAODAO platform (www.naodao.com) between 04/15/2023 and 06/06/2023. Before the experiment began, participants were required to read and fully understand the informed consent form. The experiment only commenced automatically upon their explicit confirmation of participation. To ensure participants were familiar with the procedure, a practice session preceded the formal experiment. This practice session included four practice stimuli and feedback on accuracy. Additionally, participants could adjust the headphone volume to a comfortable level.

Fig 1 displays the design of the trial in the experiment. At the start of each trial, a red fixation cross appeared in the center of the screen for 0.5 s to indicate that a stimulus was about to come. The fixation cross then turned green as the stimulus began to play for approximately 1–2 s. After listening to the stimulus, a response interface appeared, and participants were asked to click on one of the five emotional prosody labels and icons to match the emotion expressed in the speech as quickly and accurately as possible. Subsequently, a 7-point Likert item of emotion intensity (1 = very low, 7 = very high) was presented on the screen, and participants were asked to rate the intensity of the emotion in the stimulus without a time limit. To control for potential sex-related differences in vocal acoustic features [82, 83] and reduce participants’ cognitive load from sudden change in speaker sex, we presented sentences spoken by male and female speakers in two separate blocks, with 105 randomly presented sentences per block. The order of the blocks was counterbalanced across participants, with a 1–2 min break between the two blocks.

Data analysis

Before data analysis, trials with response times (RTs) falling below 0.2 s (anticipatory responses or fast guesses) or exceeding 2 SD above the mean for each participant (distraction) were excluded as outliers. Trials with errors were also omitted for RT and intensity ratings. However, in speeded decision tasks, a pervasive speed-accuracy tradeoff (SAT) [84, 85] exists. Analyzing mean RT or accuracy rates separately could, therefore, lead to contradictory or invalid conclusions [8688]. To control for this potential SAT, a composite metric incorporating speed and accuracy should be used for performance evaluation. Composite metrics offer more sensitive and reliable assessments than relying solely on accuracy or RT measures [89, 90].

Several such metrics exist, including inverse efficiency score, rate correct score, linear integrated speed-accuracy score, and balanced integration score (BIS). Among these, only the BIS exhibits relatively insensitivity to SAT [91, 92]. To minimize the SAT effect and preserve “true” effects, we opted for the BIS, assigning equal weights to speed and accuracy. The BIS is calculated by subtracting the standardized (i.e., z scores) mean correct RT from the standardized mean proportions of correct responses across all conditions and participants [9193]. Previous studies have widely adopted this metric [e.g., 94102]. Using this approach, participants achieving high accuracy at the expense of RT or fast RTs at the expense of accuracy will have BISs near zero compared to the sample average. Participants with both high accuracy and fast RTs will have higher positive BISs, while those with low accuracy and slow RTs will have lower negative BISs. Therefore, higher BIS represents faster and more accurate performance. To investigate the SAT in this study, we first analyzed accuracy and RT data. If an SAT is identified, we will subsequently analyze the data using the BIS framework. Both the frequentist and Bayesian approaches were applied to the data.

Frequentist analysis.

In frequentist analysis, effect sizes and confidence intervals are used to evaluate the magnitude and precision of estimation of an effect and to draw conclusions, instead of relying on p-values from traditional null hypothesis significance tests [103108]. This method of interpretation is endorsed by the American Psychological Association [109, 110] and several journals [111113] and researchers [114117]. If the confidence interval for an effect size does not include zero, the effect is considered to be present; while if the confidence interval (CI) does include zero, the effect is considered to be nonexistent [118121]. Additionally, effect sizes can be categorized as small, medium, or large based on established thresholds: g or d = 0.20, 0.50, and 0.80 [122], r = 0.10, 0.30, and 0.50 [123], and ω2 = 0.01, 0.06, and 0.14 [124]. This study employed effect sizes and their confidence intervals to interpret the results.

We analyzed the demographic variables of age, years of education, MaRs-IB, handedness, and Gold-MSI using Welch’s t-tests with JASP (Version 0.19.0). Hedges’ g* and their 95% CIs [125] were calculated using the deffectsize package (Version 0.0.0.9000) in R (Version 4.4.1) and RStudio (Version 2024.04.2+764). For sex, we used Fisher’s exact test in JASP and calculated log OR and its 95% CI. For years of music training, we performed a one-sample t-test using the statsExpressions package (Version 1.5.5) and calculated g and its 95% CI.

For each perceptual feature, we conducted an ANOVA with emotion (happiness, sadness, fear, anger, and/or neutrality) as a within-items variable using jamovi (Version 2.5.6). Prior to the ANOVAs, we tested for normality with the Shapiro-Wilk test and sphericity with Mauchly’s test. If assumptions were met, parametric tests were conducted. If any assumption was violated, the Greenhouse-Geisser correction () was applied to adjust the degrees of freedom.

BIS was used to assess overall task performance, but we also conducted a correlation analysis between accuracy and RT to test the SAT. We used repeated measures correlation [126128] implemented in the rmcorr package (Version 0.7.0). The effect size (correlation coefficient) and its 95% CI were estimated using the bootstrapping with the suggested number of 10000 bootstrap resamples [129132].

For accuracy, RT, BIS and emotional intensity, we conducted a two-way mixed design ANOVA with group (music and nonmusic) as the between-subjects variable and emotion (happiness, sadness, fear, anger, and neutrality) as the within-subjects variable. Data were tested for normality, homogeneity of variance with Levene’s test, and sphericity using JASP. Parametric tests were conducted if assumptions were met; otherwise, nonparametric tests were conducted with ARTool (Version 0.11.1). Post hoc comparisons with Bonferroni correction were performed using the ART-C program for significant interaction effects. We used the effectsize package (Version 0.8.9) to calculate ωp2 and its 95% CI for each ANOVA effect, and d and its 95% CI for each pairwise or post hoc comparison.

Bayesian analysis.

To assess the strength of the evidence and obtain more conservative results, we conducted Bayesian hypothesis testing. The Bayes factor (BF) was calculated to quantify the relative plausibility of the null hypothesis (H0) and alternative hypothesis (H1) given the observed data [133136]. Conventionally, a BF of 1–3 indicates weak evidence for H1, 3–10 indicates moderate evidence, 10–30 indicates strong evidence, and above 30 indicates very strong evidence for H1. Conversely, BFs between 0.33–1, 0.10–0.33, 0.03–0.10, and below 0.03 indicate weak, moderate, strong, and very strong evidence for H0, respectively [137, 138]. While acknowledging the inherent subjectivity in interpreting evidential strength, this study adopted the common practice of considering moderate evidence as a sufficient threshold for drawing meaningful conclusions. For example, a BF of 5 indicates that the data are 5 times more likely under H1 than H0, implying substantial support for the presence of an effect.

For age, years of education, MaRs-IB, handedness and Gold-MSI, we employed the Bain module in JASP to perform Bayesian Welch’s t-tests [139] and calculate BFs. For sex, we employed JASP to perform a Bayesian contingency table test based on an independent multinomial sampling scheme [140] and obtain the BF. For years of music training, we used JASP to execute a Bayesian one-sample t-test with the default prior (Cauchy r = 0.707) and determine the BF.

For perceptual features, accuracy, RT, BIS, or emotional intensity, we calculated BFs using anovaBFcalc (Version 0.1.0) with default α = −0.5 (MWS method) based on F-values and degrees of freedom obtained from frequentist analysis. For pairwise or post hoc comparisons, we computed BFs using the BayesFactor package (Version 0.9.12–4.7) with a prior scale factor of r = 0.707, based on t values and sample sizes obtained from frequentist analysis.

Results

We removed 24 music group trials (0.16%) and 86 nonmusic group trials (0.57%) with RT < 0.2 s. Subsequently, we discarded 312 music group trials (2.08%) and 368 nonmusic group trials (2.45%) with RT > (M + 2 SD) s. Finally, we eliminated 753 music group trials (5.25%) and 1742 nonmusic group trials (12.16%) due to error responses.

Accuracy and response time

Regarding the accuracy data, all dataset except for happiness in the nonmusic group (W = 0.95, p = .146) deviated from normality (Ws ≤ 0.93, ps ≤ .025). Additionally, all datasets except for happiness (F(1, 70) = 2.40, p = .126) violated the assumptions of homogeneity of variance (Fs(1, 70) ≥ 5.01, ps ≤ .028) and sphericity (W = 0.54, p < .001). As for the RT data, all datasets except for happiness (W = 0.95, p = .138) and sadness (W = 0.95, p = .121) in the nonmusic group violated the normality assumption (Ws ≤ 0.93, ps ≤ .024). Furthermore, all RT datasets violated the assumptions of homogeneity of variance (Fs(1, 70) ≥ 4.47, ps ≤ .038) and sphericity (W = 0.65, p < .001).

Fig 2 illustrates the accuracy and RTs of the music and nonmusic groups across all emotions. It is evident that the music group exhibited higher accuracy but slower RTs compared to the nonmusic group for all emotions. To validate these visual observations, separate two-way ANOVAs were conducted on the accuracy and RT data. There was very strong evidence for a large main effect of group on accuracy (F(1,70) = 62.10, ωp2 = 0.459, 95% CI = [0.290, 0.589], BF = 3.13 × 108), indicating that the music group (M = 89.47, SD = 11.67) performed better than the nonmusic group (M = 75.39, SD = 20.34). Similarly, there was very strong evidence for a large main effect of group on RTs (F(1,70) = 14.33, ωp2 = 0.156, 95% CI = [0.033, 0.312], BF = 59.08), with the music group exhibiting slower RTs (M = 1.28, SD = 0.60) than the nonmusic group (M = 0.89, SD = 0.37). These results suggest the presence of an SAT for both music and nonmusic subjects.

thumbnail
Fig 2.

Accuracy (A) and response time (B) as a function of group and emotion. These raincloud plots, combining a halved violin plot, a boxplot, and jittered individual data points, depict the data distribution. They were generated using ggplot2 (Version 3.5.1), PupillometryR (Version 0.0.5), ggpp (Version 0.5.8–1), and cowplot (Version 1.1.3) in R.

https://doi.org/10.1371/journal.pone.0309432.g002

Fig 3 depicts the speed–accuracy correlations for the combined group and each individual group across all emotions. A moderate negative correlation between RT and accuracy was observed for the combined group, r(287) = −0.397, 95% CI [−0.485, −0.306], consistent with the SAT phenomenon. This correlation remained similar for the music group, r(143) = −0.366, 95% CI [−0.504, −0.204], but was larger for the nonmusic group, r(143) = −0.510, 95% CI [−0.616, −0.392]. These findings suggest that faster RT generally come at the expense of reduced accuracy within individuals, highlighting the suitability of BIS as a more comprehensive measure of task performance.

thumbnail
Fig 3. Scatterplots depicting correlations between response time and accuracy for the combined group and each group across emotion.

Each dot represents the average response time and accuracy for a specific emotion. Participant identity is indicated by color, and colored lines show the repeated measures correlation fits for each participant’s data. These scatterplots were created using the following R packages: rmcorr (Version 0.7.0), ggplot2 (Version 3.5.1), pals (Version 1.9), and cowplot (Version 1.1.3).

https://doi.org/10.1371/journal.pone.0309432.g003

Balanced integration score

All but one dataset (happiness for the nonmusic group: W = 0.96, p = .281) deviated from normality (Ws ≤ 0.94, ps ≤ .037). Homogeneity of variance was confirmed for all datasets (Fs(1, 70) ≤ 0.76, ps ≥ .388), but the data violated the assumption of sphericity (W = 0.62, p < .001).

Fig 4A illustrates the BIS of the music and nonmusic groups across all emotions, suggesting comparable performance. A nonparametric ANOVA corroborated these observations. There was substantial evidence for the absence of a main effect of group, F(1,70) = 0.14, ωp2 = 0.000, 95% CI = [0.000, 0.000], BF = 0.103, indicating that the music group (M = 0.03, SD = 1.40) did not outperform the nonmusic group (M = −0.03, SD = 1.19). In contrast, there was very strong evidence for a large main effect of emotion, F(4, 280) = 34.62, ωp2 = 0.321, 95% CI = [0.230, 0.396], BF = 8.41 × 1019. Pairwise comparisons (see S3 Table) showed that participants performed best in detecting neutral emotional prosody (M = 0.74, SD = 1.19), followed by sadness (M = 0.19, SD = 1.20), happiness (M = 0.17, SD = 1.08), or anger (M = −0.08, SD = 1.40), and worst in detecting fear (M = −1.03, SD = 1.70). Notably, there was very strong evidence for the absence of an interaction effect between group and emotion, F(4, 280) = 2.07, ωp2 = 0.015, 95% CI = [0.000, 0.041], BF = 2.21 × 10−3. These findings suggest that instrumental training is not associated with the ability to recognize emotional prosody in tonal languages.

thumbnail
Fig 4.

Balanced integration score (A) and intensity ratings (B) as a function of group and emotion. These raincloud plots, combining a halved violin plot, a boxplot, and jittered individual data points, depict the data distribution. They were generated using ggplot2 (Version 3.5.1), PupillometryR (Version 0.0.5), ggpp (Version 0.5.8–1), and cowplot (Version 1.1.3) in R.

https://doi.org/10.1371/journal.pone.0309432.g004

Intensity ratings

All but four datasets (neutrality for the music group, and sadness, anger, and neutrality for the nonmusic group: Ws ≤ 0.90, ps ≤ .004) exhibited normality (Ws ≥ 0.94, ps ≥ .060). Homogeneity of variance was met for all datasets (Fs(1, 70) ≤ 3.02, ps ≥ .087), but the assumption of sphericity was violated (W = 0.35, p < .001).

Fig 4B demonstrates the intensity ratings of the music and nonmusic groups across all emotions, suggesting higher ratings for the music group. A nonparametric ANOVA revealed moderate evidence for a medium main effect of group, F(1,70) = 8.47, ωp2 = 0.094, 95% CI = [0.006, 0.240], BF = 4.92. This indicates that the music group (M = 5.07, SD = 0.85) assigned higher intensity ratings than the nonmusic group (M = 4.68, SD = 0.89). Additionally, a very strong evidence for a large main effect of emotion emerged, F(4, 280) = 74.55, ωp2 = 0.508, 95% CI = [0.428, 0.571], BF = 3.24 × 1039. Pairwise comparisons (see S3 Table) showed that angry emotional prosody received the highest rating (M = 5.47, SD = 0.79), followed by happy (M = 5.05, SD = 0.71) and sad (M = 4.97, SD = 0.84) or fearful (M = 4.81, SD = 0.75), with neutral receiving the lowest rating (M = 4.09, SD = 0.77). Furthermore, moderate evidence was found for a medium interaction effect between group and emotion, F(4, 280) = 6.15, ωp2 = 0.067, 95% CI = [0.014, 0.121], BF = 4.80. Post hoc comparisons (Table 3) revealed very strong evidence for higher ratings by the music group for happy emotion, moderate evidence for higher ratings for fearful and angry emotions, similar ratings for neutral emotion, and only weak evidence for higher ratings for sad emotion. These findings suggest that instrumental training is associated with heightened sensitivity to the intensity of emotional prosodies in tonal languages.

thumbnail
Table 3. Post hoc comparisons for the Group × Emotion interaction on intensity ratings.

https://doi.org/10.1371/journal.pone.0309432.t003

Discussion

This study uniquely explores the link between instrumental music training and emotional prosody identification in tonal languages, focusing on Mandarin. Utilizing robust analysis with effect size estimates and Bayesian factors, we draw strong conclusions based on our sizable sample of 72 participants. Interestingly, the results revealed no difference in emotional prosody identification between musically trained and untrained participants. However, music majors assigned higher intensity ratings to happy, fearful, and angry prosodies. These findings suggest that music training is not positively correlated with an improved ability to recognize emotional prosody in tonal languages but is instead associated with perceived emotional intensity. This adds a new dimension to our understanding of an intricate relationship between music training and emotion perception in tonal languages, distinct from research focused on nontonal languages.

Utilizing the BIS to mitigate speed-accuracy tradeoff effects in our data, we discovered that music and nonmusic majors exhibited similar proficiency in recognizing emotional prosodies. This outcome contrasts with previous studies suggesting that instrumentalists [3841, 48] and mixed musicians [45, 46, 49] exhibit superior performance on emotional prosody recognition in nontonal languages. This inconsistency may be explained by the intricate interplay between music and language experience in shaping emotional prosody recognition. Indeed, prior research has demonstrated that musicians with a nontonal language background (i.e., English) outperform their nonmusician counterparts in tasks involving linguistic pitch perception [141, 142] and lexical tone perception [14, 143145]. However, this advantage is not as pronounced among musicians with a tonal language background such as Mandarin [146], Cantonese [147], and Thai [144, 148]. The null result observed in this study may be related to the characteristics of the music subjects. Specifically, the music subjects in our study are more likely to possess absolute pitch—the ability to name a tone without reference—due to their experience with a tonal language. Although we did not directly assess absolute pitch ability, previous research indicates that 53%-72% of Chinese music students possess absolute pitch [149152]. Given that emotional prosody perception primarily relies on relative pitch processing [153, 154], which focuses on the pitch changes and patterns rather than absolute pitch, the presence of absolute pitch ability may not necessarily confer an advantage in emotional prosody recognition tasks. In contrast, among musical subjects who spoke a nontonal language, only 3%-10% have absolute pitch [149, 151, 155, 156], while the majority possess relative pitch. This heightened capacity for relative pitch processing [157, 158] likely contributes to their enhanced recognition of emotional prosody.

We also discovered that music majors gave higher intensity ratings to happy, fearful, and angry emotional prosody compared to nonmusic majors. This observation aligns with previous evidence that musicians provided higher intensity ratings for happy emotional prosody in English than did nonmusicians [159]. However, it is contradictory to the studies reporting no differences in intensity or arousal ratings for emotional prosody in Portuguese or German between musicians and nonmusicians [40, 160]. The discrepancy may be attributed to the difference in sample sizes; our study included 36 music and 36 nonmusic subjects, whereas Dibben et al.’s and Lima and Castro’s studies involved 17 music and 15 nonmusic subjects, and 20 music and 20 nonmusic subjects, respectively. The larger sample size in our study, being at least 1.8 times larger than those of the other studies, likely facilitated a more sensitive detection the effect of music training. Our findings might be explained by enhanced auditory processing and selective attention abilities in music subjects. Through extensive instrument training and practice, music subjects develop advanced auditory skills that make them more sensitive to subtle changes in acoustic features (e.g., pitch, intensity, and tempo) shared by speech and music. Studies show musicians excel in pitch [161165], intensity [166], and tempo [167, 168] discrimination compared to nonmusicians. High-arousal emotional speech (e.g., happiness, fear, and anger), often characterized by high pitch and/or large pitch variability, rapid speech rate, and/or increased intensity [18, 21], exhibits more pronounced shifts in these features. As a result, even slight increases in these features might be more salient to music subjects, potentially leading to higher intensity ratings. Furthermore, high-arousal emotional prosodies tend to capture more attention than low-arousal ones. Musicians typically demonstrate superior auditory attentional capacities relative to nonmusicians [169171]. This heightened selective attention allows them to allocate more cognitive resources to processing emotional information in high-arousal speech prosody. This deeper processing amplifies the perceived intensity of the emotion. This dual effect of enhanced perception and attentional capacity may result in a heightened emotional experience, making high-arousal emotions seem more intense than those with low arousal.

Music majors in our study displayed similar levels of emotion recognition as nonmusic majors but provided higher intensity ratings for emotional prosodies in Mandarin. This dissociation pattern may arise from differences in task demands between emotion recognition and intensity ratings. Emotion recognition tasks, such as categorizing emotions, are generally regarded as explicit tasks; intensity rating tasks are seen as more implicit [172174]. This distinction stems from the fact that categorizing emotions require direct access to a stable internal representation of a stimulus and a comparison of the presented stimulus with mental prototypes of each emotion. This process necessitates higher activation levels and a conscious representation of the stimulus. On the other hand, rating the intensity of an emotion may not require verbal categorization of a given emotion and relies on a weaker internal representation of the stimulus, with less precise knowledge about this stimulus being necessary for a more global judgement [173175]. The dissociation between emotion categorization and emotion intensity judgment has been documented in the processing of musical [173, 175177] and vocal [174, 178] emotion. Thus, our findings suggest that instrumental music training may primarily relate to the implicit, rather than explicit, processing of emotional prosody in tonal languages. This cross-domain transfer likely stems from a statistical learning mechanism employed in music and language acquisition. Statistical learning refers to a domain-general perceptual mechanism where learners unconsciously extract statistical patterns or regularities from sensory input over time, without explicit instruction [29, 179]. This type of learning is particularly relevant for implicit knowledge acquisition that relies on specific acoustic cues [180]. Therefore, even passive exposure to music allows both music and nonmusic subjects to acquire implicit knowledge about the statistical regularities between acoustic cues and musical emotions [181, 182]. On the basis of this implicit knowledge, all subjects are likely to show heightened attention to these cues in emotional prosody and make intensity judgments. Nonetheless, music training enhances music subjects’ sensitivity to the acoustic cues [183, 184], leading to a superior ability to attend to them compared to nonmusic subjects. This heightened sensitivity arises from the focused and intensive nature of music training, which makes music subjects particularly adept at detecting the underlying statistical regularities. Indeed, research demonstrates that musicians outperform nonmusicians in auditory statistical learning tasks involving pitch [161, 185187] and rhythm [188190]. Considering the implicit nature of statistical learning [191], its impact is more likely to be observed in implicit tasks rather than explicit ones.

This study has several limitations that warrant careful consideration. First, the quasi-experimental design precludes establishing a causal link between instrumental music training and emotional prosody processing. Future studies should employ randomized experimental designs to address this issue. Second, limited sample size likely contributed to the weakness of evidence for some findings. Consequently, subsequent future studies should use larger sample sizes to obtain more reliable results. Third, this investigation utilized only one tonal language, raising questions about the generalizability of the findings to other tonal languages. Future research is needed to explore this broader applicability. Finally, the mechanisms by which instrumental music training is related to the implicit processing of emotional prosody remain unclear and merit further exploration.

Conclusion

Despite the limitations outlined above, our findings do not support an association between instrumental music training and enhanced recognition of emotional prosody in tonal languages. However, they do suggest a link between music training and participants’ tendency to assign higher intensity ratings to emotional prosody. This dissociation between emotion recognition and emotion intensity evaluation underscores the complexity of the relationship between music training and emotion perception in tonal languages.

The present study is the first to investigate the relationship between music training and emotional prosody perception in Mandarin, a representative of tonal language. Future research should examine this question using different tonal languages to test the reliability of our results and deepen the understanding of the relationship. Such research has important implications for educators and clinicians, as it may provide a new method to treat impaired implicit processing of emotional prosody.

Supporting information

S1 Table. Overview of selected study characteristics for musically trained subjects.

https://doi.org/10.1371/journal.pone.0309432.s001

(XLSX)

S2 Table. Pairwise comparisons of emotional prosodies.

https://doi.org/10.1371/journal.pone.0309432.s002

(DOCX)

S3 Table. Pairwise comparisons of emotional prosodies.

All statistics were obtained from the aligned-and-ranked data.

https://doi.org/10.1371/journal.pone.0309432.s003

(DOCX)

Acknowledgments

We are grateful to Dr. Pan Liu from the Department of Psychology at University of Alberta for providing us with the necessary experimental materials.

References

  1. 1. Neves L, Martins M, Correia AI, Castro SL, Lima CF. Associations between vocal emotion recognition and socio-emotional adjustment in children. Royal Society Open Science. 2021;8(11). pmid:34804582
  2. 2. Chronaki G, Wigelsworth M, Pell MD, Kotz SA. The development of cross-cultural recognition of vocal emotion during childhood and adolescence. Scientific Reports. 2018;8(1). pmid:29904120
  3. 3. Alba-Ferrara L, Hausmann M, Mitchell RL, Weis S. The neural correlates of emotional prosody comprehension: Disentangling simple from complex emotion. PLOS ONE. 2011;6(12). pmid:22174872
  4. 4. Grandjean D. Brain networks of emotional prosody processing. Emotion Review. 2021;13(1):34–43.
  5. 5. Scherer KR, Banse R, Wallbott HG. Emotion inferences from vocal expression correlate across languages and cultures. Journal of Cross-Cultural Psychology. 2001;32(1):76–92.
  6. 6. Young AW, Frühholz S, Schweinberger SR. Face and voice perception: Understanding commonalities and differences. Trends in Cognitive Sciences. 2020;24(5):398–410. pmid:32298625
  7. 7. Fridenson-Hayo S, Berggren S, Lassalle A, Tal S, Pigat D, Bölte S, et al. Basic and complex emotion recognition in children with autism: Cross-cultural findings. Molecular Autism. 2016;7(1). pmid:28018573
  8. 8. Schelinski S, von Kriegstein K. The relation between vocal pitch and vocal emotion recognition abilities in people with autism spectrum disorder and typical development. Journal of Autism and Developmental Disorders. 2019;49(1):68–82. pmid:30022285
  9. 9. Gong B, Li Q, Zhao Y, Wu C. Auditory emotion recognition deficits in schizophrenia: A systematic review and meta-analysis. Asian Journal of Psychiatry. 2021;65. pmid:34482183
  10. 10. Hoekert M, Kahn RS, Pijnenborg M, Aleman A. Impaired recognition and expression of emotional prosody in schizophrenia: Review and meta-analysis. Schizophrenia Research. 2007;96(1):135–45. pmid:17766089
  11. 11. Icht M, Zukerman G, Ben-Itzchak E, Ben-David BM. Keep it simple: Identification of basic versus complex emotions in spoken language in individuals with autism spectrum disorder without intellectual disability: A meta-analysis study. Autism Research. 2021;14(9):1948–64. pmid:34101373
  12. 12. Pawełczyk A, Łojek E, Radek M, Pawełczyk T. Prosodic deficits and interpersonal difficulties in patients with schizophrenia. Psychiatry Research. 2021;306. pmid:34673310
  13. 13. Ong JH. Distributional learning of lexical tone and musical pitch by naïve and experienced adult learners [Doctoral dissertation]: Western Sydney University; 2016.
  14. 14. Lee C-Y, Tao L, Bond ZS. Identification of acoustically modified Mandarin tones by native listeners. Journal of Phonetics. 2008;36(4):537–63.
  15. 15. Banse R, Scherer KR. Acoustic profiles in vocal emotion expression. Journal of Personality and Social Psychology. 1996;70(3):614–36. pmid:8851745
  16. 16. Johnstone T, Scherer KR. Vocal communication of emotion. In: Lewis M, Haviland-Jones JM, editors. Handbook of emotions. 2nd ed. New York, NY, US: Guilford Press; 2000. p. 220–35.
  17. 17. Goudbeek M, Scherer K. Beyond arousal: Valence and potency/control cues in the vocal expression of emotion. The Journal of the Acoustical Society of America. 2010;128(3):1322–36. pmid:20815467
  18. 18. Juslin PN, Laukka P. Communication of emotions in vocal expression and music performance: Different channels, same code? Psychological Bulletin. 2003;129(5):770–814. pmid:12956543
  19. 19. Juslin PN, Laukka P. Impact of intended emotion intensity on cue utilization and decoding accuracy in vocal expression of emotion. Emotion. 2001;1(4):381–412. pmid:12901399
  20. 20. Leitman DI, Laukka P, Juslin PN, Saccente E, Butler P, Javitt DC. Getting the cue: Sensory contributions to auditory emotion recognition impairments in schizophrenia. Schizophrenia Bulletin. 2008;36(3):545–56. pmid:18791077
  21. 21. Scherer KR. Vocal affect expression: A review and a model for future research. Psychological Bulletin. 1986;99(2):143–65. pmid:3515381
  22. 22. Curtis ME, Bharucha JJ. The minor third communicates sadness in speech, mirroring its use in music. Emotion. 2010;10(3):335–48. pmid:20515223
  23. 23. Ilie G, Thompson WF. A comparison of acoustic cues in music and speech for three dimensions of affect. Music Perception. 2006;23(4):319–30.
  24. 24. Coutinho E, Dibben N. Psychoacoustic cues to emotion in speech prosody and music. Cognition and Emotion. 2013;27(4):658–84. pmid:23057507
  25. 25. Thorndike EL, Woodworth RS. The influence of improvement in one mental function upon the efficiency of other functions. (I). Psychological Review. 1901;8(3):247–61.
  26. 26. Patel AD. Why would musical training benefit the neural encoding of speech? The OPERA hypothesis. Frontiers in Psychology. 2011;2. pmid:21747773
  27. 27. Patel AD. Can nonlinguistic musical training change the way the brain processes speech? The expanded OPERA hypothesis. Hearing Research. 2014;308:98–108. pmid:24055761
  28. 28. Besson M, Chobert J, Marie C. Transfer of training between music and speech: Common processing, attention, and memory. Frontiers in Psychology. 2011;2. pmid:21738519
  29. 29. Patel AD. Music, Language, and the Brain: Oxford University Press; 2008 22 Mar 2012.
  30. 30. Ong JH, Burnham D, Escudero P. Distributional learning of lexical tones: A comparison of attended vs. unattended listening. PLOS ONE. 2015;10(7). pmid:26214002
  31. 31. Asaridou SS, Hagoort P, McQueen JM. Effects of early bilingual experience with a tone and a non-tone language on speech-music integration. PLOS ONE. 2015;10(12). pmid:26659377
  32. 32. Chen A, Peter V, Wijnen F, Schnack H, Burnham D. Are lexical tones musical? Native language’s influence on neural response to pitch in different domains. Brain and Language. 2018;180–182:31–41. pmid:29689493
  33. 33. Creel SC. The familiar-melody advantage in auditory perceptual development: Parallels between spoken language acquisition and general auditory perception. Attention, Perception, & Psychophysics. 2019;81(4):948–57. pmid:30635834
  34. 34. Creel SC, Obiri-Yeboah M, Rose S. Language-to-music transfer effects depend on the tone language: Akan vs. East Asian tone languages. Memory & Cognition. 2023;51(7):1624–39. pmid:37052771
  35. 35. Ip MHK, Cutler A. Universals of listening: Equivalent prosodic entrainment in tone and non-tone languages. Cognition. 2020;202. pmid:32502869
  36. 36. Peng G, Zheng H-Y, Gong T, Yang R-X, Kong J-P, Wang WSY. The influence of language experience on categorical perception of pitch contours. Journal of Phonetics. 2010;38(4):616–24.
  37. 37. Ong JH, Wong PCM, Liu F. Musicians show enhanced perception, but not production, of native lexical tones. The Journal of the Acoustical Society of America. 2020;148(6):3443–54. pmid:33379922
  38. 38. Correia AI, Castro SL, MacGregor C, Müllensiefen D, Schellenberg EG, Lima CF. Enhanced recognition of vocal emotions in individuals with naturally good musical abilities. Emotion. 2022;22(5):894–906. pmid:32718172
  39. 39. Farmer E, Jicol C, Petrini K. Musicianship enhances perception but not feeling of emotion from others’ social interaction through speech prosody. Music Perception. 2020;37(4):323–38.
  40. 40. Lima CF, Castro SL. Speaking to the trained ear: Musical expertise enhances the recognition of emotions in speech prosody. Emotion. 2011;11(5):1021–31. pmid:21942696
  41. 41. Twaite JT. Examining relationships between basic emotion perception and musical training in the prosodic, facial, and lexical channels of communication and in music [Doctoral dissertation]: City University of New York; 2016.
  42. 42. Martins M, Pinheiro AP, Lima CF. Does music training improve emotion recognition abilities? A critical review. Emotion Review. 2021;13(3):199–210.
  43. 43. Schellenberg EG, Lima CF. Music training and nonmusical abilities. Annual Review of Psychology. 2024;75:87–128. pmid:37738514
  44. 44. Nussbaum C, Schweinberger SR. Links between musicality and vocal emotion perception. Emotion Review. 2021;13(3):211–24.
  45. 45. Thompson WF, Schellenberg EG, Husain G. Perceiving prosody in speech. Annals of the New York Academy of Sciences. 2003;999(1):530–2. pmid:14681180
  46. 46. Thompson WF, Schellenberg EG, Husain G. Decoding speech prosody: Do music lessons help? Emotion. 2004;4(1):46–64. pmid:15053726
  47. 47. Fuller CD, Galvin JJ, Maat B, Free RH, Başkent D. The musician effect: does it persist under degraded pitch conditions of cochlear implant simulations? Frontiers in Neuroscience. 2014;8. pmid:25071428
  48. 48. Nussbaum C, Schirmer A, Schweinberger SR. Musicality–Tuned to the melody of vocal emotions. British Journal of Psychology. 2023. pmid:37851369
  49. 49. Schellenberg EG, Mankarious M. Music training and emotion comprehension in childhood. Emotion. 2012;12(5):887–91. pmid:22642351
  50. 50. Francisco RMC. Pianos and microphones: Does the type of musical training affect emotion recognition? [Master’s thesis]: Universidade de Lisboa; 2021.
  51. 51. Başkent D, Fuller CD, Galvin JJ III, Schepel L, Gaudrain E, Free RH. Musician effect on perception of spectro-temporally degraded speech, vocal emotion, and music in young adolescents. The Journal of the Acoustical Society of America. 2018;143(5):EL311–EL6. pmid:29857757
  52. 52. Mualem O, Lavidor M. Music education intervention improves vocal emotion recognition. International Journal of Music Education. 2015;33(4):413–25.
  53. 53. Park M, Gutyrchik E, Welker L, Carl P, Pöppel E, Zaytseva Y, et al. Sadness is unique: Neural processing of emotions in speech prosody in musicians and non-musicians. Frontiers in Human Neuroscience. 2015;8. pmid:25688196
  54. 54. Zhang JD, Susino M, McPherson GE, Schubert E. The definition of a musician in music psychology: A literature review and the six-year rule. Psychology of Music. 2020;48(3):389–409.
  55. 55. Gómez-de-Mariscal E, Guerrero V, Sneider A, Jayatilaka H, Phillip JM, Wirtz D, et al. Use of the p-values as a size-dependent function to address practical differences when analyzing large datasets. Scientific Reports. 2021;11(1). pmid:34686696
  56. 56. Whitley E, Ball J. Statistics review 4: Sample size calculations. Critical Care. 2002;6(4):335–41. pmid:12225610
  57. 57. Brown S. The "musilanguage" model of music evolution. In: Wallin NL, Merker B, Brown S, editors. The origins of music: MIT Press; 1999. p. 271–300.
  58. 58. Burnham D, Jones C. Categorical perception of lexical tone by tonal and non-tonal language speakers. 9th International Conference on Speech Science and Technology; Melbourne, VIC, Australia 2002, December 2–5. p. 515–20.
  59. 59. Hallé PA, Chang Y-C, Best CT. Identification and discrimination of Mandarin Chinese tones by Mandarin Chinese vs. French listeners. Journal of Phonetics. 2004;32(3):395–421.
  60. 60. Kitayama S, Ishii K. Word and voice: Spontaneous attention to emotional utterances in two languages. Cognition and Emotion. 2002;16(1):29–59.
  61. 61. Ishii K, Reyes JA, Kitayama S. Spontaneous attention to word content versus emotional tone: Differences among three cultures. Psychological Science. 2003;14(1):39–46. pmid:12564752.
  62. 62. Yang Y, Wang L, Wang Q. Take your word or tone for it? European American and Chinese children’s attention to emotional cues in speech. Child Development. 2021;92(3):844–52. pmid:33881172
  63. 63. Good A, Gordon KA, Papsin BC, Nespoli G, Hopyan T, Peretz I, et al. Benefits of music training for perception of emotional speech prosody in deaf children with cochlear implants. Ear and Hearing. 2017;38(4). pmid:28085739
  64. 64. Bodner E, Aharoni R, Iancu I. The effect of training with music on happiness recognition in social anxiety disorder. Journal of Psychopathology and Behavioral Assessment. 2012;34(4):458–66.
  65. 65. Chari DA, Barrett KC, Patel AD, Colgrove TR, Jiradejvong P, Jacobs LY, et al. Impact of auditory-motor musical training on melodic pattern recognition in cochlear implant users. Otology & Neurotology. 2020;41(4):e422–e31. pmid:32176126
  66. 66. Fuller CD, Galvin JJ, Maat B, Başkent D, Free RH. Comparison of two music training approaches on music and speech perception in cochlear implant users. Trends in Hearing. 2018;22. pmid:29621947.
  67. 67. Faul F, Erdfelder E, Lang A-G, Buchner A. G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods. 2007;39(2):175–91. pmid:17695343
  68. 68. Zhang JD, Schubert E. A single item measure for identifying musician and nonmusician categories based on measures of musical sophistication. Music Perception. 2019;36(5):457–67.
  69. 69. Müllensiefen D, Gingras B, Musil J, Stewart L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. PLOS ONE. 2014;9(2). pmid:24586929
  70. 70. Chierchia G, Fuhrmann D, Knoll LJ, Pi-Sunyer BP, Sakhardande AL, Blakemore S-J. The matrix reasoning item bank (MaRs-IB): Novel, open-access abstract reasoning items for adolescents and adults. Royal Society Open Science. 2019;6(10). pmid:31824684
  71. 71. Oldfield RC. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia. 1971;9(1):97–113. pmid:5146491
  72. 72. Liu P, Pell MD. Recognizing vocal emotions in Mandarin Chinese: A validated database of Chinese vocal emotional stimuli. Behavior Research Methods. 2012;44(4):1042–51. pmid:22539230
  73. 73. Pell MD, Jaywant A, Monetta L, Kotz SA. Emotional speech processing: Disentangling the effects of prosody and semantic cues. Cognition and Emotion. 2011;25(5):834–53. pmid:21824024
  74. 74. Larrouy-Maestri P, Poeppel D, Pell MD. The sound of emotional prosody: Nearly 3 decades of research and future directions. Perspectives on Psychological Science. 2024. pmid:38232303.
  75. 75. Castro SL, Lima CF. Recognizing emotions in spoken language: A validated set of Portuguese sentences and pseudosentences for research on emotional prosody. Behavior Research Methods. 2010;42(1):74–81. pmid:20160287
  76. 76. Liu P, Rigoulot S, Jiang X, Zhang S, Pell MD. Unattended emotional prosody affects visual processing of facial expressions in Mandarin-speaking Chinese: A comparison with English-speaking Canadians. Journal of Cross-Cultural Psychology. 2021;52(3):275–94. pmid:33958813
  77. 77. Paulmann S, Uskul AK. Cross-cultural emotional prosody recognition: Evidence from Chinese and British listeners. Cognition and Emotion. 2014;28(2):230–44. pmid:23862740
  78. 78. Raven J. The Raven Progressive Matrices and measuring aptitude constructs. The International Journal of Educational and Psychological Assessment. 2009;2:2–38. 2010-11467-001.
  79. 79. Zorowitz S, Chierchia G, Blakemore S-J, Daw ND. An item response theory analysis of the matrix reasoning item bank (MaRs-IB). Behavior Research Methods. 2023. pmid:37020082
  80. 80. Li J, Lin H-R, Wolf A, Lothwesen K. Measuring musical sophistication in the Chinese general population: Validation and replication of the Simplified Chinese Gold-MSI. Musicae Scientiae. 2023.
  81. 81. Lin H-R, Kopiez R, Müllensiefen D, Wolf A. The Chinese version of the Gold-MSI: Adaptation and validation of an inventory for the measurement of musical sophistication in a Taiwanese sample. Musicae Scientiae. 2019;25(2):226–51.
  82. 82. Santos AO, Godoy J, Silverio K, Brasolotto A. Vocal changes of men and women from different age decades: An analysis from 30 years of age. Journal of Voice. 2023;37(6):840–50. pmid:34284927
  83. 83. Hillenbrand JM, Clark MJ. The role of f0 and formant frequencies in distinguishing the voices of men and women. Attention, Perception, & Psychophysics. 2009;71(5):1150–66. pmid:19525544
  84. 84. Heitz RP. The speed-accuracy tradeoff: History, physiology, methodology, and behavior. Frontiers in Neuroscience. 2014;8. pmid:24966810
  85. 85. Wickelgren WA. Speed-accuracy tradeoff and information processing dynamics. Acta Psychologica. 1977;41(1):67–85.
  86. 86. Alhowaide A, Alsmadi I, Tang J. Ensemble detection model for IoT IDS. Internet of Things. 2021;16.
  87. 87. Lerche V, Voss A. Speed–accuracy manipulations and diffusion modeling: Lack of discriminant validity of the manipulation or of the parameter estimates? Behavior Research Methods. 2018;50(6):2568–85. pmid:29542062
  88. 88. Lerche V, Voss A. When accuracy rates and mean response times lead to false conclusions: A simulation study based on the diffusion model. TQMP. 2020;16(2):107–19.
  89. 89. Mueller ST, Alam L, Funke GJ, Linja A, Ibne Mamun T, Smith SL. Examining methods for combining speed and accuracy in a go/no-go vigilance task. Proceedings of the Human Factors and Ergonomics Society Annual Meeting. 2020;64(1):1202–6.
  90. 90. Bakun Emesh T, Garbi D, Kaplan A, Zelicha H, Yaskolka Meir A, Tsaban G, et al. Retest reliability of integrated speed–accuracy measures. Assessment. 2022;29(4):717–30. pmid:33522278.
  91. 91. Liesefeld HR, Janczyk M. Combining speed and accuracy to control for speed-accuracy trade-offs(?). Behavior Research Methods. 2019;51(1):40–60. pmid:30022459
  92. 92. Liesefeld HR, Janczyk M. Same same but different: Subtle but consequential differences between two measures to linearly integrate speed and accuracy (LISAS vs. BIS). Behavior Research Methods. 2023;55(3):1175–92. pmid:35595937
  93. 93. Liesefeld HR, Fu X, Zimmer HD. Fast and careless or careful and slow? Apparent holistic processing in mental rotation is explained by speed-accuracy trade-offs. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2015;41(4):1140–51. pmid:25528084
  94. 94. Bouvet L, Simard-Meilleur A-A, Paignon A, Mottron L, Donnadieu S. Auditory local bias and reduced global interference in autism. Cognition. 2014;131(3):367–72. pmid:24637103
  95. 95. Hassan EK, Jones AM, Buckingham G. A novel protocol to induce mental fatigue. Behavior Research Methods. 2023. pmid:37537491
  96. 96. Paas Oliveros LK, Cieslik EC, Pieczykolan A, Pläschke RN, Eickhoff SB, Langner R. Brain functional characterization of response-code conflict in dual-tasking and its modulation by age. Cerebral Cortex. 2023;33(18):10155–80. pmid:37540164
  97. 97. English MCW, Maybery MT, Visser TAW. Magnitude of sex differences in visual search varies with target eccentricity. Psychonomic Bulletin & Review. 2021;28(1):178–88. pmid:32875533
  98. 98. Weissbach A, Moyé J, Takacs A, Verrel J, Chwolka F, Friedrich J, et al. Perception–action integration is altered in functional movement disorders. Movement Disorders. 2023;38(8):1399–409. pmid:37315159
  99. 99. Robson SG, Tangen JM, Searston RA. The effect of expertise, target usefulness and image structure on visual search. Cognitive Research: Principles and Implications. 2021;6. pmid:33709197
  100. 100. Yu Z, Dong Y. The emergence of a complex language skill: Evidence from the self-organization of interpreting competence in interpreting students. Bilingualism: Language and Cognition. 2022;25(2):269–82. Epub 2021/09/28. https://doi.org/10.1017/S1366728921000870
  101. 101. McGarrigle R, Knight S, Hornsby BWY, Mattys S. Predictors of listening-related fatigue across the adult life span. Psychological Science. 2021;32(12):1937–51. pmid:34751602.
  102. 102. McGarrigle R, Mattys S. Sensory-processing sensitivity predicts fatigue from listening, but not perceived effort, in young and older adults. Journal of Speech, Language, and Hearing Research. 2023;66(2):444–60. pmid:36657070
  103. 103. Calin-Jageman RJ, Cumming G. Estimation for better inference in neuroscience. eNeuro. 2019;6(4). pmid:31453316
  104. 104. Calin-Jageman RJ, Cumming G. The new statistics for better science: Ask how much, how uncertain, and what else is known. The American Statistician. 2019;73(sup1):271–80. pmid:31762475
  105. 105. Cohen J. Things I have learned (so far). American Psychologist. 1990;45(12):1304–12.
  106. 106. Nakagawa S, Cuthill IC. Effect size, confidence interval and statistical significance: A practical guide for biologists. Biological Reviews. 2007;82(4):591–605. pmid:17944619
  107. 107. Wasserstein RL, Lazar NA. The ASA statement on p-values: Context, process, and purpose. The American Statistician. 2016;70(2):129–33.
  108. 108. Wasserstein RL, Schirm AL, Lazar NA. Moving to a world beyond “p < 0.05”. The American Statistician. 2019;73(sup1):1–19.
  109. 109. American Psychological Association. Publication manual of the American Psychological Association. 6th ed 2010.
  110. 110. American Psychological Association. Publication manual of the American Psychological Association. 7th ed 2020.
  111. 111. Points of significance. Nature Human Behaviour. 2023;7(3):293–4. pmid:36964301
  112. 112. Harrington DD’Agostino RB, Gatsonis C, Hogan JW, Hunter DJ, Normand S-LT, et al. New guidelines for statistical reporting in the Journal. The New England Journal of Medicine. 2019;381(3):285–6. pmid:31314974.
  113. 113. Trafimow D, Marks M. Editorial. Basic and Applied Social Psychology. 2015;37(1):1–2.
  114. 114. Cumming G. The new statistics: A how‐to guide. Australian Psychologist. 2013;48(3):161–70.
  115. 115. Cumming G. The new statistics: Why and how. Psychological Science. 2014;25(1):7–29. pmid:24220629.
  116. 116. Griffiths P, Needleman J. Statistical significance testing and p-values: Defending the indefensible? A discussion paper and position statement. International Journal of Nursing Studies. 2019;99. pmid:31442781
  117. 117. Karadaghy OA, Hong H, Scott-Wittenborn N, Sinha P, Suko J, Tait S, et al. Reporting of effect size and confidence intervals in JAMA Otolaryngology–Head & Neck Surgery. JAMA Otolaryngology–Head & Neck Surgery. 2017;143(11):1075–80. pmid:28910425
  118. 118. Cheung SF, Pesigan IJA, Vong WN. DIY bootstrapping: Getting the nonparametric bootstrap confidence interval in SPSS for any statistics or function of statistics (when this bootstrapping is appropriate). Behavior Research Methods. 2023;55(2):474–90. pmid:35292932
  119. 119. Perdices M. Null hypothesis significance testing, p-values, effects sizes and confidence intervals. Brain Impairment. 2018;19(1):70–80. Epub 2017/12/07.
  120. 120. Sohn D. Sex differences in achievement self-attributions: An effect-size analysis. Sex Roles. 1982;8(4):345–57.
  121. 121. Steiger JH. Beyond the F test: Effect size confidence intervals and tests of close fit in the analysis of variance and contrast analysis. Psychological Methods. 2004;9(2):164–82. pmid:15137887
  122. 122. Cohen J. Statistical power analysis for the behavioral sciences. 2nd ed. Hillsdale, NJ: Erlbaum; 1988.
  123. 123. Cohen J. A power primer. Psychological Bulletin. 1992;112(1):155–9. pmid:19565683
  124. 124. Kirk RE. Practical significance: A concept whose time has come. Educational and Psychological Measurement. 1996;56(5):746–59.
  125. 125. Delacre M, Lakens D, Ley C, Liu L, Leys C. Why Hedges’ g*s based on the non-pooled standard deviation should be reported with Welch’s t-test 2021.
  126. 126. Bakdash JZ, Marusich LR. Repeated measures correlation. Frontiers in Psychology. 2017;8. pmid:28439244
  127. 127. Bland JM, Altman DG. Calculating correlation coefficients with repeated observations: Part 1—correlation within subjects. BMJ. 1995;310(6977). pmid:7873953
  128. 128. Marusich L, Bakdash J. rmcorrShiny: A web and standalone application for repeated measures correlation. F1000Research. 2021;10. pmid:34621514
  129. 129. Chihara LM, Hesterberg TC. Mathematical statistics with resampling and R. 3rd ed: John Wiley & Sons; 2022.
  130. 130. Hair JF, Hult GTM, Ringle CM, Sarstedt M. A primer on partial least squares structural equation modeling (PLS-SEM). 3rd ed: SAGE Publications; 2022.
  131. 131. Streukens S, Leroi-Werelds S. Bootstrapping and PLS-SEM: A step-by-step guide to get more out of your bootstrap results. European Management Journal. 2016;34(6):618–32.
  132. 132. Hesterberg TC. What teachers should know about the bootstrap: Resampling in the undergraduate statistics curriculum. The American Statistician. 2015;69(4):371–86. pmid:27019512
  133. 133. Brydges CR, Bielak AAM. A Bayesian analysis of evidence in support of the null hypothesis in gerontological psychology (or lack thereof). The Journals of Gerontology: Series B. 2020;75(1):58–66. pmid:30877301
  134. 134. Heck DW, Boehm U, Böing-Messing F, Bürkner P-C, Derks K, Dienes Z, et al. A review of applications of the Bayes factor in psychological research. Psychological Methods. 2023;28(3):558–79. pmid:35298215
  135. 135. Schmalz X, Biurrun Manresa J, Zhang L. What is a Bayes factor? Psychological Methods. 2023;28(3):705–18. pmid:34780246
  136. 136. Wagenmakers E-J, Marsman M, Jamil T, Ly A, Verhagen J, Love J, et al. Bayesian inference for psychology. Part I: Theoretical advantages and practical ramifications. Psychonomic Bulletin & Review. 2018;25(1):35–57. pmid:28779455
  137. 137. Jeffreys H. Theory of probability. 3rd ed. Oxford, UK: Oxford University Press; 1961.
  138. 138. Lee MD, Wagenmakers E-J. Bayesian cognitive modeling: A practical course. New York, NY: Cambridge University Press; 2014. xiii, 264–xiii, p.
  139. 139. Fu Q, Hoijtink H, Moerbeek M. Sample-size determination for the Bayesian t test and Welch’s test using the approximate adjusted fractional Bayes factor. Behavior Research Methods. 2021;53(1):139–52. pmid:32632740
  140. 140. Jamil T, Ly A, Morey RD, Love J, Marsman M, Wagenmakers E-J. Default “Gunel and Dickey” Bayes factors for contingency tables. Behavior Research Methods. 2017;49(2):638–52. pmid:27325166
  141. 141. Wayland R, Herrera E, Kaan E. Effects of musical experience and training on pitch contour perception. Journal of Phonetics. 2010;38(4):654–62.
  142. 142. Wong PCM, Skoe E, Russo NM, Dees T, Kraus N. Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nature Neuroscience. 2007;10(4):420–2. pmid:17351633
  143. 143. Alexander J, Wong P, Bradlow A. Lexical tone perception in musicians and non-musicians. 6th Interspeech 2005 and 9th European Conference on Speech Communication and Technology; Lisbon, Portugal: International Speech Communication Association; 2005, September 4–8. p. 397–400.
  144. 144. Cooper A, Wang Y. The influence of linguistic and musical experience on Cantonese word learning. The Journal of the Acoustical Society of America. 2012;131(6):4756–69. pmid:22712948
  145. 145. Lee C-Y, Lekich A, Zhang Y. Perception of pitch height in lexical and musical tones by English-speaking musicians and nonmusiciansa). The Journal of the Acoustical Society of America. 2014;135(3):1607–15. pmid:24606295
  146. 146. Chen S, Zhu Y, Wayland R, Yang Y. How musical experience affects tone perception efficiency by musicians of tonal and non-tonal speakers? PLOS ONE. 2020;15(5). pmid:32384088
  147. 147. Mok PKP, Zuo D. The separation between music and speech: Evidence from the perception of Cantonese tonesa). The Journal of the Acoustical Society of America. 2012;132(4):2711–20. pmid:23039463
  148. 148. Maggu AR, Wong PCM, Liu H, Wong FCK. Experience-dependent influence of music and language on lexical pitch learning is not additive. 19th Annual Conference of the International Speech Communication Association (Interspeech 2018); Hyderabad, India. Hyderabad, India2018, September 2–6. p. 3791–4.
  149. 149. Deutsch D, Henthorn T, Marvin E, Xu H. Absolute pitch among American and Chinese conservatory students: Prevalence differences, and evidence for a speech-related critical perioda). The Journal of the Acoustical Society of America. 2006;119(2):719–22. pmid:16521731
  150. 150. Deutsch D, Li X, Shen J. Absolute pitch among students at the Shanghai Conservatory of Music: A large-scale direct-test study. The Journal of the Acoustical Society of America. 2013;134(5):3853–9. pmid:24180794
  151. 151. Ki Miyazaki, Rakowski A, Makomaska S, Jiang C, Tsuzaki M, Oxenham AJ, et al. Absolute pitch and relative pitch in music students in the East and the West: Implications for aural-skills education. Music Perception. 2018;36(2):135–55.
  152. 152. Lee C-Y, Lee Y-F. Perception of musical pitch and lexical tones by Mandarin-speaking musicians. The Journal of the Acoustical Society of America. 2010;127(1):481–90. pmid:20058993
  153. 153. Pell MD, Paulmann S, Dara C, Alasseri A, Kotz SA. Factors in the recognition of vocally expressed emotions: A comparison of four languages. Journal of Phonetics. 2009;37(4):417–35.
  154. 154. Breitenstein C, Lancker DV, Daum I. The contribution of speech rate and pitch variation to the perception of vocal emotions in a German and an American sample. Cognition and Emotion. 2001;15(1):57–79.
  155. 155. Ki Miyazaki, Makomaska S, Rakowski A. Prevalence of absolute pitch: A comparison between Japanese and Polish music students. The Journal of the Acoustical Society of America. 2012;132(5):3484–93. pmid:23145628
  156. 156. Leite RBC, Mota-Rolim SA, Queiroz CMT. Music proficiency and quantification of absolute pitch: A large-scale study among Brazilian musicians. Frontiers in Neuroscience. 2016;10. pmid:27790084
  157. 157. Ngo MK, Vu K-PL, Strybel TZ. Effects of music and tonal language experience on relative pitch performance. The American Journal of Psychology. 2016;129(2):125–34. pmid:27424415
  158. 158. Schellenberg EG, Moreno S. Music lessons, pitch processing, and g. Psychology of Music. 2010;38(2):209–21.
  159. 159. Faber S, Fiveash A. Emotion without words: A comparison study of music and speech prosody. Canadian Journal of Music Therapy. 2014;20(2):86–101.
  160. 160. Dibben N, Coutinho E, Vilar JA, Estévez-Pérez G. Do individual differences influence moment-by-moment reports of emotion perceived in music and speech prosody? Frontiers in Behavioral Neuroscience. 2018;12. pmid:30210316
  161. 161. Vasuki PRM, Sharma M, Demuth K, Arciuli J. Musicians’ edge: A comparison of auditory processing, cognitive abilities and statistical learning. Hearing Research. 2016;342:112–23. pmid:27770623
  162. 162. Micheyl C, Delhommeau K, Perrot X, Oxenham AJ. Influence of musical and psychoacoustical training on pitch discrimination. Hearing Research. 2006;219(1):36–47. pmid:16839723
  163. 163. Liang C, Earl B, Thompson I, Whitaker K, Cahn S, Xiang J, et al. Musicians are better than non-musicians in frequency change detection: behavioral and electrophysiological evidence. Frontiers in Neuroscience. 2016;10. pmid:27826221
  164. 164. Kishon-Rabin L, Amir O, Vexler Y, Zaltz Y. Pitch discrimination: Are professional musicians better than non-musicians? Journal of Basic and Clinical Physiology and Pharmacology. 2001;12(2):125–44. pmid:11605682
  165. 165. Sares AG, Foster NEV, Allen K, Hyde KL. Pitch and time processing in speech and tones: The effects of musical training and attention. Journal of Speech, Language, and Hearing Research. 2018;61(3):496–509. pmid:29466555
  166. 166. Slade T, Gascon A, Comeau G, Russell D. Just noticeable differences in sound intensity of piano tones in non-musicians and experienced pianists. Psychology of Music. 2023;51(3):924–37.
  167. 167. Madsen CK. Modulated beat discrimination among musicians and nonmusicians. Journal of Research in Music Education. 1979;27(2):57–67.
  168. 168. Nguyen T, Sidhu RK, Everling JC, Wickett MC, Gibbings A, Grahn JA. Beat perception and production in musicians and dancers. Music Perception. 2022;39(3):229–48.
  169. 169. Strait DL, Kraus N, Parbery-Clark A, Ashley R. Musical experience shapes top-down auditory mechanisms: Evidence from masking and auditory attention performance. Hearing Research. 2010;261(1):22–9. pmid:20018234
  170. 170. Weijkamp J, Sadakata M. Attention to affective audio-visual information: Comparison between musicians and non-musicians. Psychology of Music. 2017;45(2):204–15.
  171. 171. Bialystok E, DePape A-M. Musical expertise, bilingualism, and executive functioning. Journal of Experimental Psychology: Human Perception and Performance. 2009;35(2):565–74. pmid:19331508
  172. 172. Cunningham WA, Raye CL, Johnson MK. Implicit and explicit evaluation: fMRI correlates of valence, emotional intensity, and control in the processing of attitudes. Journal of Cognitive Neuroscience. 2004;16(10):1717–29. pmid:15701224
  173. 173. Lévêque Y, Teyssier P, Bouchet P, Bigand E, Caclin A, Tillmann B. Musical emotions in congenital amusia: Impaired recognition, but preserved emotional intensity. Neuropsychology. 2018;32(7):880–94. pmid:30047757
  174. 174. Pralus A, Fornoni L, Bouet R, Gomot M, Bhatara A, Tillmann B, et al. Emotional prosody in congenital amusia: Impaired and spared processes. Neuropsychologia. 2019;134. pmid:31647961
  175. 175. Pralus A, Belfi A, Hirel C, Lévêque Y, Fornoni L, Bigand E, et al. Recognition of musical emotions and their perceived intensity after unilateral brain damage. Cortex. 2020;130:78–93. pmid:32645502
  176. 176. Hirel C, Lévêque Y, Deiana G, Richard N, Cho TH, Mechtouff L, et al. Amusie acquise et anhédonie musicale. Revue Neurologique. 2014;170(8):536–40. pmid:24856610
  177. 177. Lévêque Y, Roulet-Perez E, Deonna T, Moulin A, Fornoni L, Mayor-Dubois C, et al. Music processing deficits in Landau-Kleffner syndrome: Four case studies in adulthood. Cortex. 2020;129:99–111. pmid:32442777
  178. 178. Bourgeois–Vionnet J, Moulin A, Hermier M, Pralus A, Nighoghossian N. A case of verbal and emotional prosody processing dissociation after a right temporal venous infarct. Neurological Sciences. 2020;41(6):1615–8. pmid:31902013
  179. 179. Frost R, Armstrong BC, Siegelman N, Christiansen MH. Domain generality versus modality specificity: the paradox of statistical learning. Trends in Cognitive Sciences. 2015;19(3):117–25. pmid:25631249
  180. 180. Ong JH, Burnham D, Stevens CJ, Escudero P. Naïve learners show cross-domain transfer after distributional learning: The case of lexical and musical pitch. Frontiers in Psychology. 2016;7. pmid:27551272
  181. 181. Pearce MT. Statistical learning and probabilistic prediction in music cognition: mechanisms of stylistic enculturation. Annals of the New York Academy of Sciences. 2018;1423(1):378–95. pmid:29749625
  182. 182. Bigand E, Poulin-Charronnat B. Are we “experienced listeners”? A review of the musical capacities that do not depend on formal musical training. Cognition. 2006;100(1):100–30. pmid:16412412
  183. 183. François C, Schön D. Neural sensitivity to statistical regularities as a fundamental biological process that underlies auditory learning: The role of musical practice. Hearing Research. 2014;308:122–8. pmid:24035820
  184. 184. Schön D, François C. Musical expertise and statistical learning of musical and linguistic structures. Frontiers in Psychology. 2011;2. pmid:21811482
  185. 185. Vasuki PRM, Sharma M, Ibrahim RK, Arciuli J. Musicians’ Online Performance during Auditory and Visual Statistical Learning Tasks. Frontiers in Human Neuroscience. 2017;11. pmid:28352223
  186. 186. Paraskevopoulos E, Kuchenbuch A, Herholz SC, Pantev C. Statistical learning effects in musicians and non-musicians: An MEG study. Neuropsychologia. 2012;50(2):341–9. pmid:22197571
  187. 187. François C, Schön D. Musical expertise boosts implicit learning of both musical and linguistic structures. Cerebral Cortex. 2011;21(10):2357–65. pmid:21383236
  188. 188. Daikoku T, Yumoto M. Musical expertise facilitates statistical learning of rhythm and the perceptive uncertainty: A cross-cultural study. Neuropsychologia. 2020;146. pmid:32649945
  189. 189. van Zuijen TL, Sussman E, Winkler I, Näätänen R, Tervaniemi M. Auditory organization of sound sequences by a temporal or numerical regularity—a mismatch negativity study comparing musicians and non-musicians. Cognitive Brain Research. 2005;23(2):270–6. pmid:15820634
  190. 190. Vasuki PRM, Sharma M, Ibrahim R, Arciuli J. Statistical learning and auditory processing in children with music training: An ERP study. Clinical Neurophysiology. 2017;128(7):1270–81. pmid:28545016
  191. 191. Perruchet P, Pacton S. Implicit learning and statistical learning: one phenomenon, two approaches. Trends in Cognitive Sciences. 2006;10(5):233–8. pmid:16616590