Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Improved performance with automatic sound management 3 in the MED-EL SONNET 2 cochlear implant audio processor

  • Anja Kurz ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Validation, Writing – original draft, Writing – review & editing

    Kurz_A@ukw.de

    Affiliation Department of Otorhinolaryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery, Comprehensive Hearing Center, University Hospital of Würzburg, Würzburg, Germany

  • Kristen Rak,

    Roles Conceptualization, Supervision, Writing – review & editing

    Affiliation Department of Otorhinolaryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery, Comprehensive Hearing Center, University Hospital of Würzburg, Würzburg, Germany

  • Rudolf Hagen

    Roles Conceptualization, Funding acquisition, Supervision, Writing – review & editing

    Affiliation Department of Otorhinolaryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery, Comprehensive Hearing Center, University Hospital of Würzburg, Würzburg, Germany

Abstract

Objectives

The SONNET 2 audio processor features ambient noise reduction (ANR), transient-noise reduction (TNR), and adaptive intelligence (AI). The primary aim of this study was to evaluate if using these features improves speech perception in noise, subjective listening effort, and sound quality.

Design

In this prospective longitudinal study, twenty adult SONNET users were fitted with the SONNET 2 audio processor, configured either as a default SONNET (no ANR/TNR/AI), with mild ANR/TNR, with strong ANR/TNR, with mild AI, and with strong AI. Speech perception in noise was assessed in speech and stationary noise from the front (S0N0); speech, stationary noise, and transient noise from the front (S0N0T0); and speech from the front in spatially-distributed stationary noise (S0N±45N±135). Listening effort, subjective sound quality, and device/setup preference were assessed.

Results

In the S0N0 setup, speech perception in noise was significantly better with the SONNET 2 when using ANR/TNR in the mild setup than with the SONNET or the SONNET 2 in the default SONNET configuration. In the S0N±45N±135 setup, speech understanding was significantly better in all four SONNET 2 configurations than with the SONNET or the SONNET 2 in the default SONNET configuration (a 1.26–2.55 dB SRT80 benefit). Subjects tolerated consistently lower signal-to-noise values with the SONNET 2 configurations using ANR/TNR than with the default SONNET configuration in all listening effort categories. All SONNET 2 configurations using ANR/TNR were preferred and better rated in speech in stationary and/or transient noise compared to the default SONNET configuration. Sound quality and pleasantness were better in those SONNET 2 configurations. Subjects strongly preferred the SONNET 2 configurations over the SONNET configuration.

Conclusions

The new front-end features implemented in the SONNET 2 audio processor objectively improve speech perception in noise. Subjects preferred the SONNET 2, over the SONNET, in the presence of stationary and transient noise.

Introduction

While cochlear implant (CI) users’ ability to understanding speech has continuously improved over the last two decades [1], it is still significantly more susceptible to background noise compared to that of people with normal hearing [2, 3]. CI users may achieve ceiling-level performance of sentence recognition in quiet [4], however, they need a high signal-to-noise ratio (SNR) to understand speech in the presence of background noise [5]. To reach a 50% speech reception threshold (SRT), SNRs of at least +5dB up to +25dB are necessary [1, 69]. For CI users, difficulties in understanding speech are even more pronounced in the presence of fluctuating noise than in stationary noise or in noise with low temporal variations. This results in a considerable increase in listening effort and an overall loss in quality of life [6, 7, 1012]. Consequently, reducing the amount of noise is an important and enduring issue in the CI field.

Noise reduction algorithms have been implemented in the front-end signal processing of CI audio processors by all major CI manufacturers. MED-EL (Innsbruck, Austria) introduced wind-noise reduction and fixed and adaptive directional processing (i.e., beamforming) with the SONNET audio processor in 2014. This technology enabled significantly improved speech performance in background noise compared to the predecessor system [1317]. The recently launched SONNET 2 audio processor features two further noise reduction algorithms: ambient noise reduction (ANR) and transient noise reduction (TNR). ANR aims to reduce noise in a sound signal where the target signal (e.g., speech) is contaminated with stationary or quasi-stationary background noise (with no, or comparably slow, temporal variations). Examples for such noise signals are the ones that emanate from a ventilating system or vehicle such as a car, airplane, or train [1820]. In the SONNET 2, ANR is available either with moderate noise suppression (“mild” setting) or with stronger noise suppression (“strong” setting). The “strong” setting provides larger noise reduction at the cost of potential target signal distortions.

TNR aims to decrease the amplitude of short, impulsive signal bursts in the sound signal. These transient sounds are characterized by a rapid onset to the peak in sound pressure level, a fast decay, and a short duration. The peak sound pressure level of the transient noise is well above the average sound pressure level [21]. Transient sounds can be loud (slamming doors, hammer blows, or clanking of dishes) or soft (such as the clicking of a pen or typing on a computer keyboard) [22]. They have been characterized as distracting and are a frequent source of complaint for people with impaired hearing [23]. In the SONNET 2, TNR is implemented with either moderate (“mild” mode) or stronger suppression (“strong” mode).

Both stationary-noise as well as transient-noise reduction systems have been realized in currently marketed state-of-the-art CI systems. Stationary-noise reduction is a feature of ClearVoice by Advanced Bionics (Stäfa, Switzerland), SNR-NR as part of Smart Sound iQ by Cochlear Ltd (Macquarie University, Australia), and VoiceTrack by Oticon Medical/Neurelec (Vallauris, France). Other than at MED-EL, a dedicated transient-noise reduction algorithm is presently only available by Advanced Bionics, as a component of ClearVoice based on an original Phonak technology named SoundRelax.

The growing number of front-end sound processing technologies implemented in CI systems can lead to uncertainty when to change programs. In real life, few users change from their everyday settings to try to avoid noisy situations [24], which may prevent them from achieving best hearing performance [20]. To this end, signal classifiers have been developed that can detect acoustic situations and switch between programs and pre-processing algorithms [25]. In the SONNET 2, a signal-classification-based system named Adaptive Intelligence (AI) discriminates five signal classes: Speech, Speech in Noise, Noise, Music, and Quiet. The purpose of the signal classification is to estimate the exposure time of the CI user in a certain acoustic environment and to enable an automatic selection of signal processing algorithms, e.g. directionality and/or noise reduction, depending on the detected signal class, as shown in Table 1 (see [17]), for an in-depth investigation of the directionality legacy feature). As can also be seen from Table 1, only ANR and TNR are activated or deactivated automatically in AI mild mode. The directionality remains in fixed beamforming mode, whereby the directional characteristic is tuned in such way as to approximate the natural directionality provided by the human pinna (Natural mode). ANR is further active in all signal classes, except music, whereas TNR is only active if noise is classified, either in the presence of only noise or in noise and speech. The AI strong mode provides a more intervening signal processing in that directionality is also switched between omnidirectional and adaptive directional, and ANR and TNR are activated in the strong modes. The adaptive directional mode is an adaptive beamformer (ABF) which adjusts its directional characteristic to maximally reduce the noise, i.e., it adapts to the direction and spectral characteristic of the noise sources. Note that in both AI modes, in the "Quiet" signal class, the system remains in the previous state, according to the previously detected signal class.

Automatic signal classification technologies have been recently introduced in some commercially available CI systems. Cochlear Ltd’s SmartSound iQ has implemented the automatic signal classifier SCAN in its currently marketed audio processors. Results revealed objective and subjective improvements in speech understanding for both adult and paediatric users [4, 20, 26, 27]. Similar systems are also promoted by Advanced Bionics (AutoSound™ OS) and Oticon Medical/Neurelec (Coordinated Adaptive Processing), but so far, no publicly available documentation is available on these products.

The primary aim of the present study was to evaluate the added ANR, TNR, and AI capabilities of the new SONNET 2 on speech understanding in noise by comparing them to the SONNET audio processor and to the SONNET 2 in the default SONNET configuration. The secondary aim was to access the subjective aspects of using the new device and configurations. To this end, listening effort, sound quality, and satisfaction with audio processor were collected and analysed.

Materials and methods

The study was approved by the ethics committee of the University of Würzburg, Germany and conducted following the Declaration of Helsinki. Written informed consent was obtained from all subjects prior to any study-specific procedure.

Participants

To be included in the study, participants had to meet all of the following inclusion criteria: 1) be at least 18 years old, 2) have ≥ 6 months of experience with a MED-EL CI (C40+ or later model), 3) have ≥ 6 months experience using the MED-EL SONNET audio processor, 4) have postlingual onset of bilateral severe to profound sensory-neural hearing loss, 5) be a unilateral CI users, 6) have a minimum of 10 active electrodes on their implanted array, 7) score ≥40% speech recognition on the Freiburg Monosyllables Test in quiet at 65dB SPL (at the last time tested), 8) be fluent in German (the language of the test centre), and 9) have signed and dated their informed consent before the start of any study-specific procedure.

Audio processors

The SONNET 2 processor was investigated in five configurations, as shown in Table 2. Configuration S2.Sonnet resembles a SONNET processor, which is the predecessor of the SONNET 2. The S2.Mild, S2.Strong, S2.AImild, and S2.AIstrong configurations represent the SONNET 2 processor with different combinations of ANR, TNR, and AI, together with certain settings of the already established front-end features wind-noise reduction (WNR), directionality, and automatic gain control (AGC). For details on AGC, please see Stöbich et al. [28]. S2.AImild is the clinical default setting of the SONNET 2 processor. In addition to performance with the SONNET 2 processor, the subjects’ SONNET processors were also tested as a baseline in the clinical default setting. Technical integrity was checked and microphone covers were replaced, if needed. Configurations are shown in Table 2.

thumbnail
Table 2. Audio processors and tested audio processor configurations.

https://doi.org/10.1371/journal.pone.0274446.t002

Study design

This was a prospective longitudinal monocentric study conducted at the Universitätsklinikum Würzburg, Germany, with each subject acting as his/her own control. Subjects made five visits to the clinic and completed one test session in each of the visits. See Table 3 for an overview of the study schedule and the different configurations tested at each visit. The order in which the configurations were tested within each visit was randomized between subjects (Table 3).

At the first visit, subjects were tested using their SONNET and answered the questionnaires as regards their experience with the SONNET (not the SONNET 2). They were then upgraded with a SONNET 2. Subject’s old SONNET map was used as a base for the new SONNET 2 fitting. For the new fitting, the thresholds and MCLs were taken unchanged from their original fitting and the new signal processing (ANR, TNR, AI) were activated according to the test plan. MAESTRO software suite 8 and the MAX programming interface were used for fitting.

Between each of the visits, in four take-home periods (each 1–3 weeks long), subjects used one of the following SONNET 2 configurations in an order that was randomized between subjects: S2.Mild, S2.Strong, S2.AImild, and S2.AIstrong. Subjects were asked to predominantly use the SONNET 2 configuration; however, if they found it unbearable, they could revert to using the default SONNET 1 configuration (i.e., BF = NAT & WNR = Mild). Care was taken to confirm usage of the SONNET 2 configurations via the datalogging feature of the processor.

Test setups

  1. Speech perception tests and subjective ratings were performed in five different test setups: S0 or M0: Speech or music (presented from the front).
  1. S0N0: Speech in stationary noise (both presented from the front).
  2. S0N0T0: Speech in stationary and transient noise (all presented from the front).
  3. S0N±45N±135: Speech in stationary noise (speech presented from the front, stationary noise presented from the sides, noise speakers presented uncorrelated noise).
  4. S0N±45N±135T135: Speech in stationary and transient noise (speech presented from the front, stationary noise presented from the sides, and transient noise presented ipsilaterally to the CI-side at either +135° or -135°) (Fig 1).
thumbnail
Fig 1.

Speech test setups for (a) speech or music testing in quiet (S0 or M0), (b) speech in noise S0N0, (c) speech in noise S0N0T0, (d) speech in noise S0N±45N±135, and (e) speech in noise S0N±45N±135T135 (note: transient noise presented ipsilaterally to ear with the CI). Orange speaker icons indicate target signal (speech or music), black speaker icons indicate stationary noise, and the blue speaker icons indicate transient noise.

https://doi.org/10.1371/journal.pone.0274446.g001

Speech understanding tests

Speech perception in quiet was evaluated with the Freiburg Monosyllables Test [29] in the S0 setup (see Fig 1). Speech perception in noise was assessed with the Oldenburg Sentence Test (OLSA) [30] for 80% correct word recognition in noise (SRT80 in dB SNR). SRT80 was used because this results in larger SNR values, which then limit possible speech distortions introduced by the ANR if the SNR approaches low levels, e.g., around 0dB, in high-performing users. The stationary noise-level was fixed and presented continuously at 60dB sound pressure level (SPL); the noise used was the OLSA speech-shaped noise. The level of the speech signal was varied according to the OLSA procedure [30]. Transient noise was also presented at a fixed level, which was set to be 3dB higher than the level of the stationary noise. The presentation level of the accumulated noise (stationary and transient noise) in this case was again 60dB SPL. With this setup it was guaranteed that the transient noise was loud enough but did not provoke clipping in the audio processor (due to the large peak-to-average power ratio for the transient noises). The transient noise used was hammering (an iron hammer hammering on iron). The rate was approximately 1–2 strokes per seconds. The stroke duration, measured from the onset of the signal to a signal level below 60dB of the maximum, was approx. 500ms. Due to the short duration of the hammering strokes, the strokes were inherently broadband, thus extended over the entire bandwidth. The speech signal was therefore completely masked during the hammering stroke. Speech perception in noise was tested in three setups: S0N0, S0N0T0, and S0N±45N±135 (see Fig 1). During the testing of subjects who were bimodal users, their hearing aid was switched off and the hearing aid and earmold were left on the ear to act as an additional damping.

Listening effort

The Adaptive Categorical Listening Effort Scaling (ACALES) test was used to investigate the subjectively perceived listening effort for the S2.Sonnet, S2.Mild, and S2.Strong configurations [31]. Subjects (n = 17) rated the perceived listening effort when listening to OLSA sentences in OLSA speech-shaped noise on a categorical scale ranging from 1 (no effort) to 13 (extreme effort). During the procedure, the SNRs automatically adapted according to the given ratings. It was expected that listening effort would decrease as the SNR increased (and vice versa). Listening effort scaling was performed in two different setups: S0N0 and S0N±45N±135 (see Fig 1). The sound levels were the same as for the speech tests described in the section above, i.e., the noise was presented continuously with 60dB SPL, the speech level was adapted according to the ACALES procedure.

Differences in the results with each configuration were calculated for the three categories low effort (“no effort” to “little effort”, comprising ratings from 1 to 5), moderate effort (“little effort” to “considerable effort”, comprising ratings from 5 to 9), and high effort (“considerable effort” to “extreme effort", comprising ratings from 9 to 13). A positive result (e.g. 2 dB) can then be interpreted in a way that subjects rated listening to a lower (i.e. more difficult) SNR (e.g. 6 dB) with the S2.Mild or S2.Strong configuration as equally strenuous as listening to a higher (i.e. easier) SNR (e.g. 8 dB) with the S2.Sonnet configuration. For determining these SNR differences, a sigmoid-shaped psychometric function was fitted to the rating results for each of the six conditions, i.e., the three processor configurations (S2.Sonnet vs. S2.Mild vs. S2.Strong) and for the two test conditions (S0N0 and S0N±45N±135). If the rating results did not allow a fit, for example, because of inconsistent rating, or the user was not able to exploit the entire range of the effort scale, the subject was excluded from this analysis. Due to this, 3 subjects were excluded.

Subjective sound quality and preference

Subjects rated their subjective sound quality and preference using a visual analogue scale after listening in six setups: S0, M0, S0N0, S0N±45N±135, S0N0T0, and S0N±45N±135T135 (see Fig 1 for the setups). The sound samples were presented with an acoustic level of 65dB SPL. Subjects ranked preference as well as subjective sound quality for the S2.Sonnet, S2.Mild, and S2.Strong configurations. Concerning subjective sound quality, subjects rated the six setups for four different quality aspects: “Speech and Music Quality”, “Pleasantness”, “Listening Effort”, and “How Disturbing the Noise Is”.

Questionnaires

The Audio Processor Satisfaction Questionnaire (APSQ) was used to assess user satisfaction with audio processors [32]. Further questionnaires included the Hearing Implant Sound Quality Index (HISQUI), which evaluates subjective sound quality [33], and the Speech, Spatial and Qualities of Hearing scale (SSQ12) for measuring everyday listening abilities and limitations [34].

In addition, subjects used a product-specific questionnaire to grade their preference for one of the four different SONNET 2 configurations (S2.Mild, S2.Strong, S2.AImild, and S2.AIstrong) versus the baseline SONNET configuration (S2.Sonnet) across different listening situations. The questionnaire consisted of the following 14 listening situations::

  • talking with one person in a crowded area
  • talking with one person in a quiet room
  • talking with one person in an area with high reverberation
  • talking with one person while the radio or TV is on
  • talking with one person next to a busy road
  • talking with one person when there is steady noise
  • being in an environment where the direction of the sound and the distance to the sound source both play important roles (like on a street)
  • being in a cinema or while watching TV in a quiet room
  • being outdoors (e.g., while running or cycling)
  • being in an environment with short and loud ambient noise
  • being in a place where the listening condition changes frequently (e.g., between quiet and noise)
  • talking in a place where other people are also talking
  • being in a loud environment with more or less steady ambient noise
  • listening to music.

For each situation, subjects chose their preference for the new SONNET 2 configuration, the baseline SONNET configuration, no preference (i.e., they were equal), or “not applicable” (“NA”), (i.e., if the subject is never in the specific environment or situation). Subjects could also leave a situation unanswered (“Missing”). Note that during one period the subjects’ audio processor was configured with two programs: the S2.Sonnet (baseline) and one of the four SONNET 2 configurations.

Data analysis

For analysing results of speech perception tests, configurations were compared using paired-samples t-tests (i.e., t-test) or, depending on the data distribution, Wilcoxon signed-rank tests. The Kolmogorov-Smirnov test and the Shapiro-Wilk test were conducted before to check the data distribution. If both tests confirmed that the data were normally distributed, then parametric statistical methods were applied. Otherwise, non-parametric statistical methods were conducted.

For analysing listening effort tests, for each effort category (low, moderate, high) of the ACALES rating, one-sample t-tests were carried out to test whether the difference in dB between S2.Sonnet, S2.Mild, and S2.Strong were significantly different from zero.

Significance was set to p ≤ 0.05. To control for the problem of multiplicity resulting from multiple pairwise comparisons (i.e., to avoid Type I errors), the Holm-Bonferroni correction method was applied. Missing data were treated as missing values.

IBM SPSS Statistics V24 and V25 (IBM, Armonk, NY, USA) were used for statistical analyses.

Results

Participants

Twenty adult unilaterally implanted CI users with bilateral postlingual sensorineural deafness (8 females, 12 males; (14 right ears, 6 left ears) met the inclusion criteria and took part in the study. Some participants were bimodal users, i.e., wore a hearing aid in the contralateral ear. Participants were a mean 48.2 years old (± 16.25 SD; range 17.7–70.4 years) at implantation and a mean 53.7 years old (± 16.95 SD; range 20.4–72.9 years) at enrolment. All subjects had been provided with various MED-EL CIs coupled to either a STANDARD or FLEX-type electrode array: STANDARD arrays are 31.5mm long and have 24 contacts; FLEX-type arrays range from 20–31.5mm long, have added flexibility, and 19 contacts; all are straight lateral wall arrays. All subjects had at least 6 months experience with the SONNET at the time of testing.

Speech understanding in quiet (S0)

Speech understanding scores for each device/configuration listed in Table 2 are shown in Fig 2 and Table 4.

thumbnail
Fig 2. Speech intelligibility in quiet with each configuration.

Boxplots show median values as well as first and third quartiles. Whisker bars show minimum and maximum values within 1.5 interquartile ranges below and above the first and third quartile. Diamonds show outliers.

https://doi.org/10.1371/journal.pone.0274446.g002

thumbnail
Table 4. Speech understanding scores (in mean, standard deviation, and range) in quiet and speech reception threshold (SRT80) for each device and setup.

https://doi.org/10.1371/journal.pone.0274446.t004

Pairwise comparisons (t-test) showed that there were no significant differences between the SONNET or the S2.Sonnet configurations and the other four SONNET 2 configurations (Fig 2).

Speech understanding in noise (S0N0)

Speech reception thresholds (SRT80) are given in Table 4 and shown in Figs 35 for different devices and configurations. Paired-samples t-tests revealed a significantly lower SRT (and therefore better speech perception) with S2.Mild compared to S2.Sonnet (t = 6.283, df: 19, p < 0.001). The difference between S2.Sonnet and S2.Strong was not significant, nor were the differences between S2.Sonnet and S2.AImild or S2.Sonnet and S2.AIstrong. Further, the SRT for the S2.Mild was significantly lower than for the SONNET (t = 3.44, df:18, p = 0.003). The SRT for the S2.AImild was also significantly lower than for the SONNET (t = 4.377, df:18, p < 0.001) (Fig 3).

thumbnail
Fig 3. SRT80 observed in the OLSA S0N0 for each configuration.

A cross marks a significant difference compared to S2.Sonnet. A star marks a significant difference compared to the SONNET. Boxplots show median values as well as first and third quartiles. Whisker bars show minimum and maximum values within 1.5 interquartile ranges below and above the first and third quartile. Diamonds show outliers.

https://doi.org/10.1371/journal.pone.0274446.g003

thumbnail
Fig 4. SRT80 observed in the OLSA S0N0T0 for each test configuration.

Boxplots show median values as well as first and third quartiles. Whisker bars show minimum and maximum values within 1.5 interquartile ranges below and above the first and third quartile. Diamonds show outliers.

https://doi.org/10.1371/journal.pone.0274446.g004

thumbnail
Fig 5. SRT80 observed in the OLSA S0N±45N±135 for each configuration.

A cross marks a significant difference compared to S2.Sonnet. A star marks a significant difference compared to the SONNET. Boxplots show median values as well as first and third quartiles. Whisker bars show minimum and maximum values within 1.5 interquartile ranges below and above the first and third quartile. Diamonds show outliers.

https://doi.org/10.1371/journal.pone.0274446.g005

Speech understanding in noise (S0N0T0)

See Table 4 for scores for each tested device/configuration. Results of paired-samples t-tests showed that the three tested SONNET 2 configurations were not significantly different (Fig 4).

Speech understanding in noise (S0N±45N±135)

See Table 4 for scores for each tested device/configuration. Paired-samples t-tests revealed significantly lower SRTs (and therefore better speech perception) compared to the S2.Sonnet configuration with both the S2.Mild configuration (t = 4.682, df: 19, p < 0.001) and the S2.Strong configuration (t = 4.236, df: 19, p < 0.001). Significantly better speech perception was also found for the SONNET 2 AI configurations (S2.AImild: t = 6.573, df: 19, p < 0.001; S2.AIstrong: t = 2.450, df: 19, p = 0.025). Significantly lower SRTs compared with the SONNET were found for the S2.Mild configuration (t = 3.885, df: 19, p = 0.001), for the S2.Strong configuration (t = 3.915, df: 19, p = 0.001), and for the S2.AImild configuration (t = 4.815, df: 19, p < 0.001) (Fig 5). Raw speech data can be found in S1 Dataset.

Listening effort

Results for SNR differences between S2.Sonnet, and S2.Mild, and S2.Strong are shown in Fig 6. All differences were found to be positive, thereby confirming reduced subjective listening effort when using the SONNET 2 audio processor in configurations S2.Mild and S2.Strong (Fig 6). As stated in the Methods section, 3 subjects were excluded because their rating results did not allow a fit.

thumbnail
Fig 6. ACALES results (SNR difference, see text) for a) S2.Mild vs SONNET in S0N0, b) S2.Strong vs SONNET in S0N0, c) S2.Strong vs. S2. Mild in S0N0, d) S2.Mild vs SONNET in S0N±45N±135, e) S2.Strong vs. SONNET in S0N±45N±135, f) S2.Strong vs. S2. Mild in S0N±45N±135.

Positive scores indicate a higher tolerated noise level (i.e., better performance). Boxplots show median values as well as first and third quartiles. Whisker bars show minimum and maximum values within 1.5 interquartile ranges below and above the first and third quartile. Diamonds show outliers.

https://doi.org/10.1371/journal.pone.0274446.g006

Additionally, one-sample t-tests were carried out for each effort category (low, moderate, high) to test whether the difference in dB between the S2.Sonnet and S2.Mild and S2.Strong were significantly different from zero. The differences between S2.Sonnet and S2.Mild were significantly different from zero in the S0N0 and the S0N±45N±135 test setup for all three effort categories. Likewise, the differences between the S2.Sonnet and S2.Strong were significantly different from zero in the S0N0 and the S0N±45N±135 test setup for all three effort categories (see Table 5). The differences between S2.Mild and S2.Strong were not significant. Raw ACALES data can be found in S2 Dataset.

Subjective sound quality and preference

Regarding preference ratings, pairwise comparisons were done across the auditory setups. No differences in the preference ratings between the S2.Sonnet configuration and S2.Mild or S2.Strong were found for speech in quiet (S0) or music in quiet (M0). For all four auditory setups with speech in noise (S0N0, S0N±45N±135, S0N0T0, and S0N±45N±135T135), the S2.Mild or S2.Strong were significantly more preferred than S2.Sonnet. Table 6 summarizes the inferential statistic results: Wilcoxon signed-rank tests (Z) or paired-samples t-tests (t), depending on the data distribution.

thumbnail
Table 6. Paired comparisons for subjects’ preference ratings between the S2.Sonnet and the SONNET 2 configurations S2.Mild or S2.Strong (original p-values).

https://doi.org/10.1371/journal.pone.0274446.t006

Concerning quality, pairwise comparisons did not show any differences between the S2.Sonnet configuration and S2.Mild or S2.Strong on sound quality and pleasantness for speech in quiet (S0) or music in quiet (M0). For all four setups with speech in noise (S0N0, S0N±45N±135, S0N0T0, and S0N±45N±135T135), S2.Mild or S2.Strong were rated with significantly higher quality for all four aspects (quality, pleasantness, listening effort, and noise disturbance) than S2.Sonnet. Table 7 summarizes the inferential statistic results: Wilcoxon signed-rank tests (Z) or a paired-samples t-tests, (t) depending on the data distribution (Table 7).

thumbnail
Table 7. Paired comparisons for subjects’ quality ratings between the SONNET and the SONNET 2 configurations S2.Mild or S2.Strong (original p-values).

https://doi.org/10.1371/journal.pone.0274446.t007

Questionnaires

To obtain descriptive results from the product-specific questionnaire, preferences were averaged across the fourteen listening situation items in the questionnaire. Results showed a clear preference for the S2.Mild, S2.Strong, S2.AImild, and S2.AIstrong configurations over the S2.Sonnet configuration (Fig 7).

thumbnail
Fig 7. Product-specific questionnaire results.

Figures are the S2.Sonnet vs. 7a) S2.Mild, 7b) S2.Strong, 7c) S2.AImild, 7d) S2.AIstrong. A point indicates that this percentage of users preferred either the SONNET 2 configuration given on the abscissa, had no preference (None), found the question not applicable (NA), or did not rate (Missing) for the specific listening environment. Higher scores indicate greater preference. Diamonds indicate outliers.

https://doi.org/10.1371/journal.pone.0274446.g007

The APSQ was used to assess subjects’ satisfaction with the SONNET and the SONNET 2 audio processor using the configuration from the previous take-home period. The average APSQ total score was 8.62 (± 0.83 SD; range 7.30–9.78) for the SONNET and 8.73 (± 0.81 SD; range 7.24–9.75) for the SONNET 2. The paired t-test test did not show significant differences in subjects’ satisfaction with the two audio processors [t(19) = -.881; p = 0.389].

The HISQUI was completed at each study visit. The SONNET was rated with "moderate sound quality", while all four SONNET 2 configurations (e.g., S2.mild, S2.strong, S2.AImild, S2.AIstrong) were rated with "good sound quality". The mean HISQUI score was 85 (± 15.71 SD; range 51–110) for the SONNET and 94.74 (± 16.54 SD; range 73–132) for the S2.mild, 93.80 (± 16.41 SD; range 61–116) for the S2.strong, 93.75 (± 14.80 SD; range 57–120) for the S2.AImild, and 91.40 (± 18.49 SD; range 59–131) for the S2.AIstrong configuration, respectively. Results of the repeated measure ANOVA (N = 19) did not show a significant difference between audio processors or configurations [F(4; 72) = 1.880; p = 0.123].

The SSQ12 was completed at each study visit. Total mean scores were 5.45 (± 1.28; range 3.35–7.83) for the SONNET, 6.16 (± 1.41 SD; range 3.33–8.32) for the S2.mild, 5.53 (± 1.44 SD; range 3.42–8.11) for the S2.strong, 5.61 (± 1.15 SD; range 4.00–7.47) for the S2.AImild, and 5.34 (± 1.69 SD; range 1.75–8.40) for the S2.AIstrong configuration, respectively. The repeated-measures ANOVA (N = 13) revealed significant differences in the SSQ12 total scores between the five fitting configurations [F(4; 52) = 2.557; p = 0.049]. Paired t-tests were performed but did not show significant differences after adjusting the alpha for multiple comparisons. Raw questionnaire data can be found in S3 Dataset.

Discussion

The primary goal of this study was to determine if the use of stationary-noise reduction (ANR), transient noise-reduction (TNR), and a signal-classification-based system named ‘Adaptive Intelligence (AI)’ would result in superior performance in listening conditions with stationary or transient noise. In short, results from speech perception tests confirmed superior performance of the SONNET 2 in stationary noise with ANR and AI but did not find an effect of TNR. Regarding subjective measures, users rated their self-perceived listening effort consistently lower with the SONNET 2 when using ANR and TNR. Likewise, they rated their subjective sound quality as higher with the SONNET 2 when using ANR and TNR in conditions with competing noise. Overall preference for SONNET 2 configurations using ANR and TNR was strong for most categories across listening with competing noise.

Speech understanding in quiet

No significant differences were found for speech understanding in quiet, indicating that using ANR, TNR, and AI in this situation does not degrade speech perception. Since ANR and TNR were specifically designed to reduce noise in relevant conditions, it was expected that in quiet results with the SONNET 2 would be similar to those with the SONNET (or to the SONNET 2 configured as SONNET, i.e. S2.Sonnet). This result is also in line with earlier studies on noise-reduction algorithms of other CI manufacturers. Dyballa et al. [35] tested a TNR algorithm in high and low mode in an Advanced Bionics audio processor in quiet and found that speech perception was comparable to when not using TNR. The noise-reduction algorithm ClearVoice by Advanced Bionics did not compromise listening effort in quiet [19]. A Cochlear Ltd CP900 audio processor with the noise-reduction algorithm SNR-NR performed like one without noise reduction technology [36]. Similarly, the noise-reduction algorithm VoiceTrack by Oticon Medical/Neurelec did not have significant effects on speech understanding in quiet [37].

Speech understanding in noise

SONNET 2 and SONNET mode (S2.Sonnet).

In stationary noise, speech reception thresholds were, setups compared to the SONNET 2 in SONNET mode (S2.Sonnet), significantly better for the SONNET 2 with ANR and TNR in S0N0 (1.43 dB improvement in SRT80 for S2.Mild) and in S0N±45N±135 (2.19 dB for S2.Mild, 1.41 dB for S2.Strong, 2.55 dB for S2.AImild, 1.62 dB for S2.AIstrong). Similar significant benefits of the SONNET 2 with ANR and TNR active were observed when compared with the SONNET processor in the S0N0 setup (1.33 dB improvement in SRT for S2.Mild, 1.07 dB improvement for S2.AImild) and S0N±45N±135 (2.03 dB for S2.Mild,1.26 dB for S2.Strong, 2.4 dB for S2.AImild). In general, speech reception thresholds and improvements in SRTs in the AI configurations (S2.AImild, S2.AIstrong) corresponded well to the respective ANR/TNR configurations (S2.Mild, S2.Strong), as expected with AI activating ANR and TNR accordingly in the speech-in-noise signal class (see Table 1). Thus, in the discussion below, while we concentrate on ANR and TNR, it should be kept in mind that these features are also included in the AI switching scheme.

The benefit found for lower speech reception thresholds with an active stationary noise reduction algorithm like ANR is in line with several studies on stationary noise reduction algorithms featured in CI audio processors from various manufacturers. However, benefit varied widely across manufacturers because of different noise types, source localisation, fitting configurations, and possibly also noise reduction techniques and implementations. These parameters need to be kept in mind when comparing the results of the present study with results presented in the literature.

SONNET 2 and other systems.

ClearVoice by Advanced Bionics is a single microphone noise reduction algorithm that aims to improve speech understanding by reducing stationary noise and emphasizing the dynamic channels that presumably contain more speech [38]. Kam et al. [39] found that subjects performed significantly better with ClearVoice in the medium setup (max. 12 dB attenuation) than with the control option (ClearVoice off) in the S0N0 setup. The average increase in speech perception in the Cantonese HINT test was a moderate 5.5%. No significant performance gain was measured for the ClearVoice high setup (max. 18 dB attenuation). Correspondingly, Koch et al. [19] found significantly improved speech understanding in speech-spectrum noise and multitalker babble, likewise in the S0N0 setup. In speech-spectrum noise, performance gains of 7.6%, 8.7%, and 10.6% for ClearVoice low, medium, and high setup, respectively, were measured using the AzBio sentence test with individually adjusted SNRs (aiming at achieving 40–50% in the control setup (ClearVoice off)). In multitalker babble noise, performance gains were smaller (ranging from 1.4% to 4.3%) and not significant. Holden et al. [40] found performance gains due to an active ClearVoice in the low, medium, and high setup only in the R-SPACETM setup (eight speaker setup with speech from 0°, and restaurant noise from the sides and behind the listener). The performance gains for this adaptive test were 1.22 dB for the ClearVoice low setup, 1.46 dB for the ClearVoice medium setup, and 2.51 dB for the ClearVoice high setup. Only the performance gain for the high setup was significant. No significant differences were found for the other administered tests, which were the BKB-SIN test (sentences from 0° and four-talker babble noise from 90° ipsilateral to the CI ear), the ANL test (acceptable noise level test), CUNY sentence test in speech-spectrum noise, and the AzBio sentences in multitalker babble noise.

Mauger et al. [20] and De Ceulaer et al. [41] evaluated the SNR-NR algorithm as implemented by Cochlear Ltd in the SmartSound iQ suite, and found a significant gain in the S0N0 setup. In Mauger et al. [20] a significant performance gain in terms of a reduced speech reception threshold due to the SNR-NR algorithm (within the SCAN option) was present only in the S0N0 condition with speech-weighted noise. The mean performance gains ranged from 1.8 dB to 2.3 dB, depending on the compared processor setups. Interestingly, no performance gains were measured either for the S0N3 condition with speech-weighted noise played from ±90° and 180° azimuth, or for the S0N3 condition with four-talker babble noise played from ±90° and 180° azimuth, or for the S0N0 condition with four-talker babble noise. In De Ceulaer et al. [41], a performance gain for SNR-NR in an S0N0 setup with speech-weighted noise of 1.2 dB (reduction of the 50% speech reception threshold, SRT50) was measured. Most of the SRT50 improvements due to SNR-NR within this study were due to subjects having an SRT50 of 5‒20 dB; thus, in a comparably high SNR range.

Guevara et al. [37] investigated the VoiceTrack noise reduction algorithm implemented in the Saphyr Neo audio processor (Oticon Medical/Neurelec) and found a significant improvement in speech perception (French disyllabic words) in speech-shaped noise (at +5 dB SNR) and cocktail party noise (at 0 dB SNR), both presented in the S0N0 setup. For the speech-shaped noise tests, significance could only be achieved when the preferred VoiceTrack setup (soft, medium, or strong) was considered. In this case, the mean improvement was 12.3%. Using a liberal least significant difference post-hoc test 1) the VoiceTrack soft setup achieved significantly better results compared to VoiceTrack off in speech-shaped noise and 2) the VoiceTrack medium setup achieved significantly better results compared to VoiceTrack off in cocktail-party noise at 0 dB SNR.

With setups that presented speech from the front and noise from spatially-separated sources, the benefit of the SONNET 2 paralleled results of previous studies. Noël-Petroff et al. [42] evaluated the ClearVoice noise-reduction algorithm in children fitted with Harmony (Advanced Bionics) audio processors and showed significant improvement for speech perception in stationary noise in an S0N135 setup. Geißler et al. [43] used ClearVoice in S0N0N±45N±90N±135N180 and S0N±70N±135N180 setups with stationary noise. Speech perception improved significantly. However, a test where roving sound was added to the second setup did not reveal any significant benefit from the use of ClearVoice. Wolfe et al. [4] tested the SNR-NR noise reduction algorithm implemented in the Nucleus 6 audio processor (Cochlear Ltd) in a S0N90 or S0N270 setup and found superior speech recognition compared to the previously used Nucleus 5 audio processor (without SNR-NR). Considering the SNR-NR algorithm of the Nucleus 6 audio processor, a mean significant improvement of 8.5% (AzBio sentence test at a fixed individually adjusted SNR) was shown (Nucleus 6 with Automatic sensitivity control (ASC) + adaptive dynamic range optimization (ADRO) and standard microphone mode, SNR-NR disabled vs. Nucleus 6 with ASC + ADRO and SNR-NR enabled with standard microphone mode)).

Regarding transient noise reduction, speech recognition was comparable between SONNET 2 in SONNET configuration (S2.Sonnet, no noise reduction) and the SONNET 2 with ANR/TNR (S2.Mild, S2.Strong). This result can be explained by the specific type of transient noise used in this study. Although hammering usually dominates speech signals in terms of level, it is very short temporally. Thus, with this noise signal, enough speech components remained so that word recognition was similar between the non-TNR (S2.Sonnet) and the two TNR conditions (S2.Mild, S2.Strong), with SRT values being comparable. At the same time, TNR perceptively reduced the amplitude of the transient noise signal, which consequently led to the superior subjective preference and quality ratings found for S2.Mild, S2.Strong.

Unfortunately, to the best of our knowledge few studies have investigated transient noise reduction in CI systems. This is likely due to only one other manufacturer (Advanced Bionics) offering a TNR algorithm. Dyballa et al. [35, 44] tested a TNR by Advanced Bionics and found significant improvements in SNRs when using single- or multiband TNR algorithms. However, the sound tokens (e.g. speech and hammering noise) were delivered directly via an input jacket to the audio processor. In Dingemanse et al. [21], the TNR algorithm SoundRelax was run on a Q70 audio processor (Advanced Bionics) in the S0N0T0 setup with recorded transient kitchen sounds (clinking bowls, dishes, etc.). Interestingly, and in contrast to the results found here, TNR could not prevent transient noise lowering speech understanding in noise, although it did reduce subjective annoyance from transient sounds.

In summary, the results for ANR and TNR in the SONNET 2 sound processor show significant gains in speech perception in stationary noise and compare well with similar technologies in other CI sound processors. These gains exist regardless of whether ANR and TNR are explicitly activated in the SONNET 2 processor (S2.Mild, S2 Strong) or are used as part of AI (S2.AImild, S2.AIstrong). In combined stationary and transient noise, ANR and TNR do not lead to a decrement in speech perception, which at least means that the benefits in subjective sound quality from these technologies (as discussed below) come without any penalty in speech perception.

Listening effort, subjective sound quality and preference, questionnaires

Like in the previous section, the discussion below will concentrate on ANR and TNR. It should be kept in mind that these features are also included in the AI switching scheme.

Two different tests assessed listening effort and perceived sound quality in this study. The ACALES test quantified perceived listening effort for speech perception in noise. Its results were in line with those found for the objective speech understanding in noise tests: subjects tolerated consistently lower SNR values with the SONNET 2 configurations (S2.Mild, S2.Strong) compared to the SONNET configuration (S2.Sonnet) in all hearing effort categories. In fact, the speech-in-noise tests revealed that the objective benefit in speech understanding in noise was consistent with the subjective ratings. More precisely, there was a 1–2 dB advantage for the SONNET 2 configurations S2.Mild and S2.Strong in both setups, S0N0 and S0N±45N±135, throughout all three effort categories, e.g. low, moderate, and high effort. Consistently, the speech-in-noise results also showed a similar advantage for these two SONNET 2 configurations (only the advantage in the ANR/TNR strong configuration in the S0N0 condition was not significant). Thus, the subjects’ subjective self-assessment of their perceived hearing effort closely reflected the improvements in SRT. We therefore conclude that, users not only tolerate a smaller SNR with a noise reduction algorithm in place, but this noise reduction algorithm improves speech understanding in noise, that is, it effectively lowers the SRT.

SONNET 2 and other systems.

The above findings are in contrast with those described for the ClearVoice algorithm of Advanced Bionics. Dingemanse & Goedegebure [45] used the acceptable noise level (ANL) test [46] in a similar cohort of CI recipients and found a significant improvement, i.e., higher background noise levels were tolerated. However, they did not find any corresponding improvements in speech understanding in background noise, in any of the determined SRTs (SRT50, SRT70) or at a fixed SNR (SRT50 + 11dB). Differently to our study where the SRT80 was determined, they determined the SRT50 or SRT70. While these results might highlight differences between ClearVoice and ANR as well as TNR, one needs to also consider differences between the ACALES and the ANL test. Although similar in nature, they may differ in their specificity. The ACALES test might be better suited to self-assess the speech understanding in noise because the subjects must relate their hearing effort during the test directly to their ability to understand speech in noise. With the ANL test, subjects are asked for their personally acceptable background noise level, which might be less related to their ability to still understand speech but more to their overall hearing experience (e.g. their “willingness to listen to speech in background noise”, cf. Dingemanse & Goedegebure [45]). A further difference between the two tests is that the ACALES used in the present study employed the same speech (and noise) signals as the objective speech understanding tests (OLSA sentences and OLSA speech-shaped noise). In Dingemanse & Goedegebure [45], the speech and noise material used and speech understanding tests were different in that for the ANL a recorded story in noise was administered, whereas unrelated sentences of female speech of five to nine words were used in the objective speech in noise test (a proprietary test developed by the authors).

The other subjective assessments of the present study were the preference and quality ratings of various test signals. Again, these results were in line with those of the objective speech perception tests. Firstly, the ANR and TNR were both generally found to not affect speech understanding or music perception in quiet. Thus, there was no perceived negative effect of an active ANR and TNR. This parallels the outcomes of the Freiburg monosyllabic test, administered in quiet, which showed no difference in monosyllable perception for the SONNET 2 when using ANR and TNR (S2.Mild, S2.Strong, S2.AImild, S2.AIstrong) in comparison to tests without ANR and TNR (SONNET, S2.Sonnet). This is a remarkable finding, since music signals typically contain transient components, which are left intact by the active TNR. Secondly, the SONNET 2 configurations S2.Mild and S2.Strong were consistently preferred and better rated for speech in stationary and/or transient noise. Most notably, the hearing effort and the perceived disturbance by stationary and transient noise were rated lower for the SONNET 2 configurations S2.Mild and S2.Strong. Also, the sound quality and pleasantness were graded as being better in both SONNET 2 configurations. This indicates that the reduction of the stationary and transient disturbance by an active ANR and TNR led to no perceivable loss in sound quality.

Dingemanse et al. [21] showed that the simultaneous activation of the stationary-noise reduction ClearVoice and the transient noise reduction SoundRelax by Advanced Bionics did not result in subjective improvement when comparing various sounds contaminated with stationary or transient noise. This was perhaps due to a 5% increase of M-levels when ClearVoice was active, which in turn might have led to an increased loudness perception of transient noise. Although this negative effect was corrected by activating the transient noise reduction algorithm (SoundRelax), subjects only took advantage of this functionality when ClearVoice was not active. Thus, there was a negative effect of ClearVoice in environments with transient noise signals present. Such an effect was not seen in the SONNET 2 when both ANR and TNR were activated, since both configurations that use ANR and TNR (S2.Mild, S2.Strong) were 1) preferred over the SONNET configuration (S2.Sonnet, i.e., ANR and TNR off) and 2) rated better for all presented test signals, irrespective of the presence of stationary and/or transient noise. This was also reflected in the results of the product-specific questionnaire, where subjects strongly and consistently rated all SONNET 2 configurations higher than the SONNET configuration across various listening situations. In addition, the questionnaire proved to be a relevant assessment of the AI modes (S2.AImild, S2.AIstrong) in that they showed that AI, like the configurations using ANR and TNR, was preferred over SONNET for fourteen real-life auditory scenarios. APSQ scores did not differ between SONNET and SONNET 2, but nonetheless, subjects rated their satisfaction with both processors as very high. For the HISQUI, the difference between the SONNET configuration and the four SONNET 2 configurations also reflected different sound quality levels, with moderate sound quality for the SONNET and good sound quality for the SONNET 2 configurations. Finally, the SSQ12 did not show any significant differences between the five configurations after alpha-correction. Nonetheless, these results were descriptively higher than those in a recent publication [47].

Conclusions

The MED-EL SONNET 2 audio processor features new front-end functionalities: ambient noise reduction (ANR); transient noise reduction (TNR); and Adaptive Intelligence (AI), which is a classifier-based automated program-switching scheme. All three technologies resulted in significant improvements in speech understanding in noise in a variety of different test setups. This was paralleled by significantly reduced perceived listening effort when using these functionalities as well as a strong subjective preference of these functionalities in real-life auditory scenarios.

Acknowledgments

The authors thank the following individuals (all from MED-EL, Innsbruck, Austria) for their contributions to this manuscript: Ernst Aschbacher, Ivo Machatschke, Alexander Möltner, Edda Amman, Michael Todd, and Mareike Billinger-Finke.

References

  1. 1. Zeng FG, Rebscher S, Harrison W, Sun X, Feng H. Cochlear implants: system design, integration, and evaluation. IEEE Rev Biomed Eng. 2008;1: 115–142. pmid:19946565
  2. 2. Schafer EC, Thibodeau LM. Speech recognition abilities of adults using cochlear implants with FM systems. J Am Acad Audiol. 2004;15(10): 678–691. pmid:15646666
  3. 3. Wolfe J, Schafer EC. Optimizing the benefit of sound processors coupled to personal FM systems. J Am Acad Audiol. 2008;19(8): 585–594. pmid:19323350
  4. 4. Wolfe J, Neumann S, Marsh M, Schafer E, Lianos L, Gilden J, et al. Benefits of adaptive signal processing in a commercially available cochlear implant sound processor. Otol Neurotol. 2015 Aug;36(7): 1181–1190. pmid:26049314
  5. 5. Kokkinakis K, Azimi B, Hu Y, Friedland DR. Single and multiple microphone noise reduction strategies in cochlear implants. Trends Amplif. 2012 June;16(2): 102–116. pmid:22923425
  6. 6. Nelson PB, Jin SH, Carney AE, Nelson DA. Understanding speech in modulated interference: Cochlear implant users and normal-hearing listeners. J Acoust Soc Am. 2003;113(2): 961–968. pmid:12597189
  7. 7. Stickney GS, Zeng FG, Litovsky R, Assmann P. Cochlear implant speech recognition with speech maskers. J Acoust Soc Am. 2004 Aug;116(2): 1081–1091. pmid:15376674
  8. 8. Zeng F-G. Trends in cochlear implants. Trends Amplif. 2004;8(1): 1–34. pmid:15247993
  9. 9. Hey M, Hocke T, Böhnke B, Mauger SJ. ForwardFocus with cochlear implant recipients in spatially separated and fluctuating competing signals–introduction of a reference metric. Int J Audiol. 2019;58(12): 869–878. pmid:31464542
  10. 10. Dorman MF, Loizou PC, Fitzke J. The identification of speech in noise by cochlear implant patients and normal-hearing listeners using 6-channel signal processors. Ear Hear. 1998;19(6): 481–484. pmid:9867296
  11. 11. Fu Q-J, Shannon RV. Effect of acoustic dynamic range on phoneme recognition in quiet and noise by cochlear implant users. J Acoust Soc Am. 1999;106(6): L65–L70. pmid:10615712
  12. 12. Zeng F-G, Galvin JJ III. Amplitude mapping and phoneme recognition in cochlear implant listeners. Ear Hear. 1999;20(1): 60–74. pmid:10037066
  13. 13. Wimmer W, Weder S, Caversaccio M, Kompis M. Speech intelligibility in noise with a pinna effect imitating cochlear implant processor. Otol Neurotol. 2016;37(1): 19–23. pmid:26427637
  14. 14. Dorman MF, Natale S, Loiselle L. Speech understanding and sound source localization by cochlear implant listeners using a pinna-effect imitating microphone and an adaptive beamformer. J Am Acad Audiol. 2018;29(03): 197–205. pmid:29488870
  15. 15. Honeder C, Liepins R, Arnoldner C, Šinkovec H, Kaider A, Vyskocil E, et al. Fixed and adaptive beamforming improves speech perception in noise in cochlear implant recipients equipped with the MED-EL SONNET audio processor. PLoS One, 2018;13(1): e0190718. pmid:29304186
  16. 16. Büchner AM, Schwebs M, Lenarz T. Speech understanding and listening effort in cochlear implant users–microphone beamformers lead to significant improvements in noisy environments. Cochlear Implants Int. 2020;21(1): 1–8. pmid:31590612
  17. 17. Hagen R, Radeloff A, Stark T, Anderson I, Nopp P, Aschbacher E, et al. Microphone directionality and wind noise reduction enhance speech perception in users of the MED-EL SONNET audio processor. Cochlear Implants Int. 2020. 21(1): 53–65. pmid:31524107
  18. 18. Hersbach AA, Arora K, Mauger SJ, Dawson PW. Combining directional microphone and single-channel noise reduction algorithms: a clinical evaluation in difficult listening conditions with cochlear implant users. Ear Hear. 2012. 33(4): e13–e23. pmid:22555182
  19. 19. Koch DB, Quick A, Osberger MJ, Saoji A, Litvak L. Enhanced hearing in noise for cochlear implant recipients: clinical trial results for a commercially available speech-enhancement strategy. Otol Neurotol. 2014;35(5): 803–809. pmid:24691504
  20. 20. Mauger SJ, Warren CD, Knight MR, Goorevich M, Nel E. Clinical evaluation of the Nucleus® 6 cochlear implant system: Performance improvements with SmartSound iQ. Int J Audiol. 2014. 53(8): 564–576. pmid:25005776
  21. 21. Dingemanse JG, Vroegop JL, Goedegebure A. Effects of a transient noise reduction algorithm on speech intelligibility in noise, noise tolerance and perceived annoyance in cochlear implant users. Int J Audiol. 2018;57(5): 360–369. pmid:29334269
  22. 22. Korhonen P, Kuk F, Lau C, Keenan D, Schumacher J, Nielsen J. Effects of a transient noise reduction algorithm on speech understanding, subjective preference, and preferred gain. J Am Acad Audiol. 2013;24(9): 845–858. pmid:24224991
  23. 23. Keidser G O’Brien A, Latzel Convery E Evaluation of a noise-reduction algorithm that targets non-speech transient sounds. Hear J. 2007;60(2): 29–32.
  24. 24. Buechner A, Dyballa K-H, Hehrmann P, Fredelake S, Lenarz T. Advanced beamformers for cochlear implant users: acute measurement of speech perception in challenging listening conditions. PloS One. 2014l;9(4): e95542. pmid:24755864
  25. 25. Hamacher V, Chalupper J, Eggers J, Fischer E, Kornagel U, Puder H, et al. Signal processing in high-end hearing aids: State of the art, challenges, and future trends. EURASIP J Adv Signal Process. 2005;152674.
  26. 26. Plasmans A, Rushbrooke E, Moran M, Spence C, Theuwis L, Zarowski A, et al. A multicentre clinical evaluation of paediatric cochlear implant users upgrading to the Nucleus® 6 system. Int J Pediatr Otorhinolaryngol. 2016;83: 193–199. pmid:26968076
  27. 27. Ganek H, Forde-Dixon D, Cushing SL, Papsin BC, Gordon KA. Cochlear implant datalogging accurately characterizes children’s ‘auditory scenes’. Cochlear Implants Int. 2021;22(2): 85–95.
  28. 28. Stöbich B, Zierhofer CM, Hochmair ES. Influence of automatic gain control parameter settings on speech understanding of cochlear implant users employing the continuous interleaved sampling strategy. Ear Hear. 1999;20(2): 104–16. pmid:10229512
  29. 29. Hahlbrock KH. [Speech audiometry and new word-tests]. Arch Ohren Nasen Kehlkopfheilkd, 1953;162(5): 394–431.
  30. 30. Wagener K, Brand T, Kollmeier B. Entwicklung und Evaluation eines Satztests für die deutsche Sprache. I-III: Design, Optimierung und Evaluation des Oldenburger Satztests (Development and evaluation of a sentence test for the German language. I-III: Design, optimization and evaluation of the Oldenburg sentence test). Zeitschrift für Audiologie (Audiological Acoustics), 1999;38: 4–15.
  31. 31. Krueger M, Schulte M, Brand T, Holube I. Development of an adaptive scaling method for subjective listening effort. J Acoust Soc Am. 2017;141(6): 4680–4693. pmid:28679238
  32. 32. Billinger-Finke M, Bräcker T, Weber A, Amann E, Anderson I, Batsoulis C. Development and validation of the audio processor satisfaction questionnaire (APSQ) for hearing implant users. Int J Audiol. 2020;59(5): 392–397. pmid:31944127
  33. 33. Amann E, Anderson I. Development and validation of a questionnaire for hearing implant users to self-assess their auditory abilities in everyday communication situations: the Hearing Implant Sound Quality Index (HISQUI19). Acta Otolaryngol. 2014;134(9): 915–923. pmid:24975453
  34. 34. Noble W, Jensen NS, Naylor G, Bhullar N, Akeroyd MA. A short form of the Speech, Spatial and Qualities of Hearing scale suitable for clinical use: The SSQ12. Int J Audiol. 2013;52(6): 409–412. pmid:23651462
  35. 35. Dyballa K-H, Hehrmann P, Hamacher V, Nogueira W, Lenarz T, Büchner A. Evaluation of a transient noise reduction algorithm in cochlear implant users. Audiol Res. 2015;5(2): 44–49. pmid:26779325
  36. 36. Hey M, Hocke T, Mauger S, Müller-Deile J. A clinical assessment of cochlear implant recipient performance: implications for individualized map settings in specific environments. Eur Arch Otorhinolaryngol. 2016;273(11): 4011–4020. pmid:27276990
  37. 37. Guevara N, Bozorg-Grayeli A, Bebera J-P, Ardoint M, Saaï S, Gnassia D, et al. The Voice Track multiband single-channel modified Wiener-filter noise reduction system for cochlear implants: patients’ outcomes and subjective appraisal. Int J Audiol. 2016;55(8): 431–438. pmid:27108635
  38. 38. Buechner A, Brendel M, Saalfeld H, Litvak L, Frohne-Buechner C, Lenarz T. Results of a pilot study with a signal enhancement algorithm for HiRes 120 cochlear implant users. Otol Neurotol. 2010;31(9): 1386–1390. pmid:20856160
  39. 39. Kam AC, Ng IH, Cheng MM, Wong TK, Tong MC. Evaluation of the ClearVoice strategy in adults using HiResolution Fidelity 120 sound processing. Clin Exp Otorhinolaryngol. 2012;5(Suppl 1): S89–92. pmid:22701156
  40. 40. Holden LK, Brenner C, Reeder RM, Firszt JB. Postlingual adult performance in noise with HiRes 120 and ClearVoice Low, Medium, and High. Cochlear implants Int. 2013;14(5): 276–286. pmid:23683298
  41. 41. De Ceulaer G., Swinnen F, Pascoal D, Philips B, Killian M, James C, et al. Conversion of adult Nucleus® 5 cochlear implant users to the Nucleus® 6 system. Cochlear Implants Int. 2015;16(4): 222–232. pmid:25284643
  42. 42. Noël-Petroff N, Mathias N, Ulmann C, Van Den Abbeele T. Pediatric evaluation of the ClearVoice™ speech enhancement algorithm in everyday life. Audiol Res. 2013;3(1): 57–62.
  43. 43. Geißler G, Arweiler I, Hehrmann P, Lenarz T, Hamacher V, Büchner A. Speech reception threshold benefits in cochlear implant users with an adaptive beamformer in real life situations. Cochlear Implants Int. 2015;16(2): 69–76. pmid:24993633
  44. 44. Dyballa K-H, Hehrmann P, Hamacher V, Lenarz T, Buechner A. Transient noise reduction in cochlear implant users: a multi-band approach. Audiol Res. 2016;6(2): 28–35. pmid:27942372
  45. 45. Dingemanse G, Goedegebure A. Application of Noise Reduction Algorithm ClearVoice in Cochlear Implant Processing: Effects on Noise Tolerance and Speech Intelligibility in Noise in Relation to Spectral Resolution. Ear Hear. 2015;36(3): 357–367. pmid:25479412
  46. 46. Nabelek AK, Freyaldenhoven MC, Tampas JW, Burchfiel SB, Muenchen RA. Acceptable noise level as a predictor of hearing aid use. J Am Acad Audiol. 2006;17(9): 626–639. pmid:17039765
  47. 47. Calvino M, Sánchez-Cuadrado I, Gavilán J, Lassaletta L. Does bimodal hearing increase self-assessed abilities and hearing outcomes when compared to unilateral cochlear implantation? Int J Audiol. 2020;59(9): 654–660. pmid:32174222