J Am Acad Audiol 2020; 31(05): 324-335
DOI: 10.3766/jaaa.18106
Research Article
American Academy of Audiology. All rights reserved. (2020) American Academy of Audiology

High- and Low-Performing Adult Cochlear Implant Users on High-Variability Sentence Recognition: Differences in Auditory Spectral Resolution and Neurocognitive Functioning

Terrin N. Tamati
1   Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
2   Department of Otolaryngology - Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH
,
Christin Ray
2   Department of Otolaryngology - Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH
,
Kara J. Vasil
2   Department of Otolaryngology - Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH
,
David B. Pisoni
3   Department of Psychological and Brain Sciences, Indiana University - Bloomington, Bloomington, IN
,
Aaron C. Moberly
2   Department of Otolaryngology - Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH
› Institutsangaben
Funding Collection of data was primarily supported by the National Institutes of Health, National Institute on Deafness and Other Communication Disorders (NIDCD) Career Development Award 5K23DC015539-02, and the American Otological Society Clinician-Scientist Award to Aaron C. Moberly. Preparation of the manuscript was also supported in part by VENI Grant No. 275-89-035 from the Netherlands Organization for Scientific Research (NWO) and funding from the President's Postdoctoral Scholars Program (PPSP) at The Ohio State University awarded to Terrin N. Tamati.
Weitere Informationen

Address for correspondence

Terrin N. Tamati
Department of Otolaryngology - Head and Neck Surgery
The Ohio State University, Columbus, OH 43212

Publikationsverlauf

Publikationsdatum:
09. Juni 2020 (online)

 

Abstract

Background Postlingually deafened adult cochlear implant (CI) users routinely display large individual differences in the ability to recognize and understand speech, especially in adverse listening conditions. Although individual differences have been linked to several sensory (‘‘bottom-up’') and cognitive (‘‘top-down’') factors, little is currently known about the relative contributions of these factors in high- and low-performing CI users.

Purpose The aim of the study was to investigate differences in sensory functioning and neurocognitive functioning between high- and low-performing CI users on the Perceptually Robust English Sentence Test Open-set (PRESTO), a high-variability sentence recognition test containing sentence materials produced by multiple male and female talkers with diverse regional accents.

Research Design CI users with accuracy scores in the upper (HiPRESTO) or lower quartiles (LoPRESTO) on PRESTO in quiet completed a battery of behavioral tasks designed to assess spectral resolution and neurocognitive functioning.

Study Sample Twenty-one postlingually deafened adult CI users, with 11 HiPRESTO and 10 LoPRESTO participants.

Data Collection and Analysis A discriminant analysis was carried out to determine the extent to which measures of spectral resolution and neurocognitive functioning discriminate HiPRESTO and LoPRESTO CI users. Auditory spectral resolution was measured using the Spectral-Temporally Modulated Ripple Test (SMRT). Neurocognitive functioning was assessed with visual measures of working memory (digit span), inhibitory control (Stroop), speed of lexical/phonological access (Test of Word Reading Efficiency), and nonverbal reasoning (Raven's Progressive Matrices).

Results HiPRESTO and LoPRESTO CI users were discriminated primarily by performance on the SMRT and secondarily by the Raven's test. No other neurocognitive measures contributed substantially to the discriminant function.

Conclusions High- and low-performing CI users differed by spectral resolution and, to a lesser extent, nonverbal reasoning. These findings suggest that the extreme groups are determined by global factors of richness of sensory information and domain-general, nonverbal intelligence, rather than specific neurocognitive processing operations related to speech perception and spoken word recognition. Thus, although both bottom-up and top-down information contribute to speech recognition performance, low-performing CI users may not be sufficiently able to rely on neurocognitive skills specific to speech recognition to enhance processing of spectrally degraded input in adverse conditions involving high talker variability.


#

Introduction

Cochlear implants (CIs) are auditory prosthetic devices that restore a sense of hearing, albeit degraded in nature, to individuals with moderate-to-profound sensorineural hearing loss. CIs have been very effective as a medical treatment, and many adult postlingually deafened CI recipients are generally able to achieve a high level of speech understanding in quiet listening conditions. Despite the overall success of CIs, enormous individual differences in outcomes in this patient population are routinely observed in CI centers around the world (Lazard et al;[27] Blamey et al)[6] [5], and rates of ‘‘poor’' speech understanding in CI users are high; for example, 13% of patients score <10% correct in recognizing words in sentences in quiet (Lenarz et al)[27]. The factors underlying individual differences in speech recognition have not been fully explained and integrated, posing a major clinical problem for optimizing CI outcomes.

Variability in basic auditory sensitivity (e.g., spectral and temporal resolution) may help predict individual differences in speech recognition in postlingually deafened adult CI users (e.g., Moberly et al)[30]. CI users rely on a signal that is highly degraded in spectrotemporal detail because of the limitations of the electrode-nerve interface and the relatively broad electrical stimulation of the auditory nerve. In particular, spectral information, important for phonetic perception, is poorly conveyed by the CI device, limiting speech recognition abilities (Henry et al)[19]. However, spectral resolution across the electrode array appears to vary across individual implant users (Won et al)[59]. Differences in spectral resolution and speech recognition have typically been associated with patient-related demographic characteristics, such as age, age at implantation, duration of deafness, and residual hearing, as well as device- or surgical-related factors, such as implant model, implant settings, and surgical technique. Regardless of the underlying factors, poorer spectral resolution for some CI users may contribute to increased difficulty in recognizing speech (Henry et al;[19] Won et al;[59] Moberly et al)[35]. CI users with good spectral resolution may be better able to use the increased acoustic-phonetic spectral details provided by modern CI devices (Croghan et al)[9].

Beyond differences in the initial ‘‘bottom-up’' sensory input, related to physiologic or device-related factors, there also appear to be differences in ‘‘top-down’' processing, the capacity of individual CI users to make use of neurocognitive processes and language knowledge to understand the degraded sensory information. Neurocognitive and linguistic skills have been found to be related to individual differences in speech recognition in hearing-impaired adults (e.g., Arehart et al)[2] and adult CI users (e.g., Heydebrand et al;[20] Lazard et al;[26] Holden et al)[21]. Yet, compared with other demographic and device-related factors, the involvement and relative contributions of specific neurocognitive skills in speech recognition in diverse listener populations are still unclear. Some specific neurocognitive skills, such as working memory capacity, appear to be more strongly related to speech perception and recognition and may be more strongly related than more general cognitive abilities, such as nonverbal intelligence (for a review, see Akeroyd)[1]. Working memory (Lyxell et al;[29] Tao et al)[52] as well as inhibitory control (Moberly et al[31]), verbal learning and memory (Pisoni et al)[25], and processing speed (Tinnemore et al)[53] have been linked to individual differences in speech recognition among adult CI users. In addition, although a strong relation has not been established, nonverbal reasoning skills have recently been found to be associated with individual performance among postlingually deafened adult CI users, independently of age (Mattingly et al)[30].

Individual differences among CI users may be further exacerbated in adverse listening conditions commonly encountered in daily life. Normal-hearing (NH) listeners are able to recognize and successfully understand speech under an enormously wide range of adverse and challenging conditions, such as noise, reverberation, or the presence of competing talkers. In real-world listening conditions, listeners are required to rapidly adapt not only to these diverse listening environments but also to highly variable acoustic changes in the vocal source, reflecting the speaker's gender, age, regional dialect, speaking rate, and speaking style, to facilitate speech communication (e.g., Pisoni)[41]. Although NH listeners appear to adjust quickly and almost effortlessly to acoustic differences in speech (e.g., Johnsrude et al;[23] Souza et al)[2] [2], talker variability places substantial perceptual and cognitive load on basic speech recognition processes. Increased talker variability creates additional cognitive load and requires greater cognitive processing resources than spoken materials with lower talker variability (e.g., Nusbaum and Magnuson).[39] As such, high-variability speech with multiple talkers has been shown to be more challenging to recognize for NH listeners (e.g., Mullennix et al)[36]. Furthermore, studies with NH listeners also revealed large individual differences in performance and suggested several basic perceptual and neurocognitive skills as factors underlying multitalker sentence recognition (e.g., Tamati et al;[50] Tamati and Pisoni)[49].

Auditory sensitivity and neurocognitive skills may show different relative contributions to speech recognition for CI users in general, but also, more specifically, to high- versus low-performing CI users on speech recognition in testing conditions with high talker variability. Early research on CI speech recognition focused on the high-performing CI users (or ‘‘stars’') as a way to demonstrate the efficacy or limitations of CIs, whereas some recent research works have tried to investigate speech recognition performance in low-performing CI users (e.g., Moberly et al;[31] Pisoni et al)[33]. Although no standard criteria exist for determining high and low performers, previous studies have observed substantial variability across CI users. For example, some CI users obtain near ceiling performance on sentence recognition in quiet, whereas others perform quite poorly (e.g., <10%) on tasks measuring sentence recognition in quiet (Lenarz et al)[27]. In principle, high- and low-performing CI users might represent two distinct groups of patients whose outcomes may be more specifically predicted by differences in underlying neurocognitive functioning and auditory skills. These two groups often show large differences in performance in speech recognition tasks: low-performing CI users are more susceptible to noise (e.g., Fu and Nogaki)[15], as well as other sources of signal degradations, such as speech variability (e.g., Tamati et al)[51]. Furthermore, although both groups perform more poorly under adverse listening conditions, it appears that some high-performing CI users are better able to take advantage of top-down compensatory mechanisms (Bhargava et al)[4]. These performance differences suggest that the extreme groups of listeners may demonstrate different underlying information-processing mechanisms that reflect the sensory and neurocognitive resources the listeners have with them.

Differences in auditory sensitivity to spectral and temporal properties of speech and/or neurocognitive skills between high and low performers may lead them to adopt different perceptual strategies. Indeed, there is some evidence that CI users demonstrate individual differences in their perceptual strategies, such as the weights they assign to spectral and temporal cues (Winn et al;[57] Bhargava et al;[4] Moberly et al)[34], resulting in group differences not only in speech recognition accuracy but also in the underlying mechanisms. Although previous research has assumed that both auditory sensitivity and neurocognitive functioning contribute to speech recognition performance in CI users (e.g., Bhargava et al)[4], these factors may not contribute equally to performance differences among high and low performers. The specific neurocognitive and auditory skills that are associated with high or low performance are unclear, in particular, with respect to whether these groups can be defined by more general auditory sensitivity and/or neurocognitive skills or whether they are differentiated by specific subskills related to speech perception and recognition.

The Current Study

The aim of the present study was to examine the relative contributions of spectral resolution and several neurocognitive skills that differentiate speech recognition outcomes among high- and low-performing experienced CI users, as defined by performance in the top versus bottom quartile on sentence recognition relative to our study population. For the present study, the relative speech recognition skills of the CI users were determined using scores from the Perceptually Robust English Sentence Test Open-set (PRESTO; Gilbert et al)[18]. PRESTO maximizes talker variability by incorporating multiple talkers, genders, and regional accents. The PRESTO materials have been shown to be more challenging to recognize than sentence materials with lower talker variability, such as Hearing in Noise Test (HINT; Nilsson et al)[37] and AzBio (Spahr et al)[46] sentences, for NH listeners and hearing-impaired listeners with CIs (Gilbert et al)[50]. In addition, PRESTO materials yield large individual dif-ferences in performance, which have been found to be related to several neurocognitive skills (Tamati et al;[50] Moberly et al)[30]. Taken together, these earlier studies suggest that high-variability speech recognition cannot be attributed to peripheral hearing acuity or audibility alone and reflect complex interactions of auditory sensitivity and neurocognitive processes.

A discriminant function analysis was carried out to determine the extent to which measures of auditory sensitivity, specifically auditory spectral resolution, and neurocognitive functioning discriminate the high- versus low-performing CI users on PRESTO. Auditory spectral resolution was measured with a behavioral test of spectral resolution, specifically, the Spectral-Temporally Modulated Ripple Test (SMRT; Aronoff and Landsberger)[3]. Several neurocognitive functions were measured, including working memory, inhibitory control, verbal learning and memory, lexical/phonological processing speed, and a global measure of nonverbal reasoning. Based on previous studies, we predicted that spectral resolution would discriminate the high- and low-performing groups. More specifically, if spectral resolution primarily determines whether a CI user will achieve a high or low level of speech recognition in adverse conditions, then the contribution of spectral resolution to the discriminant function should be the greatest among the measures. For the neurocognitive skills, we anticipated two possible outcomes. If poor spectral resolution and weak neurocognitive functioning combine to result in extremely low performance (and good spectral resolution and strong cognitive skills combine for high performance), then neurocognitive skills would be expected to help discriminate the groups. Alternatively, if low performers receive such a poor signal that they cannot benefit from neurocognitive skills, regardless of their respective individual abilities, then measures of neurocognitive functioning may not contribute to the discriminant function. Thus, spectral resolution was expected to be the most useful primary factor in discriminating the groups, with a potential secondary contribution of neurocognitive functioning.


#
#

Methods

Participants

Data from 21 postlingually deafened adult CI users were included in the present study. All of the CI users had participated in a larger study on speech, language, and neurocognitive skills, which, at the time of the present study, included 44 adult postlingually deafened CI users (Moberly et al;[33] Kramer et al;[25] Moberly et al)[35]. Of the 21 CI users in the present study, 11 scored within the upper quartile of the distribution (HiPRESTO) and 10 scored within the lower quartile (LoPRESTO) of a group of 44 CI users (overlapping with participant groups reported in earlier studies) on PRESTO, after adjusting for those who did not complete or were unable to complete the SMRT (n = 4). The 11 HiPRESTO CI users included 5 female and 6 male participants, aged 45-70 years (M = 60; standard deviation [SD] = 8), and the 10 LoPRESTO CI users included 4 female and 6 male participants, aged 52-81 years (M = 65; SD = 10).

All participants were native English speakers, with at least a high school diploma or equivalent, with relatively normal general language proficiency as demonstrated by word reading scores within two SD of the normative mean on the Wide Range Achievement Test (Wilkinson and Robertson)[56], and no evidence of significant cognitive impairment based on a written version of the Mini-Mental State Examination (Folstein et al)[13]. Additional demographic information for the CI participants is provided in [Table 1]. The 11 HiPRESTO CI users included five bilateral and six unilateral users (five with contralateral hearing aid use) and the ten LoPRESTO CI users included two bilateral and eight unilateral users (four with contralateral hearing aid use). The HiPRESTO CI users were on average 53.7 (SD = 9.1) years old at their first CI with a total duration of hearing loss (from age of initial hearing loss to present) of 32.8 (SD = 14.8) years, whereas the LoPRESTO CI users were on average 59.9 (SD = 11.7) years old with a total duration of hearing loss of 39.5 (SD = 19.2) years. Socioeconomic status (SES) was assessed using a measure developed by Nittrouer and Burton (2005), which incorporates occupational and educational levels; higher scores indicate higher SES (range of scores: 1-64). The HiPRESTO CI users scored on average 26.7 (SD = 8.8) using this measure, whereas the LoPRESTO CI users scored on average 28.7 (SD = 13.7). As can be seen in [Table 1], overall, HiPRESTO and LoPRESTO groups showed similar mean demographic characteristics, but with vast individual differences in these characteristics within both groups.

Table 1

Participant Demographics and Group Means (SD)

Group

Participant Number

Gender

Age (Years)

Age of First Cl (Years)

SES (out of 64)

Device

Side of Implant

Current Contralateral Hearing Aid Use

Etiology of Hearing Loss

Duration of Hearing Loss (Years)

Better Ear PTA (dB HL)

PRESTO Score (% Correct Keywords)

HiPRESTO

1

M

61

55

24

Cochlear

Right

Yes

Unknown

16

111.25

76.61

HiPRESTO

2

M

70

68

42

Cochlear

Left

Yes

Unknown, progressive loss as adult

30

93.75

76.61

HiPRESTO

3

F

67

58

9

Cochlear

Bilateral

No

Genetic, Meniere's disease

31

120.00

79.03

HiPRESTO

4

M

55

50

30

Cochlear

Bilateral

No

Unknown, progressive loss as an adult

34

120.00

79.84

HiPRESTO

5

M

66

60

18

Cochlear

Left

No

Meniere's disease

52

80.00

81.25

HiPRESTO

6

F

57

48

25

Cochlear

Right

Yes

Genetic, progressive loss as child/adult

50

82.50

81.45

HiPRESTO

7

M

59

57

24

Cochlear

Bilateral

No

Sudden loss

4

120.00

84.38

HiPRESTO

8

F

45

44

30

Advanced Bionics

Right

Yes

Unknown, progressive loss as child

39

88.75

84.68

HiPRESTO

9

F

65

54

24

Cochlear

Bilateral

No

Genetic

24

120.00

86.61

HiPRESTO

10

F

66

62

35

Cochlear

Right

Yes

Genetic, progressive loss as adult, noise exposure, sudden loss in the right ear

31

78.75

88.84

HiPRESTO

11

F

50

35

33

Bilateral

No

Unknown, progressive loss as child

50

117.50

95.16

HiPRESTO Mean (SD)

60.1 (7.8)

53.7 (9.1)

26.7 (8.8)

32.8 (14.8)

102.95 (18.06)

83.1 (5.6)

LoPRESTO

1

M

76

74

35

Cochlear

Left

Yes

Ototoxicity (amikacin)

53

96.25

4.02

LoPRESTO

2

F

52

46

42

Cochlear

Right

Yes

Unknown, born with HL

52

113.75

22.58

LoPRESTO

3

M

60

52

36

Cochlear

Bilateral

No

Measles

54

105.00

25.81

LoPRESTO

4

M

59

54

6

Cochlear

Right

No

Genetic

44

108.75

29.84

LoPRESTO

5

F

55

44

15

Cochlear

Left

No

Sudden, otosclerosis, progressive as adult

30

120.00

32.26

LoPRESTO

6

M

81

80

49

Cochlear

Left

Yes

Unknown, progressive as adult

13

75.00

33.06

LoPRESTO

7

F

64

59

24

Cochlear

Left

Yes

Unknown, progressive as child

55

95.00

33.87

LoPRESTO

8

F

62

59

14

Cochlear

Bilateral

No

Sepsis, ototoxicity

4

95.00

34.68

LoPRESTO

9

F

76

68

30

Cochlear

Left

No

Progressive loss as adult, probable autoimmune

31

108.75

37.05

LoPRESTO

10

F

65

63

36

Cochlear

Right

No

Genetic, progressive as adult

59

86.25

41.94

LoPRESTO Mean (SD)

65.0 (9.6)

59.9 (11.7)

28.7 (13.7)

39.5 (19.2)

100.38 (13.51)

29.5 (10.5)

F = female; M = male; PTA = pure-tone average.


Participants wore their typical hearing prostheses, including CIs for bilateral CI participants or any contralateral hearing aids for bimodal CI participants, except during an unaided audiogram. Testing with typical hearing prostheses was carried out to obtain more ecologically valid measures of sentence recognition and auditory functioning in which scores would provide a measure of how much linguistic and auditory information could be reasonably expected to be received and processed by the listener in a quiet listening environment. All participants provided informed written consent and received $15 per hour for participation. The local institutional review board approved the study protocol, and all testing took place at The Ohio State University. All participants provided informed written consent before participation.


#

Measures and Procedures

All measures were collected as part of a larger study on speech, language, and cognitive skills in adult postlingually deafened CI users. Testing was carried out at the Eye and Ear Institute of The Ohio State University Wexner Medical Center in sound-proof booths and acoustically insulated rooms. Visual stimuli were presented on a paper or a touch screen monitor (Keytec, Inc, Richardson, TX) placed two feet in front of the participant. Auditory stimuli were presented at 68 dB SPL, via a Roland MA-12C loudspeaker (Roland Corp, Los Angeles, CA) placed 1 m from the participant at 0° azimuth. Responses were audio-visually recorded via a Sony HDR-PH260 High Definition Handycam (Sony Corp, Tokyo, Japan) for later offline scoring. Measures used in the present study are described in the following paragraphs. For detailed information on the general approach, materials, and collection procedures, see Moberly et al (2017),[33] Kramer et al (2018),[25] and Moberly et al (2018).[35]


#

Sentence Recognition

To determine the group membership, PRESTO sentences (Gilbert et al)[18] were used to assess high- variability sentence recognition among the CI users. Scores were based on 32 PRESTO sentences, originally selected from the Texas Instruments/Massachusetts Institute of Technology Speech Corpus (TIMIT; Garofolo et al.)[17]. PRESTO maximizes talker variability by including sentences produced by multiple male and female talkers with different regional accents. Original PRESTO sentence lists were balanced for talker gender, keyword frequency, and keyword familiarity, with no repeated talkers. For the present study, participants were asked to repeat 32 sentences, with the first 2 as practice. Scores were percent keywords correct across the final 30 sentences. As discussed earlier, quartile scores from a larger distribution of 44 CI users were used to define HiPRESTO and LoPRESTO groups. For the purposes of the present study, five participants were removed before analysis because of unavailable PRESTO keyword scores (n =1) or SMRT scores (n = 4). Overall, the final group of 39 CI users had a mean score of 57.8% (SD = 21.7) keywords correct on PRESTO. The 11 HiPRESTO CI users had a mean score of 83.1% (SD = 5.6) keywords correct on PRESTO, whereas the 10 LoPRESTO CI users had a mean score of 29.5% (SD = 10.5) keywords correct on PRESTO.


#

Spectral Resolution

The CI users' spectral resolution was assessed using the SMRT (Aronoff and Landsberger)[3]. The stimuli consisted of 202 pure-tone frequency components with amplitudes spectrally modulated by a sine wave. The SMRT was carried out in a three-interval, two-alternative forced choice task in which two of the intervals contained a reference signal with 20 ripples per octave (RPO) and one contained the target signal. The target signal, set initially at 0.5 RPO, was modified using a one-up/one-down adaptive procedure with a step size of 0.2 RPO. A ripple detection threshold was calculated based on the last six reversals of each run, with the first three runs discarded as practice. Listeners selected the deviant (target) signal. A higher score represented better spectral resolution. For more details on the task, see Aronoff and Landsberger (2013).


#

Neurocognitive Functioning

Working Memory Capacity: Working memory capacity was assessed using a computerized version of a visual digit span task, based on the original auditory digit span task from the Wechsler Intelligence Scale for Children, Fourth Edition, Integrated (Wechsler)[55]. Participants were presented visually with lists of (2-7) digits and were instructed to reproduce each test list in the same order by touching the digits in a 3 × 3 matrix on the computer touch screen. The score was calculated based on the total correct items and used for final data analysis.

Inhibitory Control: A computerized Stroop test (http://millisecond.com) was used to assess inhibitory control, based on the original version by Stroop (1935). On each trial, a color word was presented on the computer screen, in either the same or different color font, and the participant was asked to press a key on a keyboard corresponding to the font color and not the name of the color word. Response times were calculated for both congruent (matching color word and font color) and incongruent trials (mismatching color word and font color). An interference score was calculated by subtracting the mean response time of the congruent trials from the mean response time of the incongruent trials. Response times from the congruent trials and interference scores were used for analysis as measures of processing speed and inhibitory control, respectively.

Verbal Learning and Memory: The California Verbal Learning Test, Version II (CVLT; Delis et al)[10], was used to assess verbal learning and memory, using a visual version created and validated in adult CI users by Pisoni et al (2018). On the computer screen, participants first saw a list of 16 familiar words, one at a time on the screen, from four semantic categories (List A). List A was presented in the same order five times, and participants recalled as many words as possible after each presentation as a mea-sure of repetition learning and free recall. The fifth presentation of List A was followed by an interference list of 16 new words from four semantic categories (List B). Participants recalled List B words as a measure of proactive interference. After the presentation and recall of List B, the participants then recalled List A again, as a measure of retroactive interference from List B. Finally, after all recall lists, a yes/no recognition memory test with the List A items was implemented to assess storage of items without retrieval demands. For the present study, three measures of CVLT performance were used for analysis: CVLT T1 and CVLT T5 (total words correctly recalled on first and fifth trials, respectively), List B (total words correctly recalled from List B presentation), and CVLT Y/N recognition discriminability (hits versus total false positives in the recognition memory test). (CVLT scores for one LoPRESTO CI user were not available. To carry out the discriminant analysis, group averages for the three CVLT scores were used for this participant.)

Lexical/Phonological Processing Speed: Lexical/Phonological processing speed was assessed with the Test of Word Reading Efficiency, Version 2 (TOWRE; Torgesen et al)[54]. TOWRE is a measure of word reading accuracy and fluency that measures the participant's ability to accurately recognize and identify familiar real words and to phonologically decode nonwords. Participants read as many real words as possible from a 108-word list or as many nonwords as possible from a 66-nonword list in 45 seconds. The two scores used in the present study were calculated based on the number of whole words correct and the number of whole nonwords correctly read aloud.

Nonverbal Reasoning: The Raven's Progressive Matrices test was used to measure nonverbal intelligence or abstract visuospatial reasoning (Raven)[44]. Participants were presented with incomplete visual patterns on a touch screen monitor and were asked to complete the visual pattern by selecting the best option from a closed set of alternatives. Participants completed as many items as possible in ten minutes; scores were number of correct items.


#

Statistical Analysis

The aim of the present study was to explore group differences between low-performing and high-performing CI users on high-variability sentence recognition assessed with PRESTO. First, means and SDs for HiPRESTO and LoPRESTO groups were calculated for each measure. An independent samples t-test was carried out to confirm group differences in PRESTO scores. Second, to examine the relations among the measures to be used in the discriminant analyses, Pearson correlations were calculated. Third, a discriminant function analysis was carried out to determine the extent to which the measures of spectral resolution and neurocognitive functioning could discriminate LoPRESTO and HiPRESTO CI users. A discriminant function analysis is a technique that uses a linear combination of independent variables to determine which of these variables help to explain between-group differences observed for a specific dependent variable.


#
#

Results

Means and SDs on each measure are displayed in [Table 2]. An independent sample t-test demonstrated that the difference in overall performance on PRESTO between the HiPRESTO and LoPRESTO groups was significant [t (19) = 14.83, p < 0.001], confirming the group selection criteria.

Table 2

Mean (SD) Scores for LoPRESTO and HiPRESTO Groups

Group

LoPRESTO (n = 10)

HiPRESTO (n = 11)

Sentence recognition

 PRESTO keywords correct (%)

29.5 (10.5)

83.1 (5.6)

Neurocognitive functioning

 Working memory capacity

  Digit span (# correct)

47.5 (20.9)

47.3 (20.6)

 Information-processing speed and inhibitory control

  Stroop control (congruent) condition (msec)

1,148.9 (250)

1,077.7 (245)

  Stroop interference (msec)

142.7 (293.0)

194.0 (118.6)

 Speed of phonological and lexical access

  TOWRE nonwords (% nonwords correct)

68.8 (11.9)

76.8 (7.3)

  TOWRE words (% words correct)

72.0 (9.6)

79.3 (5.7)

 Verbal learning and memory

  CVLT T1/T5 (# correct)

50.9 (10.5)

49.1 (10.5)

  CVLT List B (# correct)

29.3 (11.8)

36.4 (12.1)

  CVLT Y/N discriminability (hits vs. false positives)

2.9 (0.66)

3.2 (0.64)

 Nonverbal reasoning

  Raven's (# correct)

9.7 (3.6)

14.7 (4.1)

Spectral resolution

 SMRT (ripple-resolution threshold)

1.1 (0.4)

3.6 (1.6)

Correlational analyses were carried out to explore the relations between the measures, shown in [Table 3]. Overall, very few strong correlations emerged among the measures. The SMRT scores (spectral resolution) were correlated with the TOWRE word scores (r = 0.45, p = 0.039) and Raven's scores (r = 0.72, p < 0.001). Raven's scores were moderately correlated with the Stroop control scores (r = –0.52, p = 0.015). Finally, some significant relations emerged for the CVLT measures: CVLT T1/T5 scores were mod-erately correlated with CVLT List B scores (r = 0.56, p = 0.01) and CVLT Y/N recognition discriminability scores (r = 0.65, p = 0.002). Because the measures were not consistently and strongly related, all selected measures were used in the discriminant analysis.

Table 3

Pearson Correlations between Measures Across All Participants

Measure

1

2

3

4

5

6

7

8

9

10

1.

Digit span

2.

Stroop control

–0.03

3.

Stroop interference

0.03

0.11

4.

TOWRE nonwords

0.07

–0.09

0.22

5.

TOWRE words

–0.28

0.09

0.42

0.22

6.

CVLT T1/T5

0.06

–0.22

–0.19

0.02

–0.29

7.

CVLT List B

0.09

–0.34

–0.01

0.03

0.04

0.56

8.

CVLT Y/N discriminability

0.06

–0.34

0.00

0.04

–0.36

0.65

0.37

9.

Raven’s

0.18

–0.52

0.19

0.32

0.30

0.16

0.23

0.44

10.

SMRT

–0.18

–0.40

–0.01

0.18

0.45

0.18

0.28

0.39

0.72

Note: r values are reported. Statistically significant (<0.05) correlations are in bold.


Results of the discriminant analysis showed that LoPRESTO and HiPRESTO groups were significantly classified by the measures, Wilks' λ = 0.21 [X2 (10) = 22.31, p = 0.014]. The Wilks' λ statistic (provided in SPSS) represents the proportion of the variance not explained by group differences (the range of possible values is 0.0-1.0; lower values emerge when within-group variability is small compared with the total variability, indicating that the groups differ). The discriminant function had an eigenvalue, representing the discriminating ability of the discriminant function, of 3.92 (larger values are associated with a stronger discriminating ability) and a canonical correlation, representing the correlation between the discriminant scores and the groupings of the dependent variable, of 0.89 (high correlations—perfect is 1.00—are associated with a stronger discriminating ability). Overall, 100% of the original cases could be correctly classified, superior to random assignment based on prior group membership probabilities (Tabachnik and Fidell)[48]. Because classification can be quite high when sample sizes are small, a leave-one-out cross-validation classification was also used. The leave-one-out crossvalidation method classifies each case based on the functions derived from all other cases and can be used to assess a model's performance. Using cross-validation, 76.2% of the cases were correctly classified, suggesting that the model is stable and robust.

Structure matrix coefficients are listed in [Table 4], indicating the relative importance of the measures in predicting group membership. Structure matrix coefficients, representing the correlation between independent variables and the discriminant function, are generally considered to be important if greater than |10.30|. The results indicate that LoPRESTO and HiPRESTO groups are primarily discriminated by SMRT scores (0.58) and Raven's scores (0.35). [Figures 1] and [2] show the SMRT and Raven's scores, respectively, for HiPRESTO and LoPRESTO groups. Although other measures contribute to the discriminant function, they were relatively less important than these two primary measures to discriminate the LoPRESTO and HiPRESTO groups. Thus, the HiPRESTO group is characterized by high spectral resolution and stronger nonverbal reasoning skills.

Table 4

Structure Matrix Coefficients

Measure

Structure Matrix Coefficients

Rank

SMRT

0.58

1

Raven’s

0.35

2

TOWRE words

0.25

3

TOWRE nonwords

0.22

4

CVLT List B

0.16

5

CVLT Y/N discriminability

0.12

6

Stroop control

–0.08

7

Stroop interference

0.06

8

CVLT T1/T5

–0.04

9

Digit span

0.00

10

Zoom Image
Fig. 1 Average SMRT scores for HiPRESTO and LoPRESTO groups. The boxes extend from the lower to the upper quartile (the interquartile range, IQ), and the midline indicates the median. The whiskers indicate the highest and lowest values not >1.5 times the IQ. Light gray circles represent individual scores. Circles containing an ‘‘x’' indicate outliers, that is, data points >1.5 times the IQ.
Zoom Image
Fig. 2 Same as Figure 1, except average Raven's scores are shown for HiPRESTO and LoPRESTO groups.

#

Discussion

The goal of the present study was to determine whether high- and low-performing postlingually deafened adult CI users could be discriminated by measures of auditory spectral resolution and neurocognitive functioning. The results of the discriminant function analysis, using a measure of spectral resolution and several measures of neurocognitive skills, demonstrated that spectral resolution, primarily, and nonverbal reasoning, secondarily, predicted CI users' performance on PRESTO, a high-variability sentence recognition test. In particular, high-performing users were characterized by better spectral resolution and stronger nonverbal reasoning, whereas low-performing users were overall characterized by poorer spectral resolution and weaker nonverbal reasoning. These results suggest that both spectral resolution and neurocognitive functioning are important for discriminating high- and low-performing CI users in speech recognition in conditions with high talker variability, with spectral resolution as the greate st contributor to group differences.

Our first hypothesis was that spectral resolution would contribute the most to discriminating the high and low performers on PRESTO. This hypothesis was supported by our results. Spectral resolution was the primary contributor to the discriminant function. HiPRESTO users obtained significantly higher (better) SMRT thresholds than the LoPRESTO users, although the SMRT thresholds of the HiPRESTO group should be interpreted with caution, given the potential for spectral distortion of SMRT stimuli with high RPO values when presented through a CI processor (DiNino and Arenberg)[11]. The function of spectral resolution to speech recognition is well established. Studies using acoustic vocoder simulations of CI hearing have demonstrated that speech perception and recognition declines with decreasing spectral resolution (i.e., with decreasing number of spectral channels) (e.g., Fu et al;[16] Friesen et al)[14]. In addition, for individual CI users, differences in spectral resolution have been found to be related to individual differences in vowel and consonant perception (Henry et al)[19], word recognition (Won et al;[58] Drennen et al)[12], and sentence recognition (Moberly et al)[35], especially in noise (Won et al)[59]. The role of reduced spectral resolution may be further strengthened in conditions with high talker variability because listeners are forced to rely primarily on bottom-up cues when perceptual normalization for talker and accent variability is difficult (Clopper)[7]. The results from the present study replicate those previous findings, suggesting that spectral resolution is important in accounting for individual differences in CI speech recognition. The present study extends previous findings by demonstrating that spectral resolution is particularly important for determining high and low performance in speech recognition in adverse conditions with high talker variability. However, the extent to which the relative importance of spectral resolution to discriminating the high- and low-performing groups depends on the use of a high-variability talker sentence recognition test is unclear. More research studies should examine the contribution of auditory sensitivity to spectral and temporal properties of speech to speech recognition under different adverse listening conditions.

Regarding neurocognitive functioning, we predicted that neurocognitive skills may not discriminate the high- versus low-performing groups in the case that the signal is so degraded for the low-performing CI users that they could not benefit from compensatory top-down skills. However, if neurocognitive functioning still plays a role even with poor spectral resolution, such that both combine for either extremely low or extremely high performance, then neurocognitive skills would be expected to help discriminate the groups. Our results showed that neurocognitive abilities also contributed to the discriminant function, supporting the second hypothesis. In particular, nonverbal reasoning, as measured by the Raven's Progressive Matrices test, helped in discriminating the low- and high-performing CI users. However, note that because the focus of the present study was to determine whether auditory spectral resolution and neurocognitive functioning discriminate high- and low-performers on the PRESTO rather than to explain individual differences, scores from neurocognitive tests were not adjusted by demographic factors, such as age, SES, and gender, for the analyses. Given possible associations between demographic factors and cognitive skills (e.g., Kramer et al)[25], some of the differences between groups could be attributed to these factors. Moreover, had normalized scores been used, the discriminant function may have had a weaker discriminating ability.

Nonverbal reasoning has not consistently emerged as a significant predictor of speech recognition in CI users (e.g., Akeroyd;[1] but see Knutson et al;[24] Holden et al)[21]. Moberly et al (2018)[35] found that the strongest neurocognitive predictor of speech recognition among a similar group of CI users was nonverbal reasoning, using the same Raven's Progressive Matrices test as in the present study. In both studies, a timed version of the Raven's test was used—participants had to complete as many items as possible within ten minutes. Thus, this measure may incorporate processing speed, beyond nonverbal reasoning. If this were the case, we would also expect that scores on the Raven's test to be related to other neurocognitive measures involved in processing speech, including the timed Stroop and TOWRE measures. Scores on the Raven's test showed a fairly high, but not significant, cor-relation with the Stroop and TOWRE measures. Furthermore, they were more strongly related to the Stroop control measure from congruent trials, which involves processing speed, in the predicted negative direction (faster Stroop associated with fewer correct Raven's items). Although all these tasks may involve processing speed to some extent, the group means did not differ substantially and a strong relation between these measures and the discriminant function did not emerge. The lack of contribution of specific processing operations related to speech recognition, such as working memory (see Akeroyd)[1], and the discriminant function may be explained by the focus on the extreme groups. The LoPRESTO group may be receiving such a poor signal that these specific cognitive domains, which may account for some variability among a larger distribution of CI users with better spectral resolution, are not engaged in supporting speech processing in this group. By contrast, domain-general nonverbal reasoning measured by the Raven's test, which reflects the ability to reason and solve novel problems, may underlie performance across all the auditory tasks in the present study, including the SMRT. Performance on the SMRT and Raven's test was also strongly correlated in this group (r = 0.72), as shown in [Table 3]. In addition to procedural commonalities of task performance, performance on SMRT and the Raven's test may be influenced by the same underlying mechanisms; specifically, long durations of deafness without substantial auditory input may contribute to both poor spectral resolution and nonverbal reasoning. However, given that the HiPRESTO and LoPRESTO groups did not vary greatly in duration of deafness or other typical demographic factors (see [Table 1]), the potential factor(s) underlying the relation between SMRT and nonverbal reasoning is not apparent. Alternatively, because the present study involves a unique group of experienced adult postlingually deafened CI users, the observed relations between spectral resolution and nonverbal reasoning and PRESTO performance may not be representative of the larger, more diverse sample of adult CI users with varying degrees of experience with their CIs. Nonetheless, given that the present study was specifically designed to test broadly whether both auditory sensitivity and neurocognitive skills contribute to group differences among high- and low-performing CI users, and not specifically to test which domains, the results suggest that in addition to spectral resolution, a domain-general global measure of nonverbal intelligence discriminates higher and lower performance on speech recognition in adverse conditions.

Regardless of the specific neurocognitive skills involved and how they may be related to demographic factors, the results suggest that neurocognitive functioning contributes to low and high performance together with auditory spectral resolution. To better understand the contribution of these factors, we examined individual performance on the spectral resolution and nonverbal reasoning measures in more detail. The HiPRESTO group seemed to benefit from both better spectral resolution and stronger nonverbal reasoning. Of the HiPRESTO CI users, 8/11 appeared in the top 11 in spectral resolution scores, with 3/11 falling in the middle of the overall distribution. In addition, 8/11 appeared in the top 11 in nonverbal reasoning scores, with 3/11 falling in the middle of the overall distribution. The LoPRESTO group appears to be more variable in their scores, at least in nonverbal reasoning. For spectral resolution, 7/10 low-performing CI users also appeared in the bottom 10, with 3/10 falling in the middle of the overall distribution. For nonverbal reasoning, only 2/10 LoPRESTO CI users also appeared in the bottom 10 in nonverbal reasoning scores, with 7/10 in the middle and 1/10 in the top of the overall distribution. Interestingly, the LoPRESTO CI user with strong nonverbal reasoning had one of the worst spectral resolution scores (0.60 on the SMRT). Thus, the contributory role of neurocognitive functioning to speech recognition among the low performers may be relatively limited and constrained by the amount of useful acoustic-phonetic information provided by the CI for higher levels of processing.

With reduced sensory input, listeners tend to use perceptual strategies relying on top-down linguistic or contextual information (e.g., Pichora-Fuller et al;[40] Janse and Ernestus)[22]. Although most CI users may be able to engage neurocognitive resources to compensate for a degraded signal, the ability to apply these compensation mechanisms (and potentially benefit from stronger neurocognitive skills) is likely to be reduced for CI users with poor spectral resolution. The findings from the present study suggest a limited role for cognitive skills in CI users with poor auditory sensitivity (Collison et al)[8] and imply a reduced ability to rely on topdown compensation mechanisms for CI users with lower overall speech recognition skills (e.g., Chatterjee et al;[6] Bhargava et al)[4]. These limitations may lead low-performing CI users to adopt different perception strategies than high-performing users in completing complex speech recognition tasks, as further individual differences within the group reflect variability in neurocognitive skills. Differences in spectral resolution between high- and low-performing CI users, and within the low-performing group, may emerge in different types of adverse conditions, which demand varying reliance on bottom-up or top-down information. Additional research should be carried out to examine the perceptual strategies of the high- versus low-performing CI users, and their reliance on top-down compensation mechanisms in a variety of adverse conditions.

Taken together, the results of the present study emphasize the importance of both auditory sensitivity and neurocognitive functioning in accounting for individual differences in CI speech recognition in adverse conditions with high talker variability. To obtain high sentence recognition accuracy in these conditions, good spectral resolution and strong cognitive skills appear to be necessary. Overall, low-performing CI users seem to be limited by poor spectral resolution and weak cognitive skills, combined. However, spectral resolution may play the most important and limiting role for low-performing CI users; some CI users may be restricted by very poor spectral resolution, regardless of their individual cognitive capacities. These findings have some potential clinical implications in the development of more effective rehabilitative treatment protocols for adult postlingually deafened CI users. High-performing CI users, and possibly those falling in the middle of the distribution, may benefit from a broad treatment protocol involving both adjustments to the sensory input and training in neurocognitive and linguistic skills. Because the signal is already rich enough for them to effectively engage top-down mechanisms, neurocognitive or linguistic training may yield improvements in their ability to deal with degraded signals in adverse listening conditions. Conversely, low-performing CI users may more greatly benefit from improvements in sensory information. For these users, stronger neurocognitive or linguistic skills through training would not be expected to lead to improved recognition without at least some improvement in sensory information. These users may primarily benefit from improvements in signal processing, through new processing or speech coding strategies, or auditory training specifically targeting spectral resolution, such as training involving spectral ripple noise. Future research should be carried out to further examine the contributions of auditory sensitivity and neurocognitive functioning in a variety of challenging and adverse conditions to help predict speech recognition outcomes and to inform rehabilitative treatments for postlingually deafened adult CI users.


#

Abbreviations

CI: cochlear implant
CVLT: California Verbal Learning Test
HiPRESTO: group of CI users in the upper quartile of performance distribution on PRESTO
LoPRESTO: group of CI users in the lower quartile of performance distribution on PRESTO
NH: normal hearing
PRESTO: Perceptually Robust English Sentence Test Open-set
RPO: ripples per octave
SD: standard deviation
SMRT: Spectral-Temporally Modulated Ripple Test
TOWRE: Test of Word Reading Efficiency
WRAT: Wide Range Achievement Test


#

Die Autoren geben an, dass kein Interessenkonflikt besteht.

  • References

  • 1 Akeroyd MA. Are individual differences in speech reception related to individual differences in cognitive ability? A survey of twenty experimental studies with normal and hearing-impaired adults. Int J Audiol 2008; 47: S53-S71
  • 2 Arehart KH, Souza P, Baca R, Kates JM. Working memory, age and hearing loss: susceptibility to hearing aid distortion. Ear Hear 2013; 34: 251-260
  • 3 Aronoff JM, Landsberger DM. The development of a modified spectral ripple test. J Acoust Soc Am 2013; 134: EL217-EL222
  • 4 Bhargava P, Gaudrain E, Basxkent D. Top–down restoration of speech in cochlear-implant users. Hear Res 2014; 309: 113-123
  • 5 Blamey P, et al. Factors affecting auditory performance of postlingually deaf adults using cochlear implants: an update with 2251 patients. Audiol Neuro Otol 2013; 18: 36-47
  • 6 Chatterjee M, Peredo F, Nelson D, Basxkent D. Recognition of interrupted sentences under conditions of spectral degradation. J Acoust Soc Am 2010; 127: EL37-EL41
  • 7 Clopper CG. Effects of dialect variation on the semantic predictability benefit. Lang Cogn Process 2012; 27: 1002-1020
  • 8 Collison EA, Munson B, Carney AE. Relations among linguistic and cognitive skills and spoken word recognition in adults with cochlear implants. J Speech Lang Hear Res 2004; 47: 496-508
  • 9 Croghan NBH, Duran SI, Smith ZM. Re-examining the relationship between number of cochlear implant channels and maximal speech intelligibility. J Acoust Soc Am 2017; 142: EL537
  • 10 Delis DC, Kramer J, Kaplan E, Ober BA. California Verbal Learning Test, 2. San Antonio, TX: The Psychological Corporation; 2000
  • 11 DiNino M, Arenberg JG. Age-related performance on vowel identification and the spectral-temporally modulated ripple test in children with normal hearing and with cochlear implants. Trends Hear 2018; 22: 1-20
  • 12 Drennen WR, Anderson ES, Won JH, Rubinstein JT. Validation of a clinical assessment of spectral ripple resolution for cochlear-implant users. Ear Hear 2014; 35: E92-E98
  • 13 Folstein MF, Folstein SE, McHugh PR. “Mini-mental state” A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res 1975; 12: 189-198
  • 14 Friesen LM, Sannon RV, Basxkent D, Wang X. Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. J Acoust Soc Am 2001; 110: 1150-1163
  • 15 Fu QJ, Nogaki G. Noise susceptibility of cochlear implant users: the role of spectral resolution and smearing. J Assoc Res Otolaryngol 2005; 6: 19-27
  • 16 Fu QJ, Shannon RV, Wang X. Effects of noise and spectral resolution on vowel and consonant recognition: acoustic and electric hearing. J Acoust Soc Am 1998; 104: 3586-3596
  • 17 Garofolo JS, Lamel LF, Fisher WM, Fiscus JG, Pallett DS, Dahlgren NL. The DARPA TIMIT Acoustic-Phonetic Continuous Speech Corpus. Philadelphia, PA: Linguistic Data Consortium; 1993
  • 18 Gilbert JL, Tamati TN, Pisoni DB. Development, reliability and validity of PRESTO: a new high-variability sentence recognition test. J Am Acad Audiol 2013; 24: 1-11
  • 19 Henry BA, Turner CW, Behrens A. Spectral peak resolution and speech recognition in quiet: normal hearing, hearing impaired and cochlear implant listeners. J Acoust Soc Am 2005; 118: 1111-1121
  • 20 Heydebrand G, Hale S, Potts L, Gotter B, Skinner M. Cognitive predictors of improvements in adults’ spoken word recognition six months after cochlear implant activation. Audiol Neuro Otol 2007; 12: 254-264
  • 21 Holden LK, Finley CC, Firszt JB, Holden TA, Brenner C, Potts LG, Gotter BD, Vanderhoff SS, Mispagel K, Heydebrand G, Kinner MW. Factors affecting open-set recognition in adults with cochlear implants. Ear Hear 2013; 34: 342-360
  • 22 Janse E, Ernestus M. The roles of bottom-up and top-down information in the recognition of reduced speech: evidence from listeners with normal and impaired hearing. J Phonet 2011; 39: 330-343
  • 23 Johnsrude IS, Mackey A, Hakyemez H, Alexander E, Trang HP, Carlyon RP. Swinging at a cocktail party: voice familiarity aids speech perception in the presence of a competing voice. Psychol Sci 2013; 24: 1995-2004
  • 24 Knutson JF, Gantz BJ, Hinrichs JV, Schartz HA, Tyler RS, Woodsworth G. Psychological predictors of audiological outcomes of multichannel cochlear implants: preliminary findings. Ann Otol Rhinol Laryngol 1991; 100: 817-822
  • 25 Kramer S, Vasil KJ, Adunka OF, Pisoni DB, Moberly AC. Cognitive functions in adult cochlear implant users, cochlear implant candidates, and normal-hearing listeners. Laryngoscope Investig Otolaryngol 2018; 3: 304-310
  • 26 Lazard DS, Lee HJ, Gaebler M, Kell CA, Truy E, Giraud AL. Phonological processing in post-lingual deafness and cochlear implant outcome. Neuroimage 2010; 49: 3443-3451
  • 27 Lazard DS, et al. Pre-, per- and postoperative factors affecting performance of postlingually deaf adults using cochlear implants: a new conceptual model over time. PLoS One 2012; 7: e48739
  • 28 Lenarz M, Sönmez H, Joseph G, Buchner A, Lenarz T. Cochlear implant performance in geriatric patients. Laryngoscope 2012; 122: 1361-1365
  • 29 Lyxell B, Andersson J, Andersson U, Arlinger S, Bredberg G, Harder H. Phonological representation and speech understanding with cochlear implants in deafened adults. Scand J Psychol 1998; 39: 175-179
  • 30 Mattingly JK, Castellanos I, Moberly AC. Nonverbal reasoning as a contributor to sentence recognition outcomes in adults with cochlear implants. Otol Neurotol 2018; 39 (10) e956-e963
  • 31 Moberly AC, Bates C, Harris MS, Pisoni DB. The enigma of poor performance by adults with cochlear implants. Otol Neurotol 2016; a 37: 1522-1528
  • 32 Moberly AC, Houston DM, Castellanos I. Non-auditory neurocognitive skills contribute to speech recognition in adults with cochlear implants. Laryngoscope Investig Otolaryngol 2016; b 1: 154-162
  • 33 Moberly AC, Castellanos I, Vasil KJ, Adunka OF, Pisoni DB. “Product” versus “process” measures in assessing speech recognition outcomes in adults with cochlear implants. Otol Neurotol 2017; 39: E195-E202
  • 34 Moberly AC, Lowenstein JH, Tarr E, Caldwell-Tarr A, Welling DB, Shahin AJ, Nittrouer S. Do adults with cochlear implants rely on different acoustic cues for phoneme perception than adults with normal hearing?. J Speech Lang Hear Res 2014; 57: 566-582
  • 35 Moberly AC, Vasil KJ, Wcinich T, Safdar N, Boyce L, Roup C, Holt R, Adunka O, Castellanos I, Sharifo V, Houston D, Pisoni DB. How does aging affect recognition of spectrally degraded speech?. Laryngoscope 2018; 128 (05) DOI: 10.1002/lary.27457.
  • 36 Mullennix JW, Pisoni DB, Martin CS. Some effects of talker variability on recall of spoken word recognition. J Acoust Soc Am 1989; 85: 365-378
  • 37 Nilsson M, Soli SD, Sullivan JA. Development of the Hearing in Noise Test for the measurement of speech reception thresholds in quiet and in noise. J Acoust Soc Am 1994; 95: 1085-1099
  • 38 Nittrouer S, Burton LT. The role of early language experience in the development of speech perception and phonological processing abilities: evidence from 5-year-olds with histories of otitis media with effusion and low socioeconomic status. J Commun Disord 2005; 38: 29-63
  • 39 Nusbaum HC, Magnuson JS. Talker normalization: phonetic constancy as a cognitive process. In: Johnson K, Mullennix JW. , ed. Talker Variability in Speech Processing. San Diego, CA: Academic Press; 1997: 109-132
  • 40 Pichora-Fuller MK, Schneider B, Daneman M. How young and old listeners listen to and remember speech in noise. J Acoust Soc Am 1995; 97: 593-608
  • 41 Pisoni DB. Some thoughts on “normalization” in speech perception. In: Johnson K, Mullennix JW. , ed. Talker Variability in Speech Processing. San Diego, CA: Academic Press; 1997: 9-32
  • 42 Pisoni DB, Broadstock A, Wucinich T, Safdar N, Miller K, Hernandez LR, Vasil K, Boyce L, Davies A, Harris MS, Castellanos I, Xu H, Kronenberger WG, Moberly AC. Verbal learning and memory after cochlear implantation in postlingually deaf adults: some new findings with the CVLT-II. Ear Hear 2018; 39: 720-745
  • 43 Pisoni DB, Kronenberger WG, Harris MS, Moberly AC. Three challenges for future research on cochlear implants. World J Otorhinolaryngol Hear Neck Surg 2017; 3: 240-254
  • 44 Raven J. The Raven’s progressive matrices: change and stability over culture and time. Cogn Psychol 2000; 41: 1-48
  • 45 Souza P, Gehani N, Wright R, McCloy D. The advantage of knowing the talker. J Am Acad Audiol 2013; 24: 689-700
  • 46 Spahr AJ, Dorman MF, Litvak LM, Van Wie S, Gifford RH, Loizou PC, Loiselle LM, Oakes T, Cook S. Development and validation of the AzBio sentence lists. Ear Hear 2012; 33: 112-117
  • 47 Stroop JR. Studies of interference in serial verbal reactions. J Exp Psychol 1935; 18: 643-662
  • 48 Tabachnik BG, Fidell LS. Using Multivariate Statistics. 4th ed. Boston, MA: Allyn and Bacon; 2001
  • 49 Tamati TN, Pisoni DB. Non-native speech recognition in high-variability listening conditions using PRESTO. J Am Acad Audiol 2014; 25: 869-892
  • 50 Tamati TN, Gilbert JL, Pisoni DB. Some factors underlying individual differences in speech recognition on PRESTO: a first report. J Am Acad Audiol 2013; 24: 616-634
  • 51 Tamati TN, Janse E, Basxkent D. Perceptual discrimination of speaking style under cochlear implant simulation. Ear Hear 2019; 40: 63-76
  • 52 Tao D, Deng R, Jiang Y, Galvin JJ, Fu QJ, Chen B. Contribution of auditory working memory to speech understanding in Mandarin-speaking cochlear implant users. PLoS One 2014; 9: E99096
  • 53 Tinnemore AR, Zion DJ, Kulkarni AM, Chatterjee M. Children’s recognition of emotional prosody in spectrally degraded speech is predicted by their age and cognitive status. EarHear 2018; 39: 874-880
  • 54 Torgesen JK, Wagner RK, Rashotte CA. Test of Word Reading Efficiency. Austin, TX: Pro-Ed; 1999
  • 55 Wechsler D. WISC-IV: Wechsler Intelligence Scale for Children, Integrated: Technical and Interpretive Manual. San Antonio, TX: Harcourt Brace and Company; 2004
  • 56 Wilkinson GS, Robertson GJ. Wide Range Achievement Test. 4th ed. Lutz, FL: Psychological Assessment Resources; 2006
  • 57 Winn MB, Won JH, Moon IJ. Assessment of spectral and temporal resolution in cochlear implant users using psychoacoustic discrimination and speech cue categorization. Ear Hear 2012; 37: E377-E390
  • 58 Won JH, Drennan WR, Kang RS, Rubinstein JT. Psychoacoustic abilities associated with music perception in cochlear implant users. Ear Hear 2010; 31: 796-805
  • 59 Won JH, Drennan WR, Rubinstein JT. Spectral-ripple resolution correlates with speech reception in noise in cochlear implant users. J Assoc Res Otolaryngol 2007; 8: 384-392

Address for correspondence

Terrin N. Tamati
Department of Otolaryngology - Head and Neck Surgery
The Ohio State University, Columbus, OH 43212

  • References

  • 1 Akeroyd MA. Are individual differences in speech reception related to individual differences in cognitive ability? A survey of twenty experimental studies with normal and hearing-impaired adults. Int J Audiol 2008; 47: S53-S71
  • 2 Arehart KH, Souza P, Baca R, Kates JM. Working memory, age and hearing loss: susceptibility to hearing aid distortion. Ear Hear 2013; 34: 251-260
  • 3 Aronoff JM, Landsberger DM. The development of a modified spectral ripple test. J Acoust Soc Am 2013; 134: EL217-EL222
  • 4 Bhargava P, Gaudrain E, Basxkent D. Top–down restoration of speech in cochlear-implant users. Hear Res 2014; 309: 113-123
  • 5 Blamey P, et al. Factors affecting auditory performance of postlingually deaf adults using cochlear implants: an update with 2251 patients. Audiol Neuro Otol 2013; 18: 36-47
  • 6 Chatterjee M, Peredo F, Nelson D, Basxkent D. Recognition of interrupted sentences under conditions of spectral degradation. J Acoust Soc Am 2010; 127: EL37-EL41
  • 7 Clopper CG. Effects of dialect variation on the semantic predictability benefit. Lang Cogn Process 2012; 27: 1002-1020
  • 8 Collison EA, Munson B, Carney AE. Relations among linguistic and cognitive skills and spoken word recognition in adults with cochlear implants. J Speech Lang Hear Res 2004; 47: 496-508
  • 9 Croghan NBH, Duran SI, Smith ZM. Re-examining the relationship between number of cochlear implant channels and maximal speech intelligibility. J Acoust Soc Am 2017; 142: EL537
  • 10 Delis DC, Kramer J, Kaplan E, Ober BA. California Verbal Learning Test, 2. San Antonio, TX: The Psychological Corporation; 2000
  • 11 DiNino M, Arenberg JG. Age-related performance on vowel identification and the spectral-temporally modulated ripple test in children with normal hearing and with cochlear implants. Trends Hear 2018; 22: 1-20
  • 12 Drennen WR, Anderson ES, Won JH, Rubinstein JT. Validation of a clinical assessment of spectral ripple resolution for cochlear-implant users. Ear Hear 2014; 35: E92-E98
  • 13 Folstein MF, Folstein SE, McHugh PR. “Mini-mental state” A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res 1975; 12: 189-198
  • 14 Friesen LM, Sannon RV, Basxkent D, Wang X. Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. J Acoust Soc Am 2001; 110: 1150-1163
  • 15 Fu QJ, Nogaki G. Noise susceptibility of cochlear implant users: the role of spectral resolution and smearing. J Assoc Res Otolaryngol 2005; 6: 19-27
  • 16 Fu QJ, Shannon RV, Wang X. Effects of noise and spectral resolution on vowel and consonant recognition: acoustic and electric hearing. J Acoust Soc Am 1998; 104: 3586-3596
  • 17 Garofolo JS, Lamel LF, Fisher WM, Fiscus JG, Pallett DS, Dahlgren NL. The DARPA TIMIT Acoustic-Phonetic Continuous Speech Corpus. Philadelphia, PA: Linguistic Data Consortium; 1993
  • 18 Gilbert JL, Tamati TN, Pisoni DB. Development, reliability and validity of PRESTO: a new high-variability sentence recognition test. J Am Acad Audiol 2013; 24: 1-11
  • 19 Henry BA, Turner CW, Behrens A. Spectral peak resolution and speech recognition in quiet: normal hearing, hearing impaired and cochlear implant listeners. J Acoust Soc Am 2005; 118: 1111-1121
  • 20 Heydebrand G, Hale S, Potts L, Gotter B, Skinner M. Cognitive predictors of improvements in adults’ spoken word recognition six months after cochlear implant activation. Audiol Neuro Otol 2007; 12: 254-264
  • 21 Holden LK, Finley CC, Firszt JB, Holden TA, Brenner C, Potts LG, Gotter BD, Vanderhoff SS, Mispagel K, Heydebrand G, Kinner MW. Factors affecting open-set recognition in adults with cochlear implants. Ear Hear 2013; 34: 342-360
  • 22 Janse E, Ernestus M. The roles of bottom-up and top-down information in the recognition of reduced speech: evidence from listeners with normal and impaired hearing. J Phonet 2011; 39: 330-343
  • 23 Johnsrude IS, Mackey A, Hakyemez H, Alexander E, Trang HP, Carlyon RP. Swinging at a cocktail party: voice familiarity aids speech perception in the presence of a competing voice. Psychol Sci 2013; 24: 1995-2004
  • 24 Knutson JF, Gantz BJ, Hinrichs JV, Schartz HA, Tyler RS, Woodsworth G. Psychological predictors of audiological outcomes of multichannel cochlear implants: preliminary findings. Ann Otol Rhinol Laryngol 1991; 100: 817-822
  • 25 Kramer S, Vasil KJ, Adunka OF, Pisoni DB, Moberly AC. Cognitive functions in adult cochlear implant users, cochlear implant candidates, and normal-hearing listeners. Laryngoscope Investig Otolaryngol 2018; 3: 304-310
  • 26 Lazard DS, Lee HJ, Gaebler M, Kell CA, Truy E, Giraud AL. Phonological processing in post-lingual deafness and cochlear implant outcome. Neuroimage 2010; 49: 3443-3451
  • 27 Lazard DS, et al. Pre-, per- and postoperative factors affecting performance of postlingually deaf adults using cochlear implants: a new conceptual model over time. PLoS One 2012; 7: e48739
  • 28 Lenarz M, Sönmez H, Joseph G, Buchner A, Lenarz T. Cochlear implant performance in geriatric patients. Laryngoscope 2012; 122: 1361-1365
  • 29 Lyxell B, Andersson J, Andersson U, Arlinger S, Bredberg G, Harder H. Phonological representation and speech understanding with cochlear implants in deafened adults. Scand J Psychol 1998; 39: 175-179
  • 30 Mattingly JK, Castellanos I, Moberly AC. Nonverbal reasoning as a contributor to sentence recognition outcomes in adults with cochlear implants. Otol Neurotol 2018; 39 (10) e956-e963
  • 31 Moberly AC, Bates C, Harris MS, Pisoni DB. The enigma of poor performance by adults with cochlear implants. Otol Neurotol 2016; a 37: 1522-1528
  • 32 Moberly AC, Houston DM, Castellanos I. Non-auditory neurocognitive skills contribute to speech recognition in adults with cochlear implants. Laryngoscope Investig Otolaryngol 2016; b 1: 154-162
  • 33 Moberly AC, Castellanos I, Vasil KJ, Adunka OF, Pisoni DB. “Product” versus “process” measures in assessing speech recognition outcomes in adults with cochlear implants. Otol Neurotol 2017; 39: E195-E202
  • 34 Moberly AC, Lowenstein JH, Tarr E, Caldwell-Tarr A, Welling DB, Shahin AJ, Nittrouer S. Do adults with cochlear implants rely on different acoustic cues for phoneme perception than adults with normal hearing?. J Speech Lang Hear Res 2014; 57: 566-582
  • 35 Moberly AC, Vasil KJ, Wcinich T, Safdar N, Boyce L, Roup C, Holt R, Adunka O, Castellanos I, Sharifo V, Houston D, Pisoni DB. How does aging affect recognition of spectrally degraded speech?. Laryngoscope 2018; 128 (05) DOI: 10.1002/lary.27457.
  • 36 Mullennix JW, Pisoni DB, Martin CS. Some effects of talker variability on recall of spoken word recognition. J Acoust Soc Am 1989; 85: 365-378
  • 37 Nilsson M, Soli SD, Sullivan JA. Development of the Hearing in Noise Test for the measurement of speech reception thresholds in quiet and in noise. J Acoust Soc Am 1994; 95: 1085-1099
  • 38 Nittrouer S, Burton LT. The role of early language experience in the development of speech perception and phonological processing abilities: evidence from 5-year-olds with histories of otitis media with effusion and low socioeconomic status. J Commun Disord 2005; 38: 29-63
  • 39 Nusbaum HC, Magnuson JS. Talker normalization: phonetic constancy as a cognitive process. In: Johnson K, Mullennix JW. , ed. Talker Variability in Speech Processing. San Diego, CA: Academic Press; 1997: 109-132
  • 40 Pichora-Fuller MK, Schneider B, Daneman M. How young and old listeners listen to and remember speech in noise. J Acoust Soc Am 1995; 97: 593-608
  • 41 Pisoni DB. Some thoughts on “normalization” in speech perception. In: Johnson K, Mullennix JW. , ed. Talker Variability in Speech Processing. San Diego, CA: Academic Press; 1997: 9-32
  • 42 Pisoni DB, Broadstock A, Wucinich T, Safdar N, Miller K, Hernandez LR, Vasil K, Boyce L, Davies A, Harris MS, Castellanos I, Xu H, Kronenberger WG, Moberly AC. Verbal learning and memory after cochlear implantation in postlingually deaf adults: some new findings with the CVLT-II. Ear Hear 2018; 39: 720-745
  • 43 Pisoni DB, Kronenberger WG, Harris MS, Moberly AC. Three challenges for future research on cochlear implants. World J Otorhinolaryngol Hear Neck Surg 2017; 3: 240-254
  • 44 Raven J. The Raven’s progressive matrices: change and stability over culture and time. Cogn Psychol 2000; 41: 1-48
  • 45 Souza P, Gehani N, Wright R, McCloy D. The advantage of knowing the talker. J Am Acad Audiol 2013; 24: 689-700
  • 46 Spahr AJ, Dorman MF, Litvak LM, Van Wie S, Gifford RH, Loizou PC, Loiselle LM, Oakes T, Cook S. Development and validation of the AzBio sentence lists. Ear Hear 2012; 33: 112-117
  • 47 Stroop JR. Studies of interference in serial verbal reactions. J Exp Psychol 1935; 18: 643-662
  • 48 Tabachnik BG, Fidell LS. Using Multivariate Statistics. 4th ed. Boston, MA: Allyn and Bacon; 2001
  • 49 Tamati TN, Pisoni DB. Non-native speech recognition in high-variability listening conditions using PRESTO. J Am Acad Audiol 2014; 25: 869-892
  • 50 Tamati TN, Gilbert JL, Pisoni DB. Some factors underlying individual differences in speech recognition on PRESTO: a first report. J Am Acad Audiol 2013; 24: 616-634
  • 51 Tamati TN, Janse E, Basxkent D. Perceptual discrimination of speaking style under cochlear implant simulation. Ear Hear 2019; 40: 63-76
  • 52 Tao D, Deng R, Jiang Y, Galvin JJ, Fu QJ, Chen B. Contribution of auditory working memory to speech understanding in Mandarin-speaking cochlear implant users. PLoS One 2014; 9: E99096
  • 53 Tinnemore AR, Zion DJ, Kulkarni AM, Chatterjee M. Children’s recognition of emotional prosody in spectrally degraded speech is predicted by their age and cognitive status. EarHear 2018; 39: 874-880
  • 54 Torgesen JK, Wagner RK, Rashotte CA. Test of Word Reading Efficiency. Austin, TX: Pro-Ed; 1999
  • 55 Wechsler D. WISC-IV: Wechsler Intelligence Scale for Children, Integrated: Technical and Interpretive Manual. San Antonio, TX: Harcourt Brace and Company; 2004
  • 56 Wilkinson GS, Robertson GJ. Wide Range Achievement Test. 4th ed. Lutz, FL: Psychological Assessment Resources; 2006
  • 57 Winn MB, Won JH, Moon IJ. Assessment of spectral and temporal resolution in cochlear implant users using psychoacoustic discrimination and speech cue categorization. Ear Hear 2012; 37: E377-E390
  • 58 Won JH, Drennan WR, Kang RS, Rubinstein JT. Psychoacoustic abilities associated with music perception in cochlear implant users. Ear Hear 2010; 31: 796-805
  • 59 Won JH, Drennan WR, Rubinstein JT. Spectral-ripple resolution correlates with speech reception in noise in cochlear implant users. J Assoc Res Otolaryngol 2007; 8: 384-392

Zoom Image
Fig. 1 Average SMRT scores for HiPRESTO and LoPRESTO groups. The boxes extend from the lower to the upper quartile (the interquartile range, IQ), and the midline indicates the median. The whiskers indicate the highest and lowest values not >1.5 times the IQ. Light gray circles represent individual scores. Circles containing an ‘‘x’' indicate outliers, that is, data points >1.5 times the IQ.
Zoom Image
Fig. 2 Same as Figure 1, except average Raven's scores are shown for HiPRESTO and LoPRESTO groups.