Académique Documents
Professionnel Documents
Culture Documents
http://tia.sagepub.com
Methods and Applications of the Audibility Index in Hearing Aid Selection and Fitting
Amyn M. Amlani, Jerry L. Punch and Teresa Y. C. Ching
Trends Amplif 2002; 6; 81
DOI: 10.1177/108471380200600302
Published by:
http://www.sagepublications.com
Additional services and information for Trends in Amplification can be found at:
Subscriptions: http://tia.sagepub.com/subscriptions
Reprints: http://www.sagepub.com/journalsReprints.nav
Permissions: http://www.sagepub.com/journalsPermissions.nav
During the first half of the 20th century, communications engineers at Bell Telephone
Laboratories developed the articulation model for predicting speech intelligibility transmitted
through different telecommunication devices under varying electroacoustic conditions. The
profession of audiology adopted this model and its quantitative aspects, known as the
Articulation Index and Speech Intelligibility Index, and applied these indices to the prediction
of unaided and aided speech intelligibility in hearing-impaired listeners. Over time, the calcu-
lation methods of these indices-referred to collectively in this paper as the Audibility
Index-have been continually refined and simplified for clinical use. This article provides
(1) an overview of the basic principles and the calculation methods of the Audibility Index,
the Speech Transmission Index and related indices, as well as the Speech Recognition
Sensitivity Model, (2) a review of the literature on using the Audibility Index to predict
speech intelligibility of hearing-impaired listeners, (3) a review of the literature on the applic-
ability of the Audibility Index to the selection and fitting of hearing aids, and (4) a discussion
of future scientific needs and clinical applications of the Audibility Index.
From the Department of Audiology and Speech Sciences, Michigan State University, East Lansing,
Michigan; and the -rNational Acoustic Laboratories, 126 Greville Street, Chatswood, NSW 2067,
Australia
Correspondence: Amyn M. Amlani, Department of Audiology and Speech Sciences, Michigan State
University, East Lansing, MI 48824 E-mail: amlaniam@msu.edu
C)2002 Westminster Publications, Inc., 708 Glen Cove Avenue, Glen Head, NY 11545, U.S.A.
81
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
cues in a given frequency band determine the fy the frequency-gain characteristics of hearing
amount of information in the band, and thus its aids. Aside from its use in hearing aid selection
contribution to the overall intelligibility of speech. and verification (Popelka and Mason, 1987;
As a way to quantify the relationship between Pavlovic, 1988; Rankovic, 1991; Mueller, 1992;
the audible speech cues and intelligibility, the Studebaker, 1992; Studebaker and Sherbecoe,
Articulation Index was developed as an inter- 1993), the Articulation Index has also been used
mediate step. The Articulation Index is defined to differentiate between sensory and neural au-
as "a weighted fraction representing, for a given ditory lesions (Gates and Popelka, 1992) and to
speech channel and noise condition, the effec- predict speech intelligibility of listeners who wear
tive proportion of the normal speech signal personal hearing-protection devices (Wilde and
which is available to a listener for conveying Humes, 1990).
speech intelligibility" (Kryter, 1962 [a], p. 1689). Because the term articulation suggests the act
This calculated proportion ranges from 0.0 to of speech production rather than audition, it is
1.0 and is used subsequently to predict the misleading in today's vernacular. Thus it has been
speech intelligibility. suggested that articulation be substituted by au-
The Articulation Index has had, and contin- dibility, and reference to this index be termed
ues to have, major practical implications for the Audibility Index (Studebaker, 1992; Killion et al.,
practice of audiology. Clinical use of the 1993; Killion, 2002). Despite this recommenda-
Articulation Index allows audiologists to predict tion, the American National Standards Institute
speech intelligibility under unaided listening con- (ANSI) adopted Speech Intelligibility Index, in
ditions, as well as the benefit to be derived from recognition of the critical concept of predicting
a hearing aid, or aids, by comparing predicted in- speech intelligibility (ANSI S3.5-1997).
telligibility performance under unaided and aided To minimize confusion, Articulation Index
conditions (eg, Rankovic, 1991; Studebaker and will be used in this paper to refer to the princi-
Sherbecoe, 1993). Many audiologists use count- ples and procedures described in the original
the-dot-audiograms (Mueller and Killion, 1990; ANSI S3.5-1969 standard, measures derived di-
Humes, 1991; Pavlovic, 1991), a simplification of rectly from this procedure, and a set of general
the Articulation Index, to demonstrate to their pa- methods in which articulation theory is applied.
tients the effects of hearing impairment on speech The Speech Intelligibility Index (SII), on the other
understanding without and with amplification. hand, will refer to those principles and proce-
The Articulation Index has also played an im- dures described in the most recent standard
portant historical role in procedures designed to (ANSI S3.5-1997) or to derivative measures of
prescribe the frequency-gain characteristics of this newer method. Because the Articulation
linear hearing aids (eg, Humes, 1986; Rankovic, Index and Speech Intelligibility Index are based,
1991). The basic goal of these threshold-based in essence, on determining the amount of audi-
methods is to amplify the speech spectrum so that bility available to a listener, we refer to these in-
the long-term average speech in each band is 15 dices collectively as the Audibility Index through-
to 18 dB above threshold at each frequency. out the remainder of the paper. Incidentally,
According to articulation theory, this allows the to avoid introducing a new abbreviation, we
full 30-dB range of speech to be audible (ie, de- will use the abbreviation AI to refer interchange-
tectable) at each frequency and provides maxi- ably to the Audibility Index and the Articulation
mum speech intelligibility (Humes, 1986; Index.
Rankovic, 1991; Humes and Halling, 1994). The Based on its clinical potential and the amount
Articulation Index model was adopted by the of ongoing research aimed at improving its pre-
National Acoustic Laboratories in Australia in the dictive validity, the Al promises to play a signifi-
derivation of one of the most recent methods of cant role in the practice of audiology for many
fitting nonlinear hearing aids (Dillon, 1999; years to come. It behooves audiologists, therefore,
Byrne et al., 2001). to understand the relevant concepts. Determining
Articulation Index-based methods are incor- the Al has been viewed as a complicated and con-
porated into various commercially available hear- fusing procedure. In this paper, we take on the
ing aid analyzers, real-ear probe-microphone sys- decidedly difficult task of providing a somewhat
tems, and hearing aid fitting software, and thus simplified, yet detailed, explanation of articula-
they provide a primary means to select and veri- tion theory as it relates to clinical audiologic prac-
82
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
tice. While mathematical formulas are necessari- audibility of weighted speech bands in quiet and
ly a part of any such description, our goal is to in the presence of competing noise measured at a
provide sufficient narrative explanation of the for- listener's ear. It ranges in value from 0.0 to 1.0. A
mulas to give even mathematically challenged value of 0.0 suggests that none of the speech cues
readers a general understanding of the various are audible, thereby making zero contribution to
methods and applications. speech intelligibility. A value of 1.0, on the other
This paper is divided into three major sec- hand, indicates that the proportion of available,
tions. In the first section, the fundamental princi- or audible, speech cues contributes maximally to
ples and calculation methods of the AI as speci- speech intelligibility. For values between 0.0 and
fied in the ANSI S3.5-1969 and ANSI S3.5-1997 1.0, a proportionate growth of audibility is as-
standards are described, along with clinical sim- sumed to occur, and with it, a proportionate
plifications of these procedures and other related growth in intelligibility performance. That is,
methods used to establish the audibility of speech within any frequency band, every decibel of au-
under diverse conditions. The second section pro- dibility is weighted equally.
vides a review of the literature on the use of the The basic equation for the Articulation Index
Al to predict speech intelligibility for individuals is
with various degrees of hearing impairment. The AI = EIIAi (1)
third section presents a discussion about the use
of the Al to predict speech intelligibility in clinical In the formula, Al is equal to the sum of products
hearing aid applications. resulting from multiplying I by A at each fre-
quency band (i). The frequency-importance func-
tion (Ii) represents the relative contribution of dif-
ferent frequency bands (ie, fractions of 1.0) to
Principles and Calculation Methods speech intelligibility. The variable Ai, or audibili-
ty function, which ranges from 0 to 30 dB, refers
to the amount of speech energy that is above the
A. Articulation Index listener's threshold and any competing noise in a
(ANSI S3.5-1969 Standard) given frequency band. Before illustrating the cal-
culation of the Al, we will first review how the
Descriptions of the basic model and calculation frequency-importance function, or Ii is derived.
method of the Al (French and Steinberg, 1947) Historically, the number of bands used to de-
were based on a compendium of studies that rive the frequency-importance function has var-
aimed to determine the factors affecting speech ied from procedure to procedure, ranging from 4
intelligibility. These included the ideal or optimal to 21 (ANSI S3.5-1997). Initially, French and
speech spectrum (Dunn and White, 1940); the ef- Steinberg (1947) defined the importance function
fects of masking (Fletcher and Munson, 1937); by dividing the frequency spectrum into 20 oc-
the frequency of steady-state noise and the inten- tave bands ranging in frequency from 250 to
sities of noises from several sources, including 7000 Hz, such that each band made an equal con-
ambient noise (Waring, 1946; French and tribution (0.05) to speech intelligibility. This con-
Steinberg, 1947); zero loudness-contour curves cept was adopted in the original (ANSI S3.5-
for one and two ears (ASA Z24.2, 1942); and the 1969) standard and modified to incorporate the
number of frequency bands needed to estimate addition of one-third octave bands. Of these two
speech intelligibility (French and Steinberg, procedures, the one-third octave-band method is
1947). known to be more sensitive to variations in
Over the next 15 years, investigators amend- speech and noise (Kryter, 1962[a]; ANSI S3.5-
ed and modified the Al. These addenda, however, 1969).
were not generally accepted because of insuffi- The value of Ii is some proportion between
cient evidence of their validity (Kryter, 1962 [b]). 0.0 and 1.0, and it is chosen so that Al = 1.0
As a means to validate the AI, Kryter (1962[a], when Ai = 30. The values are consistent with the
1962 [b]) published a series of papers that result- number of bands and bandwidth used in the de-
ed in the ANSI S3.5-1969 standard. rivation, and the relative importance of each fre-
According to the original standard, the Al is a quency band varies with the speech material. As
proportional index based simply on the summed shown in Figure 1, the one-third octave bands
83
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
84
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
Threshold
(or when the noise level exceeds the speech spec-
trum level). When the entire 30-dB range of v *l Frequency
speech is above the listener's hearing threshold, Audibility (A,): 30 30 20 0
the band makes maximal contribution to speech x x x x
intelligibility (ie, 1.0). The AI, therefore, is the Importance (1,): 0.001 0.002 0.003 0.002
sum of the weighted audibility across all fre-
quency bands.
The use of the idealized speech peaks in the lAi x I,: + 0.03 + 0.06 + 0.06 + 0.0 =
Al calculation suggests that accurate results can
best be obtained if it is assumed that the short-
term speech distribution is uniform over time. Figure 3. An illustration of audibility and band
Specifically, the speech spectrum of the CVC syl- importance as a function of threshold in the calculation
lables of the Bell Telephone Laboratories was of the Al.
used as the basis for the original Al (ANSI S3.5-
1969) standard.
Dunn and White (1940) used the average
sound pressure level within a band 1 Hz wide, dynamic range" (Boothroyd, 1986, as referenced
over the integration time of the ear (ie, 125 mil- in Pavlovic, 1987 and 1989).
liseconds). The accuracy of predictions of speech The idealized speech spectrum is assumed to
intelligibility over an integration time of 125 mil- be based on speech stimuli presented in quiet,
liseconds has since been verified in other studies with normal vocal effort, and at a distance of 1
(French and Steinberg, 1947; Kryter, 1962[b]; meter from the talker. The overall presentation
Pavlovic and Studebaker, 1984). Dunn and White level of speech produced by normal vocal effort
found that within this integration time, the dis- is further assumed to be 65 dB SPL in the orig-
tribution of the speech root-mean-square (RMS) inal AI standard and 63 dB SPL for the SII,1 as
values is approximately linear over a 30-dB range measured in a sound field. Under everyday lis-
in any given frequency band. This 30-dB range tening conditions, however, various factors
extends from +12 dB to -18 dB between 1000 may change the amplitude characteristics of
and 1400 Hz relative to the long-term average speech. For instance, if a high level of back-
speech spectrum of the CVC syllables in the orig- ground noise is present, a talker is more likely
inal standard. to raise the voice level due to the Lombard ef-
This range was modified to 15 dB in the re- fect. That is, for every decibel increase in noise
vision of the ANSI S3.5-1969 standard, which was above 50 dBA, there is an increase in speech
drafted as the ANSI S3.5-1993 interim standard level of 0.46 dB (ANSI S3.14-1977). Pearson et
(ANSI, 1993). The dynamic range of 15 dB in al. (1976) found that increased vocal effort is
the ANSI S3.5-1997 standard resulted from associated not only with higher overall level,
speech-recognition studies reflecting differing but also with variations in the frequency-ampli-
speech stimuli in speech-weighted noise at vari- tude spectrum. A second factor to consider clin-
ous signal-to-noise ratios (SNRs) (eg, Studebaker ically is that a given hearing-impaired listener
et al., 1993) and the methods used in the calcu- may participate in conversational speech at lis-
lation of the Speech Transmission Index, which is tener-to-talker distances that vary considerably
described later in this paper. in different conversational situations.
Thus, for CVC syllables, the overall RMS
speech level minimally required for recognition
of weak consonants is 30 dB higher than that re- 'The AI accounts for variations in vocal effort using a correc-
quired minimally for recognition of strong vow- tion factor (Figure 13, ANSI S3.5-1969). Conversely, the SII
assumes values of 69, 75, and 82 dB SPL for raised, loud,
els. This range has also been termed "perceptual and shouted vocal efforts, respectively.
85
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
The following conclusions can be drawn re- whether related to hearing threshold, compet-
garding the ANSI S3.5-1969 standard: ing noise, or listener-to-talker differences, will
result in concomitant changes in signal audi-
1. The AI is based on two components: the impor- bility for the listener.
tance of each frequency band (Ii) and the 4. The dynamic range of speech that is regarded
amount of signal audibility within a given band to be effective in maximizing speech intelligi-
(A,). bility is 30 dB. This range was originally found
2. The frequency-importance (Ii) function is a to be between + 12 and -18 dB, and was re-
mathematical weighting that represents the cently modified to 15 dB.
importance of different frequencies for speech
intelligibility. The importance of these weight- B. Calculation of the Al
ings differs with respect to the speech material,
and the sum of these weightings must equal To calculate the Al, the audiometric thresholds of
1.0. a listener are converted from dB HL to an equiv-
3. The audibility (A,) function quantifies how alent level of a fictitious internal noise in the lis-
much of the signal within a given frequency tener's ear. This is done by first converting dB HLs
band is being delivered to the listener. When to the reference equivalent threshold sound pres-
the idealized speech spectrum is used in the sure levels. An illustrative example of the calcu-
calculation, this function assumes normal lation of the Al is shown in Table 1, which uses
vocal effort and a conversational level of 65 transformations from Pavlovic (1987) and Bentler
dB SPL in the original standard and 63 dB SPL and Pavlovic (1989) to convert HLs of a hypo-
in the newer standard, at a distance of 1 meter thetical hearing-impaired listener tested with
from the talker. Changes in these levels, TDH-49 earphones to equivalent SPLs in free
'Bentler and Pavlovic (1989); 2Pavlovic (1987); 3ANSI S3.5-1997, Table 4; 4ANSI S3.5-1969, Table 8;
5ANSI S3.5-1969, Table 7. These values are based on nonsense syllables developed at Bell Telephone Laboratories.
*When this value is greater than 30, the assigned value is terminated at 30.
**When this value is less than 0, the assigned value is terminated at 0.
86
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
87
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
ods, suggested that hearing levels ranging be- spectively. The number of audible decibels is 20
tween 20 and 50 dB HL and for the frequencies of at 1000 Hz and 10 at 2000 Hz. Note that at 4000
500, 1000, 2000, and 4000 Hz form the bound- Hz, the threshold is 50 dB, resulting in 0 dB of
aries for all speech cues important to speech audibility. To determine the amount of available
recognition. In his original method, termed AO speech information, one simply sums the number
and later renamed AO(4), clinicians can easily cal- of decibels audible to the listener, which in this
culate an Al for the audible portion of the dy- case is 60 dB (30+20+10+0), and divides the
namic range of speech (ie, 30 dB) across the four total by 120, or the maximum number of audible
frequencies using a standard audiogram. As seen decibels (30 dB x 4 frequencies). This results in
in Figure 5, our hypothetical patient exhibits es- an unaided Al of 0.50 for this patient's right ear,
sentially a mild-to-moderate sloping hearing loss suggesting that exactly one half of the speech
in the right ear. At 500 Hz, a threshold of 20 dB spectrum is audible.
HL is observed. Therefore, all 30 dB of speech en- Suppose this individual is fit with a hearing
ergy around 500 Hz is considered audible. For the aid and sound-field thresholds are measured ei-
frequencies of 1000 and 2000 Hz, unaided ther by traditional audiometric methods or de-
thresholds are measured at 30 and 40 dB HL, re- rived by using real-ear aided gain (REAG) values
88
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
89
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
the listener. As with any other method designed plification device and real-ear insertion gain
to take inter-octave frequencies into account, (REIG) is measured, thresholds are 5, 10, 15, 20,
thresholds at these frequencies must be measured and 30 dB at the frequencies between 250 and 4000
or interpolated from audiometric data. Hz, and the number of dots below the aided curve
For aided conditions, the impact of output increases to 30, resulting in an aided Al of 0.90.
limiting must also be considered. Measurement In 1991, Pavlovic concluded that the Mueller
of uncomfortable loudness levels (UCLs) plays an and Killion (1990) and Humes (1991) methods
important role in calculating the AI. In other might not sufficiently predict performance in
words, the Al is calculated based on the number everyday listening situations. In essence, his view-
of dots (in the case of the count-the-dot methods), point was that an importance function for average
or audible decibels at each frequency between speech is more appropriate for AI calculations de-
aided threshold and aided UCL. When speech rived with hearing aids. Termed the Ad method
cues are deemed to be above the aided UCL, they (the d being a mnemonic for "dot"), this count-
are assumed to be limited by the hearing aid and the-dot procedure is similar to Mueller and
therefore inaudible. Using Figure 5 as an exam- Killion's (1990) method in every respect (includ-
ple, assume that an unaided UCL of 90 dB HL is ing calculation), except that the speech spectrum
measured for the audiometric frequency of 1000 used is that of conversational speech in quiet,
Hz. Based on the 15 dB of gain provided by the measured at 63 dB SPL and at a distance of 1
hearing aid, the aided UCL is assumed to be 75 meter from the speaker. Thus the Ad method is
dB HL, or the unaided UCL minus the amount of applicable directly to continuous discourse.
gain (90 - 15). Here, the aided UCL would not Lundeen (1996) noted the need to make two
impact the Al calculation. If the unaided UCL corrections in the published version of Pavlovic's
were 60 dB HL, as seen potentially with a hyper- Ad method (1991). Specifically, the physical
acusic patient, the derived Al would indeed be af- width of an octave on the frequency scale of the
fected. The examples presented here are relevant audiogram (X-axis) did not equal the length of a
to linear signal processing. In cases where non- 20-dB interval on the ordinate, or Y-axis, as spec-
linear signal processing is used, the gain values ified by ANSI S3.21-1978 (R1997). Also, the dots
determined at the high input levels should be ap-
plied when calculating aided UCL values.
As illustrated in Figure 7, Humes (1991) pro- Frequency (Hz)
posed a revision to the Mueller and Killion (1990) 125 250 500 1000200040008000
method. Specifically, Humes' count-the-dot pro-
cedure differs from Mueller and Killion's (1990) 0
in that (1) it uses an octave-band resolution, as 10
opposed to a one-third octave-band resolution;
(2) the dynamic range of speech varies between -0
20
30 and 40 dB across frequencies, based on the 30 ....IIt
90
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
91
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
Frequency (Hz)
. ..I----- -
m 90 . 0 0 1 A udibility
.
80 Unaided =0.52
-
> 70 CD ..........
-J 60 . ..
-
~0n
z
4
: 50
cz 40
m
I 30
.5 1 0 ." . .. (> ._ ..~~~~~~~~~~~~~..... ....
'a
-
m
O~ ~ ~ ~
~~~~~~~~~~._
~
_.
~~~~~~~~~~~~~~~~~~~~
_._. . . .._,,.._._....
(A -1 0 ..:........
92
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
advantage is that these methods enable clinicians Intelligibility Index, presumably to focus on the
to describe the extent to which speech cues are SII's objective of speech intelligibility prediction.
available, independently of the need to perform As in the previous standard, audibility is ex-
speech-intelligibility testing. In addition, count- pressed by the equation
the-dot audiograms provide a means whereby dif-
ferent hearing aids and types of circuits can be SIIH==II (3)
compared as a function of the speech input level.
While these methods are a simplified, effec- where Ii and Ai are the frequency-important and
tive, and efficient means of estimating the pro- audibility functions, respectively.
portion of speech cues available to a given lis- Several modifications and additions to the
tener, clinicians must be aware of several factors original ANSI S3.5-1969 standard characterize
that could yield inaccurate results. Some but not the 1997 standard, and these are detailed in this
all of these inaccuracies stem from shortcomings subsection. In the ANSI S3.5-1997 standard, for
inherent in the AI itself. Clinicians, for example, instance, the calculation of the audibility function
may mistake audibility for intelligibility. As de- has been modified to consider the spread of
scribed earlier in this paper, audibility and intelli- masking and the standard speech spectrum level.
gibility are related, but they are not synonymous. That is, because the world is a noisy place, target
For instance, 0.6 audibility does not equate to 60% sounds listeners want to hear are mixed with en-
intelligibility, but rather indicates only that 60% of vironmental, background noises. When speech is
the speech cues are available to the listener. The mixed with noise, some parts of the speech spec-
prediction of speech intelligibility can be deter- trum become inaudible due to masking-a phe-
mined based on the transfer functions for various nomenon with which audiologists are very famil-
speech stimuli, as seen in Figure 4, for example. iar. Masking is also problematic with respect to
While speech-intelligibility performance is assumed speech intelligibility when speech masks itself,
to increase monotonically with increasing sensa- or when higher energy vowels make lower en-
tion level or audibility, this is generally true only ergy consonants inaudible. Empirical evidence
for those individuals with normal hearing or with suggests that a reduction in consonant informa-
mild-to-moderate hearing losses. As discussed else- tion can yield decreased speech-intelligibility
where in this paper, for individuals with more se- performance because about 90% of the acoustic
vere hearing losses, an increase in sensation level information important for understanding is pro-
(ie, audibility) may not result in improved intelli- vided by consonants (French and Steinberg,
gibility, but may actually degrade intelligibility. 1947; Licklider and Miller, 1951). The effect of
Pavlovic (1991) further warns that these sim- masking depends on the various parameters of
plifications might mislead clinicians and patients noise: (a) its long-term spectrum, (b) its inten-
in several other ways. First, a value of 1.0 does sity fluctuation over time, and (c) its average in-
not necessarily mean that the aided auditory sys- tensity relative to the intensity of speech.
tem is functioning normally. Rather, it indicates To account for masking, the 1997 standard in-
that the hearing aid is matched optimally to the cludes formulas for self-speech masking and up-
impaired auditory system for maximizing speech ward spread of masking. These variables are ref-
intelligibility. Second, improved thresholds, as erenced and detailed in Appendix A. The revised
measured by REIG, may suggest that the listener's calculation method also provides a framework for
threshold has changed when, in actuality, it is the determining the speech, noise, and threshold spec-
speech spectrum that has been shifted to a more trum levels based on measurements of modulation
audible range. Lastly, and perhaps most impor- transfer functions in reverberation.
tantly, these simplified Al methods cannot predict In addition, the revised standard takes into
a patient's ability to understand speech in noise consideration the empirical findings of French
(Killion and Christensen, 1998; Killion, 2002). and Steinberg (1947) and Fletcher (1953) that
suggested a decrease in speech performance for
E. Speech Intelligibility Index normal-hearing listeners at high sound pressure
(ANSI S3.5-1997 Standard) levels (ie, rollover). As a means to counter the
phenomenon of rollover, the audibility function
In the ANSI S3.5-1997 standard, the term has been modified to include a level distortion
Articulation Index was changed to Speech factor (LDF). This factor has a maximum of 1
93
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
when there is believed to be no distortion, typi- (Studebaker and Sherbecoe, 1991), NU-6 mono-
cally at low presentation levels. This factor, how- syllables (Studebaker et al., 1993), Diagnostic
ever, is reduced when the overall sound pressure Rhyme Test (Duggirala et al., 1988), short pas-
level exceeds 73 dB SPL, and reaches a minimum sage of easy reading material (Studebaker et al.,
value of 0 only at extremely high presentation 1987), the monosyllables of the Speech in Noise
levels (ie, 233 dB SPL). The modified audibility Test (Bell et al., 1992) and average speech
function Ai is expressed as (Pavlovic, 1987). These importance functions are
shown in Table 3.
Ai = KiLi (4) In 1996, DePaolis et al. attempted to deter-
mine the effects of experimental methodology
where Ki is the proportion of speech that is above on the frequency-importance functions for dif-
the listener's hearing threshold or masking noise, fering stimuli. Specifically, they were interested
whichever is higher; and Li is the level distortion in quantifying the difference between the fre-
factor. The formula for calculating Li is given in quency-importance functions for words, mean-
Appendix B. ingful sentences, and continuous discourse test-
In Annex A of the ANSI S3.5-1997 standard, ed under a similar methodology. Recall that fre-
modifications allow for the fundamental AI to be quency-importance functions differ, depending
extended to individuals exhibiting conductive on the stimuli, equipment, and the procedures
hearing losses. Accounting for this type of hear- used to determine them. Twenty-four normal-
ing loss is accomplished by modifying Li with a hearing adults served as subjects and made in-
loss factor determined by the conductive hearing telligibility estimates of PB-50 monosyllabic
loss component of the hearing threshold level. words (ANSI S3.2-1989), sentences from the re-
The relative importance of various frequen- vised Speech Perception in Noise (SPIN) test
cies to speech intelligibility for different speech (Bilger, 1985), and continuous discourse spoken
material is provided in Annex B of the ANSI S3.5- by a male talker. During data collection, the
1997 standard. The types of speech test material methodology of low-pass and high-pass filtering
include various nonsense syllables tests (Fletcher at various SNRs closely approximated the proce-
and Steinberg, 1929; Pavlovic and Studebaker, dures reported by French and Steinberg (1947).
1984; Humes et al., 1986), CID-W22 or PB-words Results agreed well with other published works,
Table 3. Band-importance Functions Using Common Audiological Octave and One-half Octave Frequencies for Various Speech
Tests: NNS (Various Nonsense-syllable Tests Where Most of the English Phonemes Occur Equally Often), CID-W22 Words, NU-6
Monosyllabic Words, DRT (Diagnostic Rhyme Test), Short Passages of Easy Reading Material, SPIN Monosyllables, and Average
Speech. Adopted from Pavlovic (1994) and Reprinted with Permission of Lippincott Williams & Wilkins.
Band Center Average
Frequency (Hz) NNS CID-W22 NU-6 DRT Short Passage* SPIN Speech
250 0.0437 0.1549 0.0853 0.0960 0.1004 0.0871 0.0617
500 0.1043 0.1307 0.1586 0.1659 0.2197 0.1190 0.1344
750 0.0841 0.0836 0.0973 0.1142 0.1013 0.1079 0.1035
1000 0.1056 0.1157 0.1068 0.1187 0.0966 0.1078 0.1235
1500 0.1297 0.1349 0.1407 0.1377 0.1092 0.1447 0.1321
2000 0.1664 0.1401 0.1631 0.1288 0.1177 0.1436 0.1328
3000 0.1542 0.1134 0.1244 0.1077 0.1213 0.1433 0.1285
4000 0.1227 0.0648 0.0724 0.0640 0.0676 0.0942 0.1039
6000 0.0893 0.0619 0.0514 0.0670 0.0662 0.0542 0.0796
*This importance function is also known as the importance function for "continuous discourse" (Studebaker et al., 1987)
or as the importance function for "easy speech" (Pavlovic, 1987).
94
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
95
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
96
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
97
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
sented in a broad-band competing noise having a To determine the validity of these viewpoints,
spectral shape that varied from flat to a slope of attempts to use the STI to predict the perfor-
12 dB/octave, and (c) several forms of temporal- mance of amplitude-compressed speech have
ly distorted broad-band speech presented against been undertaken (Steeneken and Houtgast, 1980;
a broad-band noise. Conversely, the AI weights Plomp, 1988; Festen et al., 1990; Hohmann and
were derived with extreme and abrupt spectral Kollmeier, 1995). In general, findings have
distortion (ie, filtered speech). An example calcu- demonstrated that the STI fails to describe the ef-
lation of the STI can be found in Appendix C. fect of amplitude compression on the intelligibil-
Amplitude compression, such as that found in ity of speech. This shortcoming has been attrib-
many nonlinear hearing aids, is also a character- uted to the stimuli used and the choice of input
istic whose effects could possibly be assessed by level, compression threshold, compression ratio,
the STI (Humes, 1993). Specifically, amplitude and attack/release times (Humes, 1993;
compression decreases the dynamic range of nat- Hohmann and Kollmeier, 1995).
ural acoustic signals into the reduced range be- In 1986, Humes et al. retrospectively ana-
tween a given hearing-impaired listener's thresh- lyzed the results of studies conducted by Moncur
old and loudness discomfort level. In a given and Dirks (1967), Finitzo-Heiber and Tillman
hearing aid, this compression can be achieved by (1978), and Duquesnoy and Plomp (1980) to de-
techniques that filter in the frequency and/or termine the effectiveness of the STI and AI in pre-
time domains or through digital signal processing dicting the speech-intelligibility performance of
(DSP) algorithms that attempt to enhance speech normal-hearing and hearing-impaired listeners
relative to the competing noise. under conditions of temporally and spectrally dis-
Experimental findings on amplitude com- torted speech. Findings revealed that the STI pro-
pression have shown no consistent positive effect, vided a better description of speech-intelligibility
and in some cases have shown a negative effect data under conditions of temporal distortion (re-
on speech intelligibility, particularly in the pres- verberation) and conditions not involving abrupt
ence of reverberation and competing noise (for a and/or considerable changes in the speech signal.
review, see Braida et al., 1979 and Plomp, 1988). Conversely, the Al was found to provide a superi-
The STI and MTF are conceptualized as preserv- or description of intelligibility for spectrally dis-
ing envelope fluctuations of the target signal that torted speech. When the investigators undertook
contribute to speech intelligibility. a prospective study that examined the intelligi-
Plomp (1988), for instance, hypothesized that bility of the Nonsense Syllable Test (Resnick et
the effect of amplitude compression on speech in- al., 1975) under the conditions of multitalker
telligibility is more detrimental as the number of babble, filtering, and reverberation, their conclu-
channels are increased and the compression time sions confirmed the findings of the retrospective
constants are decreased (ie, shorter attack/release analysis.
times). He further argued that the flattening of the Based on these results, Humes et al. (1986)
envelope fluctuations caused by amplitude com- conjectured that a hybrid model characterizing
pression increases the negative effect of poor fre- the best attributes of both acoustical indices
quency resolution on intelligibility. Plomp (1988) might provide a better description of the data ob-
suggested that just as the STI/MTF reflects the loss tained in the prospective study. This led to the
of lower level information due to reverberation and creation of the modified Speech Transmission
noise, so too it could reflect the loss of spectral and Index, or mSTI. Specifically, the mSTI combined
temporal information caused by compression. the STI's MTF approach of determining SNR with
Consequently, he speculated that the reduced MTF the AI's approach of using 15 one-third octave
caused by compressing the troughs and peaks of the bands centered at frequencies between 250 and
signal modulation could be used to predict the loss 6300 Hz and the weighting factors originally de-
of speech intelligibility due to compression. scribed by French and Steinberg (1947).
In a retort to Plomp's statements, Villchur Humes et al. (1986) rationalized use of the
(1989) indicated that the STI concept does not one-third octave bands by suggesting that octave
hold for nonlinear processing. That is, "the weak bands are not as sensitive to the abrupt changes
speech elements that are lost when the MTF is re- in signal spectrum and that the STI's weighting
duced by noise are preserved when the MTF is re- factors showed no appreciable changes across
duced by compression" (Villchur, 1989, p. 425). frequency. In addition to these changes, the mSTI
98
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
decreased to six the number of modulation fre- 649 milliseconds). For normal-hearing listeners,
quencies for each one-third octave band. This they found that the speech-like signal degraded
change also resulted in a wider range of modula- considerably as distance increased from the
tion frequencies, which include 0.5, 1.0, 2.0, 4.0, source. Specifically, a value of 0.83 was deter-
8.0, 16.0 Hz. mined for the front-row center seat (2.65 meters),
The application of the mSTI was then com- 0.66 for the middle-most center seat (6.76 me-
pared to the results obtained for the STI using ters), and 0.55 for a student seated 10.88 meters
data from their normal-hearing group and data from the source, which equated to the center seat
acquired from a group of hearing-impaired lis- of the last row. In fact, a RASTI value of 1.0 was
teners under quiet and babble conditions. Results achieved only at the 6-inch reference position.
revealed that the mSTI offered the best descrip- This finding strongly suggests the need to consid-
tion of speech-intelligibility performance of tem- er distance when fitting amplification systems.
porally and spectrally distorted speech in both Pekkarinen and Viljanen (1991) found that
groups of listeners. RASTI values decreased as background noise and
The Rapid Speech Transmission Index reverberation increased in occupied classrooms.
(RASTI) is a manufacturer-based derivation of the Despite increases in vocal effort by the instructor,
STI. The RASTI is typically used to predict speech speech intelligibility was reported as only fair.
intelligibility in various rooms and other enclosed The authors suggest use of the RASTI as a mech-
environments such as auditoriums, theaters, and anism in the design engineering of classrooms.
concert halls. Manufactured by Briiel and Kjaer
(1985) as a portable instrument that can make G. Speech Recognition Sensitivity Model
rapid measurements, the RASTI is a simplification
of the more complex STI. The simplification is Recently, Musch and Buus (2001 [a], 2001 [b])
that the RASTI is measured at only two octave developed a new intelligibility theory based on a
bands centered at 500 and 2000 Hz. macroscopic model incorporating statistical-deci-
Because of this simplification, it has been in- sion theory. This theory, called the Speech
corporated as a standard in a number of Recognition Sensitivity (SRS) model, predicts
European architectural codes. Unfortunately, this speech-intelligibility performance based on the
simplification has forced reevaluation of those long-term average speech spectrum, the masking
standards. For example, because the RASTI tests excitation in the listener's ear, the linguistic en-
in only two frequency bands, it assumes that the tropy5 of speech stimuli, and the number of re-
output response actually extends in a reasonably sponse alternatives available to the listener. The
flat fashion for frequencies lower than 100 Hz underlying basis for the creation of this model is
and greater than 8000 Hz. While this might be that redundant and synergetic interactions among
the case in a well-designed auditorium, the per- the spectral components of speech account for its
formance of many types of paging systems falls intelligibility.
short. In these cases, the RASTI almost invariably The Al was developed to predict intelligibility
underestimates the effects of room acoustics on performance in telecommunication devices,
speech intelligibility. Furthermore, compression where each frequency band contributes to the
or limiting in the system can cause artificially low overall signal based on its weighting factor. Each
STI and RASTI values by reducing the modula- frequency band, therefore, is assumed to con-
tions. This is problematic when, in fact, such pro- tribute independently to phoneme recognition.
cessing might enhance intelligibility. It should The SRS, on the other hand, predicts speech in-
also be noted that the RASTI does not take system telligibility by combining spectral information
distortion into account. from all frequency bands. According to the au-
To date, the RASTI has received little clinical thors, the SRS model allows for synergetic and
attention because experimentation on hearing-im- redundant information to be simulated, even in
paired listeners is lacking. Two studies, though, the absence of band interactions that may result
have used this index as a measure of intelligibili- from the upward spread of masking.
ty in classrooms (Leavitt and Flexer, 1991;
Pekkarinen and Viljanen, 1991). 5Linguistic entropy is defined as the information content (eg,
word frequency, lexical density, ambiguity, sentence
Leavitt and Flexer (1991) measured RASTI "depth", recognition points) of linguistic stimuli (van Rooij
values in a moderately reverberant room (Rt - and Plomp, 1991).
99
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
To accomplish the predictive task, the authors Most filter conditions included two or more
use the concept of statistical-decision theory sharply filtered narrow bands of speech. Using
(Green and Birdsall, 1958), which implies the ex- statistical-decision theory, findings revealed that
istence of an ideal signal. An ideal signal is best listeners extract speech cues in distinct spectral
described by means of an analogy. Assuming that bands based on redundant, independent, or in-
a talker and listener share the same linguistic teractive synergy. This finding was based on the
background, the talker will attempt to provide the relationship between the observed error rates
listener with an ideal signal. Because the accura- and the error rates obtained for the individual
cy of speech is variable, actual speech usually narrow bands. It was concluded that listeners are
only approximates this ideal signal. Stated differ- more likely to combine speech cues synergisti-
ently, articulated speech typically produces vari- cally when a spectral separation occurs between
able degrees of production-related noise, as op- bands and listeners are afforded a closed set of
posed to ideal speech. This production-related alternatives.
noise, along with any other competing noises, is
delivered to the listener's auditory system and de-
termines the intelligibility of the signal. In
essence, intelligibility depends on the relation be- Prediction of Speech Intelligibility for
tween the intensity of the ideal signal and the in- Hearing-Impaired Listeners
tensity of the noises delivered to the ear.
In theory, the SRS model must account for
three factors. First, the model assumes that the A. Hearing Loss and Desensitization
auditory periphery's ability to decode the incom-
ing signal is correlated with an internal template Research has shown the AI to be a valid predictor
for every expected word. Second, a linguistic en- of speech intelligibility for most listeners with
tropy-or knowledge of the phonological, se- hearing loss in the mild-to-moderate range.
mantic, contextual, and pragmatic variables-is Unfortunately, this is not the case for listeners
necessary for speech recognition to occur. Third, with more severe hearing impairment. A study by
speech-intelligibility performance is dependent on Kamm et al. (1985) exemplifies these differences.
the number of response alternatives. That is, the Kamm et al. obtained performance-intensity
smaller the number of alternatives-such as those functions on 5 normal-hearing listeners and 10
available in a closed-set listening task-the listeners with mild-to-moderate sensorineural
greater the likelihood that speech intelligibility hearing loss. Each of the listeners had word-
will increase. recognition scores greater than 88%, as measured
Research has shown that a listener's ability to clinically in quiet using the NU-6 monosyllabic-
discriminate among words is related to the type word lists (Tillman and Carhart, 1966). An addi-
of response format. This relationship is particu- tional listener with moderate hearing loss also
larly pronounced in closed-set formats consisting participated.
of phonemic variations in voicing, place, and During the experiment, participants were also
manner for each alternative foil (eg, bad, had, administered the closed-set Nonsense Syllable
sad, mad, pad) and in the number of foils within Test (NST) (Resnick et al., 1975) processed
each test item. Listener performance may be in- through a hearing aid capable of simulating both
fluenced substantially by the response option and a flat-frequency and high-frequency response at
number of alternative foils (Studebaker and various presentation levels. AIs were computed
Sherbecoe, 1993). Thus, predicting scores on ma- for each subject at each presentation level and
terials presented in different formats could result under both frequency responses using the one-
in poorly correlated Al values. third-octave-band method for the NST stimulus
To determine the validity of the SRS, Musch items.
and Buus (2001 [b]) assessed the consonant-dis- Results revealed comparable performance be-
crimination ability of 5 normal-hearing listen- tween the observed behavioral measurements
ers. The listener's task was to discriminate among and the predicted intelligibility scores for the 5
18 consonants in a CV context, filtered under 58 normal-hearing listeners and the 10 hearing-im-
conditions using either an open-set response for- paired listeners, each of whom exhibited high
mat or closed-set response format with 9 foils. word-recognition scores. This finding is in good
100
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
agreement with results obtained by other re- Braida (1980), reflecting on the findings of
searchers (eg, French and Steinberg, 1947; Dugal et al. (1980), pointed out that although a
Pavlovic, 1984; Pavlovic and Studebaker, 1984; constant proficiency factor could accurately ana-
Humes et al., 1986; Pavlovic et al., 1986; Turner lyze data, such a method does not account for the
and Robb, 1987; Zurek and Delhorne, 1987; individual degradations in spectral resolution. As
Dubno et al., 1989 [b]; Ching et al., 1997). a result, Pavlovic (1984) investigated the magni-
For the lone moderately hearing-impaired lis- tude of P in accounting for the effects of desensi-
tener with reduced word-recognition scores, the tization by representing the proficiency factor as
AI was a poor predictor of performance. Similar a function of frequency rather than as a constant
findings have been observed in listeners with value, as originally proposed by Fletcher and Galt
moderate, severe, and profound hearing losses (1950). This was accomplished by determining
(eg, Dugal et al., 1980; Pavlovic, 1984; Dubno et the speech-recognition scores achieved for high-
al., 1989[a]; Ching et al., 1998) and listeners pass and low-pass filtered Harvard Psycho-
with steeply sloping high-frequency losses acoustic Laboratory (PAL) PB words (Egan, 1948)
(Skinner, 1980; Rankovic, 1991). In general, as by 14 normal-hearing and 16 mild-to-moderately
hearing impairment increases, the Al overesti- hearing-impaired presbycusic listeners. P was de-
mates actual speech intelligibility, and the ability fined as the sum of practice effects and desensiti-
to predict Al is overestimated in such cases. zation. In the calculation of P, the practice effect
To improve the accuracy of predicting speech proficiency factor was held constant for both
intelligibility for these hearing-impaired listeners, groups, while the desensitization proficiency fac-
various attempts have been made to modify the tor was averaged for subjects' scores on the high-
AI calculation procedure. Fletcher and Galt pass and low-pass filtered speech task. Findings
(1950), for instance, proposed a proficiency fac- revealed that the low-frequency desensitization
tor to account for variations in the enunciation of proficiency factor was not significantly different
the talker and in the familiarity of the listener between listener groups. The high-frequency de-
with the talker, using the equation sensitization proficiency factor, however, was sig-
nificantly different for the two groups. This find-
Al = P I,Ai (7) ing indicated that desensitization varied as a
function of frequency.
where P refers to the proficiency factor. P reach- To improve prediction based on audibility,
es its maximum value of 1.0 when communica- Pavlovic et al. (1986) suggested a desensitization
tion involves normal-hearing listeners and when factor to reduce the deficits in frequency discrim-
the talker and listener speak the same dialect. ination and temporal coding ability in persons ex-
Because laboratory-recorded materials do not in- hibiting cochlear impairment. Basically, this ap-
clude large variations in talker skills (Studebaker proach, which was derived from earlier work
et al., 1995), P is often used as a generalized vari- (Pavlovic, 1984), multiplies the frequency-impor-
able to account for differences between individual tance function, or Ii, by a correction factor Di:
listeners that are not explained by audibility
(Magnusson, 1996; Magnusson et al., 2001). AI = EIiAiDi (8)
Fletcher (1952) proposed in a series of exper-
iments that P should be explored as a means to In the equation, Di ranges from 0.0 to 1.0 and is
describe the variability of suprathreshold derived by means of multiplying the hearing
speech-recognition performance of hearing-im- threshold by a factor ranging from 1.0 to 0.0 for
paired listeners, as predicted by the Al. Dugal et hearing thresholds between 15 dB HL and 95 dB
al. (1980) assessed the efficacy of Fletcher's HL, respectively. A Di of 1.0 indicates no desensi-
(1952) proposed proficiency factor by compar- tization, while a D, of 0.0 indicates maximal de-
ing the observed speech scores of 6 hearing- sensitization. The effect of desensitization as a
impaired listeners with those predicted on the function of hearing loss is illustrated in Figure 15.
basis of the Al with and without the proficiency Using this modified procedure, Pavlovic et al.
factor. They concluded that the Al that incorpo- (1986) found substantial improvements, relative
rated the proficiency factor improved prediction to the unmodified Al (1969) method, in the ac-
accuracy for performance at intermediate pre- curacy of absolute predicted values in 4 listeners
sentation levels. having various audiometric configurations.
101
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
102
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the AI
scores of each subject. Ultimately, they recom- hearing-impaired listeners under 70 years of age,
mended use of formulas that improved the rela- and a group of 12 elderly adult hearing-impaired
tionship between observed and predicted scores listeners aged 70 years and older. Listeners were
by employing the standard level distortion factor presented NU-6 words previously digitized by
(see Appendix B) and individual frequency-de- Studebaker et al. (1993) and bandlimited be-
pendent proficiency factors. That formula is ex- tween 447 and 2239 Hz. The stimuli were pre-
pressed as sented at 8 long-term RMS levels and 10 SNRs
relative to a talker-spectrum-matched noise that
SII = E Ii Ai's LiPn,s (11) was band-pass filtered between 282 Hz and 2818
Hz.
where Li is the level distortion factor, or LDF, and Raw data obtained from each subject was
Pn,s is the proficiency factor for the nth band for transformed into rau units (Studebaker, 1985)
subject s. and plotted graphically as a function of speech
To evaluate the effectiveness of the unmodi- level. Using a method of least-squares regression
fied ANSI S3.5-1969 procedure, the ANSI S3.5- at each SNR, findings across groups showed that
1997 procedure, and the modified ANSI S3.5- speech intelligibility in noise decreased as the
1997 procedures that incorporate an individual speech level exceeded 69 dB SPL. As a means to
frequency-dependent proficiency factor (Equation predict these outcomes, Al values were derived
11) for predicting speech-intelligibility perfor- using an adjusted method. Specifically, SNRs
mance, the same subjects were tested using BKB were calculated for each one-third octave band
sentence lists and VCV syllables that were low- between 125 and 10000 Hz by taking the differ-
pass filtered at 5600 Hz. The nonsense syllables ence between the speech level and either the ap-
were comprised of 3 vowels and 24 consonants plied noise level or the subjects' average internal
with 6 repetitions. In the calculation of the SII, noise level (ie, threshold).
the importance functions of sentences and non- Fifteen decibels were added to this SNR, re-
sense syllables were used to weight the audibility sulting in a speech peaks-to-noise ratio
function (Pavlovic, 1994). (Steeneken and Houtgast, 1980; ANSI S3.5-
For 17 of the 19 listeners with mild and mod- 1997). Each speech peaks-to-noise ratio was then
erate hearing losses, observed performance was divided by 40, or the effective dynamic range of
found to be better than predicted at the low SLs speech, and multiplied by the frequency-impor-
of 6, 12, and 18 dB for each of the predicted SII tance function for the NU-6 test (Studebaker et
scores. At the higher SLs, observed performance al., 1993). The effective dynamic range of speech
scores were slightly better than the predicted value of 40 dB was based on data provided by the
scores for these listeners. For subjects exhibiting normal-hearing group for speech presented in
severe or profound losses, all three SII proce- quiet at a level of 69 dB SPL. It was derived by
dures, on average, were found to predict poorly varying repeatedly the assumed range until the
the observed performance scores at every sensa- mean and standard deviation between the pre-
tion level. dicted and observed scores showed essentially no
Despite a slight improvement achieved by in- variability.
cluding the effect of a frequency-dependent pro- Two optimal estimates of dynamic range
ficiency factor, results also showed considerable were determined (39 and 41 dB) and then aver-
unexplained variance. Overall, these results sug- aged. For the normal-hearing group, the Al did
gested that audibility, as quantified by the SII, not predict accurately observed scores in noise.
might not be adequate in predicting the speech For the hearing-impaired group, whose data were
intelligibility of all hearing-impaired listeners, es- pooled, results revealed an improvement in the
pecially those with severe hearing losses. predictability of their observed scores. As a result,
Amplification that aims to maximize audibility it was concluded that the effective dynamic range
does not optimize speech intelligibility for these of speech might be larger than the assumed value
listeners. of 30 dB.
Studebaker and colleagues (1999) compared As a means to understand better the relative
the effects of high-level listening on monosyllab- importance to speech intelligibility of different in-
ic word recognition in a group of 72 young adult tensities within the dynamic range of speech,
normal-hearing listeners, a group of 32 adult Studebaker and Sherbecoe (2002) investigated
103
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
the validity of the 30-dB dynamic range associat- held presently with regard to the various AI meth-
ed with the various AI procedures, as well as the ods, which assumes that intensity is uniformly
STI. Specifically, they assessed the assumption distributed across each frequency band.
that the intensity importance assigned to each fre- Hogan and Turner (1998) assessed the effect
quency band is uniformly distributed across the of increasing audibility in high-frequency regions
dynamic range and that the dynamic range is the on speech-intelligibility scores for normal-hearing
same for all frequency bands. and hearing-impaired listeners exhibiting high-
One hundred normal-hearing listeners were frequency impairments. Fourteen participants, 5
divided into 5 groups of 20 listeners each. Each normal hearing and 9 hearing impaired, served
group provided speech intelligibility data for NU- as subjects. Each hearing-impaired subject
6 monosyllabic-word lists (Studebaker et al., demonstrated no more than a mild hearing loss
1993), sharply filtered into one of 5 frequency through 500 Hz. Varying degrees of hearing loss,
bands (141 to 562, 562 to 1122, 1122 to 1778, however, were noted in the higher frequencies.
1778 to 2818, and 2818 to 8913 Hz). These fre- Subjects listened to nonsense syllables low-
quency bands were then mixed with filtered pass filtered at cutoff frequencies corresponding
speech-weighted noise and presented alone or in to one-third octave bands with center frequencies
pairs at 19 SNRs ranging from -25 to 47 dB. In of 400, 500, 630, 800, 1000, 1250, 1600, 2000,
essence, each frequency band was presented 2500, 3150, and 4000 Hz at 2 different supra-
alone, and also in conjunction with a second band threshold intensity levels (Figure 16). Each sub-
differing in frequency but consisting of the same ject was presented a minimum of 7 low-pass cut-
speech stimuli. For bands presented as pairs, one off frequencies (1120 to 4500 Hz), in addition to
band was kept at a fixed level, while the other a broad-band condition (9000 Hz cutoff).
was varied in level. Adjacent bands were never Audibility values were calculated for each by
paired. methods in the ANSI S3.5-1969 and ANSI S3.5-
Data were converted into importance values 1997 standards. Both calculation methods were
by determining the amount of change in speech used to compare the effect of high presentation
recognition that occurred with and without the levels when the level distortion factor was used
inclusion of particular auditory areas of interest.
This was accomplished using the equation
120
v ,,I I..
I=
(AX - AY) (12) 100 F
0
AT 80 H
me.
c
104
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
(ANSI S3.5-1997 standard) versus when it was The authors noted no improvement in intelli-
not used (ANSI S3.5-1969 standard). No profi- gibility scores with increase in bandwidth for
ciency or desensitization factors were applied to hearing-impaired participants when hearing sen-
either method. Speech recognition was quanti- sitivity increased above 55 dB HL. For listeners
fied, however, using an efficiency measure that exhibiting hearing levels below 55 dB HL, infor-
compared the relative performance of hearing- mation between 4000 and 8000 Hz did not im-
impaired and normal-hearing listeners as audi- prove speech-intelligibility scores.
bility was increased by an additional high-fre- Hogan and Turner (1998) concluded that in-
quency band of speech. Efficiency was defined as creasing bandwidth may result in a decrease in
"a measure of how well the hearing-impaired lis- speech intelligibility for some listeners with high-
teners used speech information presented at au- frequency hearing loss, a finding supported first
dible levels and at various frequencies compared by the work of Murray and Byrne (1986), and
to the normal-hearing listeners" (Hogan and more recently by Ching et al. (1998) and Turner
Turner, 1998, p. 437). and Cummings (1999). In addition, Hogan and
Efficiency was computed using the equation Turner (1998) noted a "diminishing return in am-
plifying high-frequency speech information, in
A Scorel A Al (hearing impaired) that as amplification is provided in high-frequen-
Efficiency cy regions of moderate and severe impairment,
A Scorel A AI (normal) particularly when hearing loss exceeds 55 dB HL,
the result may be a decrease in speech recogni-
The formula compares the ratio, in hearing-im- tion performance" (p. 440).
paired and normal-hearing listeners, of change in Investigators have attempted to understand
measured intelligibility to the change in Al pre- the implications of the reduction in speech-intel-
diction occurring at higher signal levels. ligibility performance as gain is increased in the
Quantitatively, efficiency was equal to 1.0 if the high frequencies. Van Tasell (1993), reporting on
hearing-impaired listener was able to use the in- the work of Liberman and Dodds (1984), pointed
crement in speech audibility from a given band out that a decrease in speech-intelligibility per-
to improve speech intelligibility to the same ex- formance as hearing loss increased beyond 60 dB
tent as the normal-hearing group. An efficiency HL was evidence of cochlear damage not only to
score of < 1.0 indicated that the hearing-impaired the outer hair cells, but also to the inner hair
listener was less able to use the available infor- cells. In fact, it has been conjectured that individ-
mation to improve the score. uals with similar hearing losses perform differ-
If the increase in bandwidth resulted in ently with hearing aids as a consequence of dif-
rollover (ie, decrease in speech-intelligibility per- ferent degrees of inner hair cell damage (Killion,
formance), efficiency would result in a negative 1997). Moore (2000, 2001) and Moore and
value. Findings demonstrated an increase in in- Alcaintara (2001) reported that the cochlea of an
telligibility scores as audibility increased for nor- individual exhibiting a severe hearing impairment
mal-hearing listeners and listeners with mild could, in fact, have dead regions in which inner
high-frequency hearing losses. They revealed also hair cells or neurons are not functioning.
that listeners with more severe high-frequency Unfortunately, such a diagnosis cannot be deter-
hearing losses performed more poorly than nor- mined presently based on traditional audiomet-
mal-hearing, mildly impaired, or moderately im- ric test measures.
paired listeners. To circumvent this shortcoming, Moore
In addition, comparisons were plotted as a (2000) developed the Threshold Equalizing Noise
function of the Al methods. Findings from this (TEN) test. Briefly, this test assesses thresholds
comparison showed that for a given condition, for tones in the presence of an ipsilateral broad-
the newer standard produced lower values, but band noise masker that produces essentially equal
resulted in more values from the hearing-im- masked thresholds, in dB SPL, across a wide fre-
paired data falling within the intervals predicted quency range in both normal listeners and listen-
for normal-hearing listeners. This suggests that ers with hearing impairment that is unaccompa-
speech intelligibility can be estimated more ade- nied by dead regions. A dead region is indicated
quately from audibility when the effect of level at frequencies where masked thresholds are 10
distortion is taken into account. dB or more above absolute threshold and 10 dB
105
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
or more above the equivalent rectangular band- equately predict the intelligibility performance of
width (ERB) level of the masker. listeners with severe hearing impairments.
According to Moore (2001), the TEN test is a Effective audibility is expressed as
clinically feasible substitute for psychophysical
tuning curves (PTCs), and is useful in isolating Effective Audibilityi =
regions of the basilar membrane where inner hair Desensitized Audibilityi xLi (15)
cells are not functioning. This test procedure is
based on the principle that a cochlear dead re- Recall that Li refers to the level distortion factor,
gion will produce absolute thresholds of the tonal LDF. HLD is expressed in terms of desensitized
signal when the intensity of the signal is loud audibility at the sensation level of speech peaks.
enough to produce a detectable spread of excita- Thus
tion at an adjacent region where there are sur-
viving inner hair cells and neurons. Results of the m
TEN test could shed light on the functioning of Desensitized Audibilityi = (16)
the inner hair cells and indicate regions in which [1 + (30/SL)Pi] '/Pi
amplification is not likely to be useful in improv-
ing speech intelligibility. Still in its infancy, the where SLi is the difference between the maximum
TEN test appears to have potential for future clin- level of the signal and the hearing threshold level
ical use. at the ith frequency band. The factor mi is the
Both the Ching et al. (1998) and Hogan and maximum value of desensitized audibility. The
Turner (1998) findings revealed that level distor- rate at which effective audibility changes with
tion and desensitization affect the ability of hear- audibility at low sensation levels is equal to
ing-impaired listeners to understand speech. mi/30. The parameter pi is used to control the
Ching et al. (2001) addressed the implications of nonlinear function that relates sensation level to
allowing, or failing to allow, for hearing loss de- effective audibility. For large values of pi, and
sensitization when prescribing hearing aids. They when mi = 1, effective audibility is equal to the
suggested that attempts to maximize audibility at band audibility function as defined in the ANSI
the high frequencies, where the hearing loss is se- S3.5-1997 standard. Both pi and mi vary with
vere, might be inappropriate because the in- hearing loss and frequency, and the relationship
creased gain in regions of severe impairment re- between desensitized audibility and audibility for
sults in greater loudness, but not necessarily en- different degrees of hearing losses at different fre-
hanced intelligibility. As a result, clinicians and quencies varies. Figure 17 demonstrates the rela-
researchers should aim to maximize effective au- tionship at 3000 and 4000 Hz.
dibility, or the contribution of audibility to speech After deriving effective audibility, Ching et al.
intelligibility. Effective audibility was incorporat- (2001) compared the observed speech scores of
ed into the SII formula (Equation 3) as follows hearing-impaired listeners at each of 6 sensation
(Ching et al., 2001): levels when speech was filtered in a low-frequen-
cy band (0 to 700 Hz) and a high-frequency band
(1400 to 5600 Hz), and when speech was broad-
SII(LDF HLD) = XI, x Effective Audibilityi (14) band filtered (0 to 5600 Hz) to predict speech
scores based on the original AI (ANSI S3.5-1969),
As seen in this equation, the SII was modified to
SII(LDF), and SII(LDF, HLD) (ie, effective audibility)
combine the effects of LDF and a modified desen- methods. Findings from the low-frequency band
sitization factor, or HLD.6 These factors were revealed that all methods adequately estimated
combined based on the findings of previous em- the performance of listeners with normal hearing
pirical evidence demonstrating that audibility and those with mild or moderate flat hearing loss-
alone or audibility combined with LDF do not ad- es. Performance for individuals with sloping mild
or moderate hearing losses was underestimated
6Hearing loss desensitization (HLD) is similar to the term
desensitization used throughout this paper. As seen in by both SII(LDF) and SII(LDF, HLD) except at the high
Figure 15, the difference in terms lies in the fact that sensation levels, where effective audibility was
Studebaker and Sherbecoe (1994) and Studebaker et al. found to be a good estimate. For severe and pro-
(1997) modified this factor based on speech performance found flat hearing losses, both procedures that con-
data suggesting that desensitization occurs nonlinearly when
hearing loss exceeds 120 dB. sidered LDF adequately estimated performance at
106
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
3000 Hz 0dB H l
ers, provided less high-frequency gain than most
0.8
other prescriptive procedures. The data were used
-c , 08<60dBHL in the derivation of the new NAL prescription for
<0) 6 0.6 nonlinear hearing aids (NAL-NL1) (Dillon, 1999;
N 0.4 , 80 dB Hl 0.4
Byrne et al., 2001).
0.2 - 100ldBHL 0.2 B. Age
G
2
O r- Another factor to consider in the context of the
0 10 20 30 40 0 10 20
AI is the age of hearing-impaired listeners. Age is
Audibility or Sensation Level (dB) known to affect speech understanding (eg, Plomp
and Mimpen, 1979; Dubno et al., 1984; Hargus
Figure 17. Hearing loss desensitization at 3000 and and Gordon-Salant, 1995; Abel et al., 2000). In
4000 Hz as a function of audibility for different degrees fact, some reports suggest that speech-intelligi-
of hearing loss (calculated using unpublished data, bility performance decreases in listeners as early
Ching et al., 2001). as age 40, while others have indicated steady per-
formance for listeners through 60 years of age
(for a review, see Willot, 1991).
low sensation levels and overestimated perfor- For the elderly population, explanations for
mance at high sensation levels. Performance was decreased speech-intelligibility performance have
well predicted by both procedures that considered varied. Overall, this decline in performance can
LDF, but was not well predicted by the AI proce- be attributed to a reduction in auditory sensitivi-
dure as specified in the ANSI S3.5-1969 standard. ty (Humes and Roberts, 1990; Humes et al.,
For the 1400 to 5600 Hz (high-frequency) 1994), peripheral processing changes (Jerger et
band, each of the three methods provided ade- al., 1989; Schum et al., 1991; Hargus and
quate descriptions of performance for persons Gordon-Salant, 1995), and changes in the central
with mild-to-moderate hearing losses. For indi- auditory nervous system (Humes et al., 1996). As
viduals with severe or profound hearing loss, both a result, an age-correction factor might be needed
the standard Al and SII(LDF) methods were found to compensate ". . . for the fact the AI does not
to overestimate considerably the performance at predict the performance of older listeners as well
high sensation levels. For this group, effective au- as it predicts the performance of younger listen-
dibility was found to be a better predictor of in- ers" (Studebaker et al., 1997, p. 151).
telligibility performance. Investigators have assessed differences in
In comparison, for the broad-band speech speech-intelligibility performance of groups dif-
low-pass filtered at 5600 Hz, the authors found fering in age. Magnusson (1996) investigated the
comparable predicted and observed scores for in- systematic influence of presbycusis and desensiti-
dividuals exhibiting a mild or moderate hearing zation for 57 elderly subjects, aged 61 to 88
loss. At high sensation levels, effective audibility years, exhibiting a sloping mild-to-moderately se-
was found to provide a better estimate of perfor- vere hearing loss. To accomplish this task, the
mance for subjects demonstrating severe or pro- newer AI (ANSI S3.5-1997) procedure was com-
found hearing losses. pared to two modified versions of the same pro-
These findings contribute significantly to cedure using a regression analysis. The first mod-
hearing aid fitting applications, particularly as re- ified version simply took into account desensiti-
lated to the prescriptive fitting of the frequency- zation (Equation 8), while the second modified
gain response and loudness. The clinician must version incorporated desensitization and an age-
be aware of the implications of providing high- correction proficiency factor. Age was corrected
frequency amplification for persons exhibiting se- by the addition of the individual proficiency fac-
vere or profound hearing loss. These recent find- tor, or Pa. This resulted in the formula of
ings helped to explain why the National Acoustic
Laboratories-Revised for profound linear pre- SII = ZIiAiDiP, (17)
scriptive method, or NAL-RP (Byrne et al., 1990),
which was based on intelligibility judgments and where Pa equates to a proficiency factor of 0.919
performance of severely hearing-impaired listen- for individuals up to 83 years of age. For indi-
107
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
viduals above 83 years, P, took the regression addition, intersubject variability was found to be
form of greater for subjects over age 40 than for those
under age 40. No statistically significant differ-
Pa = -0.0614 x a + 6.0733 (18) ences attributable to age were noted in learning
effects or intrasubject test-retest reliability.
where a is the age. Studebaker et al. (1997) concluded that the ef-
The author (Magnusson, 1996) indicates that fect of age on the AI, independent of hearing loss,
these age-correction factors suggest no age de- was relatively small.
pendence below 83 years of age and suggest a
performance decrease of about 4%, or 2 words in C. Reverberation
a standard 50-item list, per year for individuals
above 83 years of age. The proficiency factor, Humes and Roberts (1990) assessed the effective-
however, should not be discounted entirely for ness of the mSTI in predicting speech performance
persons below age 83 years, as it accounts for in- in noise for hearing-impaired listeners. They ex-
dividual differences not explained by audibility. amined the speech-intelligibility performance of
Magnusson (1996) compared differences be- young normal-hearing adults, elderly hearing-im-
tween measured and predicted intelligibility paired adults, and young normal-hearing adults
scores obtained with the unmodified Al method with simulated sensorineural hearing loss using
(Equation 3) and the two age-corrected modified the mSTI. Thresholds for the elderly hearing-im-
procedures described previously. Results showed paired group consisted of a sloping high-frequen-
a 14% improvement in speech intelligibility for cy hearing loss.
the Al method incorporating the desensitization For the group of subjects tested under the
correction factor when compared to the unmodi- simulated condition, hearing loss was simulated
fied Al method. The modified AI incorporating by introducing a spectrally shaped masking noise
both desensitization and age-correction factors into the ear receiving the target signal, which re-
was found to exhibit 21% and 7% improvements, sulted in an equivalent loss of audibility seen in
respectively, over the original and initially modi- the elderly subject group. Subjects listened to
fied Al procedures. Furthermore, an overall age nonsense syllables, presented in a closed-set for-
effect was not apparent for the modified Al in- mat, recorded in quiet and noise. Speech was pre-
corporating the proficiency factor Pa This, how- sented at 0 azimuth through the left and right
ever, was not the case for individuals above 83
years of age, who demonstrated a decrease in in- ears of an acoustic manikin (KEMAR) (Burkhard
telligibility performance. and Sachs, 1975), while noise was presented at 0
Studebaker et al. (1997) also investigated the and 90 azimuth relative to KEMAR.
impact of age on speech recognition by compar- Recordings were made under anechoic and
ing the efficacy of an age-dependent correction reverberant (Rt = 3.1 sec) conditions with no
factor on the Al. Holding audibility constant, the noise, noise at 00, and noise at 90. Listening con-
investigators compared performance across seven ditions for the normal-hearing group and elderly
age groups (20 to 80 years) using digitized NU-6 hearing-impaired group included stimulus pre-
monosyllabic words. Audibility was fixed by pre- sentation in the right ear only, left ear only, and
senting band-pass-filtered stimuli at a constant binaurally at a presentation level of 75 dB SPL.
SNR and limiting threshold losses to 25 dB HL for Data were collected under the left ear only con-
frequencies between 250 and 2000 Hz. Eight 50- dition for the simulated hearing loss group.
word lists were presented to each listener on two Results revealed that the mSTI, derived using
different days. Raw data were converted to rau an octave-band method, overestimated the per-
scores by using an arcsine transform (Studebaker, formance of listeners with sloping high-frequency
1985). Results revealed that performance did not hearing loss, either real or simulated. This was
vary considerably with age, except for subjects not found to be the case with young normal-hear-
over 70 years. For the 70- to 80-year-old group, a ing listeners in either the anechoic or reverberant
modest reduction in scores was noted relative to conditions. The authors concluded that the use of
the 30-year-old group only. A significant reduc- octave bands and nonsense syllables might have
tion in scores was noted for subjects over 80 years contributed to the overestimation of speech-intel-
of age in comparison to all other age groups. In ligibility performance.
108
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
120
1,_ 1 _ L .. jIII I
109
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
110
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
11 0 AS
100 Discomfort
Level
Similarly, incorrect conclusions can also be 90
drawn when considering only the absolute val-
co
111
teristics. Although these investigations address in- As a result, subjects were expected to receive
directly the intelligibility of speech, their direct "the minimum amount of gain that assured max-
focus is on audibility. Humes (1986), for exam- imum audibility of the speech spectrum in each
ple, compared prescribed insertion-gain values frequency region" (Rankovic, 1991, p. 392), in
derived from 10 linear hearing aid selection pro- principal, an AI of 1.0. Twelve hearing-impaired
cedures as applied to 9 different degrees of au- subjects, each exhibiting various degrees and con-
diometric configurations. The hypothetical au- figurations of hearing loss, participated in the
diograms for each of the 9 cases were grouped by study. No subject achieved an Al of 1.0 because of
their slope configurations of sloping, flat, and ris- the noise floor, limitations of the dynamic range
ing, resulting in 90 sets of prescribed insertion- of the equipment, and/or subjects' UCLs. The re-
gain data. These insertion-gain data were then lationship between the percentage of correct
compared to determine how well each procedure scores on a nonsense syllable test and audibility
achieved the goal of maximizing speech intelligi- was compared for all subjects. Overall, findings
bility. The results of this analysis showed that showed AIMax to be similar to or slightly better
when volume-control adjustments were not al- than POGO and NAL-R, possibly a result of its re-
lowed, 8 out of the 10 prescriptive methods pre- quiring more prescribed gain than the other pre-
scribed sufficient gain to yield optimal speech-in- scriptive methods. AIMax appeared to provide ad-
telligibility performance. The insertion-gain data equate gain for individuals exhibiting relatively
were then compared to hearing-impaired listener flat hearing losses, although data did not reveal if
preferences based on the work of Leijon et al. these listeners found the AIMax condition to be
(1984). Interestingly, findings revealed that the comfortable during everyday listening conditions.
same two procedures found not to provide opti- While this increase did not always yield an im-
mal speech-intelligibility performance were most provement in observed scores, it also did not de-
preferred by these listeners. It was also surmised grade performance for most listeners. Four of the
that when the listener is given the ability to adjust subjects, each demonstrating a sloping high-fre-
the overall gain with the use of a volume control, quency hearing loss with near-normal low-fre-
any of the procedures could produce optimal quency thresholds, demonstrated a decrease in
aided speech-intelligibility performance. speech-intelligibility performance for AIMax. This
In 1991, Rankovic compared the frequency- decrease in performance occurred as a result of
gain response of two linear prescriptive proce- the considerable amount of gain needed in the
dures and a prescription that aimed to maximize higher frequencies to achieve maximum intelligi-
the AI (AIMax). The hearing aid prescriptive pro- bility. Skinner (1988) also noted a similar find-
cedures were the Prescription of Gain/Output ing for persons with high-frequency hearing loss.
(POGO) (McCandless and Lyregaard, 1983) and Linear prescriptive approaches fail to address
the National Acoustic Laboratories-Revised pro- the fact that changes in speech spectra, ambient
cedure (NAL-R) (Byrne and Dillon, 1986). POGO noise, and other environmental factors are com-
is essentially one of a handful of half-gain meth- monplace in everyday listening situations. Hou
ods, while NAL-R was derived from a consider- and Thornton (1994) developed a method for in-
able amount of clinical and laboratory research tegrating the AI as a means to predict speech in-
aimed at amplifying the long-term average speech telligibility across a range of listening conditions.
spectrum to a comfortable equal-loudness con- This integrated AI model, termed IAI, takes into
tour, predicted from pure-tone thresholds (Byrne, account hearing threshold, masking of noise, self-
1986[a], 1986[b]). masking of speech, high-level cochlear distortion,
AIMax was aimed at maximizing the frequen- and peak-clipping effects of a hearing aid. The IAI
cy-gain response by positioning the 30-dB dy- evaluates separately Al values of hearing aid per-
namic range of the short-term speech levels 15 to formance across several listening conditions.
18 dB above the listener's pure-tone thresholds These derived Al values are first grouped accord-
and below the level of discomfort. This was ing to such factors as listening conditions and
achieved by selectively amplifying the speech input levels, and then weighted based on impor-
spectrum so that the long-term average one-third tance. The sum of these groupings and weightings
octave-band level of speech in each band was 18 is then integrated to provide an estimate of the
dB above the pure-tone threshold and below a optimal performance, or optimal IAI (OIAI), for a
formulated UCL. specific hearing aid or hearing aid characteristic.
112
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
Hou and Thornton (1994) validated their ap- tion and hearing loss desensitization (as specified
proach by using the integrated Al across a range in Ching et al., 2001), within the constraint of an
of listening conditions as a criterion for evaluat- overall loudness that is either the same or lower
ing a specific hearing aid response characteristic than that experienced by a normal-hearing lis-
and calculating an optimal frequency-gain char- tener at the same input level (Dillon, 1999; Byrne
acteristic that maximizes the IAI. For a high-fre- et al., 2001). This prescriptive procedure can be
quency hearing loss, the OIAI was compared to the implemented via standalone software or via some
NAL-R (Byrne and Dillon, 1986) and the POGO manufacturers' fitting software. Like any other
(McCandless and Lyregaard, 1983) prescriptions. such procedure, the prescription is based on
The results in quiet showed no difference among group data, and does not account for individual
the three procedures in predicted performance, but differences. It should be viewed, therefore, as pro-
in a noisy condition, superior performance was viding a starting point, with fine-tuning per-
predicted for the OIAI procedure when compared formed as needed for individual hearing aid
to the two prescriptive procedures. users. Because NAL-NL1 is still in its infancy, its
As a result of the examples presented in this effectiveness in predicting the speech-intelligibil-
section, the reader should readily appreciate that ity performance of aided listeners has not yet
the selection and fitting of the electroacoustic been validated fully.
characteristics of hearing aids should be carefully
considered. Simply stated, the highest AI value D. Evaluation of Amplification Strategies
may not provide the listener with optimal speech
intelligibility. Empirically, Byrne (1986[b]) found In 1990, Fabry and Van Tasell used the Al to eval-
this to be true. He assessed judgments of quality uate a hearing aid incorporating adaptive fre-
and intelligibility in quiet and noise for 14 ears quency response (AFR). AFR is a hearing aid cir-
of 11 participants, 7 of whom exhibited precipi- cuit that can be enabled to reduce automatically
tous hearing losses. Frequency responses yielding the low-frequency gain in noisy backgrounds. The
greater high-frequency emphasis produced the authors chose to investigate this feature due to its
highest Al values. However, six of the seven lis- widespread application throughout the industry
teners with precipitous loss judged the frequency at the time, despite their acknowledgement that
responses yielding the highest Al value to be the articulation theory generally has not supported
poorest in three of four trials based on paired- this type of signal processing as an effective noise-
comparison preference judgments. In 1991, reduction technique (French and Steinberg, 1947;
Rankovic confirmed this finding when partici- Fletcher, 1952). Using connected discourse, a
pants with high-frequency hearing losses did not transfer function relating rated speech intelligi-
achieve the best speech recognition with a fre- bility to the Al was derived for 12 normal-hearing
quency response that maximized AI. subjects. This transfer function was then used to
predict aided speech-intelligibility ratings by 12
C. Prescription of Frequency-Gain Response hearing-impaired listeners wearing a master hear-
ing aid with an adaptive filter circuit that could
Researchers have also attempted to use the Al be activated (AFR on) or deactivated (AFR off).
model to specify a frequency-gain characteristic For all subjects, results showed that the Al pre-
for optimizing speech-intelligibility scores. Humes dicted no improvements in speech-intelligibility
(1986) and Rankovic (1991) have both reported performance for the AFR-on versus AFR-off con-
the use of procedures that aimed to amplify the dition. In addition, no significant improvements
full 30-dB range of speech to be audible at each in rated intelligibility were observed. For every
frequency for hearing-impaired listeners. These hearing-impaired participant, however, ratings of
procedures have not been adopted for clinical use. speech intelligibility were related monotonically
At present, the only prescriptive procedure to Al.
that uses an Al model for deriving optimal fre-
quency-gain characteristics is the NAL-NL1 pro- E. Prediction of Hearing Aid
cedure for fitting nonlinear hearing aids (Dillon, Outcomes Speech Gain
1999). This prescription aims to maximize speech
intelligibility using a modified AI procedure In an attempt to overcome the need for deriving
(ANSI S3.5-1997) that incorporates level distor- proficiency factors and the various transfer func-
113
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
114
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
115
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
level. During pilot testing, Souza and Turner ing losses showed similar improvements in intel-
(1999) found this level to provide maximum au- ligibility, given comparable increases in audibili-
dibility without discomfort for listeners with sim- ty. At the high-intensity level, however, listeners
ilar degrees of hearing loss. As a result, the LIMIT with high-frequency hearing loss showed smaller
and DF factors were omitted for both AAI calcu- improvements in intelligibility with nonlinear am-
lations (Equation 19). A two-channel compressor plification when compared to listeners exhibiting
with a 1500 Hz center frequency was used for the flat audiometric configurations.
compression-amplified condition. Each channel of The authors offered two possible explanations
the WDRC device compressed signals above 45 for the difference between groups. First, empirical
dB SPL and at compression ratios of 2:1 and 5:1 evidence has shown that listeners with sloping
in the low-frequency and high-frequency chan- hearing losses have a compromised ability to use
nels, respectively. temporal cues at high intensity levels (Bacon and
Sixteen hearing-impaired subjects heard 16 Viemeister, 1985; Bacon and Gleitman, 1992).
VCV stimuli spoken by 2 male and 2 female talkers Second, recalling that a greater compression ratio
at presentation levels of 55, 70, and 85 dB SPL for of 5:1 was used in the high-frequency channel,
both amplified conditions. Each subject's task was time-amplitude cues in this region may have been
to indicate recognition of the correct consonant in affected the most, which in turn, might have led
a 16-alternative forced-choice paradigm. Using the to a further reduction in audibility.
appropriate AAI measurement, the relationship In 2001, Souza and Kitch compared preferred
between audibility and intelligibility was exam- volume-control settings for (1) a linear peak-clip-
ined for each type of amplification. ping, (2) compression-limiting, and (3) WDRC
Souza and Turner (1999) found that at low multimemory, programmable behind-the-ear
and moderate input levels, AAI and intelligibility (BTE) device. For each amplification type, speech
scores were higher for compression-amplified audibility was quantified at the listener's pre-
speech relative to linear-amplified speech. At the ferred volume setting. Ten listeners with mild-to-
higher input level, however, AAIs and intelligibil- moderate cochlear hearing loss were fitted
ity scores were similar for the two types of ampli- monaurally. The frequency response, OSPL90,
fication. Because intelligibility scores increased and WDRC ratio were set individually based on
monotonically for both types of amplification, no the DSL [i/o] (Cornelisse et al., 1995) prescrip-
significant differences were found between com- tive procedure.
pression-amplified speech and linear-amplified For the compression-limiting hearing aid con-
speech. This finding suggests that a given increase dition, the aid was set to an output-compression
in audibility resulted in the same increase in in- mode, and compression thresholds were pro-
telligibility for both amplification conditions. grammed between 65 and 85 dB SPL, with a com-
Clinically, this outcome offers preliminary evi- pression ratio ranging between 8:1 and 20:1 rel-
dence that compression does not introduce detri- ative to the listener-preferred volume-control set-
mental changes to the speech signal that might ting. An input-compression mode was used for
otherwise reduce audibility. the WDRC condition, with the compression
Recently, Souza and Bishop (2000) took the threshold fixed at 50 dB SPL and compression ra-
work of Souza and Turner (1999) one step fur- tios ranging from 1:1 to 2.7:1.
ther. Specifically, they attempted to determine Subjects listened to 20 passages of the Speech
whether increases in audibility with nonlinear Intelligibility Rating (SIR) test (Cox and
amplification improved speech intelligibility to a McDaniel, 1989) presented at input levels of 50,
comparable degree for 10 listeners with sloping 65, and 80 dB SPL in quiet and against the SIR's
sensorineural hearing loss relative to a group of accompanying cafeteria noise (+ 7 dB SNR).
10 listeners with a flat sensorineural configura- Audibility was calculated using linear and WDRC
tion. Except for the selection of a frequency-gain versions of the AAI. Subjects were required to ad-
response that maximized high-frequency audibil- just the volume control of their instrument until
ity, the methods and AAI calculations were simi- maximum clarity (intelligibility) was determined.
lar to those used in the Souza and Turner (1999) Once the listener had indicated that a preferred
study. volume-control setting had been achieved, the
For linear-amplified speech, overall results re- real-ear hearing aid output was measured using a
vealed that listeners with flat and sloping hear- probe-microphone system.
116
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
117
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
aids. They measured the effectiveness of global The amount of attenuation, however, can vary
satisfaction and hearing aid adherence using two relative to microphone configuration and the aid's
hearing-specific surveys and self-reported ratings. quantitatively determined directionality charac-
Subjects consisted of 115 patients seen through teristics (Preves, 1997; Ricketts and Mueller,
the Department of Veterans Affairs. All subjects 1999; Valente, 1999, 2000; Ricketts and Ditt-
were fit binaurally with custom analog, nonpro- berner, 2002). An electroacoustic method that
grammable instruments incorporating either can be used to account for these variations is the
peak-clipping or compression-limiting schemes. directivity index (DI). The DI is typically derived
Target, unaided, and aided AI values were deter- by presenting frequency-specific stimuli (often
mined during the hearing aid fitting by means of 500 to 4000 Hz) directly in front of the hearing
real-ear analyzer software. The AI value was aid as the aid is rotated either in free field or on a
based on the AJ(4) procedure developed by manikin at discrete angles, or azimuths. This elec-
Pavlovic (1988) and modified to incorporate con- troacoustic measure is found in many manufac-
versational speech levels (Pavlovic, 1991). The ef- turers' technical specification manuals.
fectiveness of hearing aids was determined by the To provide a reasonable estimate of the im-
amount of change in listening ability without and provement in speech recognition in noise for
with hearing aids. The Abbreviated Profile of DMHAs under laboratory conditions, Killion and
Hearing Aid Benefit (APHAB) (Cox and colleagues (1998) report a means by which audi-
Alexander, 1995) and the Satisfaction with bility can be estimated from DI values. Using
Amplification in Daily Life (SADL) (Cox and speech weighting from the Mueller and Killion
Alexander, 1999) were the two standardized tests (1990) count-the-dot audiogram, importance
used. APHAB data were obtained unaided, prior functions of 0.20, 0.23, 0.33, and 0.24 are as-
to hearing aid fitting, and aided at the 1-month signed to the frequencies of 500, 1000, 2000, and
follow-up visit. Data for the SADL were obtained 4000 Hz, respectively, to achieve audibility. This
1 year after fitting as a means to evaluate hear- modification has been termed the Articulation
ing aid quality. Adherence data were also gath- Index-Directivity Index (AI-DI) and is calculated
ered 1 year after fitting through the use of ques- using the equation
tions regarding the number of hours per day and
days per week the aids were worn. Results re- AI-DI = (0.2 x DI5o) + (0.23 x DI,ood)
vealed no systematic relationship between mea- + (0.33 x DI200) + (0.24 x DI400) (20)
sures of improved audibility and patient ratings
on communication ability. The authors hypothe- The AI-DI is calculated first by determining each
sized that this finding may be the result of the manufacturer-reported frequency-specific DI
volume-control level at which measurements value. For example, assume a manufacturer re-
were initially made. A similar conclusion was ports values of 2.1, 2.5, 3, and 2.5, measured at
also drawn for improved audibility and overall 500, 1000, 2000, and 4000 Hz, respectively, for
hearing aid satisfaction. A major outcome, how- an in-the-ear instrument. Second, multiply each
ever, was that a moderate relationship was found DI value by its frequency-specific importance
between hearing aid use and achievement of im- function. Lastly, sum the importance functions
proved audibility. across frequencies, resulting in an AI-DI value of
2.59 (0.42 + 0.58 + 0.99 + 0.60).
G. Articulation Index-Directivity Index (AI-DI) In contrast, an unweighted DI value is deter-
mined by adding the manufacturer-reported DI
The Al can also be applied to directional-micro- values and dividing by 4, or the number of fre-
phone hearing aids (DMHAs). Studies have quencies. Using the same values in the earlier ex-
shown that DMHAs effectively provide listeners ample, the calculation results in an unweighted
with increased speech intelligibility in noise over DI value of 2.53 ([2.1 + 2.5 + 3 + 2.5]/4). Note
omnidirectional-microphone devices (for a re- that the AI-DI provides a benefit of 0.06 dB (2.59-
view, see Valente, 1999 and Amlani, 2001). This 2.53) over the unweighted DI method as a result
improvement in speech intelligibility occurs as a of the greater importance given to the high fre-
result of the DMHA's ability to attenuate sounds quencies, predominately the 2000 Hz region.
from the rear and sides of the listener with re- Despite its intuitive and clinical appeal for com-
spect to sounds originating from directly in front. paring the improvement in speech recognition in
118
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
noise for DMHAs, the AI-DI has not been investi- al., 2001), is the only formal prescriptive proce-
gated systematically. dure developed specifically to use the principles
of articulation theory to prescribe the hearing aid
H. Summary frequency-response characteristics that maximize
speech intelligibility.
Based on the evidence provided in this section, it Great strides have been made to incorporate
can be concluded that: the AI in assessments of unaided and aided
1. No single prescriptive formula, either linear or speech audibility for hearing-impaired listeners
nonlinear, has been determined to best maxi- and to overcome inaccuracies inherent in these
mize speech intelligibility. To date, recently de- indices by attempting to account for a variety of
veloped prescriptive formulas have not been in- confounding factors, such as severity of hearing
vestigated empirically. loss, desensitization, age, distortion, and upward
spread of masking. This review has made it obvi-
2. Almost any change to a hearing aid's frequen- ous that these applications, to be useful for pre-
cy response results in a change in the Al value. dicting intelligibility in hearing-impaired listen-
Changes in the input level or hearing aid gain ers, will require the incorporation of such correc-
will also result in changes in the Al. Various tions because they account for much of the vari-
electroacoustic features of a hearing aid inter- ability in the speech intelligibility of hearing-im-
act with each other, making a determination of paired listeners with moderately severe-to-pro-
which electroacoustic features optimize the Al found hearing impairments.
very difficult. Determining which electroa- Researchers, to achieve their desired research
coustic features maximize speech-intelligibili- objectives, have promulgated piecemeal formulas
ty performance, therefore, is an issue that will that have received rather limited use beyond the
not be easily resolved. scope of their particular studies. For clinical use,
3. Use of the Al as an outcomes-assessment tool it is desirable that corrections to the Al be com-
seems enticing, provided the AI calculation is bined, insofar as possible, into a single formula,
performed at the volume-control setting used and that such formulas be implemented through
in everyday life. We believe that further em- software applications that audiologists can readi-
pirical evidence is needed, however, before ly apply. Clinically useful applications based on
clinical implementation can be validated. such refinements are likely to become a reality
only if researchers, clinicians, and manufacturers
of hearing aids, hearing aid analyzers, and real-
ear analyzers collaborate to contribute their
unique perspectives.
Conclusion As Licklider and Miller (1951) noted-in the
quote in our introduction-it is ostensibly more
Devised as a tool for predicting the intelligibility efficient to predict speech intelligibility than to
of speech transmitted by telecommunication de- measure it directly. Our review of the literature
vices, the Articulation Index and its successor, the leads to the observation that research efforts to
Speech Intelligibility Index, have been refined in develop suitable models for predicting speech in-
ways that make them clinically useful to audiolo- telligibility also require substantial time and re-
gists. Using these indices, audiologists can theo- sources. As long as such efforts continue to show
retically predict unaided and aided speech intel- promise for utilization in unaided and aided ap-
ligibility in their hearing-impaired patients with- plications for the hearing impaired, investiga-
out having to perform time-consuming measure- tions of more efficient methods to determine
ments of intelligibility. It is mainly through speech audibility and to predict intelligibility ap-
count-the-dot audiograms that these predictive pear justified.
methods have found their way into the routine Except for the various count-the-dot audio-
practice of clinical audiology. Implementation of grams and the NAL-NL1 fitting approach, current
these procedures as a primary tool in the fitting measurement strategies remain relatively ineffi-
of hearing aids and in the management of hear- cient and difficult to implement clinically in
ing loss, however, has been limited. To date, the demonstrating, either before or after a fitting,
NAL-NL1 fitting method (Dillon, 1999; Byrne et that a particular hearing aid or set of electroa-
119
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
coustic characteristics has potential for optimiz- of speech stimuli (eg, syllables and words). In ad-
ing speech intelligibility for a given hearing-im- dition, we view the development of appropriate
paired listener. Because contemporary fitting psychoacoustic measures to augment audiomet-
strategies essentially allow the direct measure- ric thresholds as essential to improving the
ment of intelligibility only after the fitting process prospects of developing a clinical tool that can ac-
is completed, the need will continue for clinical curately predict speech intelligibility in individ-
procedures that predict or estimate speech-intel- ual hearing-impaired listeners. It is hoped that
ligibility outcomes for alternative hearing aids or these attempts to produce an accurate predictor
signal-processing schemes. With further valida- of intelligibility will be accompanied by parallel
tion, the AI could also be used following the fit- attempts to develop short-and scientifically
ting of hearing aids as an outcome-assessment valid-direct tests of speech intelligibility. We be-
tool. lieve, though, that the many potential uses of an
More research is needed to develop the tradi- accurate predictor of intelligibility would ulti-
tional (auditory-only) Al before it can be used mately make it a more efficient tool for wide-
with confidence to predict speech-intelligibility spread clinical use than any given direct test of
performance under real-life conditions in individ- intelligibility.
ual hearing-impaired listeners. Investigation of In the near future, it is hoped that research
the integration of auditory and speech-reading efforts will provide the kinds of information that
cues appears to be a natural extension of such re- will allow practical applications of AI measures
search. Furthermore, the principles and proce- to take a more prominent role in the practice of
dures of the Al may prove to be extremely bene- clinical audiology. These applications can be ex-
ficial in the development of environmentally pected to improve significantly the diagnostic and
adaptive signal-processing schemes, as well as in rehabilitation capabilities of audiologists, allow-
future prescriptive formulas that optimize aided ing them to serve their hearing-impaired patients
speech intelligibility. more effectively.
Successful implementation of the Al in the
audiology clinic depends not only on success in
the laboratory but also upon the ability to build
bridges between scientific theory and clinical ap- Acknowledgments
plications. Count-the-dot audiograms are a good
example of success in bridging the laboratory and The authors gratefully acknowledge the contribu-
the clinic. As is true of the past, such efforts in tions of Harvey Dillon, Todd Ricketts, and two
the future will likely evolve through the work of a anonymous reviewers for their comments on pre-
core of interested researchers who tend to func- liminary versions of the manuscript. We also ex-
tion somewhat independently. Ideally, a more press gratitude to our colleagues who unselfishly
systematic research approach is needed in which contributed their original figures.
well-funded, collaborative research is aimed at
developing relatively simple techniques that au-
diologists can use confidently in hearing aid se-
lection and fitting. We would encourage the de- Glossary
velopment and refinement of simplified proce-
dures such as that described by Dillon (1993). Articulation Index (AI)-The quantitative as-
Investigations should, of course, aim at dis- pect of articulation theory that accounts for the
covery of all those factors that affect the accuracy contribution of audible speech cues in given fre-
of intelligibility prediction, including those cov- quency bands to speech intelligibility. Refers
ered in this review (ie, hearing thresholds, speech specifically to the principles and procedures de-
and noise spectra, reverberation, desensitization, scribed in the ANSI S3.5-1969 standard. Also
distortion, masking, visual cues). Indeed, if all or known as Audibility Index.
most of these factors can be addressed adequate-
ly in the predictive scheme, it would be reason- Articulation Index-Directivity Index (Al-
able to base predictions of intelligibility on either DI)-A means by which audibility can be esti-
sentences or connected discourse only, as op- mated from directivity index (DI) values using
posed to attempting to account for all other types speech weighting from the Mueller and Killion
120
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
121
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
122
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
Appendix C
An Example of the Calculation Method for the STI
The Speech Transmission Index (STI) can be determined either by direct measurement or through a calculation method.
Only the calculation method is described here. Readers interested in details regarding the direct-measurement proce-
dure are referred to the Steeneken and Houtgast (1980) article.
To calculate the STI, modulation transfer functions (MTFs) are determined for the octave frequencies 125 Hz to 8000
Hz at each of the following modulation frequencies: 0.63, 0.80, 1.00, 1.25, 1.60, 2.00, 2.50, 3.15, 4.00, 5.00, 6.30, 8.00,
10.00, and 12.50 Hz. These MTFs are then converted to signal-to-noise ratios (SNRs), as described in the text and
shown in the table. In the example presented, these values are hypothetically derived. Summing the values and divid-
ing by the denominator (ie, the number of MTFs) will result in the average MTF for a given octave. Next, the average
MTFs per octave are used to calculate the dynamic range of speech available. This is achieved by adding 15 at each oc-
tave's average MTF. The average MTFs for the nominal dynamic range of speech is then divided by 30, or the absolute
dynamic range of speech. This yields a proportion, or the Transmission Index per octave. To calculate the predictive
transmission of speech within this acoustical environment, the Transmission Index at each octave is multiplied by a
weighting factor shown to be optimal for correlating the STI to monosyllabic-word scores. Notice that the weighting
factors across frequency equal to 1 and are essentially equal across all octave bands. The last step in the STI calcula-
tion method requires the summing of the octave-weighted values. In this case, the STI resulted in a value of 0.85.
123
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
124
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
Byrne D, Dillon H, Ching T, Katsch R, Keidser G. NAL- high-frequency hearing loss. II: Articulation index
NL1 procedure for fitting nonlinear hearing aids: predictions. J Acoust Soc Am 85:355-364, 1989b.
Characteristics and comparisons with other proce- Dugal RL, Braida, LD, Durlach NI. Implications of previ-
dures. J Am Acad Audiol 12:37-51, 2001. ous research for the selection of frequency-gain char-
Cavanaugh WJ, Farrell WR, Hirtle PW, Walters GB. acteristics. In: Studebaker GA, Hochberg I, eds:
Speech privacy in buildings. J Acoust Soc Am 34:475- Acoustical Factors Affecting Hearing Aid
492, 1962. Performance. Baltimore: University Park Press, 1980,
Ching T, Dillon H, Byrne D. Prediction of speech recog- pp. 379-403.
nition from audibility and psychoacoustic abilities of Duggirala V, Studebaker GA, Pavlovic CV, Sherbecoe RL.
hearing impaired listeners. In: Jesteadt W, ed: Frequency importance functions for a feature recog-
Modeling Sensorineural Hearing Loss. Mahwah, New nition test material. J Acoust Soc Am 83:2372-2382,
Jersey: Lawrence Erlbaum, 1997, pp. 433-446. 1988.
Ching TY, Dillon H, Byrne D. Speech recognition of hear- Dunn HK, White SD. Statistical measurements on con-
ing-impaired listeners: Predictions from audibility versational speech. J Acoust Soc Am 11:278-288,
and the limited role of high-frequency amplification. 1940.
J Acoust Soc Am 103:1128-1140, 1998. Duquesnoy AJ, Plomp R. Effect of reverberation and
Ching TYC, Dillon H, Katsch R, Byrne D. Maximizing ef- noise on the intelligibility of sentences in cases of
fective audibility in hearing aid fitting. Ear Hear presbycusis. J Acoust Soc Am 68:537-544, 1980.
22:212-224, 2001. Egan JP. Articulation testing methods. Laryngoscope
Cornelisse LE, Seewald RC, Jamieson DG. Wide dynam- 58:955-991, 1948.
ic range compression hearing aids: The DSL[i/o] ap- Erber NP. Interaction of audition and vision in the recog-
proach. Hear J 47(10):23-29, 1994. nition of oral speech stimuli. J Speech Hear Res
Cornelisse LE, Seewald RC, Jamieson DG. The input/out- 12:423-425, 1969.
put formula: A theoretical approach to the fitting of Fabry DA, Van Tasell DJ. Evaluation of an articulation-
personal amplification devices. J Acoust Soc Am index based model for predicting the effects of adap-
97:1854-1864, 1995. tive frequency response hearing aids. J Speech Hear
Cox R, Alexander G. The abbreviated profile of hearing
Res 33:676-689, 1990.
aid benefit. Ear Hear 16:176-183, 1995. Festen JM, Van Dijkhuizen JN, Plomp P. Considerations
on adaptive gain and frequency response in hearing
Cox RM, Alexander GC. Measuring satisfaction with am- aids. Acta Otolaryngol 469(Suppl) :193-201, 1990.
plification in daily life: The SADL scale. Ear Hear
20:306-320, 1999. Finitzo-Heiber T, Tillman TW. Room acoustic effects on
monosyllabic word discrimination ability for normal
Cox RM, McDaniel DM. Development of the speech in- and hearing-impaired children. J Speech Hear Res
telligibility rating (SIR) test for hearing aid compar- 21:440-458, 1978.
isons. J Speech Hear Res 32:347-352, 1989.
Fletcher H. The perception of speech sounds by deafened
Crain TR, Van Tasell D. The effect of peak clipping on persons. J Acoust Soc Am 24:490-497, 1952.
the speech recognition threshold. Ear Hear 15:443-
453, 1994. Fletcher H. Speech and Hearing in Communication. New
York: Krieger, 1953.
DePaolis RA, Janota CP, Frank T. Frequency importance
functions for words, sentences, and continuous dis- Fletcher H, Galt RH. The perception of speech and its re-
course. J Speech Hear Res 39:714-723, 1996. lation to telephony. J Acoust Soc Am 22:89-151,
1950.
Dillon H. Hearing aid evaluation: Predicting speech gain Fletcher H, Munson WA. Masking and critical bands. J
from insertion gain. J Speech Hear Res 36:621-633, Acoust Soc Am 9:1, 1937.
1993.
Fletcher H, Steinberg JC. Articulation testing methods.
Dillon H. NAL-NL1: A new procedure for fitting non-lin- Bell Sys Tech J 8:806-854, 1929.
ear hearing aids. Hear J 52(4):10-16, 1999.
French NR, Steinberg JC. Factors governing the intelligi-
Dubno JR, Dirks DD, Ellison DE. Stop-consonant recog- bility of speech sounds. J Acoust Soc Am 19:90-119,
nition for normal-hearing listeners with high-fre- 1947.
quency hearing loss: I. The contributions of selected
frequency regions. J Acoust Soc Am 85:347-354, Gates GA, Popelka GR. Neural presbycusis: A diagnostic
1989a. dilemma. Am J Otol 13:313-317, 1992.
Dubno JR, Dirks DD, Morgan DE. Effects of age and mild Grant KW, Braida LD. Evaluating the articulation index
hearing loss on speech recognition in noise. J Acoust for auditory-visual input. J Acoust Soc Am 89:2952-
Soc Am 14:275-284, 1984. 2960, 1991.
Dubno JR, Dirks DD, Schaefer AB. Stop-consonant recog- Green DM, Birdsall TG. The effect of vocabulary size on
nition for normal-hearing listeners and listeners with articulation score. Technical Memorandum No. 81
125
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
and Technical Note AFCRC-TR-57-58, University of Factors Affecting Hearing Aid Performance (2nd ed).
Michigan: Electronic Defense Group, 1958. Boston: Allyn and Bacon, 1993, pp. 205-220.
Hargus SE, Gordon-Salant S. Accuracy of speech intelli- Humes LE, Coughlin M, Talley L. Evaluation of the use of
gibility index predictions for noise-masked young lis- a new compact disc for auditory perceptual assessment
teners with normal hearing and for elderly listeners in the elderly. J Am Acad Audiol 7:419-427, 1996.
with hearing impairment. J Speech Hear Res 38:234- Humes LE, Dirks DD, Bell TS, Ahlstrom C, Kincaid GE.
243, 1995. Application of the articulation index to the recogni-
Hawkins DB, Montgomery AA, Mueller HG, Sedge RK. tion of speech by normal-hearing and hearing-im-
Assessment of speech intelligibility by hearing-im- paired listeners. J Speech Hear Res 29:447-462,
paired listeners. In: Berglund B, Berglund U, Karlson 1986.
J, Lindwall T, eds: Noise as a Public Health Problem. Humes LE, Halling DC. Overview, rationale, and com-
Stockholm: Swedish Council for Building Research, parison of suprathreshold-based gain prescription
1988, pp. 241-246. methods. In: Valente M, ed: Strategies for Selecting
Hirsh IJ, Davis H, Silverman SR, Reynolds EG, Eldert E, and Verifying Hearing Aid Fittings. New York:
Benson RW. Development of materials for speech au- Thieme, 1994, pp. 19-37.
diometry. J Speech Hear Disord 17:321-337, 1952. Humes LE, Riker S. Evaluation of two clinical versions of
Hogan CA, Turner CW. High-frequency audibility: the articulation index. Ear Hear 13:406-409, 1992.
Benefits for hearing-impaired listeners. J Acoust Soc Humes LE, Roberts L. Speech-recognition difficulties of
Am 104:432-441, 1998. the hearing-impaired elderly: The contributions of
Hohmann V, Kollmeier B. The effect of multichannel dy- audibility. J Speech Hear Res 33:726-735, 1990.
namic compression on speech intelligibility. J Acoust Humes LE, Watson BU, Christensen LA, Cokely CG,
Soc Am 97:1191-1195, 1995. Halling DC, Lee L. Factors associated with individual
Holcomb LM, Nerbonne MA, Konkle DF. Articulation differences in clinical measures of speech recognition
index and hearing handicap. J Am Acad Audiol among the elderly. J Speech Hear Res 37:465-473,
11:224-229, 2000. 1994.
Hou Z, Thornton AR. A model to evaluate and maximize Jerger J, Jerger S, Oliver T, Pirozzolo F. Speech under-
hearing aid performance by integrating the articula- standing in the elderly. Ear Hear 10:79-89, 1989.
tion index across listening conditions. Ear Hear Kamm CA, Dirks DD, Bell TS. Speech recognition and the
15:106-112, 1994. articulation index for normal and hearing-impaired
Houtgast T, Steeneken HJM. Evaluation of speech trans- listeners. J Acoust Soc Am 77:281-288, 1985.
mission channels by using artificial signals. Acustica Killion M. The SIN report: Circuits haven't solved the
25:355-367, 1971. hearing-in-noise problem. Hear J 50(10):28-30, 32,
Houtgast T, Steeneken HJM. The modulation transfer 34, 1997.
function in room acoustics as a predictor of speech Killion M, Christensen L. The case of the missing dots:
intelligibility. Acustica 28:66-73, 1973. Al and SNR. Hear J 51(5):32-47, 1998.
Houtgast T, Steeneken HJM. Experimental verification Killion M, Schulerin R, Christensen L, Fabry D, Revit L,
of the STI: A valid basis for setting indoor noise level Niquette P, Chung K. Real-world performance of an
criteria. In Rossi G, ed: Proceedings of the Fourth ITE directional microphone. Hear J 51(4):24-26, 30,
International Congress on Noise as a Public Health 32, 35, 36, 38, 1998.
Problem. Turin, Italy: CSRA, 1983, pp. 477-487.
Houtgast T, Steeneken HJM. A review of the MTF con-
Killion MC. The noise problem: There's hope. Hear Inst
cept in room acoustics and its use for estimating
36(11):26, 28, 30, 32, 1985.
speech intelligibility in auditoria. J Acoust Soc Am Killion MC. New thinking on hearing in noise: A gener-
77:1069-1077, 1985. alized articulation index. Seminars in Hearing 23:57-
Houtgast T, Steeneken HJM, Plomp R. Predicting speech
75, 2002.
intelligibility in rooms from the modulation transfer Killion MC, Fikret-Pasa S. The 3 types of sensorineural
function: I. General room acoustics. Acustica 46:60- hearing loss: Loudness and intelligibility considera-
72, 1980. tions. Hear J 46(11):31-36, 1993.
Humes LE. An evaluation of several rationales for select- Killion MC, Mueller HG, Pavlovic CV, Humes LE. A is for
ing hearing aid gain. J Speech Hear Disord 51:272- audibility. Hear J 46(4):29-32, 1993.
281, 1986. Knudsen VO. The hearing of speech in auditoriums. J
Humes LE. Understanding the speech-understanding Acoust Soc Am 1:56-82, 1929.
problems of the hearing impaired. J Am Acad Audiol Kringlebotn M. A graphical method for calculating the
2:59-69, 1991. speech intelligibility index and measuring hearing
Humes LE. Some temporal factors affecting speech recog- disability from audiograms. Scand Audiol 28:151-
nition. In Studebaker GA, Hochberg I, eds: Acoustical 160, 1999.
126
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amiani et al. Methods and Applications of the Al
Kryter KD. Methods for the calculation of the articula- Murray N, Byrne D. Performance of hearing-impaired
tion index. J Acoust Soc Am 34:1689-1697, 1962a. and normal hearing listeners with various high-fre-
Kryter KD. Validation of the articulation index. J Acoust quency cutoffs in hearing aids. Aust J Audiol 8:21-
Soc Am 34:1698-1702, 1962b. 28, 1986.
Leavitt R, Flexor C. Speech degradation as measured by Musch H, Buus S. Using statistical decision theory to pre-
the rapid speech transmission index (RASTI). Ear dict speech inteligibility. I. Model structure. J Acoust
Hear 12:115-118, 1991. Soc Am 109:2896-2909, 2001a.
Leijon A, Eriksson-Mangold M, Beck-Karisen A. Preferred Musch H, Buus S. Using statistical decision theory to pre-
hearing aid gain and bass-cut of speech by hearing-im- dict speech intelligibility. II. Measurement and pre-
paired listeners. J Acoust Soc Am 77:281-288, 1984. diction of consonant-discrimination performance. J
Acoust Soc Am 109:2910-2920, 2001b.
Liberman MC, Dodds LW. Single-neuron labeling and
chronic cochlear pathology. III. Stereocilia damage Nabelek AK, Letowski TR, Tucker FM. Reverberant over-
and alterations of threshold tuning curves. Hear Res lap- and self-masking in consonant identification. J
16:55-74, 1984. Acoust Soc Am 86:1259-1265, 1989.
Licklider JCR, Miller GA. The perception of speech. In: Nabelek AK, Nabelek IV. Room acoustics and speech per-
Stevens SS, ed: Handbook of Experimental Psychol- ception. In: Katz J, ed: Handbook of Clinical
ogy. New York: John Wiley & Sons, Inc, 1951, pp. Audiology (4th Ed.). Baltimore: Williams & Wilkins,
1040-1074. 1994, pp. 624-637.
Lundeen C. Count-the-dot audiogram in perspective. Am Nabelek AK, Pickett JM. Monaural and binaural speech
J Audiol 5:57-58, 1996. perception through hearing aids under noise and re-
verberation with normal and hearing-impaired lis-
Magnusson L. Predicting the speech recognition perfor- teners. J Speech Hear Res 17:724-739, 1974.
mance of elderly individuals with sensorineural im-
pairment: A procedure based on the speech intelligi- Nabelek AK, Robinson PK. Monaural and binaural speech
bility index. Scan Audiol 25:215-222, 1996. perception in reverberation for listeners of various
ages. J Acoust Soc Am 71:1242-1248, 1982.
Magnusson L, Karlsson M, Leijon A. Predicted and mea-
sured speech recognition performance in noise with Neuman AC, Hochberg I. Children's perception of speech
linear amplification. Ear Hear 22:46-57, 2001. in reverberation. J Acoust Soc Am 73:2145-2149,
1983.
McCandless GA, Lyregarrd PE. Prescription of gain/out-
put (POGO) for hearing aids. Hear Inst 35(1):16-21, Pascoe DP. Clinical implications of nonverbal method of
1983. hearing aid selection and fitting. Seminars Speech
Hear Res 1:217-229, 1980.
McCarthy PA. Self-assessment inventories: They're not just
for aural rehab anymore. Hear J 45(6):41-43, 1994. Pavlovic CV. Use of the articulation index for assessing
residual auditory function in listeners with sen-
McCarthy PA. Self-assessment revisited. Hear J 51(3):10- sorineural hearing impairment. J Acoust Soc Am
18, 1998. 75:1253-1258, 1984.
Moncur JP, Dirks D. Binaural and monaural speech in- Pavlovic CV. Derivation of primary parameters and pro-
telligibility in reverberation. J Speech Hear Res cedures for use in speech intelligibility predictions. J
10:186-195, 1967. Acoust Soc Am 82:413-422, 1987.
Moore BCJ. Perceptual consequences of cochlear hear- Pavlovic CV. Articulation index predictions of speech in-
ing loss and their implications for the design of hear-
ing aids. Ear Hear 17:133-161, 1996. telligibility in hearing aid selection. Asha 30
June-July:63-65, 1988.
Moore BCJ. A test for the diagnosis of dead regions in
the cochlea. Brit J Audiol 34:205-224, 2000. Pavlovic CV. Speech spectrum considerations and speech
intelligibility predictions in hearing aid evaluations. J
Moore BCJ. Dead regions in the cochlea: Diagnosis, per- Speech Hear Disord 54:3-8, 1989.
ceptual consequences, and implications for the fitting
of hearing aids. Trends Ampl 5:1-34, 2001. Pavlovic CV. Speech recognition and five articulation in-
dexes. Hear Inst 42(9):20, 22, 23, 1991.
Moore BCJ, Alcaintara JI. The use of psychophysical tun-
ing curves to explore dead regions in the cochlea. Ear Pavlovic CV. Band importance functions for audiological
Hear 22:268-278, 2001. applications. Ear Hear 15:100-104, 1994.
Mueller HG. Insertion gain measurements. In: Mueller Pavlovic CV, Studebaker GA. Evaluation of some as-
HG, Hawkins DB, Northern JL, eds: Probe Micro- sumptions underlying the articulation index. J Acoust
phone Measurements: Hearing Aid Selection and Soc Am 75:1606-1612, 1984.
Assessment. San Diego: Singular Publishing, 1992, Pavlovic CV, Studebaker GA, Sherbecoe RL. An articula-
pp. 113-144. tion index based procedure for predicting the speech
Mueller HG, Killion MC. An easy method for calculating recognition performance of hearing-impaired indi-
the articulation index. Hear J 43(9):14-17, 1990. viduals. J Acoust Soc Am 80:50-57, 1986.
127
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Trends In Amplification Volume 6, Number 3, 2002
Payton KL, Uchanski RM, Braida LD. Intelligibility of con- Skinner MW. Hearing Aid Evaluation. Englewood Cliffs,
versational and clear speech in noise and reverbera- New Jersey: Prentice Hall, 1988.
tion for listeners with normal and impaired hearing. Souza PE, Bishop RD. Improving audibility with nonlin-
J Acoust Soc Am 95:1581-1592, 1994. ear amplification for listeners with high-frequency
Pearson KS, Bennett RL, Fidell S. Speech Levels in loss. J Am Acad Audiol 11:214-223, 2000.
Various Environments (BBN Report No. 3281). Souza PE, Kitch VJ. Effect of preferred volume setting on
Cambridge, MA: Bolt, Beranek and Newman, 1976. speech audibility in different hearing aid circuits. J
Am Acad Audiol 12:415-422, 2001.
Pekkarinen E, Viljanen V. Acoustic conditions for speech
communication in classrooms. Scand Audiol 20:257- Souza PE, Turner CW. Multichannel compression, tem-
263, 1991. poral cues and audibility. J Speech Hear Res 41:315-
326, 1998.
Pluvinage V. Rationale and development of the Resound
system. In: Sandlin RE, ed: Understanding Digitally Souza PE, Turner CW. Quantifying the contribution of
Programmable Hearing Aids. Boston: Allyn and audibility to recognition of compressed-amplified
speech. Ear Hear 20:12-20, 1999.
Bacon, 1994, pp. 15-39.
Souza PE, Yueh B, Sarubbi M, Loovis CF. Fitting hearing
Popelka GR, Mason D. Factors which affect the measures aids with the articulation index: Impact on hearing
of speech audibility with hearing aids. Ear Hear aid effectiveness. J Rehabil R D 37:473-481, 2000.
8(5S):109-118, 1987. Steele JA, Binnie CA, Cooper WA. Combining auditory
Plomp R, Mimpen AM. Speech reception threshold for and visual stimuli in the adaptive testing of speech
sentences as a function of age and noise level. J discrimination. J Speech Hear Disord 43:115-122,
Acoust Soc Am 66:1333-1342, 1979. 1978.
Plomp R. The negative effect of amplitude compression Steeneken HJM, Houtgast T. A physical method for mea-
suring speech-transmission quality. J Acoust Soc Am
in multichannel hearing aids in the light of the mod- 67:318-326, 1980.
ulation-transfer function. J Acoust Soc Am 83:2322-
2327, 1988. Steeneken HJM, Houtgast T. Mutual dependence of oc-
tave-band weights in predicting speech intelligibili-
Preves D. Directional microphone use in ITE hearing in- ty. Speech Commun 28:109-123, 1980.
struments. Hear Rev 4(7):21, 22, 24-27, 1997. Stelmachowicz P, Lewis D, Kalberer A, Creutz T.
Rankovic CM. An application of the articulation index to Situational Hearing-Aid Response Profile for Users
hearing aid fitting. J Speech Hear Res 34:391-402, (SHARP, v 2.0). Omaha, Nebraska: Boy's Town
1991. National Research Hospital, 1994.
Stelmachowicz PG, Dalzell S, Petersen D, Kopun J, Lewis
Resnick SB, Dubno JR, Hoffnung S, Levitt H. Phoneme DL, Hoover BE. A comparison of threshold-based fit-
errors on a nonsense syllable test. J Acoust Soc Am ting strategies for nonlinear hearing aids. Ear Hear
58(Suppl):S114, 1975. 19:131-138, 1998.
Ricketts T, Mueller HG. Making sense of directional mi- Studebaker GA. A 'rationalized' arcsine transform. J
crophone hearing aids. Am J Audiol 8:117-127, Speech Hear Res 28:455-462, 1985.
1999. Studebaker GA. The effect of equating loudness on audi-
Ricketts TA, Dittberner AB. Directional amplification for bility-based hearing aid selection procedures. J Am
improved signal-to-noise ratio: Strategies, measure- Acad Audiol 3:113-118, 1992.
ment, and limitations. In: Valente M, ed: Strategies Studebaker GA, McDaniel DM, Sherbecoe RL. Evaluating
for Selecting and Verifying Hearing Aid Fittings (2nd relative speech recognition performance using the
ed). New York: Thieme Medical Publishers, 2002, proficiency factor and rationalized arcsine differ-
pp. 274-346. ences. J Am Acad Audiol 6:173-182, 1995.
Schow RL, Nerbonne MA. Communication screening pro- Studebaker GA, Pavlovic CV, Sherbecoe RL. A frequency
file: Use with elderly clients. Ear Hear 3:135-147, importance function for continuous discourse. J
1982. Acoust Soc Am 81:1130-1138, 1987.
Schum DJ, Matthews LJ, Lee F. Actual and predicted Studebaker GA, Sherbecoe RL. Frequency-importance
and transfer functions for recorded CID W-22 word
word recognition performance of elderly hearing-im- lists. J Speech Hear Res 34:427-438, 1991.
paired listeners. J Speech Hear Res 34:636-642,
1991. Studebaker GA, Sherbecoe RL. Frequency-importance
functions for speech recognition. In: Studebaker GA,
Skinner MW. Speech intelligibility in noise in noise-in- Hochberg I, eds: Acoustical Factors Affecting Hearing
duced hearing loss: Effects of high frequency com- Aid Performance (2nd ed). Boston: Allyn and Bacon,
pensation. J Acoust Soc Am 67:306-317, 1980. 1993, pp. 185-204.
128
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.
Amlani et al. Methods and Applications of the Al
Studebaker GA, Sherbecoe RL. Intensity-importance Valente M. Use of microphone technology to improve
functions for bandlimited monosyllabic words. J user performance in noise. Trends in Amplification
Acoust Soc Am 111: 1422-1436, 2002. 4:112-135, 1999.
Studebaker GA, Sherbecoe RL, Gilmore C. Frequency-im- Valente M. Use of microphone technology to improve
portance and transfer functions for the Auditec of user performance in noise. In: Sandlin RE, ed:
Saint Louis recording of the NU-6 word test. J Speech Textbook of Hearing Aid Amplification: Technical
Hear Res 36:799-807, 1993. and Clinical Considerations (2nd ed). San Diego:
Singular Thomason Learning, 2000, pp. 247-284.
Studebaker GA, Sherbecoe RL. Evaluating the Speech
Recognition Performance of Hearing-Impaired van Rooij JCGM, Plomp R. The effect of linguistic en-
Subjects. Paper presented at the Lake Arrowhead tropy on speech perception in noise in young and el-
International Conference on Issues on Advanced derly listeners. J Acoust Soc Am 90:2985-2991, 1991.
Hearing Aid Research, Lake Arrowhead, California, Van Tasell DJ. Hearing loss, speech, and hearing aids. J
May 30-June 4, 1994. Speech Hear Res 36:228-244, 1993.
Studebaker GA, Sherbecoe RL, McDaniel DM, Gwaltney Villchur E. Comments on "The negative effect of ampli-
CA. Monosyllabic word recognition at higher-than- tude compression on multichannel hearing aids in
normal speech and noise levels. J Acoust Soc Am the light of the modulation transfer function"
105:2431-2444, 1999. (Plomp. J Acoust Soc Am 83:2322-2327, 1988).
J Acoust Soc Am 86:425-427, 1989.
Studebaker GA, Sherbecoe RL, McDaniel DM, Gray GA.
Age-related changes in monosyllabic word recogni- Waring CE. Transformation and Reception of Sounds
tion when audibility is held constant. J Am Acad Under Combat Conditions. Office of Scientific
Audiol 8:150-162, 1997. Research and Development #TIPU 1130. Washington
DC: Office of Naval Research, 1946.
Sumby WH, Pollack I. Visual contributions to speech in-
telligibility in noise. J Acoust Soc Am 26:212-215, Webster J. Interpretations of speech and noise charac-
1954. teristics of NTID learning centers. J Acoust Soc Am
66(Suppl):S37, 1979.
Tillman TW, Carhart R. An Expanded Test for Speech Wilde G, Humes LE. Applications of the articulation
Discrimination Utilizing CNC Monosyllabic Words. index to the speech recognition of normal and im-
Northwestern University Auditory Test No.6. (USAF paired listeners wearing hearing protection. J Acoust
School of Aerospace Medicine Technical Report). Soc Am 87:1192-1199, 1990.
Brooks Air Force Base, Texas, 1966.
Willot JF. Aging and the Auditory System: Anatomy,
Turner CW, Cummings KJ. Speech audibility for listeners Physiology, and Psychophysics. San Diego: Singular
with high-frequency hearing loss. Am J Audiol 8:150- Publishing, Inc, 1991.
162, 1999.
Zurek PM, Delhorne LA. Consonant reception in noise by
Turner CW, Robb MP. Audibility and recognition of stop listeners with mild and moderate sensorineural hear-
consonants in normal and hearing-impaired subjects. ing impairment. J Acoust Soc Am 82:1548-1559,
J Acoust Soc Am 81:1566-1573, 1987. 1987.
129
Downloaded from http://tia.sagepub.com at PENNSYLVANIA STATE UNIV on April 16, 2008
2002 SAGE Publications. All rights reserved. Not for commercial use or unauthorized distribution.