Professional Documents
Culture Documents
Volume 12 Number 4
December 2008 316-331
Pitch is important for speech and music perception, and less access to individual harmonics, and the coding
may also play a crucial role in our ability to segregate schemes currently employed in cochlear implants pro-
sounds that arrive from different sources. This article vide little or no representation of individual harmonics.
reviews some basic aspects of pitch coding in the normal These deficits in the coding of harmonic sounds may
auditory system and explores the implications for pitch underlie some of the difficulties experienced by people
perception in people with hearing impairments and with hearing loss and cochlear implants, and may point
cochlear implants. Data from normal-hearing listeners to future areas where sound representation in auditory
suggest that the low-frequency, low-numbered harmon- prostheses could be improved.
ics within complex tones are of prime importance in
pitch perception and in the perceptual segregation of
competing sounds. The poorer frequency selectivity Keywords: Pitch; Hearing Loss; Cochlear Implants;
experienced by many hearing-impaired listeners leads to Auditory Streaming
P
eople with normal hearing can pick out the vio- simultaneous sounds, such as onset and offset asyn-
lin melody in an orchestral piece, or listen to chrony (whether components start and stop at the
someone speaking next to them in a crowded same time) and harmonicity (whether components
room, with relative ease. Despite the fact that we share the same fundamental frequency or F0); sounds
achieve these tasks with little or no effort, we know sur- that start or stop at the same time and share the same
prisingly little about how the brain and the ears achieve F0 are more likely to be heard as a single object. The
them. The process by which successive sounds from other stage is to bind together across time those sound
one source (such as a violin or a person talking) elements that belong to the same source and to avoid
are perceptually grouped together and separated from combining sequential sound elements that do not—a
other competing sounds is known as stream segrega- process known as sequential grouping or segregation.
tion, or simply “streaming” (Bregman, 1990; Darwin & Again, a number of different acoustic cues play a role,
Carlyon, 1995). Streaming involves at least two stages, such as the temporal proximity of successive sounds, as
which in some cases can influence and interact with well as their similarity in F0 and spectrum, with more
each other. One stage is to bind together across fre- similar and more proximal sounds being more likely to
quency those elements emitted from one source and to be heard as belonging to a single stream. Beyond these
separate them from other sounds that are present at acoustic, or bottom-up, cues, there are many top-down
the same time. This process is known as simultaneous influences on perceptual organization, including atten-
grouping or segregation. There are a number of acoustic tion, expectations, and prior exposure, all of which can
cues that influence the perceptual organization of also play an important role (e.g., Marrone, Mason, &
Kidd, in press; Shinn-Cunningham & Best, in press).
Returning to acoustic cues, F0 and its perceptual
From University of Minnesota, Minneapolis, Minnesota. correlate—pitch—have been shown to influence both
sequential and simultaneous grouping; in fact, poorer
Address correspondence to: Andrew J Oxenham, PhD, Department
of Psychology, University of Minnesota, 75 E. River Parkway, N218, pitch perception in people with hearing impairment
Minneapolis, MN 55455; e-mail: oxenham@umn.edu. and users of cochlear implants may underlie some of
316
Pitch Perception and Auditory Stream Segregation / Oxenham 317
via preformed “harmonic templates” (e.g., Goldstein, and more accurate pitch than high-numbered harmon-
1973). ics, suggesting that the place or time information pro-
Another representation to consider involves the vided by resolved harmonics is more important than
temporal waveforms at the output of the auditory fil- the temporal envelope information provided by unre-
ters. Some examples are shown in the bottom panel of solved harmonics (Bernstein & Oxenham, 2003; Dai,
Figure 3. For the low-numbered resolved harmonics, 2000; Houtsma & Smurzynski, 1990; Moore,
filters centered at or near one of the harmonic fre- Glasberg, & Peters, 1985; Plomp, 1967; Ritsma, 1967;
quencies respond primarily to that single component, Shackleton & Carlyon, 1994). Nevertheless, some
producing a response that approximates a pure tone. pitch information is conveyed via the temporal enve-
As the harmonics become more unresolved, they inter- lope of unresolved harmonics (Kaernbach & Bering,
act within each filter producing a complex waveform 2001), as well as temporal-envelope fluctuations in
with a temporal envelope that repeats at a rate corre- random noise (Burns & Viemeister, 1981), suggesting
sponding to the F0. Temporal models of pitch use dif- that at least some aspects of pitch are derived from the
ferent aspects of these temporal waveforms to extract timing information in the auditory nerve, and not solely
the pitch. The earliest temporal theories were based place information. It is interesting to note that, when
on the regular time intervals between high-amplitude only unresolved harmonics are present, the phase rela-
peaks of the temporal fine structure in the waveform tionships between the harmonics affect pitch percep-
produced by unresolved harmonics (de Boer, 1956; tion discrimination, as one might expect if performance
Schouten, Ritsma, & Cardozo, 1962). Other temporal were based on the temporal envelope. The filter out-
models use the timing information from resolved har- puts shown in the bottom panel of Figure 3 illustrate
monics (Srulovicz & Goldstein, 1983) to extract the the well-defined and temporally compact envelope that
F0, but perhaps the most influential model, based on is produced when components are in sine phase with
the autocorrelation function, extracts periodicities one another. When the phase relationships are
from the temporal fine structure of resolved harmon- changed, the envelope can become much less modu-
ics, and the temporal envelope from high-frequency lated, which usually leads to poorer pitch perception
unresolved harmonics, by pooling temporal informa- and discrimination, but only when purely unresolved
tion from all channels to derive the most dominant harmonics are present in the stimulus (Bernstein &
periodicity, which in most cases corresponds to the F0 Oxenham, 2006a; Houtsma & Smurzynski, 1990;
(Cariani & Delgutte, 1996; Licklider, 1951; Meddis & Shackleton & Carlyon, 1994).
Hewitt, 1991a, 1991b; Meddis & O’Mard, 1997). Despite the ability of temporal models to explain
Just as with pure tones, complex-tone pitch percep- many aspects of pitch perception, timing does not
tion has been explained with a place–time model, in seem to be the whole story either: When the tempo-
addition to place or time models (Shamma & Klein, ral information that would normally be available via
2000), and recent physiological studies have explored low-frequency resolved harmonics is presented to a
this possibility (Cedolin & Delgutte, 2007; Larsen, high-frequency region of the cochlea, it does not
Cedolin, & Delgutte, 2008). However, no studies have produce a pitch of the missing fundamental, sug-
yet been able to definitively rule out or confirm any par- gesting that place information may be of importance
ticular model in the auditory periphery. More centrally, (Oxenham, Bernstein, & Penagos, 2004). Along the
Bendor and Wang (2005, 2006) have identified neu- same lines, auditory frequency selectivity and the
rons in the auditory cortex of the marmoset, which resolvability of harmonics can predict pitch discrim-
seem to respond selectively to certain F0s, as opposed ination accuracy to some extent, suggesting that
to frequencies or spectral regions, and roughly homol- peripheral filtering is important in understanding
ogous regions have been identified in humans from pitch coding (Bernstein & Oxenham, 2006a,
imaging studies (Penagos, Melcher, & Oxenham, 2006b). On the other hand, artificially improving
2004). However, these higher-level responses do not the resolvability of harmonics by presenting alter-
provide much of a clue regarding the peripheral code nating harmonics to opposite ears does not lead to
that produces the cortical responses. an improvement in pitch perception (Bernstein &
What can data from human perception experi- Oxenham, 2003). It may be that, in normal hearing,
ments tell us about the validity of the various models? all the information in the auditory periphery (place,
At present, there are no conclusive arguments in favor time, and place–time) is used to some degree to
of any one theory over others. Many studies have found increase the redundancy, and hence, the robustness
that low-numbered harmonics produce a more salient of the pitch code.
Pitch Perception and Auditory Stream Segregation / Oxenham 321
Pitch Perception of Multiple Sounds & Grose, 2004) have led to the conclusion that at least
some hearing-impaired listeners may suffer from a
Surprisingly little research has been done on the pitch
deficit in temporal coding. An extreme example of
perception of multiple simultaneous harmonic complex
deficits in temporal coding seem to be exhibited
tones. The little work that exists also points to a poten-
by subjects diagnosed with so-called auditory neuropa-
tially important role for resolved harmonics. Beerends
thy, based on normal otoacoustic emissions but
and Houtsma (1989) presented listeners with two
abnormal or absent auditory brain-stem responses
simultaneous complex tones, with each complex com-
(Sininger & Starr, 2001). Both low-frequency
prised of only two adjacent harmonics, giving a total of
pure-tone frequency discrimination and modula-
four components. They found that listeners were able to
tion perception are typically highly impaired in
identify the pitches of both complexes if at least one of
people diagnosed with auditory neuropathy (Zeng,
the four components was resolved. Carlyon (1996)
Kong, Michalewski, & Starr, 2005).
found that when two complexes, each containing many
When it came to complex tones, Moore and Peters
harmonics, were in the same spectral region and were
(1992) also found quite variable results, with one inter-
presented at the same overall level, it was only possible
esting exception: No listeners with broader-than-
to hear the pitch of one when the complexes contained
normal auditory filters showed normal F0 discrimination
some resolved harmonics; when the stimuli were fil-
of complex tones. Bernstein and Oxenham (2006b)
tered so that only unresolved harmonics were present,
also investigated the link between auditory filter band-
the mixture sounded more like a crackle than two
widths and complex pitch perception. They specifically
pitches, suggesting that listeners are not able to extract
targeted the abrupt transition from good to poor pitch
two periodicities from a single temporal envelope.
perception that is typically observed in normal-hearing
listeners as the lower (resolved) harmonics are succes-
sively removed from a harmonic complex tone
Pitch Perception in Hearing Impairment (Houtsma & Smurzynski, 1990). The main prediction
Pitch perception—both with pure tones and complex was that poorer frequency selectivity should lead to
tones—is often poorer than normal in listeners with fewer resolved harmonics. If resolved harmonics are
hearing impairment but, as with many auditory capa- important for good pitch perception, then pitch percep-
bilities, there is large variability between different lis- tion should deteriorate more rapidly in people with
teners (Moore & Carlyon, 2005). Moore and Peters broader auditory filters as the lowest harmonics in each
(1992) measured pitch discrimination in young and old stimulus are successively removed. In other words, the
normal-hearing and hearing-impaired listeners using point at which there are no resolved harmonics left in
pure tones and complex tones filtered to contain the the stimulus should be reached more quickly with
first 12 harmonics, only harmonics 1-5, or harmonics hearing-impaired listeners than with normal-hear-
4-12 or 6-12. They found a wide range of performance ing listeners, and that should be reflected in the
among their hearing-impaired listeners, with some pitch percept elicited by the stimuli. Bernstein and
showing near-normal ability to discriminate small Oxenham found a significant correlation between
differences in the frequency of pure tones, whereas the lowest harmonic number present at the point
others were severely impaired. They also found poorer- where the F0 difference limens (F0DLs) became
than-normal frequency discrimination in their older markedly poorer and the bandwidth of the auditory
listeners with near-normal hearing, suggesting a possi- filters in individual hearing-impaired subjects. The
ble age effect that is independent of hearing loss. point at which F0DLs became poorer also correlated
Hearing impairment is often accompanied by poorer well with the point at which the phase relations
frequency selectivity, in the form of broader auditory between harmonics started to have an effect on
filters. According to the place theory of pitch, poorer F0DLs. Remember that phase effects are thought to
frequency selectivity should lead to poorer pitch per- occur only when harmonics are unresolved and
ception and larger (worse) frequency difference limens interact within single auditory filters. Overall, the
(FDLs). Interestingly, however, FDLs were only weakly data were consistent with the idea that peripheral
correlated with auditory filter bandwidths (Glasberg & frequency selectivity affects the pitch perception of
Moore, 1990; Patterson, 1976), suggesting perhaps complex tones and that hearing-impaired listeners
that place coding cannot fully account for pure-tone with poorer frequency selectivity suffer from
pitch perception. These and other results (Buss, Hall, poorer complex pitch perception.
322 Trends in Amplification / Vol. 12, No. 4, December 2008
basic idea is that envelope fluctuations are impor- have an F0 of 100 Hz in the low-frequency region,
tant for speech intelligibility (Shannon, Zeng, but an F0 of 106 Hz in the high-frequency region,
Kamath, Wygonski, & Ekelid, 1995), but pitch may whereas the interferer might have an F0 of 106 Hz in
help decide which fluctuations belong to the target both regions. In this example, there is an F0 differ-
speech and which belong to the interferer (Qin & ence between the target and interferer in the low-
Oxenham, 2003). frequency region, but not in the high-frequency
region. In this way, it was possible to determine
Pitch and Source Segregation which frequency region was most important in provid-
ing a benefit of F0 difference between the target and
in Normal Hearing
interferer. The 800-Hz crossover frequency ensured
The effect of F0 differences between simultaneous that resolved harmonics were present in the lower
sources has been examined most thoroughly using region, but did not rule out the possibility that some
pairs of simultaneously presented synthetic vowels resolved harmonics were present in the higher region.
(Assmann & Summerfield, 1987; de Cheveigné, 1997; Bird and Darwin’s (1998) results showed that
Parsons, 1976; Scheffers, 1983). The task of listeners the low region was dominant in determining
is to identify both vowels present in the mixture. The whether a benefit of F0 difference between the tar-
general finding is that performance improves as the F0 get and interferer was observed: When F0 differ-
separation between the two vowels increases, in line ences were introduced only in the high region,
with expectations from F0-based perceptual segrega- listeners did not benefit from them; when the F0 dif-
tion. However, a number of studies have pointed out ferences were present only in the low region, the
other potential cues that enable listeners to hear out benefit was almost as great as when the F0 differ-
the two vowels, which are not directly related to per- ences were present in both spectral regions.
ceived pitch differences between the two vowels, such However, their results do not imply that F0 informa-
as temporal envelope fluctuations (or beating) caused tion in the high region had no effect. In fact, when
by closely spaced harmonics (de Cheveigné, 1999; the F0s in the low and high regions were swapped
Summerfield & Assmann, 1991). (so that the F0 of target in the low region was the
The influence of F0 differences between voiced same as that of the interferer in the high region, and
segments of speech has also been examined using vice versa), performance deteriorated for F0 differ-
longer samples, such as sentences. In one of the ear- ences of 5 semitones or more, suggesting that in
liest such studies, Brokx and Nooteboom (1982) normal situations F0 differences and similarities
measured sentence intelligibility in the presence of across frequency are used to segregate or group
competing speech, spoken by the same talker and components together, at least when the F0 differ-
resynthesized on a monotone. They found that the ences are large. It is not clear whether the same
number of errors in reporting the monotone target results would have been observed if the high region
speech decreased steadily as the F0 difference had been further restricted to contain only unre-
between the target and interfering speech was solved harmonics, which are known to produce a
increased from 0 to 3 semitones. A similar benefit of weaker pitch percept, and so may contribute less to
an overall F0 difference between the target and inter- grouping by F0.
ferer was also observed when the speech contained For nonspeech stimuli, the role of pitch in sequen-
the F0 fluctuations of normal intonation. More tial streaming has been long established. Some early
recently, Bird and Darwin (1998) revisited this issue studies used repeating sequences of two pure tones,
and examined the relative contributions of F0 differ- termed A and B. The sequences were created either by
ences in low- and high-frequency regions. In their presenting the A and B tones in alternation (ABAB;
control conditions, they too found that increasing Miller & Heise, 1950) or in triplets (ABA ABA; van
the F0 difference between two monotone utterances Noorden, 1975). When the frequency separation
improved the intelligibility of the target, out to at between the two tones is small, listeners typically hear
least 10 semitones. In their second experiment, the a single stream of alternating tones (Figure 4, upper
target and interfering speech were divided into low- panel). When the frequency separation is larger, the
pass and high-pass segments, with a crossover fre- percept changes, with the two tones splitting into two
quency of 800 Hz. The F0s of the target and separate perceptual streams (Figure 4, lower panel).
interferer were manipulated independently in the The exact frequency separation at which the percept
two spectral regions. For instance, the target might changes from one to two streams depends on many
324 Trends in Amplification / Vol. 12, No. 4, December 2008
perceive two streams at small frequency differences may aspects of cochlear-implant processing in normal-
correlate with the ability to make use of F0 differences hearing listeners, to investigate the link between F0
between the male and female talkers. perception and the segregation of simultaneous vowels
Finally, just as in normal-hearing listeners (Qin & Oxenham, 2005). In the noise-vocoder tech-
(Roberts, Glasberg, & Moore, 2002; Vliegen et al., nique, sound is split up into a number of contiguous
1999; Vliegen & Oxenham, 1999), auditory stream frequency channels, and the temporal envelope is
segregation has been found to occur in hearing- extracted from each channel, as in a cochlear implant.
impaired listeners for stimuli that do not differ in However, instead of using the channel envelopes to
their spectral envelope but do differ in their tempo- modulate electrical pulses, the envelopes are used to
ral envelope (Stainsby, Moore, & Glasberg, 2004). modulate the amplitudes of band-limited noise bands,
This suggests that hearing-impaired listeners are each with the same center frequency and bandwidth
able to use the temporal envelope pitch cues associ- as the original channel filter. After modulation, the
ated with unresolved harmonics in auditory stream noise bands are added together and presented acousti-
segregation. cally to normal-hearing listeners. In this way, listeners
Because of the reduced availability and poorer are presented with an approximation of the original
coding of resolved harmonics, hearing-impaired peo- temporal envelopes, but with the spectral resolution
ple may well rely more on the temporal-envelope degraded (depending on the number and bandwidths
cues for pitch perception than do people with normal of the channels) and the original temporal fine struc-
hearing. Although the evidence suggests that tempo- ture replaced by the fine structure of the noise bands.
ral-envelope coding is not necessarily impaired in The presentation of temporal-envelope information,
hearing-impaired listeners, the reliance on such cues without the original temporal fine structure, approxi-
may negatively impact hearing-impaired people’s mates the processing used in most current cochlear
ability to segregate sounds in natural environments implants.
for a number of reasons. First, temporal-envelope Qin and Oxenham (2005) found that F0 discrimi-
cues provide a weaker pitch cue than resolved har- nation was considerably poorer with harmonic com-
monics, even under ideal conditions. Second, in plex tones that were processed through an envelope
real situations, such as in rooms or halls, reverber- vocoder. More importantly, performance was worsened
ation smears rapidly fluctuating temporal-envelope further still by adding reverberation that approximated
pitch cues, meaning that performance will be worse either a classroom or a large hall, in line with expecta-
than measured under anechoic laboratory condi- tions based on the smearing effects of reverberation on
tions (Qin & Oxenham, 2005). Third, as mentioned the temporal envelope (Houtgast, Steeneken, &
above, normal-hearing listeners do not seem able to Plomp, 1980). However, in the absence of reverbera-
extract multiple pitches when two stimuli containing tion, or when the number of channels was relatively
unresolved harmonics are presented to the same spec- large, F0DLs were still less than 6%, corresponding to
tral region (Carlyon, 1996; Micheyl, Bernstein, & one musical semitone. Is such sensitivity enough to
Oxenham, 2006), suggesting that hearing-impaired allow the use of F0 cues in segregating competing
people may experience problems hearing more than speech sounds? Qin and Oxenham tested this ques-
one pitch at a time if they are relying solely on unre- tion by measuring the ability of normal-hearing lis-
solved harmonics. This presumed inability to extract teners to identify of pairs of simultaneous vowels
multiple pitches (which has not been directly tested with or without envelope-vocoder processing. They
yet), may explain at least some of the difficulties found that no benefits were found for increasing the
experienced in complex environments by people F0 difference between pairs of vowels, even with 24
with hearing impairment. This question is addressed channels of spectral information (far greater resolu-
further in the section below on masking release. tion than that found in current cochlear implants).
This suggests that even in the best situations (with
Findings and Implications high spectral resolution and no reverberation), tem-
poral envelope cues may not be sufficient to allow
for Cochlear Implant Users
the perceptual segregation of simultaneous sounds
To our knowledge there have been no studies that have based on F0 cues.
tested double-vowel recognition in cochlear-implants Similar conclusions have been reached using more
users. One study has, however, used a noise-excited natural speech stimuli in actual cochlear-implant
envelope-vocoder technique, which simulates certain users, as well as envelope-vocoder simulations. Deeks
326 Trends in Amplification / Vol. 12, No. 4, December 2008
and Carlyon (2004) used a six-channel envelope Others showed a very different pattern of results, with
vocoder. Rather than modulating noise bands with the performance hardly being affected by the frequency
speech envelopes, they modulated periodic pulse trains separation between tones. This pattern of results is par-
to more closely simulate cochlear-implant processing. ticularly interesting, because the implant users actually
When two competing speech stimuli were presented, performed relatively better than normal when the tones
performance was better when the target was presented were widely separated in frequency. Thus, it appears
at the higher of two pulse rates (140 Hz, rather than 80 that at least some implant listeners experience stream-
Hz), but there was no evidence that a difference in ing, at least with pure tones, whereas others may not.
pulse rates between the masker and target improved The difference may relate to differences in effective fre-
the ability of listeners to segregate the two, whether the quency discrimination—if a frequency difference
target and masker were presented in all six channels, or between two tones is not perceived, then no streaming
whether the masker and target were presented in alter- would be expected. Unfortunately, Hong and Turner
nating channels. Similarly, Stickney et al. (2007) found did not measure frequency discrimination in their sub-
that F0 differences between competing sentences jects. However, they did find a correlation between
did not improve performance in real and simulated speech perception in noise or multitalker babble and
cochlear-implant processing. Interestingly, they found performance on their streaming task, with better
that reintroducing some temporal fine structure infor- speech perception being related to stream segregation
mation into the vocoder-processed stimuli led to results that were closer to normal. To our knowledge,
improved performance overall, as well as a benefit for no streaming studies have been carried out with
F0 differences between the competing sentences. acoustic stimuli other than pure tones, so it is not
Restoring temporal fine structure has the effect of rein- clear to what extent F0 differences in everyday com-
troducing some of the resolved harmonics. Although plex stimuli (such as voiced speech or music) lead to
the results are encouraging, it is far from clear how stream segregation in cochlear-implant users.
these resolved harmonics (or temporal fine structure in
general) can be best presented to real cochlear-implant Masking Release in Normal, Impaired,
users (Sit, Simonson, Oxenham, Faltys, & Sarpeshkar,
and Electric Hearing
2007).
A few studies have now investigated perceptual The phenomenon of “masking release” in speech per-
organization in cochlear-implant users using non- ception has been attributed, at least in part, to suc-
speech stimuli. Using repeating sequences of ABA cessful source segregation. Masking release is defined
triplets, Chatterjee, Sarampalis, and Oba (2006) found as the improvement in performance when a stationary
some, albeit weak, evidence that differences in elec- (steady-state) noise masker is replaced by a different
trode location between A and B, as well as differences (usually modulated) masker. In normal-hearing lis-
in temporal envelope, may be used by cochlear-implant teners, performance almost always improves when
users to segregate sequences of nonoverlapping sounds. temporal and/or spectral fluctuations are introduced
In contrast, Cooper and Roberts (2007) suggested that into a masking stimulus, either by modulating the
self-reported stream segregation in cochlear-implant noise or by using a single talker as the masker (Festen
users may in fact just reflect their ability to discrimi- & Plomp, 1990; Peters, Moore, & Baer, 1998). The
nate differences in pitch between electrodes, rather ability to make use of brief periods of low masker
than the automatic streaming experienced by normal- energy in a temporally fluctuating masker has been
hearing listeners when alternating tones of very differ- termed “listening in the dips” or “glimpsing.” To make
ent frequencies are presented at high rates. To use of these dips, it is necessary to know which fluc-
overcome the potential limitations of subjective reports tuations belong to the masker and which belong to
of perception, Hong and Turner (2006) used a measure the target. This is where source segregation comes
of temporal discrimination that becomes more difficult into play.
when the tones fall into different perceptual streams. Hearing-impaired listeners generally show much
They found a wide range of performance among their less masking release than normal-hearing listeners
cochlear-implant users. Some showed near-normal (Festen & Plomp, 1990; Peters et al., 1998). This
performance, with thresholds in the discrimination has often been ascribed to reduced spectral and/
task increasing (indicating worse performance) as or temporal resolution in hearing-impaired listen-
the frequency separation between tones increased. ers, leading to lower audibility in the spectral and
Pitch Perception and Auditory Stream Segregation / Oxenham 327
temporal dips of the masker. An alternative hypoth- represent just one facet of the multiple forms of
esis is that the problem lies not with audibility, but information, including other bottom-up and top-
with a failure to segregate the target information down segregation mechanisms, that are used by nor-
from that of the masker because of poorer F0 per- mal-hearing people when listening to speech in
ception (Summers & Leek, 1998) and/or a reduction complex backgrounds.
in the ability to process temporal fine structure cues
(Lorenzi, Gilbert, Carn, Garnier, & Moore, 2006). A
similar explanation could underlie the general fail-
Summary
ure of cochlear-implant users to show masking
A number of theories have been proposed to explain
release; in fact, studies of cochlear-implant users
pitch perception in normal hearing. The three main
(Nelson & Jin, 2004; Nelson, Jin, Carney, & Nelson,
strands can be described as place, time, and place–time
2003; Stickney, Zeng, Litovsky, & Assmann, 2004)
theories. So far, neither experimental nor modeling
and envelope-vocoder simulations (Qin & Oxenham,
studies have been able to completely rule out any of
2003) have shown that in many cases, performance
these potential approaches. Peripheral frequency selec-
is actually poorer in a fluctuating background than
tivity has been shown to be linked to aspects of pitch
in a stationary noise, perhaps because the masker
perception, in line with the expectations of place-based
fluctuations mask those of the target speech (Kwon
models; however, there are many other aspects of pitch
& Turner, 2001).
perception that are better understood in terms of tim-
The idea that accurate F0 information from the
ing-based mechanisms. The question is of more than
temporal fine structure of resolved harmonics
just theoretical value—understanding what aspects of
underlies masking release in normal-hearing listen-
the peripheral auditory code are used by the brain to
ers was tested by Oxenham and Simonson (2009).
extract pitch could help in selecting the optimum pro-
The rationale was that high-pass filtering the target
cessing strategy for cochlear implants.
and masker to remove all resolved harmonic compo-
Relatively little is known about our ability to per-
nents should lead to a greater reduction in masking
ceive multiple pitches at once, even though this skill
release than low-pass filtering the same stimuli,
is clearly important in music appreciation, and may
which would reduce some speech information but
also play a vital role in the perceptual segregation of
would retain most of the information important for
competing sounds, such as two people speaking at
F0 perception. The cutoff frequencies for the high-
once. Based on the finding that resolved harmonics
pass and low-pass filters were selected to produce
seem to be important in hearing more than one
equal performance (proportion of sentence words
pitch from within a single spectral region, it seems
correctly reported) in the presence of stationary
likely that both hearing-impaired people and people
noise. In other conditions, the stationary noise was
with cochlear implants would experience great diffi-
replaced by a single-talker interferer.
culties in such situations. More generally, pitch
The results did not match expectations. Both
seems to play an important role in the perceptual
low-pass and high-pass conditions produced less
organization of both sequential and simultaneous
masking release than was found in the broadband
sounds. Better representation of pitch and temporal
(unfiltered) conditions. However, contrary to the
fine structure in cochlear implants and hearing aids
predictions based on pitch strength and temporal
remain an important, if elusive, goal.
fine structure, the amount of masking release was
the same for the low-pass as for the high-pass con-
dition, with no evidence found for superior perform- Acknowledgments
ance in the low-pass condition with any of the
different fluctuating maskers tested. The results of Dr. Steven Neely provided the code to generate Figure 1.
Oxenham and Simonson (2009) suggest that the Dr. Marjorie Leek, one anonymous reviewer and the edi-
temporal fine structure information associated with tor, Dr. Arlene Neuman, provided helpful comments on
resolved harmonics may not play the crucial role in an earlier version of this article. The research described
masking release that was suggested by studies in from the Auditory Perception and Cognition Lab was
cochlear-implant users and vocoder simulations. undertaken in collaboration with Drs. Joshua Bernstein,
This is not to say that temporal fine structure and Christophe Micheyl, Michael Qin, and Andrea
accurate pitch coding are not important; they may Simonson, and was supported by the National Institute
328 Trends in Amplification / Vol. 12, No. 4, December 2008
on Deafness and other Communication Disorders Brokx, J. P., & Nooteboom, S. G. (1982). Intonation and the
(NIDCD grant R01 DC 05216). perceptual separation of simultaneous voices. Journal of
Phonetics, 10, 23-36.
Burns, E. M., & Viemeister, N. F. (1981). Played again SAM:
References Further observations on the pitch of amplitude-modu-
lated noise. Journal of the Acoustical Society of America,
Anstis, S., & Saida, S. (1985). Adaptation to auditory stream- 70, 1655-1660.
ing of frequency-modulated tones. Journal of Experimental Buss, E., Hall, J. W., & Grose, J. H. (2004). Temporal fine-
Psychology: Human Perception and Performance, 11, structure cues to speech and pure tone modulation in
257-271. observers with sensorineural hearing loss. Ear and
Arehart, K. H., Rossi-Katz, J., & Swensson-Prutsman, J. Hearing, 25, 242-250.
(2005). Double-vowel perception in listeners with Cariani, P. A., & Delgutte, B. (1996). Neural correlates of
cochlear hearing loss: Differences in fundamental fre- the pitch of complex tones. I. Pitch and pitch salience.
quency, ear of presentation, and relative amplitude. Journal of Neurophysiology, 76, 1698-1716.
Journal of Speech, Language, and Heading Research, 48, Carlyon, R. P. (1996). Encoding the fundamental frequency
236-252. of a complex tone in the presence of a spectrally over-
Assmann, P. F., & Summerfield, Q. (1987). Perceptual seg- lapping masker. Journal of the Acoustical Society of
regation of concurrent vowels. Journal of the Acoustical America, 99, 517-524.
Society of America, 82, S120. Carlyon, R. P., Cusack, R., Foxton, J. M., & Robertson, I. H.
Beerends, J. G., & Houtsma, A. J. M. (1989). Pitch identifi- (2001). Effects of attention and unilateral neglect on audi-
cation of simultaneous diotic and dichotic two-tone tory stream segregation. Journal of Experimental Psychology:
complexes. Journal of the Acoustical Society of America, Human Perception and Performance, 27, 115-127.
85, 813-819. Cedolin, L., & Delgutte, B. (2007). Spatio-temporal representa-
Bendor, D., & Wang, X. (2005). The neuronal representation of tion of the pitch of complex tones in the auditory nerve. In
pitch in primate auditory cortex. Nature, 436, 1161-1165. B. Kollmeier, G. Klump, V. Hohmann, U. Langemann,
Bendor, D., & Wang, X. (2006). Cortical representations of M. Mauermann, S. Uppenkamp, et al. (Eds.), Hearing:
pitch in monkeys and humans. Current Opinion in From basic research to applications (pp. 61-70). New
Neurobiology, 16, 391-399. York/Berlin: Springer-Verlag.
Bernstein, J. G., & Oxenham, A. J. (2003). Pitch discrimina- Chatterjee, M., Sarampalis, A., & Oba, S. I. (2006). Auditory
tion of diotic and dichotic tone complexes: Harmonic stream segregation with cochlear implants: A prelimi-
resolvability or harmonic number? Journal of the nary report. Hearing Research, 222, 100-107.
Acoustical Society of America, 113, 3323-3334. Cohen, M. A., Grossberg, S., & Wyse, L. L. (1995). A spec-
Bernstein, J. G., & Oxenham, A. J. (2006a). The relationship tral network model of pitch perception. Journal of the
between frequency selectivity and pitch discrimination: Acoustical Society of America, 98, 862-879.
Effects of stimulus level. Journal of the Acoustical Cooper, H. R., & Roberts, B. (2007). Auditory stream segre-
Society of America, 120, 3916-3928. gation of tone sequences in cochlear implant listeners.
Bernstein, J. G., & Oxenham, A. J. (2006b). The relationship Hearing Research, 225, 11-24.
between frequency selectivity and pitch discrimination: Dai, H. (2000). On the relative influence of individual har-
Sensorineural hearing loss. Journal of the Acoustical monics on pitch judgment. Journal of the Acoustical
Society of America, 120, 3929-3945. Society of America, 107, 953-959.
Bird, J., & Darwin, C. J. (1998). Effects of a difference in Darwin, C. J., & Carlyon, R. P. (1995). Auditory organization
fundamental frequency in separating two sentences. In and the formation of perceptual streams. In B. C. J.
A. R. Palmer, A. Rees, A. Q. Summerfield, & R. Meddis Moore (Ed.), Handbook of perception and cognition: Vol.
(Eds.), Psychophysical and physiological advances in 6. Hearing. San Diego: Academic.
hearing (pp. 263-269). London: Whurr. de Boer, E. (1956). On the “residue” in hearing. Unpublished
Boex, C., Baud, L., Cosendai, G., Sigrist, A., Kos, M. I., & doctoral dissertation submitted to University of Amsterdam,
Pelizzone, M. (2006). Acoustic to electric pitch compar- Netherlands.
isons in cochlear implant subjects with residual hearing. de Cheveigné, A. (1997). Concurrent vowel identification. III. A
Journal of the Association for Research in Otolaryngology, neural model of harmonic interference cancellation. Journal
7, 110-124. of the Acoustical Society of America, 101, 2857-2865.
Bregman, A. S. (1990). Auditory scene analysis: The percep- de Cheveigné, A. (1999). Waveform interactions and the seg-
tual organization of sound. Cambridge, MA: MIT Press. regation of concurrent vowels. Journal of the Acoustical
Bregman, A. S., Ahad, P., Crum, P. A., & O’Reilly, J. (2000). Society of America, 106, 2959-2972.
Effects of time intervals and tone durations on auditory Deeks, J. M., & Carlyon, R. P. (2004). Simulations of
stream segregation. Perception and Psychophysics, 62, cochlear implant hearing using filtered harmonic com-
626-636. plexes: Implications for concurrent sound segregation.
Pitch Perception and Auditory Stream Segregation / Oxenham 329
Journal of the Acoustical Society of America, 115, or modulation interference? Journal of the Acoustical Society
1736-1746. of America, 110, 1130-1140.
Dorman, M. F., Spahr, T., Gifford, R., Loiselle, L., McKarns, Larsen, E., Cedolin, L., & Delgutte, B. (2008). Pitch repre-
S., Holden, T., et al. (2007). An electric frequency-to- sentations in the auditory nerve: Two concurrent com-
place map for a cochlear implant patient with hearing in plex tones. Journal of Neurophysiology, 100, 1301-1319.
the nonimplanted ear. Journal of the Association for Liberman, M. C. (1978). Auditory-nerve response from cats
Research in Otolaryngology, 8, 234-240. raised in a low-noise chamber. Journal of the Acoustical
Evans, E. F. (2001). Latest comparisons between physiological Society of America, 63, 442-455.
and behavioural frequency selectivity. In J. Breebaart, A. J. Licklider, J. C. R. (1951). A duplex theory of pitch percep-
M. Houtsma, A. Kohlrausch, V. F. Prijs, & R. Schoonhoven tion. Experientia, 7, 128-133.
(Eds.), Physiological and psychophysical bases of auditory Licklider, J. C. R. (1954). “Periodicity” pitch and “place”
function (pp. 382-387). Maastricht: Shaker. pitch. Journal of the Acoustical Society of America,
Festen, J. M., & Plomp, R. (1990). Effects of fluctuating 26, 945.
noise and interfering speech on the speech-reception Loeb, G. E., White, M. W., & Merzenich, M. M. (1983).
threshold for impaired and normal hearing. Journal of Spatial cross correlation: A proposed mechanism for
the Acoustical Society of America, 88, 1725-1736. acoustic pitch perception. Biological Cybernetics, 47,
Fletcher, H. (1940). Auditory patterns. Reviews of Modern 149-163.
Physics, 12, 47-65. Lorenzi, C., Gilbert, G., Carn, H., Garnier, S., & Moore, B.
Glasberg, B. R., & Moore, B. C. J. (1990). Derivation of C. (2006). Speech perception problems of the hearing
auditory filter shapes from notched-noise data. Hearing impaired reflect inability to use temporal fine structure.
Research, 47, 103-138. Proceedings of the National Academy of Sciences (USA),
Goldstein, J. L. (1973). An optimum processor theory for the 103, 18866-18869.
central formation of the pitch of complex tones. Journal Mackersie, C. L. (2003). Talker separation and sequential
of the Acoustical Society of America, 54, 1496-1516. stream segregation in listeners with hearing loss:
Gutschalk, A., Oxenham, A. J., Micheyl, C., Wilson, E. C., & Patterns associated with talker gender. Journal of
Melcher, J. R. (2007). Human cortical activity during Speech, Language, and Hearing Research, 46, 912-918.
streaming without spectral cues suggests a general neu- Mackersie, C. L., Prida, T. L., & Stiles, D. (2001). The role
ral substrate for auditory stream segregation. Journal of of sequential stream segregation and frequency selectiv-
Neuroscience, 27, 13074-13081. ity in the perception of simultaneous sentences by lis-
Heinz, M. G., Colburn, H. S., & Carney, L. H. (2001). teners with sensorineural hearing loss. Journal of
Evaluating auditory performance limits: I. One-parame- Speech, Language, and Hearing Research, 44, 19-28.
ter discrimination using a computational model for the Marrone, N., Mason, C. R., & Kidd, G. D., Jr. (2008).
auditory nerve. Neural Computation, 13, 2273-2316. Evaluating the benefit of hearing aids in solving the cock-
Hong, R. S., & Turner, C. W. (2006). Pure-tone auditory tail party problem. Trends in Amplification, 12, 300-315.
stream segregation and speech perception in noise in McDermott, J. H., & Oxenham, A. J. (2008). Spectral
cochlear implant recipients. Journal of the Acoustical completion of partially masked sounds. Proceedings of the
Society of America, 120, 360-374. National Academy of Sciences (USA), 105, 5939-5944.
Houtgast, T., Steeneken, H. J. M., & Plomp, R. (1980). McKay, C. M., McDermott, H. J., & Carlyon, R. P. (2000).
Predicting speech intelligibility in rooms from the mod- Place and temporal cues in pitch perception: Are they
ulation transfer function. I. General room acoustics. truly independent? Acoustics Research Letters Online, 1,
Acustica, 46, 60-72. 25-30.
Houtsma, A. J. M., & Smurzynski, J. (1990). Pitch identifi- Meddis, R., & Hewitt, M. (1991a). Virtual pitch and phase
cation and discrimination for complex tones with many sensitivity studied of a computer model of the auditory
harmonics. Journal of the Acoustical Society of America, periphery. I: Pitch identification. Journal of the
87, 304-310. Acoustical Society of America, 89, 2866-2882.
Kaernbach, C., & Bering, C. (2001). Exploring the temporal Meddis, R., & Hewitt, M. (1991b). Virtual pitch and phase
mechanism involved in the pitch of unresolved harmon- sensitivity studied of a computer model of the auditory
ics. Journal of the Acoustical Society of America, 110, periphery. II: Phase sensitivity. Journal of the Acoustical
1039-1048. Society of America, 89, 2882-2894.
Kim, D. O., & Molnar, C. E. (1979). A population study of Meddis, R., & O’Mard, L. (1997). A unitary model of pitch
cochlear nerve fibres: Comparison of spatial distributions perception. Journal of the Acoustical Society of America,
of average-rate and phase-locking measures of responses 102, 1811-1820.
to single tones. Journal of Neurophysiology, 42, 16-30. Micheyl, C., Bernstein, J. G., & Oxenham, A. J. (2006).
Kwon, B. J., & Turner, C. W. (2001). Consonant identification Detection and F0 discrimination of harmonic complex
under maskers with sinusoidal modulation: Masking release tones in the presence of competing tones or noise.
330 Trends in Amplification / Vol. 12, No. 4, December 2008
Journal of the Acoustical Society of America, 120, human auditory cortex revealed with fMRI. Journal of
1493-1505. Neuroscience, 24, 6810-6815.
Miller, G. A., & Heise, G. A. (1950). The trill threshold. Peters, R. W., Moore, B. C. J., & Baer, T. (1998). Speech
Journal of the Acoustical Society of America, 22, reception thresholds in noise with and without spectral
637-638. and temporal dips for hearing-impaired and normally
Moore, B. C. J. (1973). Frequency difference limens for hearing people. Journal of the Acoustical Society of
short-duration tones. Journal of the Acoustical Society of America, 103, 577-587.
America, 54, 610-619. Pfingst, B. E., Holloway, L. A., Poopat, N., Subramanya, A.
Moore, B. C. J., & Carlyon, R. P. (2005). Perception of pitch R., Warren, M. F., & Zwolan, T. A. (1994). Effects of
by people with cochlear hearing loss and by cochlear stimulus level on nonspectral frequency discrimination
implant users. In C. J. Plack, A. J. Oxenham, R. Fay, & by human subjects. Hearing Research, 78, 197-209.
A. N. Popper (Eds.), Pitch: Neural coding and percep- Pijl, S., & Schwarz, D. W. (1995). Melody recognition and
tion. New York: Springer. musical interval perception by deaf subjects stimulated
Moore, B. C. J., Glasberg, B. R., & Peters, R. W. (1985). with electrical pulse trains through single cochlear
Relative dominance of individual partials in determining implant electrodes. Journal of the Acoustical Society of
the pitch of complex tones. Journal of the Acoustical America, 98, 886-895.
Society of America, 77, 1853-1860. Plack, C. J., & Oxenham, A. J. (2005). Psychophysics of
Moore, B. C. J., & Peters, R. W. (1992). Pitch discrimination pitch. In C. J. Plack, A. J. Oxenham, A. N. Popper, &
and phase sensitivity in young and elderly subjects and R. Fay (Eds.), Pitch: Neural coding and perception. New
its relationship to frequency selectivity. Journal of the York: Springer.
Acoustical Society of America, 91, 2881-2893. Plomp, R. (1967). Pitch of complex tones. Journal of the
Moore, B. C. J., & Rosen, S. M. (1979). Tune recognition Acoustical Society of America, 41, 1526-1533.
with reduced pitch and interval information. Quarterly Qin, M. K., & Oxenham, A. J. (2003). Effects of simulated
Journal of Experimental Psychology, 31, 229-240. cochlear-implant processing on speech reception in
Narayan, S. S., Temchin, A. N., Recio, A., & Ruggero, M. A. fluctuating maskers. Journal of the Acoustical Society of
(1998). Frequency tuning of basilar membrane and America, 114, 446-454.
auditory nerve fibers in the same cochleae. Science, Qin, M. K., & Oxenham, A. J. (2005). Effects of envelope-
282, 1882-1884. vocoder processing on F0 discrimination and concur-
Nelson, P. B., & Jin, S. H. (2004). Factors affecting speech rent-vowel identification. Ear and Hearing, 26, 451-460.
understanding in gated interference: Cochlear implant Rhode, W. S. (1971). Observations of the vibration of the
users and normal-hearing listeners. Journal of the basilar membrane in squirrel monkeys using the
Acoustical Society of America, 115, 2286-2294. Mössbauer technique. Journal of the Acoustical Society
Nelson, P. B., Jin, S. H., Carney, A. E., & Nelson, D. A. of America, 49, 1218-1231.
(2003). Understanding speech in modulated interfer- Ritsma, R. J. (1967). Frequencies dominant in the percep-
ence: Cochlear implant users and normal-hearing listen- tion of the pitch of complex sounds. Journal of the
ers. Journal of the Acoustical Society of America, 113, Acoustical Society of America, 42, 191-198.
961-968. Roberts, B., Glasberg, B. R., & Moore, B. C. (2002). Primitive
Nelson, D. A., van Tasell, D. J., Schroder, A. C., Soli, S., & stream segregation of tone sequences without differences
Levine, S. (1995). Electrode ranking of “place pitch” in fundamental frequency or passband. Journal of the
and speech recognition in electrical hearing. Journal of Acoustical Society of America, 112, 2074-2085.
the Acoustical Society of America, 98, 1987-1999. Rose, J. E., Brugge, J. F., Anderson, D. J., & Hind, J. E.
Oxenham, A. J., Bernstein, J. G. W., & Penagos, H. (2004). (1967). Phase-locked response to low-frequency tones
Correct tonotopic representation is necessary for com- in single auditory nerve fibers of the squirrel monkey.
plex pitch perception. Proceedings of the National Journal of Neurophysiology, 30, 769-793.
Academy of Sciences (USA), 101, 1421-1425. Rose, M. M., & Moore, B. C. J. (1997). Perceptual grouping
Oxenham, A. J., & Simonson, A. M. (2009). Masking release of tone sequences by normally hearing and hearing-
for low- and high-pass filtered speech in the presence of impaired listeners. Journal of the Acoustical Society of
noise and single-talker interference. Journal of the America, 102, 1768-1778.
Acoustical Society of America. Ruggero, M. A., Rich, N. C., Recio, A., Narayan, S. S., &
Parsons, T. W. (1976). Separation of speech from interfering Robles, L. (1997). Basilar-membrane responses to tones
speech by means of harmonic selection. Journal of the at the base of the chinchilla cochlea. Journal of the
Acoustical Society of America, 60, 911-918. Acoustical Society of America, 101, 2151-2163.
Patterson, R. D. (1976). Auditory filter shapes derived with Scheffers, M. T. M. (1983). Simulation of auditory analysis of
noise stimuli. Journal of the Acoustical Society of pitch: An elaboration on the DWS pitch meter. Journal of
America, 59, 640-654. the Acoustical Society of America, 74, 1716-1725.
Penagos, H., Melcher, J. R., & Oxenham, A. J. (2004). A neu- Schouten, J. F. (1940). The residue and the mechanism of
ral representation of pitch salience in non-primary hearing. Proc. Kon. Akad. Wetenschap, 43, 991-999.
Pitch Perception and Auditory Stream Segregation / Oxenham 331
Schouten, J. F., Ritsma, R. J., & Cardozo, B. L. (1962). Pitch sentences. Journal of the Acoustical Society of America,
of the residue. Journal of the Acoustical Society of 122, 1069-1078.
America, 34, 1418-1424. Stickney, G. S., Zeng, F. G., Litovsky, R., & Assmann, P.
Shackleton, T. M., & Carlyon, R. P. (1994). The role of (2004). Cochlear implant speech recognition with
resolved and unresolved harmonics in pitch perception speech maskers. Journal of the Acoustical Society of
and frequency modulation discrimination. Journal of the America, 116, 1081-1091.
Acoustical Society of America, 95, 3529-3540. Summerfield, A. Q., & Assmann, P. F. (1991). Perception of
Shamma, S. A. (1985). Speech processing in the auditory concurrent vowels: Effects of pitch-pulse asynchrony
system. I: The representation of speech sounds in the and harmonic misalignment. Journal of the Acoustical
responses in the auditory nerve. Journal of the Acoustical Society of America, 89, 1364-1377.
Society of America, 78, 1612-1621. Summers, V., & Leek, M. R. (1998). F0 processing and the
Shamma, S. A., & Klein, D. (2000). The case of the missing separation of competing speech signals by listeners with
pitch templates: How harmonic templates emerge in the normal hearing and with hearing loss. Journal of Speech,
early auditory system. Journal of the Acoustical Society of Language, and Hearing Research, 41, 1294-1306.
America, 107, 2631-2644. Terhardt, E., Stoll, G., & Seewann, M. (1982). Algorithm for
Shannon, R. V., Zeng, F. G., Kamath, V., Wygonski, J., & extraction of pitch and pitch salience from complex
Ekelid, M. (1995). Speech recognition with primarily tonal signals. Journal of the Acoustical Society of
temporal cues. Science, 270, 303-304. America, 71, 679-688.
Shera, C. A., Guinan, J. J., & Oxenham, A. J. (2002). Tong, Y. C., Blamey, P. J., Dowell, R. C., & Clark, G. M.
Revised estimates of human cochlear tuning from otoa- (1983). Psychophysical studies evaluating the feasibility
coustic and behavioral measurements. Proceedings of of a speech processing strategy for a multiple-channel
the National Academy of Sciences Online (USA), 99, cochlear implant. Journal of the Acoustical Society of
3318-3323. America, 74, 73-80.
Shinn-Cunningham, B., & Best, V. (2008). Selective attention Townsend, B., Cotter, N., van Compernolle, D., & White, R. L.
in normal and impaired hearing. Trends in Amplification, (1987). Pitch perception by cochlear implant subjects.
12, 283-299. Journal of the Acoustical Society of America, 82, 106-115.
Singh, P. G. (1987). Perceptual organization of complex-tone Tsuji, J., & Liberman, M. C. (1997). Intracelluar labeling of audi-
sequences: A trade-off between pitch and timbre? tory nerve fibers in guinea pig: Central and peripheral pro-
Proceedings of the National Academy of Sciences Online jections. Journal of Comparative Neurology, 381, 188-202.
(USA), 82, 886-895. van Noorden, L. P. A. S. (1975). Temporal coherence in the
Sininger, Y., & Starr, A. (Eds.). (2001). Auditory neuropathy. perception of tone sequences. Unpublished doctoral dis-
San Diego, CA: Singular. sertation submitted to Eindhoven University of
Sit, J.-J., Simonson, A. M., Oxenham, A. J., Faltys, M. A., & Technology, Netherlands.
Sarpeshkar, R. (2007). A low-power asynchronous inter- Vliegen, J., Moore, B. C. J., & Oxenham, A. J. (1999). The
leaved sampling algorithm for cochlear implants that role of spectral and periodicity cues in auditory stream
encodes envelope and phase information. IEEE segregation, measured using a temporal discrimination
Transactions on Biomedical Engineering, 54, 138-149. task. Journal of the Acoustical Society of America, 106,
Snyder, J. S., Alain, C., & Picton, T. W. (2006). Effects of 938-945.
attention on neuroelectric correlates of auditory stream Vliegen, J., & Oxenham, A. J. (1999). Sequential stream seg-
segregation. Journal of Cognitive Neuroscience, 18, 1-13. regation in the absence of spectral cues. Journal of the
Snyder, J. S., Carter, O. L., Lee, S. K., Hannon, E. E., & Acoustical Society of America, 105, 339-346.
Alain, C. (2008). Effects of context on auditory stream von Bismark, G. (1974). Sharpness as an attribute of the
segregation. Journal of Experimental Psychology: Human timbre of steady sounds. Acustica, 30, 159-172.
Perception and Performance, 34, 1007-1016. Wightman, F. L. (1973). The pattern-transformation model
Srulovicz, P., & Goldstein, J. L. (1983). A central spectrum of pitch. Journal of the Acoustical Society of America, 54,
model: A synthesis of auditory-nerve timing and place cues 407-416.
in monaural communication of frequency spectrum. Zeng, F. G. (2002). Temporal pitch in electric hearing.
Journal of the Acoustical Society of America, 73, 1266-1276. Hearing Research, 174, 101-106.
Stainsby, T. H., Moore, B. C., & Glasberg, B. R. (2004). Zeng, F. G., Kong, Y. Y., Michalewski, H. J., & Starr, A. (2005).
Auditory streaming based on temporal structure in hear- Perceptual consequences of disrupted auditory nerve
ing-impaired listeners. Hearing Research, 192, 119-130. activity. Journal of Neurophysiology, 93, 3050-3063.
Stickney, G. S., Assmann, P. F., Chang, J., & Zeng, F. G. Zwicker, E., Flottorp, G., & Stevens, S. S. (1957). Critical band-
(2007). Effects of cochlear implant processing and fun- width in loudness summation. Journal of the Acoustical
damental frequency on the intelligibility of competing Society of America, 29, 548-557.