Professional Documents
Culture Documents
Behavioral/Systems/Cognitive
Pooling and synthesizing signals across different senses often enhances responses to the event from which they are derived. Here, we
examine whether multisensory response enhancements are attributable to a redundant target effect (two stimuli rather than one) or if
there is some special quality inherent in the combination of cues from different senses. To test these possibilities, the performance of
animals in localizing and detecting spatiotemporally concordant visual and auditory stimuli was examined when these stimuli were
presented individually (visual or auditory) or in cross-modal (visual–auditory) and within-modal (visual–visual, auditory–auditory)
combinations. Performance enhancements proved to be far greater for combinations of cross-modal than within-modal stimuli and
support the idea that the behavioral products derived from multisensory integration are not attributable to simple target redundancy.
One likely explanation is that whereas cross-modal signals offer statistically independent samples of the environment, within-modal
signals can exhibit substantial covariance, and consequently multisensory integration can yield more substantial error reduction than
unisensory integration.
kibble; Hill’s Science Diet). Each animal was first trained to stand at the
center of the arena (the start position) and fixate directly ahead at 0°, with
the experimenter providing gentle head restraint. The animal was then
required to orient toward and approach (within 4 cm) a briefly illumi-
nated (40 ms) but highly visible LED (3 foot-candles) at that perimetric
location. Its nose had to be displaced no more than 5° to the left or right
of this location within 3 s of stimulus onset to receive a reward. After
mastering this task, the number of possible target LED locations was
expanded to include increasingly more eccentric locations (these were
randomly interleaved between ⫾45° and always involved the left LED in
each stimulus complex). Given that 15° separated the perimetric loca-
tions, the animal had to choose among seven locations: three locations on
the left (⫺45°, ⫺30°, ⫺15°), the center location (0°), and three locations
on the right (⫹15°, ⫹30°, ⫹45°). Once the animal mastered the visual
localization task, it was trained on the auditory localization task using the
same general procedure (the auditory stimulus was a 40 ms, 60 dB sound-
pressure level A-weighted broadband noise). At this time, “catch” trials
Figure 1. The apparatus and task. The orientation/approach task was performed in a 90-cm- (no stimulus) were introduced, during which the animal was required to
diameter perimetry apparatus containing a complex of LEDs and speakers at 15° increments remain at the start position looking directly ahead to receive a reward.
from 90° to the left and right of a central fixation point (0°). Each complex consisted of three Training was completed when animals would accurately approach the
LEDs and two speakers, each separated by 4 cm. In the present experiments, only the two stimulus on at least 85 of 100 trials at all seven locations and have ⬍15%
outermost LEDs at any location were used. Trials consisted of randomly interleaved modality- erroneous responses to catch trials.
specific (visual or auditory), cross-modal (visual–auditory), and within-modal (visual–visual or Testing. To equilibrate the effectiveness of the stimuli among the dif-
auditory–auditory) stimuli at each of the seven perimetric locations between ⫾45° (⫺45°, ferent perimetric locations, the intensity of the modality-specific stimu-
⫺30°, ⫺15°, 0°, ⫹15°, ⫹30°, ⫹45°), as well as catch trials. lus (visual or auditory) was reduced at each location to a level eliciting
⬃25% correct responses. There was one stimulus/trial/location, but the
stimulus could be an individual modality-specific stimulus (visual or
sory and unisensory integration and suggest that a simple RTE auditory) or a pair of within-modal (visual–visual, auditory–auditory) or
model is insufficient to explain the data. cross-modal (visual–auditory) stimuli. In addition, catch trials (no stim-
ulus was presented) were included. All the presentations of the pairs of
Materials and Methods stimulus were spatially and temporally coincident, unless otherwise
All procedures were conducted in accordance with the Guide for the Care noted (see below). All animals were tested with ⬃150 trials per day (i.e.,
and Use of Laboratory Animals (National Institutes of Health publication until satiety) and 5 d per week. To ensure a reasonable number of trials
86-23) and an approved Institutional Animal Care and Use Committee per stimulus condition each day, cross-modal (visual–auditory) versus
protocol at Wake Forest University School of Medicine, an Association within-modal (visual–visual or auditory–auditory) comparisons were
for Assessment and Accreditation of Laboratory Animal Care-accredited divided into two experiments separated by ⬃9 months. The first exper-
institution. The apparatus, training, and testing procedures were similar iment compared performance in response to within-modal visual (V1V2)
to those described previously (Stein et al., 1988, 1989; Wilkinson et al., and cross-modal (V1A1) stimuli; and the second experiment compared
1996; Jiang et al., 2002; Burnett et al., 2004). performance in response to within-modal auditory (A1A2) and cross-
Apparatus. Four adult male cats (4 – 6 kg) were trained in a spatial local- modal (V1A1) stimuli. Because animals were run to satiety each day using
ization task within a 90-cm-diameter perimetry apparatus (Fig. 1). The random trial selection, there were often unequal numbers of trials/day/
apparatus contained stimulus complexes of light-emitting diodes (LEDs) location. Because the random selection procedure was used during all
and speakers placed at 15° intervals to the left (from ⫺90°) and right (to days of testing, there were some locations that exceeded the minimum
⫹90°) of a central fixation point (0°). Each stimulus complex consisted of number of trials.
two horizontally displaced (4 cm) LEDs (Lumex Opto/Components; Pairs of visual stimuli consisted of 2 LEDs separated by 5°. The paired
model 67-1102-ND) located 4 cm beneath two horizontally displaced auditory stimuli were also separated by 5°, but to ensure that two sounds
speakers (Panasonic; model 4D02C0). This displacement ensured that all were distinguishable in the second experiment, the first sound (see
stimulus pairings were likely to fall within the receptive fields of many of above) was presented first to the left speaker (30 ms duration) and 15 ms
the same multisensory SC neurons (Stein and Meredith, 1993). As a later to the second (right) speaker (35 ms duration). Thus, they over-
general convention, we reference the leftmost stimuli within each indi- lapped for 15 ms. For this reason, the stimuli used in Experiment 2 lasted
vidual stimulus complex with a “1” and the rightmost stimuli with a “2” 50 ms as opposed to 40 ms as in Experiment 1. The two sounds were
(e.g., V1 is the leftmost visual stimulus). The perimetry apparatus was distinguishable to human listeners and had differing impacts on the an-
housed in a sound-attenuating chamber (Industrial Acoustics Company) imals’ behavioral responses (see Results).
with a constant 22 dB background noise. Stimuli were controlled with Within each experiment, all five stimulus conditions were randomly
custom software, triggered by the experimenter, and unless otherwise interleaved: (1) a single visual stimulus (V1); (2) a single auditory stim-
stated, consisted of brief (40 ms) LED illuminations or bursts of broad- ulus (A1 or A2); (3) a spatiotemporally coincident visual–auditory stim-
band noise from the speakers. During the testing phase (see below), ulus pair (V1A1); (4) a spatiotemporally coincident within-modal stim-
stimulus intensities were reduced to levels undetectable to the experi- ulus pair (visual, V1V2 or auditory, A1A2); or (5) a catch trial. All stimulus
menter. The experimenter was informed of the target location (i.e., the locations were randomly selected. Orientation and approach to each
stimulus complex), and of the stimulus combination on any given trial target stimulus (see above) and maintaining fixation during a catch trial
after the categorical response, location was judged. The animal was either were rewarded. Because V1 and V2 were identical, only one of them (V1)
within 5° (4 cm) of the target or outside of this range. This proved to be was used to obtain an index of unisensory visual performance and the
an extremely easy judgment to make as animals very rarely went between intensity level of V2 was matched to V1 during the visual within-modal
stimulus complexes (each separated by 15°) because all stimuli com- stimulus combinations (V1V2).
plexes were always visible to them. When making errors, they almost Control experiment. The physical limitations of the apparatus dis-
always either went to a wrong location or failed to respond (“no-go”). placed visual–visual (V1V2) and auditory–auditory (A1A2) stimuli by 4
Rare circumstances in which the animal began to approach the target, but cm in the horizontal dimension, whereas visual–auditory stimuli (V1A1)
then returned to the starting point, were scored as no-go errors. were at the same azimuthal position, one above the other. To examine
Training. Responses were shaped using a small food reward (175 mg whether the directionality of displacement had a significant effect on the
Gingras et al. • Comparisons of Multisensory and Unisensory Integration J. Neurosci., April 15, 2009 • 29(15):4897– 4902 • 4899
Results
All animals learned to localize and ap-
proach the targets at all perimetric posi-
tions. Differences in the speed with which
different animals (n ⫽ 4) reached criterion
performance were noted, but these were
minor and in keeping with interanimal
variations noted in previous studies (Stein
et al., 1988, 1989; Wilkinson et al., 1996;
Jiang et al., 2002, Burnett et al., 2004). Per-
formance patterns for each stimulus con-
dition during testing were highly consis-
tent across all animals, and thus the data
were pooled for the general analysis.
ity only marginally enhanced localization compared with cross- data suggest that there is something inherently different about
modal stimulus combinations. Although they made localization multisensory and unisensory integration that is evident in a sim-
responses somewhat more likely, they did not make them any ple detection/localization task. Whether these differences would
more accurate. be equally evident in tasks requiring the use of higher-order cog-
These behavioral observations closely parallel recent physio- nitive processes remains to be determined.
logical findings from single neurons in the SC. Alvarado et al.
(2007) found that SC neurons responded to pairs of adjacent
visual stimuli with only a modest increase (not statistically signif- References
Alvarado JC, Vaughan JW, Stanford TR, Stein BE (2007) Multisensory ver-
icant) in their mean number of evoked impulses, numbers that sus unisensory integration: contrasting modes in the superior colliculus.
were far below those predicted by summing responses to the J Neurophysiol 97:3193–3205.
individual component stimuli. In contrast, SC neurons showed Amlôt R, Walker R, Driver J, Spence C (2003) Multimodal visual-
highly significant response enhancements to cross-modal (visua- somatosensory integration in saccade generation. Neuropsychologia
l–auditory) stimulus pairs, response enhancements that were of- 41:1–15.
Burnett LR, Stein BE, Chaponis D, Wallace MT (2004) Superior colliculus
ten greater than, or equal to, the sum of responses to the compo-
lesions preferentially disrupt multisensory orientation. Neuroscience
nent stimuli. These physiological and behavioral observations are 124:535–547.
consistent with suggestions that multisensory and unisensory in- Calvert GA, Spence C, Stein BE (2004) The handbook of multisensory pro-
tegration engage different underlying computations. cesses. Cambridge, MA: MIT.
As noted by Ernst and Banks (2002), these differences can be Diederich A, Colonius H (2004) Bimodal and trimodal multisensory en-
understood in the context of a probabilistic (Bayesian) model of hancement: effects of stimulus onset and intensity on reaction time. Per-
integration. In a Bayesian model of spatial locations (Rowland et cept Psychophys 66:1388 –1404.
Driver J, Noesselt T (2008) Multisensory interplay reveals crossmodal influ-
al., 2007a), each stimulus is assumed to generate a sensory report ences on ‘sensory-specific’ brain regions, neural responses, and judg-
of its location that may be accurate or randomly inaccurate, de- ments. Neuron 57:11–23.
pending on the fidelity of the sensory system. These reports are Ernst MO, Banks MS (2002) Humans integrate visual and haptic informa-
filtered by previous expectations such that, if two or more stimuli tion in a statically optimal fashion. Nature 415:429 – 433.
co-occur, they belong to the same event and thus should have the Forster B, Cavina-Pratesi C, Aglioti SM, Berlucchi G (2002) Redundant tar-
same location. The action of the filter is to fuse the sensory re- get effect and intersensory facilitation from visual–tactile interactions in
simple reaction time. Exp Brain Res 143:480 – 487.
ports, generating an estimate of location that is a compromise
Frens MA, Van Opstal AJ, Van der Willigen RF (1995) Spatial and temporal
between them. Sensory reports that are inaccurate in the same factors determine auditory–visual interactions in human saccadic eye
direction (e.g., both biased to the left or both to the right) do not movements. Percept Psychophys 57:802– 816.
generate more accurate estimates when fused because they do not Giard MH, Peronnet F (1999) Auditory-visual integration during multimo-
contradict the previous expectation. However, sensory reports dal object recognition in humans: a behavioral and electrophysiological
that are inaccurate in opposite directions (e.g., one biased left and study. J Cogn Neurosci 11:473– 490.
the other one biased right) are fused to generate estimates that are Goldring JE, Dorris MC, Corneil BD, Ballantyne PA, Munoz DP (1996)
Combined eye-head gaze shifts to visual and auditory targets in humans.
more consistent with the previous expectation that the stimulus Exp Brain Res 111:68 –78.
locations should be the same. These circumstances produce esti- Gondan M, Niederhaus B, Rösler F, Röder B (2005) Multisensory process-
mated stimulus locations between the two incorrect sensory re- ing in the redundant-target effect: a behavioral and event-related poten-
ports, which are closer to the actual stimulus location than the tial study. Percept Psychophys 67:713–726.
flanking sensory reports. Hecht D, Reiner M, Karni A (2008a) Enhancement of response times to bi-
Cross-modal stimulus combinations represent multiple stim- and tri-modal sensory stimuli during active movements. Exp Brain Res
185:655– 665.
uli that are conveyed by different forms of energy and transduced
Hecht D, Reiner M, Karni A (2008b) Multisensory enhancement: gains in
by independent sensory systems. Consequently, the sensory re- choice and in simple response times. Exp Brain Res 189:133–143.
ports evoked by two cross-modal stimuli are independent of one Hughes HC, Reuter-Lorenz PA, Nozawa G, Fendrich R (1994) Visual–au-
another and equally as likely to be inaccurate in the same or ditory interactions in sensorimotor processing: saccades versus manual
different directions. Instances where the sensory reports are in- responses. J Exp Psychol Hum Percept Perform 20:131–153.
accurate in different directions yield more accurate estimates (by Jiang W, Wallace MT, Jiang H, Vaughan JW, Stein BE (2001) Two cortical
the above logic) and thus enhance localization responses. In con- areas mediate multisensory integration in superior colliculus neurons.
J Neurophysiol 85:506 –522.
trast, multiple within-modal stimuli that travel by the same me- Jiang W, Jiang H, Stein BE (2002) Two corticotical areas facilitate multisen-
dium and are transduced by the same sensory system can be sory orientation behavior. J Cogn Neurosci 14:1240 –1255.
influenced by the same noise sources. Consequently, the sensory Leo F, Bertini C, di Pellegrino G, Làdavas E (2008) Multisensory integration
reports they generate are likely to covary substantially, that is, be for orienting responses in humans requires the activation of the superior
more likely to be inaccurate in the same direction than in differ- colliculus. Exp Brain Res 186:67–77.
ent directions. The lower incidence of sensory reports that are Lippert M, Logothetis NK, Kayser C (2007) Improvement of visual contrast
detection by a simultaneous sound. Brain Res 1173:102–109.
inaccurate in opposite directions reduces the incidence where
Marzi CA, Smania N, Martini MC, Gambina G, Tomelleri G, Palamara A,
estimates are improved through their fusion. The intuitive nature Alessandrini F, Prior M (1996) Implicit redundant-targets effect in vi-
of this result is apparent in the consideration of the most extreme sual extinction. Neuropsychologia 34:9 –22.
case: two sensory reports that covary 100% of the time are obvi- Meredith MA, Stein BE (1983) Interactions among converging sensory in-
ously no better than one alone. Thus, although not pursued puts in the superior colliculus. Science 221:389 –391.
quantitatively here, the Bayesian model can help explain how Meredith MA, Stein BE (1986) Visual, auditory, and somatosensory conver-
different products result when information is integrated across or gence on cells in superior colliculus results in multisensory integration.
J Neurophysiol 56:640 – 662.
within a sensory modality as a consequence of the relative statis- Miller J (1982) Divided attention: evidence for coactivation with redundant
tics of the stimuli. signals. Cogn Psychol 14:247–279.
Whether the Bayesian model or some other model is used to Murray MM, Foxe JJ, Higgins BA, Javitt DC, Schroeder CE (2001) Visuo-
explain the current results (see also Rowland et al., 2007b), the spatial neural response interactions in early cortical processing during a
4902 • J. Neurosci., April 15, 2009 • 29(15):4897– 4902 Gingras et al. • Comparisons of Multisensory and Unisensory Integration
simple reaction time task: a high-density electrical mapping study. Neu- Stanford TR, Stein BE (2007) Superadditivity in multisensory integration:
ropsychologia 39:828 – 844. putting the computation in context. Neuroreport 18:787–792.
Perrault TJ Jr, Vaughan JW, Stein BE, Wallace MT (2005) Superior collicu- Stanford TR, Quessy S, Stein BE (2005) Evaluating the operations underly-
lus neurons use distinct operational modes in the integration of multisen- ing multisensory integration in the cat superior colliculus. J Neurosci
sory stimuli. J Neurophysiol 93:2575–2586. 25:6499 – 6508.
Rowland B, Stanford T, Stein B (2007a) A Bayesian model unifies multisen- Stein BE, Meredith MA (1993) The merging of the senses. Cambridge, MA:
sory spatial localization with the physiological properties of the superior MIT.
colliculus. Exp Brain Res 180:153–161. Stein BE, Stanford TR (2008) Multisensory integration: current issues from
Rowland BA, Stanford TR, Stein BE (2007b) A model of the neural mecha- the perspective of the single neuron. Nat Rev Neurosci 9:255–266.
nisms underlying multisensory integration in the superior colliculus. Per- Stein BE, Huneycutt WS, Meredith MA (1988) Neurons and behavior: the
ception 36:1431–1443. same rules of multisensory integration apply. Brain Res 448:355–358.
Sakata S, Yamamori T, Sakurai Y (2004) Behavioral studies of auditory-visual
Stein BE, Meredith MA, Huneycutt WS, McDade L (1989) Behavioral indi-
spatial recognition and integration in rats. Exp Brain Res 159:409 – 417.
ces of multisensory integration: orientation to visual cues is affected by
Savazzi S, Marzi CA (2002) Speeding up the reaction time with invisible
auditory stimuli. J Cogn Neurosci 1:12–24.
stimuli. Curr Biol 12:403– 407.
Savazzi S, Marzi CA (2004) The superior colliculus subserves interhemi- Teder-Sälejärvi WA, Di Russo F, McDonald JJ, Hillyard SA (2005) Effects of
spheric neural summation in both normals and patients with a total sec- spatial congruity on audio-visual multimodal integration. J Cogn Neuro-
tion or agenesis of the corpus callsum. Neuropsychologia 42:1608 –1618. sci 17:1396 –1409.
Schröter H, Ulich R, Miller J (2007) Effects of redundant auditory stimuli Wallace MT, Wilkinson LK, Stein BE (1996) Representation and integra-
on reaction time. Psychon Bull Rev 14:39 – 44. tion of multiple sensory inputs in primate superior colliculus. J Neuro-
Senkowski D, Molholm S, Gomez-Ramirez M, Foxe JJ (2006) Oscillatory physiol 76:1246 –1266.
beta activity predicts responses speed during a multisensory audiovisual Wallace MT, Meredith MA, Stein BE (1998) Multisensory integration in the
reaction time task: a high-density electrical mapping study. Cereb Cortex superior colliculus of the alert cat. J Neurophysiol 80:1006 –1010.
16:1556 –1565. Wilkinson LK, Meredith MA, Stein BE (1996) The role of anterior ectosyl-
Sinnett S, Soto-Faraco S, Spence C (2008) The co-occurrence of multisen- vian cortex in cross-modality orientation and approach behavior. Exp
sory competition and facilitation. Acta Psychol (Amst) 128:153–161. Brain Res 112:1–10.