14
Cognitive Brain Research 14 (2002) 115–128 www.elsevier.com / locate / bres Research report Multisensory auditory–visual interactions during early sensory processing in humans: a high-density electrical mapping study a,b a,b a a,e Sophie Molholm , Walter Ritter , Micah M. Murray , Daniel C. Javitt , a,c a,b,c,d, * Charles E. Schroeder , John J. Foxe a Cognitive Neurophysiology Laboratory, Program in Cognitive Neuroscience and Schizophrenia, Nathan Kline Institute for Psychiatric Research, 140 Old Orangeburg Road, Orangeburg, NY 10962, USA b Department of Psychology, The City College of the City University of New York, 1338th St., Convent Avenue, New York, NY 10031, USA c Department of Neuroscience, 1300 Morris Park Avenue, Albert Einstein College of Medicine, Bronx, NY 10461, USA d Department of Psychiatry and Behavioral Science, 1300 Morris Park Avenue, Albert Einstein College of Medicine, Bronx, NY 10461, USA e Department of Psychiatry, New York University Medical Center, 550 1st Avenue, New York, NY 10016, USA Abstract Integration of information from multiple senses is fundamental to perception and cognition, but when and where this is accomplished in the brain is not well understood. This study examined the timing and topography of cortical auditory–visual interactions using high-density event-related potentials (ERPs) during a simple reaction-time (RT) task. Visual and auditory stimuli were presented alone and simultaneously. ERPs elicited by the auditory and visual stimuli when presented alone were summed (‘sum’ ERP) and compared to the ERP elicited when they were presented simultaneously (‘simultaneous’ ERP). Divergence between the ‘simultaneous’ and ‘sum’ ERP indicated auditory–visual (AV) neural response interactions. There was a surprisingly early right parieto-occipital AV interaction, consistent with the finding of an earlier study [J. Cogn. Neurosci. 11 (1999) 473]. The timing of onset of this effect (46 ms) was essentially simultaneous with the onset of visual cortical processing, as indexed by the onset of the visual C1 component, which is thought to represent the earliest cortical visual evoked potential. The coincident timing of the early AV interaction and C1 strongly suggests that AV interactions can affect early visual sensory processing. Additional AV interactions were found within the time course of sensory processing (up to 200 ms post stimulus onset). In total, this system of AVeffects over the scalp was suggestive of both activity unique to multisensory processing, and the modulation of ‘unisensory’ activity. RTs to the stimuli when presented simultaneously were significantly faster than when they were presented alone. This RT facilitation could not be accounted for by probability summation, as evidenced by violation of the ‘race’ model, providing compelling evidence that auditory–visual neural interactions give rise to this RT effect. 2002 Elsevier Science B.V. All rights reserved. Theme: Neural basis of behavior Topic: Cognition Keywords: Multisensory; Visual; Auditory; Electrophysiology; ERPs 1. Introduction nation of inputs from different senses can function to reduce perceptual ambiguity (e.g. Ref. [32]) and enhance Everyday tasks involve the seemingly automatic integra- stimulus detection (e.g. Ref. [53]). Despite the fundamental tion of information from multiple sensory modalities. For role that sensory integration plays in performance and instance, driving a car involves the synthesis of visual perception, how and when information from separate (seeing the road), auditory (hearing the car engine; a sensory modalities comes together in the human neocortex passing car), somatosensory (feeling the steering wheel), is not well understood. The bulk of our knowledge on the and motor (depressing the gas pedal) activity. The combi- mechanisms of multisensory integration in the brain comes from the pioneering research of Stein, Meredith, and co- workers (see Ref. [54] for a review) on multisensory *Corresponding author. Tel.: 11-845-398-6547; fax: 11-845-398- processing in the superior colliculus (of anaesthetized cats), 6545. E-mail address: [email protected] (J.J. Foxe). a sub-cortical structure involved in orienting to auditory, 0926-6410 / 02 / $ – see front matter 2002 Elsevier Science B.V. All rights reserved. PII: S0926-6410(02)00066-6

Multisensory auditory–visual interactions during early sensory

Embed Size (px)

Citation preview

Page 1: Multisensory auditory–visual interactions during early sensory

Cognitive Brain Research 14 (2002) 115–128www.elsevier.com/ locate /bres

Research report

Multisensory auditory–visual interactions during early sensoryprocessing in humans: a high-density electrical mapping study

a,b a,b a a,eSophie Molholm , Walter Ritter , Micah M. Murray , Daniel C. Javitt ,a,c a,b,c,d ,*Charles E. Schroeder , John J. Foxe

aCognitive Neurophysiology Laboratory, Program in Cognitive Neuroscience and Schizophrenia, Nathan Kline Institute for Psychiatric Research,140 Old Orangeburg Road, Orangeburg, NY 10962, USA

bDepartment of Psychology, The City College of the City University of New York, 1338th St., Convent Avenue, New York, NY 10031, USAcDepartment of Neuroscience, 1300 Morris Park Avenue, Albert Einstein College of Medicine, Bronx, NY 10461, USA

dDepartment of Psychiatry and Behavioral Science, 1300 Morris Park Avenue, Albert Einstein College of Medicine, Bronx, NY 10461, USAeDepartment of Psychiatry, New York University Medical Center, 550 1st Avenue, New York, NY 10016, USA

Abstract

Integration of information from multiple senses is fundamental to perception and cognition, but when and where this is accomplished inthe brain is not well understood. This study examined the timing and topography of cortical auditory–visual interactions usinghigh-density event-related potentials (ERPs) during a simple reaction-time (RT) task.Visual and auditory stimuli were presented alone andsimultaneously. ERPs elicited by the auditory and visual stimuli when presented alone were summed (‘sum’ ERP) and compared to theERP elicited when they were presented simultaneously (‘simultaneous’ ERP). Divergence between the ‘simultaneous’ and ‘sum’ ERPindicated auditory–visual (AV) neural response interactions. There was a surprisingly early right parieto-occipital AV interaction,consistent with the finding of an earlier study [J. Cogn. Neurosci. 11 (1999) 473]. The timing of onset of this effect (46 ms) wasessentially simultaneous with the onset of visual cortical processing, as indexed by the onset of the visual C1 component, which is thoughtto represent the earliest cortical visual evoked potential. The coincident timing of the early AV interaction and C1 strongly suggests thatAV interactions can affect early visual sensory processing. Additional AV interactions were found within the time course of sensoryprocessing (up to 200 ms post stimulus onset). In total, this system of AV effects over the scalp was suggestive of both activity unique tomultisensory processing, and the modulation of ‘unisensory’ activity. RTs to the stimuli when presented simultaneously were significantlyfaster than when they were presented alone. This RT facilitation could not be accounted for by probability summation, as evidenced byviolation of the ‘race’ model, providing compelling evidence that auditory–visual neural interactions give rise to this RT effect. 2002Elsevier Science B.V. All rights reserved.

Theme: Neural basis of behavior

Topic: Cognition

Keywords: Multisensory; Visual; Auditory; Electrophysiology; ERPs

1. Introduction nation of inputs from different senses can function toreduce perceptual ambiguity (e.g. Ref. [32]) and enhance

Everyday tasks involve the seemingly automatic integra- stimulus detection (e.g. Ref. [53]). Despite the fundamentaltion of information from multiple sensory modalities. For role that sensory integration plays in performance andinstance, driving a car involves the synthesis of visual perception, how and when information from separate(seeing the road), auditory (hearing the car engine; a sensory modalities comes together in the human neocortexpassing car), somatosensory (feeling the steering wheel), is not well understood. The bulk of our knowledge on theand motor (depressing the gas pedal) activity. The combi- mechanisms of multisensory integration in the brain comes

from the pioneering research of Stein, Meredith, and co-workers (see Ref. [54] for a review) on multisensory*Corresponding author. Tel.: 11-845-398-6547; fax: 11-845-398-processing in the superior colliculus (of anaesthetized cats),6545.

E-mail address: [email protected] (J.J. Foxe). a sub-cortical structure involved in orienting to auditory,

0926-6410/02/$ – see front matter 2002 Elsevier Science B.V. All rights reserved.PI I : S0926-6410( 02 )00066-6

Page 2: Multisensory auditory–visual interactions during early sensory

116 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

visual, and somatosensory stimuli. However, the extent to The finding of an AV effect that onsets at 40 ms overwhich the multisensory mechanisms defined in the superior parieto-occipital scalp [24] suggests that AV effects cancolliculus generalize to cortical processes remains to be occur at about the same time that initial activation offully elucidated. Knowledge of the timing and anatomical primary visual cortex (V1) is usually assumed to occurdistribution of cortical multisensory processing is essential (45–55 ms as represented by the onset of the earliestto determining the roles that it plays in information cortical visual evoked potential, C1: e.g. Refs. [10,11,21]).processing. This surprisingly early latency finding suggests a model of

Generally, it has been assumed that cortical multisen- auditory–visual interaction in the cortex where auditorysory processing occurs relatively late, following extensive input, which reaches the cortex in less than half the time ofprocessing of sensory inputs, and that it occurs in higher visual input (9 to 15 ms: [9,59]), is transmitted fromorder cortical areas specialized for this purpose (e.g. Ref. auditory cortices to visual or nearby visually dominant[33]). This assumption can be partially attributed to: (1) a cortical areas, and consequently affects the early sensorybias resulting from the tradition of studying sensory processing of visual input.systems in isolation, and (2) animal studies that reveal The purpose of the present study was to advance themultisensory convergence in higher-order regions of the understanding of cortical multisensory processing by plac-parietal (e.g. Refs. [17,28,50]), temporal (e.g. Refs. ing early AV interactions within the temporal and topog-[5,12,27]), and frontal lobes (e.g. Refs. [3,61]) along with raphical framework of cortical sensory processing of thea general lack of corresponding studies demonstrating individual auditory and visual inputs. We first endeavoredconvergence in lower-tier cortical areas. However, recent to determine if the early AV effect reported in Giard andevidence suggests that multisensory processing occurs Peronnet [24] would be elicited using a simple task andduring initial sensory transmission, and in cortical areas basic stimuli. In Giard and Peronnet a relatively compli-that are usually held to be unisensory. An investigation by cated task and stimulus set were employed: on each trial,Schroeder et al. [47] in the caudomedial (CM) belt area of one of two tones was presented, and/or a permanentlythe auditory association cortex of awake behaving placed circle morphed into a horizontal or vertical ellipse.macaque monkeys, which gets direct input from primary Subjects made forced two-choice classifications of the sixauditory cortex (A1), showed auditory–somatosensory co- randomly occurring stimulus conditions. In contrast, in therepresentation. Critically, both the auditory and somato- present study, single visually presented disks and auditorysensory inputs to CM had characteristic feed-forward pure tones were presented either alone or simultaneously,patterns, with both inputs arriving first in layer 4 at about and subjects performed a speeded simple reaction-time12 ms post stimulus onset, strongly suggesting bottom-up task. Elicitation of the effect under these conditions, inmultisensory integration that occurs early in the sensory conjunction with its elicitation in the very different con-processing hierarchy. Functional imaging studies have ditions of Giard and Peronnet [24], would suggest thatsuggested multisensory effects in what have been classical- early cortical multisensory processing for auditory andly considered unisensory cortical areas [7,31], although the visual stimuli that onset simultaneously may be present forprevailing opinion is that these interactions represent a variety of stimuli and tasks. We then compared the onsetfeedback from higher-tier multisensory onto the lower-tier of the earliest AV effect to the onset of C1 in response tounisensory areas. Direct empirical evidence of feedback- the visual stimulus alone. We expected that the initialmediated multisensory convergence in classical sensory cortical response to the visual stimulus would precede anycortex is sparse but supports this possibility (see Ref. [49]). AV effects, reflecting that cortical unisensory processing ofVery recently two event-related potential (ERP) studies the visual input began prior to cortical multisensoryfound surprisingly early multisensory effects that, in light processing.of their scalp topography, appear to indicate the early The technique of high-density electrical mapping (fromintegration of sensory information in traditionally held 128 scalp electrodes) was used to establish the spatio–unisensory cortex. In Giard and Peronnet [24], auditory– temporal dynamics of auditory–visual multisensory pro-visual (AV) effects were found to onset at just 40 ms over cessing in relation to activation across a distributed sensoryright parieto-occipital scalp; this is consistent with processing network. To assess multisensory processing,generators in early visual cortices, although the spatial ERPs to the ‘visual alone’ and ‘auditory alone’ stimulusresolution of ERPs does not allow the contribution of the conditions were summed (hereafter referred to as the ‘sum’abutting multisensory areas in posterior parietal cortex or ERP) and compared to the ERP to the simultaneouslysuperior temporal sulcus (STS), to be ruled out. And in presented auditory and visual stimuli (the ‘simultaneous’Foxe et al. [20], auditory–somatosensory effects onset at ERP). If neural responses to the auditory and visual inputsabout 50 ms over central /post-central scalp, consistent were processed in the same way when they were presentedwith generators in early somatosensory cortex; and at 70 simultaneously as when they were presented alone, then,ms over scalp areas consistent with neural activity from based on the principle of superposition of electrical fields,posterior auditory areas, in line with the findings of the ‘simultaneous’ ERP would be equivalent to the ‘sum’Schroeder et al. [47]. ERP. However, if the neural responses to the simultaneous-

Page 3: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 117

ly presented auditory and visual stimuli interacted during 2.2.2. Visual aloneprocessing, the ‘simultaneous’ and ‘sum’ ERPs would A disk (60 ms duration), subtending 1.28 in diameterdiverge. This method of measuring multisensory process- (143 cm viewing distance) and appearing red on a blacking is valid when neural responses reflect sensory process- background, was presented on a monitor (Iiyama Visioning unique to the stimulus, and do not reflect processes Master Pro 502, model [A102GT), peripheral to centralcommon to all three stimulus types such as target (e.g. the fixation. The location of the circle was chosen on anP3) or response (e.g. motor cortex activity) related neural individual subject basis after completion of the followingactivity. Several forms of interaction effects have been procedure. In a pre-test session, the red circle was pre-reported from this comparison (e.g. Refs. sented 100 times to each of 16 different locations across[20,24,38,40,55]). Although our primary focus was on the the visual field. The center of the stimulus was located 1.6earliest AV interaction, AV interactions up to 200 ms were or 2.6 degrees lateral from fixation and 1.6 or 2.0 degreesconsidered. vertical from fixation. For each subject, the location was

We also tested whether multisensory processing was chosen from which the best C1–P1 visual evoked po-reflected in our behavioral data. Simple reaction-times are tentials (VEPs) were observed. We sought to optimize ourgenerally facilitated when location concordant stimuli are sensitivity to the obligatory VEPs because cortical AVsimultaneously presented. This has been called the ‘re- effects have been previously reported in the latency rangedundant signal effect’ (RSE) (e.g. Refs. [29,37]). There are and general scalp location of C1 and P1 (40 to 90 mstwo classes of models to explain this effect: race models post-stimulus; [24]). There were four different locationsand coactivation models. In race models each stimulus of a used across our cohort (see Fig. 1).pair independently competes for response initiation, andthe faster of the two mediates the response for any trial. 2.2.3. Auditory and visual simultaneousAccording to this model probability summation produces The auditory-alone and visual-alone conditions de-the RSE, since the likelihood of either of the two stimuli scribed above were presented simultaneously. The auditoryyielding a fast reaction-time is higher than that from one and visual stimuli were in close proximity, with thestimulus alone. In coactivation models, the interaction of speaker placed on top of the monitor in vertical alignmentneural responses to the simultaneously presented stimuli with the visual stimulus.facilitates response initiation and produces the RSE. Wetested whether the RSE exceeded the statistical facilitation 2.3. Procedurepredicted by the race model, and thereby provided evi-dence for the contribution of AV neural interactions to RT Participants were seated in a comfortable chair in afacilitation. dimly lit and electrically shielded (Braden Shielding

2. Materials and methods

2.1. Subjects

Twelve neurologically normal, paid volunteers partici-pated (mean age 23.862.69 S.D.; five female; 11 right-handed), all reported normal hearing and normal orcorrected-to-normal vision. The Institutional Review Boardof the Nathan Kline Institute for Psychiatric Researchapproved the experimental procedures, and each subjectprovided written informed consent. Data from two addi-tional subjects were excluded, one for excessive blinking,and the other for failure to maintain central fixation.

2.2. Stimuli

Fig. 1. The 16 possible locations of the visual stimuli, with the number ofsubjects for whom a given location was used in the main experiment. The2.2.1. Auditory aloneeccentricities of the stimulus locations with respect to central locationA 1000 Hz tone (60 ms duration; 75 dB SPL, 5 mswere 1.568 for the upper left and right locations just above fixation, 2.338

rise / fall times) was presented from a single JBL speaker for the higher upper left location, and 3.288 for the far lower left location.located atop the monitor on which the visual stimulus was The far lower right location was used for all subjects in the follow-uppresented. experiment, and had an eccentricity of 3.288.

Page 4: Multisensory auditory–visual interactions during early sensory

118 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

Systems) room and asked to keep head and eye movements termine if a co-activation explanation of the RSE wasto a minimum, while maintaining central fixation. Eye necessary, Miller’s test of the race model [37] wasposition was monitored with horizontal and vertical elec- implemented. The model places an upper limit on thetro-oculogram (EOG) recordings. Subjects were instructed cumulative probability (CP) of RT at a given latency forto make a button-press response with their right index stimulus pairs. For any latency, t, the race model holdsfinger when a stimulus in either sensory modality was when this CP value is less than or equal to the sum of thedetected, as quickly as possible without making errors. The CP from each of the single stimuli minus an expression ofthree stimulus conditions were presented with equal prob- their joint probability (CP #((CP 1(t )simultaneous (t )alone 1

ability in random order. Stimulus onset asynchrony (SOA) CP )2(CP 3CP )). For each subject(t )alone 2 (t )alone 1 (t )alone 2

varied randomly between 750 and 3000 ms. Stimuli were the RT range within the valid RTs (100–800 ms) wasblocked into sequences of 150 trials, and each subject calculated over the three stimulus conditions and dividedcompleted a minimum of 18 blocks. Breaks were encour- into quantiles from the first to the hundredth percentile inaged between blocks to maintain high concentration and 5% increments (1%, 5%, . . . , 95%, 100%). T-tests com-prevent fatigue. paring the actual (CP ) and predicted(t )simultaneous

((CP 1CP )2(CP 3CP )) facili-(t )alone 1 (t )alone 2 (t )alone 1 (t )alone 2

2.4. Data acquisition and analysis tation were performed to assess the reliability of violationsof the race model.

Continuous EEG was acquired from 128 scalp elec-trodes (impedances ,5 kV), referenced to the nose, band-pass filtered from 0.05 to 100 Hz, and digitized at 500 Hz. 2.5.2. Event-related potentials: AV effectsThe continuous EEG was divided into epochs (2100 ms To date only Giard and Peronnet [24] have measured AVpre- to 800 ms post-stimulus onset) and baseline corrected multisensory processing by comparing ‘simultaneous’ andover the full 900 ms. Trials with blinks and eye move- ‘sum’ ERPs. In the first phase of our analysis we used theirments were rejected off-line on the basis of the EOG. An findings to predefine the latencies and scalp regions atartifact criterion of 660 mV was used at all other scalp which we tested for AV interactions. As such, we tested forsites to reject trials with excessive EMG or other noise an early AV interaction over parieto-occipital scalp withintransients. The average number of accepted sweeps per the latency range of the C1, AV interactions within thestimulus condition was 681 (ranging from 661 to 702 latency range of the auditory N1 and P2 and the visual P1sweeps over the three conditions). and N1 components, over corresponding scalp areas, and

EEG epochs were sorted according to stimulus condition an AV effect over right fronto-temporal scalp at about 150and averaged from each subject to compute the event- ms. For each of the six tests the mean amplitude of therelated potential (ERP). Baseline was defined as the epoch ‘simultaneous’ and ‘sum’ ERPs over a 10 ms window wasfrom 250 ms to 10 ms post-stimulus onset. In addition, submitted to a three-way repeated measures analysis ofERPs from the auditory-alone and visual-alone conditions variance (ANOVA) with factors of stimulus type (‘simulta-were summed for statistical comparison with the ERP neous’ versus ‘sum’), hemisphere, and electrode (3 repre-response to the simultaneous auditory–visual condition. senting each hemisphere). The center of each of the latencyGroup-averaged ERPs for each of the three stimulus windows was chosen on the basis of the maximumconditions and the summed auditory and visual alone ERPs difference between the ‘simultaneous’ and ‘sum’ ERPswere calculated for display purposes and for identification within the prescribed latency range, and is indicated inof the auditory N1 and P2 and the visual C1, P1, and N1. parentheses in the Results section. Geisser–Greenhouse

Button press responses to the three stimulus conditions corrections were used in reporting P values when appro-were acquired during the recording of the EEG and priate. An alpha level of less than 0.05 was used for allprocessed offline. Responses falling between 100 and 800 statistical tests. For the reliable AV interactions, thems post stimulus onset were considered valid. This win- topography and distribution of the effect was inspected indow was used to avoid the double categorization of a the difference voltage map; when the focus of the effectresponse. differed from the scalp region tested, a follow-up test was

conducted.2.5. Statistical analyses The second phase of analysis was conducted to render a

full description of the spatio–temporal properties of the AV2.5.1. Behavioral interactions, and was considered exploratory. The ‘simulta-

For individual subjects, the percent hits and average neous’ and ‘sum’ ERPs were compared using point-wiseresponse-time were calculated and their reaction-time running paired t-tests (two-tailed), with an AV interactiondistributions were recorded for each stimulus condition. To defined as at least 10 consecutive data points meeting atest for the presence of an RSE, planned comparisons 0.05 alpha criterion (10 data points520 ms at a 500 Hzbetween each of the alone stimulus conditions and the digitization rate) (for use of running t-tests see Refs.simultaneous stimulus condition were performed. To de- [26,40]). This criterion meets stringent standards for

Page 5: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 119

Table 1assessing reliable effects, as posed by Guthrie and Buch-t-tests comparing the CP for the ‘simultaneous’ stimulus condition to thewald [26], when a large number of t-tests are calculatedCP predicted by the race model over the first six quantiles, in which the

across the electrode montage and epoch. This is a suitable ‘simultaneous’ CP exceeded the predicted CPalternative to Bonferroni correction for multiple compari-

Quantile CP: simultaneous CP: predicted t P,11sons, which would increase the likelihood of Type II errors1 0.001 0.000 1.49 0.08through overcompensation for Type I errors. This ‘post-2 0.002 0.001 1.23 0.12hoc’ phase of the analysis was critical because (1) it3 0.017 0.007 2.76 0.009

provided a description of the time course of the AV effects, 4 0.118 0.066 3.23 0.004and (2) the full description of the spatio–temporal prop- 5 0.328 0.253 2.61 0.012erties of the AV interactions allows these data to be used asa hypothesis generation tool for future multisensorystudies.

neous condition (for auditory alone condition compared to2.5.3. Event-related potentials: C1 onset the simultaneous condition: t 57.972, P,0.001; for the11

C1 onset was measured by comparing the amplitude of visual alone condition compared to the simultaneousthe ERP to the visual alone condition to baseline using condition: t 511.057, P,0.001).11

point-wise running paired t-tests (two-tailed); onset wasdefined as the first point where a 0.05 alpha criterion wasmet and followed by at least 10 consecutive significant 3.1.2. Test of the race modeldata points. There was violation of the race model for each of the 12

subjects. The CP at each quantile were group averagedseparately for each stimulus condition to form an aggregate

3. Results distribution that preserved the shape of the individuals’data (also termed Vincent averaging; [60]). For the group

3.1. Behavioral averaged data, the race model was violated over the firstsix quantiles (see Fig. 2). These violations were reliable

3.1.1. Redundant signal effect across subjects, with statistically significant differencesMean reaction-times to the simultaneous condition (255 between the actual (CP ) and predicted(t )simultaneous

ms) were faster than mean reaction-times to either the ((CP 1CP )2(CP 3CP )) facili-(t )alone 1 (t )alone 2 (t )alone 1 (t )alone 2

visual or the auditory alone conditions (305 and 297, tation for three of these quantiles, and approaching signifi-respectively). An RSE was confirmed with planned com- cance for two (see Table 1). Violation of the race modelparisons of each of the alone conditions to the simulta- strongly supports the contention that the interaction of

Fig. 2. (a) Mean (N512) reaction times (S.E.M. indicated) for simultaneous (AV: red bar), and auditory (A: blue bar) and visual (V: green bar) alonestimulus conditions. (b) Cumulative probability (CP) distributions for the simultaneous (red trace) and visual (green trace) and auditory (blue trace) alonestimulus conditions, and the CP predicted by the race model (black trace). (c) Miller inequality: values greater than zero signify violation of the race model(seen here over the first six quantiles).

Page 6: Multisensory auditory–visual interactions during early sensory

120 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

auditory and visual stimulus information during neural 3.2.1. AV interactionsprocessing contributed to the RSE. Differences between the group averaged ‘simultaneous’

and ‘sum’ ERPs could be readily observed. For theelectrode sites over parieto-occipital to occipital scalp, and

3.2. Event-related potentials in the posterior temporal regions, the ‘simultaneous’ ERPwas more positive in amplitude from about 45 to 80 ms

In Fig. 3 the group averaged ERPs to the alone compared to the ‘sum’ ERP, and had a smaller peakconditions (auditory and visual), and the simultaneous amplitude for the positive going wave at about 100 ms andcondition at selected electrode sites are displayed. The the negative going wave at about 160 ms. At the centralERP elicited by the auditory alone condition was char- and fronto-central electrode sites the ‘simultaneous’ ERPacterized by the typical auditory components Pa peaking at was of greater amplitude than the ‘sum’ ERP for the40 ms, P1 peaking at 56 ms, N1 peaking at 105 ms, and P2 negative going wave at about 120 ms, and for the positivepeaking at about 200 ms [43,57,58]. Pa and N1 appeared going wave at about 180 ms.maximal over fronto-central scalp and inverted at themastoids; P1 appeared maximal at fronto-central sites but 3.2.2. Posterior AV interactionsdid not invert at the mastoids; and P2 appeared maximal The early right parieto-occipital difference between theover centro-parietal scalp. The ERP elicited by the visual ‘simultaneous’ and ‘sum’ ERPs was confirmed with aalone condition was characterized by the visual C1 from significant stimulus type by hemisphere interaction (cen-about 40 to 65 ms, and a typical P1 and N1 peaking at 110 tered at 58 ms: F(1,11)55.73, P50.036); the ‘simulta-and 160 ms, respectively [10,21]; P1 appeared maximal neous’ ERP was more positive than the ‘sum’ ERP (seeover occipital scalp and N1 over occipito-temporal scalp. Fig. 4a). A follow up focused contrast over the rightThe ERP elicited by the ‘simultaneous’ condition dis- electrode sites, with factors of stimulus type and electrode,played classic auditory and visual componentry, in general showed a main effect of stimulus type (F(1,11)58.83,as described for the alone conditions. P50.013).

Fig. 3. The ‘simultaneous’ (red trace), and visual (green trace) and auditory (blue trace) alone ERPs at left (a) and right (b) fronto-central electrode sitesand at left (c) and right (d) occipital electrode sites. The placement of the electrode sites are indicated in magenta in top and back views of the electrodemontage.

Page 7: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 121

Fig. 4. Auditory–visual interactions are illustrated in ‘simultaneous’ versus ‘sum’ difference voltage maps of the focus of the AV interactions, at their peaklatencies. The color scale is to the left of Fig. 4a, and the voltage scale used for each of the maps is indicated below it, in gray. These are accompanied byERP traces to the ‘simultaneous’ (red), ‘sum’ (blue), and ‘simultaneous’ versus ‘sum’ difference (green) from corresponding electrode sites. The electrodelocation is indicated with a black dot on the voltage map. A vertical black line is drawn through the traces at the latency of the corresponding voltage map.For each AV interaction voltage map, a unisensory voltage map (visual or auditory) at the same latency is illustrated, to the far right. Comparison of the AVinteractions and corresponding unisensory voltage maps reveals the similarity or difference of the interaction to unisensory processing.

In the latency range and scalp region of the occipitally concluded that the AV effects in this latency range mainlyfocused visual P1, the ‘simultaneous’ ERP was signifi- reflected interactions over left centro-parietal scalp.cantly less positive going than the ‘sum’ ERP (centered at In the latency range and scalp area of the occipito-120 ms: F(1,11)55.89, P50.034). However, inspection of temporally focused visual N1 the ‘simultaneous’ ERP wasthe voltage map suggested that the AV interaction in this less negative going than the ‘sum’ ERP (see Fig. 4f), withlatency range was largely focused over left centro-parietal a two-way interaction of stimulus type by electrode thatscalp (see Fig. 4c). A post-hoc two-way repeated measures approached significance (centered at 165 ms: F(1,11)5ANOVA on left centro-parietally placed electrodes sup- 4.22, P50.061). The ‘simultaneous’ versus ‘sum’ differ-ported this contention (F(1,11)515.69, P50.002) and we ence was focused over the right scalp. Given the strong

Page 8: Multisensory auditory–visual interactions during early sensory

122 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

Fig. 4. (continued)

Page 9: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 123

trend toward significance we did a post hoc contrast on the 3.2.4. Exploratory analysis of the spatio–temporalright hemisphere electrode sites, with factors of stimulus properties of the AV interactionstype and electrode; this showed a main effect of stimulus The P-values from the running t-test analysis comparingtype (F(1,11)56.31, P50.029). the amplitudes of the ‘simultaneous’ and ‘sum’ ERPs

across the electrode montage and epoch provide a com-3.2.3. Fronto-central /central AV interactions prehensive picture of the spatio–temporal properties of the

In the latency range and scalp region of the fronto- AV interactions (Fig. 5). The earliest AV interaction onsetcentrally focused auditory N1, the amplitude of the at 46 ms, was evident only over the right parieto-occipital‘simultaneous’ ERP was significantly more negative going region of the scalp, and lasted for 20 ms (also see Fig. 4a).than the amplitude of the ‘sum’ ERP (centered at 120 ms: This was followed by an AV interaction, not identified inF(1,11)510.20, P50.009). However, inspection of the our first phase of analysis, focused over frontal scalp thattopography of the effect suggested that it was largely due onset at 86 ms and lasted for up to 26 ms (also see Fig.to the left centro-parietally focused AV interaction (see Fig. 4b). Next, there were simultaneous AV interactions focused4c, and the results from the post-hoc ANOVA on left over left centro-parietal and right posterior temporalcentro-parietal sites, centered at 120 ms, reported above). regions of the scalp. Left centro-parietally it onset at 102

In the latency range and scalp area of the centrally ms, and right posterior temporally it onset at 100 ms; thesefocused auditory P2 the ‘simultaneous’ ERP was sig- evolved into AV interactions focused over left centro-nificantly more positive going than the ‘sum’ ERP (see Fig. parietal scalp and right superior temporal scalp that peaked4g) (centered at 180 ms: F(1,11)55.23, P50.043). Inspec- at 128 ms (also see Fig. 4c, d, and e). These effects did nottion of the simultaneous and sum ERPs over fronto- appear to be a modulation of the auditory N1. Starting attemporal scalp suggested that there were no amplitude 160 ms there was an AV interaction focused over the rightdifferences focused over this area (centered at 150 ms: occipito-temporal scalp that lasted until at least 180 ms,F(1,11)51.03, P50.331). and this difference had a topography and time course

Fig. 5. Significant P values over time for 119 electrodes from running t-tests comparing the ‘simultaneous’ and ‘sum’ ERPs (this excluded four mastoidelectrodes, four non-scalp electrodes that recorded ocular activity, and one bad electrode). P values are differentiated with a color scale. Grey indicates anabsence of significant P values. Time is plotted on the x axis from 0 to 200 ms. Electrodes are plotted on the y axis. Starting from the bottom of the graph,the electrodes are divided into sections from posterior to anterior scalp (O, occipital; P-O, parieto-occipital; P, parietal; C, central; F-C, fronto-central; F,frontal). Within a section the electrodes are arranged from the right most lateral to the left most lateral sites (indicated to the right of the graph for eachsection). The sections are labeled based on the mid line of the scalp (e.g. ‘central’ includes electrodes over temporal scalp).

Page 10: Multisensory auditory–visual interactions during early sensory

124 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

consistent with the visual N1 (also see Fig. 4f). Finally, 3.3. C1 onsetthere was a central AV interaction focused over the leftscalp that onset at 174 ms and extended out to 200 ms post The early AV interaction fell within the time range ofstimulus onset (also see Fig. 4g). We should note that the when the initial response in primary visual cortices,two latter AV interactions were present at only 3 to 4 indexed by C1, generally occurs (e.g. Refs. [10,11,21]). Weelectrode sites, while the other AV interactions were therefore set out to compare the onset of the early AVpresent at many electrode sites clustered within a scalp effect with the onset of C1. However, the C1 in the groupregion. averaged ERP to the visual alone stimulus condition was

of relatively small amplitude, and activation did notreliably onset until the P1 portion of the ERP, at 76 ms (at

3.2.5. The present multisensory effects are not an occipital sites). The relatively small amplitude C1 wasartifact of slow wave anticipatory potentials likely due to our use of multiple visual stimulus locations

The performance of a task in our experimental paradigm across subjects. That is, because of the anatomy of theintroduced the possibility of slow wave anticipatory po- retinotopically arranged human visual cortex, the scalptentials (e.g. the contingent negative variation; see Ref. topography of C1 changes as a function of stimulus[62]). These slow potentials begin several hundred milli- position [10]. Thus, C1s to the same visual stimulusseconds prior to stimulus onset and continue for a short presented to different locations would have different scalptime thereafter. If present, anticipatory potentials would be topographies, and when averaged together the C1 wouldrepresented twice in the ‘sum’ ERP and only once in the appear diminished.‘simultaneous’ ERP, resulting in an artificial difference To get a more accurate measure of the onset of C1 tobetween the two. Problematically, this difference would our visual stimulus, we presented it to a single location forappear to begin post-stimulus onset, due to baseline an additional seven subjects (mean age 2567.9 S.D.; fivecorrection on the pre-stimulus portion of the ERPs, and female; five right-handed; one also participated in the mainwould be indistinguishable from a genuine AV interaction experiment). Only differences in procedure between this

1[56]. and the main experiment are noted. The stimulus wasThe current experiment was designed to minimize the always presented to the lower right quadrant of the visual

incidence of anticipatory potentials. SOA was varied field (2.6 degrees lateral and 2.6 degrees vertical fromrandomly between 750 and 3000 ms such that subjects fixation). To maintain arousal a simple oddball paradigmcould not accurately predict the arrival of the next was used in which subjects made a right-handed button-stimulus, and hence would be discouraged from anticipat- press response to an infrequently occurring change in coloring its onset (see Ref. [13]). To confirm that anticipatory (5% of trials). SOA varied randomly between 750 andpotentials did not contribute to the reported AV effects we 1000 ms. Stimuli were blocked into sequences of 100compared the onset of our earliest effect for two different trials, and the first three subjects completed four blocks ofbaseline criteria. If the effect was a consequence of trials while the remaining subjects each completed eightanticipatory potentials, then its onset latency should shift blocks of trials. The continuous EEG was divided into 100for different baseline criteria [56]. The early right parieto- ms pre-stimulus to 200 ms post-stimulus onset epochs.occipital effect onset at the same latency of 46 ms for the These were averaged for each subject to compute the ERP.two baseline criteria: 100 to 50 ms pre-stimulus onset; and Baseline was defined as the pre-stimulus epoch. Theour original 50 ms pre-stimulus to 10 ms post-stimulus average number of accepted sweeps was 329629 for theonset. This analysis strongly suggests that anticipatory first three subjects and 721621 for the remaining subjects.potentials did not contribute significantly to our AV effects. The group averaged ERP exhibited a C1 that was largest

To further rule out the contribution of slow wave over occipital scalp and appeared to diverge from baselineanticipatory potentials to our AV interactions, we inspected at |40 ms. The t-test analysis showed that C1 onset at 44the grand mean ‘simultaneous’ and ‘sum’ ERPs for ms over left occipital scalp and at 46 ms over central anddifferences in the pre-stimulus portion of the waves. The right occipital scalp. Thus, with a temporal resolution ofdata were reprocessed to look at a longer pre-stimulus essentially 2 ms, the onset latencies of C1 and the early AVepoch (250 ms pre-stimulus to 200 ms post-stimulus interaction were the same.onset). No pre-stimulus differences were observed atanterior electrode sites, and only minimal pre-stimulusdifferences were observed at posterior electrode sites, 4. Discussionagain suggesting that anticipatory potentials did not con-tribute to our AV effects. The present study examined the spatial and temporal

properties of cortical AV interactions to basic stimuli,while subjects performed a simple reaction-time task. In

1 the behavioral data there was a significant reaction-timeWe thank Dr. Steven Hillyard for pointing out this potential problem atthe 2001 CNS conference. advantage when the visual and auditory stimuli were

Page 11: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 125

presented simultaneously compared to when they were ions. Further, this earliest AV interaction has a differentpresented alone—the so-called redundant sensory effect scalp topography to that of the C1, with a focus that is(RSE). There was substantial violation of the race model. more dorsal over the right parieto-occipital scalp. Thus, onHence the RSE could not be accounted for by simple the basis of both timing and topography, we suggest thatprobability summation, providing strong evidence that an early stage of the dorsal visual stream is the likely sitemultisensory neural interactions are responsible for this of this AV interaction. Visual activation of such areas issizeable RSE. essentially simultaneous with activation of V1 [46]. The

auditory input, on the other hand, conceivably arrived4.1. The early AV effect through either feedforward or feedback connections, as

auditory cortices are activated by tones of moderateThe electrophysiological data exhibited surprisingly intensity by at most 15 ms [9].

early auditory–visual multisensory interactions, over right This finding of an early AV effect contributes to recentparieto-occipital scalp. The combination of the early evidence [20,24,47,49] of multisensory interactions duringtiming of this effect and its scalp topography over early early sensory processing. Models of sensory integrationvisual areas is consistent with modification of very early need to account for such early multisensory operations invisual sensory processing by auditory inputs. As unisen- addition to later multisensory processing that occurssory transmission to auditory cortex is considerably faster following the extensive sensory analysis of the individualthan it is to visual cortex, this suggested the following sensory inputs [33,39]. For example, based on functionalpossible interpretation to us. Firstly, auditory input acti- imaging data, Calvert et al. [7] and Driver and Spence [16]vates primary auditory cortex within 15 ms of stimulus suggest feedback accounts of multisensory processing inpresentation and is then transmitted up the auditory unisensory cortex, although they do not specify the point atprocessing stream. This input is then projected to ‘un- which this occurs in information processing.isensory’ visual areas through one of two possible path-ways. 4.1.2. Candidate areas for initial integration of

The first possibility is a direct, feedforward projection multisensory inputsfrom primary auditory or auditory association cortices to The right parieto-occipital focus of the earliest AVearly visual cortices. Until recently, it was largely believed interaction is consistent with sources originating from onethat such connectivity did not exist. However, very recent- of several possible underlying brain areas: regions of the

2ly, anatomic tracer studies have provided evidence for visual motion processing system (referred to here asdirect projections from both primary auditory cortex (A1) MT1, which includes V5 and related areas), and knownand auditory association areas to both V1 and V2 in multisensory regions of the posterior parietal cortex andmacaques [18,44]. The second possibility is a more the superior temporal sulcus (STS). Of these three, STS isindirect pathway, in which feedforward auditory input the most parsimonious with the detailed neurophysiologyreaches areas of AV multisensory convergence (e.g. the of visual processing in the macaque monkey. In thesuperior temporal polysensory region—STS), and is trans- monkey, STS receives visual inputs only milliseconds aftermitted via feedback connections to earlier ‘unisensory’ V1 (|3 ms; [46,49], and there are cells in STS that receivevisual areas. The critical issue is one of timing. The converging input from auditory cortices [3,5,25]. Further,question is whether there is sufficient time for auditory human fMRI studies implicate STS in auditory–visualinput to reach early visual areas to result in modulation of multisensory processing (e.g. Refs. [8,41]). However,the later arriving visual input? Given the standard timing while its auditory–visual response characteristics makeoffset between the initial auditory and visual inputs to their STS a strong candidate for the source of the early AVrespective primary cortices, there is a more than sufficient interaction, the topography of the interaction observed herewindow of |25–30 ms in which this process could occur and in Giard and Perronet [24] seems posterior to what one(see e.g. Refs. [46,48,49]). might predict for a generator in STS.

MT1 is consistent with the topography of the early4.1.1. Timing and what it tells us interaction, and there is strong evidence from monkey

The onset of an AV interaction at 46 ms in our study is intracranial data that the arrival of the fastest inputs to V5much earlier in information processing than multisensory occurs at the same time as, and sometimes even earlierprocessing is generally assumed to occur, and reinforces than, initial inputs to V1 [45,46,48]; and human data havethe view that visual processing is modified by auditory found similar timing [6,22]. Further supporting the possi-inputs well before visual sensory analysis is complete. The bility that MT1 contributed to the early AV interaction,C1 component of the VEP, our measure of initial input to multisensory processing in human MT1 for auditory–visual cortices, and the AV effect, were found to onsetessentially simultaneously. This timing of AV interactions

2relative to C1 onset suggests that the visual input to this Transient stationary stimuli, such as ours, activate the visual motionmultisensory process arrives through feedforward connect- processing system (e.g. Schroeder et al. [46]).

Page 12: Multisensory auditory–visual interactions during early sensory

126 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

visual speech has been shown with fMRI (Ref. [7]; also response used in this study would be expected to elicit asee Ref. [4]). Although monkey studies have failed to left lateralized motor response.convincingly show that MT1 responds to auditory stimuli In addition to the above, post hoc analyses suggested(e.g. unpublished data from our lab), it is possible that such two further AV interactions. One was over frontal scalpmultisensory cells in the macaque have not been sampled. (|100 ms), and was unique to the ‘simultaneous’ stimulus

Also consistent with the topography of the early inter- condition (i.e. did not correspond to activity in response toaction is posterior parietal cortex, which is activated by a either the visual or auditory alone stimulus conditions).visual stimulus just a few milliseconds following V1 Multisensory processing in the frontal cortices has beenactivation [46,49]. Human and monkey studies have shown shown previously in both human and monkey studies [3,8].that posterior parietal areas are involved in the multisen- The other was focused over right occipito-temporal areassory representation of location and motion [2,30]. Further, (|165 ms). The timing and scalp topography of this effectmonkey studies indicate that these areas include neurons was consistent with those of the visual N1, elicited by thethat respond to both auditory and visual inputs early in visual alone stimulus condition. Auditory-based modula-information processing (e.g. Ref. [34]; and see Ref. [2] for tion of the visual N1 would suggest that auditory inputsreview). A caveat is that such convergence of auditory and can affect the visual object recognition system, as thevisual inputs has only been found when monkeys have visual N1 has been implicated in visual object recognitionbeen extensively trained to make saccades to the auditory processes by many studies (e.g. Refs. [1,14,15]).stimuli [25].

4.1.3. Functional significance of early AV integration 5. Summary and conclusionsMT1 and the posterior parietal cortices are involved in

the processing of motion, and are implicated in exogenous- Our electrophysiological data are in general agreemently (stimulus) driven attentional orienting to moving stimuli. with those reported by Giard and Peronnet [24], the onlyConsistent with the right hemisphericity of the early AV comparable ERP study to date. Both studies revealed AVinteraction, the exogenous attentional orienting systems in interactions of the same polarity over right parieto-occipi-posterior parietal cortices are generally localized to the tal scalp (|40–50 ms), over occipito-temporal scalp (|165right hemisphere (e.g. Refs. [36,42]). The rapid detection ms), and over fronto-central (in Giard and Peronnet [24])and localization of moving objects is advantageous to and central (the present study) scalp (|180 ms). However,survival, for example allowing one to dodge a hurtling our AV effects were of considerably shorter duration thanobject whistling through the air, and would be enhanced by those reported in Giard and Peronnet. Additionally, whilethe early integration of temporally coincident auditory and we also found multisensory effects in the latency range ofvisual information in motion sensitive areas. Behavioral the auditory N1 and visual P1 components, high-densitystudies have shown that detection and localization of visual topographic mapping showed that these effects had differ-objects are improved when there is a temporally co- ent topographies than the unisensory components, sug-incident, location concordant, sound [52,53]. gesting different underlying source configurations for these

multisensory effects. Furthermore, our data did not evi-4.2. The later AV effects dence fronto-temporal activity at about 150 ms as was

found in Giard and Peronnet. Of course, due to theA centro-parietal AV interaction peaked at 120 ms. This considerable differences in the paradigms used, it is not

interaction overlapped with the time course of both the surprising that some differences were seen. Rather, givenauditory N1 and the visual P1. However, its centro-parietal the general similarities in findings across these two studies,topography suggested a different configuration of we suggest that AV multisensory processing comprises agenerators. This evolved into more lateralized activity base set of cortical interactions that will be present for afocused over right superior temporal and left centro-pariet- variety of stimuli and tasks.al scalp that peaked at about 128 ms; these AV interactions These data contribute to the growing literature on thewere distinct from the activity in response to the auditory spatio–temporal properties of cortical multisensory pro-or visual alone stimulus conditions within the same latency cessing in humans. The extent to which multisensoryrange (see Fig. 4d for comparisons of the interactions with activity is unique to particular combinations of the sensesthe activity to the auditory alone condition at the same versus common to various sensory combinations, and thelatency). impact of task effects on the observed activities, remains to

A later AV interaction, over left central scalp (|180 ms), be seen. For example, it has been well-documented thatwas consistent with generators in motor cortex and may there can be profound crossmodal influences on attentionalreflect sensory motor integration. The timing of the effect orienting (e.g. Refs. [19,23,35,51]). Further, the currentis consistent with such an interpretation, with the mean RT data highlight the importance of a detailed consideration ofto the simultaneously presented stimuli occurring |75 ms the time course of sensory processing within the individuallater, as is the scalp topography since the right handed sensory modalities. The inherent differences in transmis-

Page 13: Multisensory auditory–visual interactions during early sensory

S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128 127

ping of perceptual closure processes, J. Cogn. Neurosci. 12 (2000)sion time to cortex of stimulation within the different615–621.sensory modalities will likely have considerable effects on

[15] G.M. Doniger, J.J. Foxe, C.E. Schroeder, M.M. Murray, B.A.the time course and potential neural areas at which Higgins, D.C. Javitt, Visual perceptual learning in human objectmultisensory interactions will occur. based recognition areas: a repetition priming study using high-

density electrical mapping, Neuroimage 13 (2001) 305–313.[16] J. Driver, C. Spence, Multisensory perception: beyond modularity

and convergence, Curr. Biol. 10 (2000) R731–735.[17] J.R. Duhamel, C.L. Colby, M.E. Goldberg, Congruent representationAcknowledgements

of visual and somatosensory space in single neurons of monkeyventral intraparietal cortex (VIP), in: J. Paillard (Ed.), Brain and

Sincere appreciation to Ms. Beth Higgins and Ms. Space, Oxford University Press, 1991, pp. 223–236.[18] A. Falchier, L. Renaud, P. Barone, H. Kennedy, Extensive projec-Deirdre Foxe for excellent technical assistance. Work

tions from the primary auditory cortex and polysensory area stp tosupported in part by grants from the NIH—NS30029-23peripheral area v1 in the macaque, Soc. Neurosci. Abs. 27 (2001),(WR), MH63434 (JJF) and MH61989 (CES). Program no. 511.21.

[19] J.J. Foxe, G.V. Simpson, S.P. Ahlfors, Cued shifts of intermodalattention: parieto-occipital |10 Hz activity reflects anticipatory stateof visual attention mechanisms, Neuroreport 9 (1998) 3929–3933.

[20] J.J. Foxe, I.A. Morocz, B.A. Higgins, M.M. Murray, D.C. Javitt,ReferencesC.E. Schroeder, Multisensory auditory–somatosensory interactionsin early cortical processing revealed by high density electrical

[1] T. Allison, A. Puce, D. Spencer, G. McCarthy, Electrophysiologicalmapping, Cogn. Brain Res. 10 (2000) 77–83.

studies of human face perception I: potentials generated in oc- [21] J.J. Foxe, G.V. Simpson, Flow of activation from V1 to frontal cortexcipitotemporal cortex by face and non-face stimuli, Cereb. Cortex 9 in humans: a framework for defining ‘early’ visual processing, Exp.(1999) 415–430. Brain Res. 142 (2002) 139–150.

[2] R.A. Andersen, L.H. Snyder, D.C. Bradley, J. Xing, Multimodal [22] D.H. ffytche, C.N. Guy, S. Zeki, The parallel visual motion inputsrepresentation of space in the posterior parietal cortex and its use in into areas V1 and V5 of human cerebral cortex, Brain 118 (1995)planning movements, Annu. Rev. Neurosci. 20 (1997) 303–330. 1375–1394.

[3] L.A. Benevento, J. Fallon, B.J. Davis, M. Rezak, Auditory–visual [23] K.G. Fu, J.J. Foxe, M.M. Murray, B.A. Higgins, D.C. Javitt, C.E.interaction in single cells in the cortex of the superior temporal Schroeder, Cross-modality cued attention-dependent suppression ofsulcus and the orbital frontal cortex of the macaque monkey, Exp. distracter visual input indexed by anticipatory parieto-occipitalNeurol. 57 (1977) 849–872. alpha-band oscillations, Cogn. Brain Res. 12 (2001) 145–152.

[4] R.A. Berman, C.L. Colby, Both auditory and visual attention [24] M.H. Giard, F. Peronnet, Auditory–visual integration during mul-modulate motion processing in area MT1, Cogn. Brain Res. (this timodal object recognition in humans: a behavioral and electro-issue). physiological study, J. Cogn. Neurosci. 11 (1999) 473–490.

[5] C. Bruce, R. Desimone, C.G. Gross, Visual properties of neurons in [25] A. Grunewald, J.F. Linden, R.A. Andersen, Responses to auditorya polysensory area in superior temporal sulcus of the macaque, J. stimuli in macaque lateral intraparietal area I. Effects of training,Neurophysiol. 46 (1981) 369–384. Neurophysiology 82 (1999) 330–342.

[6] H. Buchner, R. Gobbele, M. Wagner, M. Fuchs, T.D. Waberski, R. [26] D. Guthrie, J.S. Buchwald, Significance testing of difference po-Beckmann, Fast visual evoked potential input into human area V5, tentials, Psychophysiology 28 (1991) 240–244.Neuroreport 28 (1997) 2419–2422. [27] K. Hikosaka, E. Iwai, H. Saito, K. Tanaka, Polysensory properties of

[7] G.A. Calvert, M.J. Brammer, E.T. Bullmore, R. Campbell, S.D. neurons in the anterior bank of the caudal superior temporal sulcusIversen, A.S. David, Response amplification in sensory-specific of the macaque monkey, J. Neurophysiol. 60 (1988) 1615–1637.cortices during crossmodal binding, Neuroreport 10 (1999) 2619– [28] J. Hyvarinen, Y. Shelepin, Distribution of visual and somatic2623. functions in the parietal associative area 7 of the monkey, Brain Res.

[8] G.A. Calvert, P.C. Hansen, S.D. Iversen, M.J. Brammer, Detection 169 (1979) 561–564.of audio–visual integration sites in humans by application of [29] R. Kinchla, Detecting target elements in multielement arrays: aelectrophysiological criteria to the BOLD effect, Neuroimage 14 confusability model, Percept. Psychophys. 15 (1974) 149–158.(2001) 427–438. [30] J.W. Lewis, M.S. Beauchamp, E.A. DeYoe, A comparison of visual

[9] G.G. Celesia, F. Puletti, Auditory input to the human cortex during and auditory motion processing in human cerebral cortex, Cereb.states of drowsiness and surgical anesthesia, Electroencephalogr. Cortex 10 (2000) 873–888.Clin. Neurophys. 31 (1971) 603–609. [31] E. Macaluso, C.D. Frith, J. Driver, Modulation of human visual

[10] V.P. Clark, S. Fan, S.A. Hillyard, Identification of early visual cortex by crossmodal spatial attention, Science 289 (2000) 1206–evoked potential generators by retinotopic and topographic analyses, 1208.Hum. Brain Map. 2 (1995) 170–187. [32] A. MacLeod, Q. Summerfield, A procedure for measuring auditory

[11] V.P. Clark, S.A. Hillyard, Spatial selective attention affects early and audio–visual speech-reception thresholds for sentences in noise:extrastriate but not striate components of the visual evoked potential, rationale, evaluation, and recommendations for use, Br. J. Audiol.J. Cogn. Neurosci. 8 (1996) 387–402. 24 (1990) 29–43.

[12] R. Desimone, C.G. Gross, Visual areas in the temporal cortex of the [33] D.W. Massaro, Speechreading: illusion or window into patternmacaque, Brain Res. 178 (1979) 363–380. recognition, Trends Cogn. Sci. 3 (1999) 310–317.

[13] E. Donchin, P. Tueting, W. Ritter, M. Kutas, E. Heffley, On the [34] P. Mazzoni, R.M. Bracewell, S. Barash, R.A. Andersen, Spatiallyindependence of the CNV and the P300 components of the human tuned auditory responses in area LIP of macaques performingaveraged evoked potential, Electroencephalogr. Clin. Neurophys. 38 delayed memory saccades to acoustic targets, J. Neurophysiol. 75(1975) 449–461. (1996) 1233–1241.

[14] G.M. Doniger, J.J. Foxe, M.M. Murray, B.A. Higgins, J.G. Snod- [35] J.J. McDonald, L.M. Ward, Involuntary listening aids seeing:grass, C.E. Schroeder, D.C. Javitt, Activation time-course of ventral evidence from human electrophysiology, Psychol. Sci. 11 (2000)visual stream object-recognition areas: high density electrical map- 167–171.

Page 14: Multisensory auditory–visual interactions during early sensory

128 S. Molholm et al. / Cognitive Brain Research 14 (2002) 115 –128

[36] M.M. Mesulam, A cortical network for directed attention and [50] B. Selzer, D.N. Pandya, Converging visual and somatic sensoryunilateral neglect, Ann. Neurol. 10 (1981) 309–325. cortical input to the intraparietal sulcus of the rhesus monkey, Brain

[37] J. Miller, Divided attention: evidence for coactivation with re- Res. 192 (1980) 339–351.dundant signals, Cogn. Psychol. 14 (1982) 247–279. [51] C. Spence, J. Ranson, J. Driver, Cross-modal selective attention: on

[38] C. Miniussi, M. Girelli, C.A. Marzi, Neural site of the redundant the difficulty of ignoring sounds at the locus of visual attention,target effect electrophysiological evidence, J. Cogn. Neurosci. 10 Percept. Psychophys. 62 (2000) 410–424.(1998) 216–230. [52] B.E. Stein, M.A. Meredith, W.S. Huneycutt, L. McDade, Behavioral

[39] J.R. Movellan, J.L. McClelland, The Morton–Massaro law of indices of multisensory integration: orientation to visual cues isinformation integration: implications for models of perception, affected by auditory stimuli, J. Cogn. Neurosci. 1 (1989) 12–24.Psychol. Rev. 108 (2001) 113–148. [53] B.E. Stein, N. London, L.K. Wilkinson, D.D. Price, Enhancement of

[40] M.M. Murray, J.J. Foxe, B.A. Higgins, D.C. Javitt, C.E. Schroeder, perceived visual intensity by auditory stimuli: a psychophysicalVisuo–spatial response interactions in early cortical processing analysis, J. Cogn. Neurosci. 8 (1996) 497–506.during a simple reaction time task: a high density electrical mapping [54] B.E. Stein, M.A. Meredith, The Merging of the Senses, MIT Press,study, Neuropsychologia 39 (2001) 828–844. Cambridge, MA, 1993.

[41] I.R. Olson, J.C. Gatenby, J.C. Gore, A comparison of bound and [55] S. Supek, C.J. Aine, D. Ranken, E. Best, E.R. Flynn, C.C. Wood,unbound audio–visual information processing in the human cerebral Single vs. paired visual stimulation: superposition of early neuro-cortex, Cogn. Brain Res. (this issue). magnetic responses and retinotopy in extrastriate cortex in humans,

[42] J.V. Pardo, P.T. Fox, M.E. Raichle, Localization of a human system Brain Res. 830 (1999) 43–55.¨ ¨for sustained attention by positron emission tomography, Nature 349 [56] W.A. Teder-Salejarvi, J.J. McDonald, F. Di Russo, S.A. Hillyard,

(1991) 61–64. An analysis of audio–visual crossmodal integration by means of[43] T.W. Picton, S.A. Hillyard, H.I. Krausz, R. Galambos, Human event-related potential (ERP) recordings, Cogn. Brain Res. (this

auditory evoked potentials. I. Evaluation of components, Elec- issue).troencephalogr. Clin. Neurophys. 36 (1974) 179–190. [57] H.G. Vaughan, W. Ritter, The sources of auditory evoked responses

[44] K.S. Rockland, H. Ojima, Calcarine area V1 as a multimodal recorded from the human scalp, Electroencephalogr. Clin. Neuro-convergence area, Soc. Neurosci. Abs. 27 (2001), Program no. phys. 28 (1970) 360–367.511.20. [58] H.G. Vaughan, W. Ritter, R. Simson, Topographic analysis of

[45] M.T. Schmolesky, Y. Wang, D.P. Hanes, K.G. Thompson, S. auditory event-related potentials, Prog. Brain Res. 54 (1980) 279–Leutgeb, J.D. Schall, A.G. Leventhal, Signal timing across the 285.macaque visual system, J. Neurophysiol. 76 (1998) 3272–3278. [59] H.G. Vaughan, J.C. Arezzo, The neural basis of event-related

[46] C.E. Schroeder, A.D. Mehta, S.J. Givre, A spatiotemporal profile of potentials, in: T.W. Picton (Ed.), Human Event-related Potentials,visual system activation revealed by current source density analysis Handbook of Electroencephalography and Clinical Neurophysiolo-in the awake macaque, Cereb. Cortex 8 (1998) 575–592. gy. Revised Series, Vol. 3, Elsevier, Amsterdam, 1988, pp. 45–96.

[47] C.E. Schroeder, R.W. Lindsley, C. Specht, A. Marcovici, J.F. [60] S.B. Vincent, The function of the vibrissae in the behavior of theSmiley, D.C. Javitt, Somatosensory input to auditory association white rat, Behav. Monogr. 1 (1912).cortex in the macaque monkey, J. Neurophysiol. 85 (2001) 1322– [61] M.T. Wallace, M.A. Meredith, B.E. Stein, Integration of multiple1327. sensory modalities in cat cortex, Exp. Brain Res. 91 (1992) 484–

[48] C.E. Schroeder, A.D. Mehta, J.J. Foxe, Determinants and mecha- 488.nisms of attentional modulation of neural processing, Front. Biosci. [62] W.G. Walter, R. Cooper, V.J. Alderidge, W.C. McCallum, A.L.6 (2001) 672–684. Winter, Contingent negative variation: an electric sign of sen-

[49] C.E. Schroeder, J.J. Foxe, The timing and laminar profile of sorimotor association and expectancy in the human brain, Natureconverging inputs to multisensory areas of the macaque neocortex, 203 (1964) 380–384.Cogn. Brain Res. (this issue).